The Logic of Hebbian Learning
DOI :
https://doi.org/10.32473/flairs.v35i.130735Mots-clés :
Neurosymbolic AI, Hebbian Learning, Dynamic Logics, Knowledge Representation and Reasoning, Nonmonotonic Reasoning, Preference UpgradeRésumé
We present the logic of Hebbian learning, a dynamic logic
whose semantics1 are expressed in terms of a layered neural
network learning via Hebb’s associative learning rule. Its lan-
guage consists of modality Tφ (read “typically φ,” formalized
as forward propagation), conditionals φ ⇒ ψ (read “typi-
cally φ are ψ”), as well as dynamic modalities [φ+]ψ (read
“evaluate ψ after performing Hebbian update on φ”). We give
axioms and inference rules that are sound with respect to the
neural semantics; these axioms characterize Hebbian learning
and its interaction with propagation. The upshot is that this
logic describes a neuro-symbolic agent that both learns from
experience and also reasons about what it has learned.
Téléchargements
Publié-e
Comment citer
Numéro
Rubrique
Licence
(c) Tous droits réservés Caleb Kisby, Saúl Blanco, Lawrence Moss 2022
Cette œuvre est sous licence Creative Commons Attribution - Pas d'Utilisation Commerciale 4.0 International.