HOME

TheInfoList



OR:

Hebbian theory is a
neuropsychological Neuropsychology is a branch of psychology concerned with how a person's cognition and behavior are related to the brain and the rest of the nervous system. Professionals in this branch of psychology focus on how injuries or illnesses of the brai ...
theory claiming that an increase in synaptic efficacy arises from a presynaptic cell's repeated and persistent stimulation of a postsynaptic cell. It is an attempt to explain
synaptic plasticity In neuroscience, synaptic plasticity is the ability of synapses to Chemical synapse#Synaptic strength, strengthen or weaken over time, in response to increases or decreases in their activity. Since memory, memories are postulated to be represent ...
, the adaptation of
neuron A neuron (American English), neurone (British English), or nerve cell, is an membrane potential#Cell excitability, excitable cell (biology), cell that fires electric signals called action potentials across a neural network (biology), neural net ...
s during the learning process. Hebbian theory was introduced by Donald Hebb in his 1949 book '' The Organization of Behavior.'' The theory is also called Hebb's rule, Hebb's postulate, and cell assembly theory. Hebb states it as follows:
Let us assume that the persistence or repetition of a reverberatory activity (or "trace") tends to induce lasting cellular changes that add to its stability. ... When an
axon An axon (from Greek ἄξων ''áxōn'', axis) or nerve fiber (or nerve fibre: see American and British English spelling differences#-re, -er, spelling differences) is a long, slender cellular extensions, projection of a nerve cell, or neuron, ...
of cell ''A'' is near enough to excite a cell ''B'' and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that ''A''’s efficiency, as one of the cells firing ''B'', is increased.
The theory is often summarized as "Neurons that fire together, wire together." However, Hebb emphasized that cell ''A'' needs to "take part in firing" cell ''B'', and such causality can occur only if cell ''A'' fires just before, not at the same time as, cell ''B''. This aspect of causation in Hebb's work foreshadowed what is now known about spike-timing-dependent plasticity, which requires temporal precedence. Hebbian theory attempts to explain
associative In mathematics, the associative property is a property of some binary operations that rearranging the parentheses in an expression will not change the result. In propositional logic, associativity is a valid rule of replacement for express ...
or ''Hebbian learning'', in which simultaneous activation of cells leads to pronounced increases in synaptic strength between those cells. It also provides a biological basis for
errorless learning Errorless learning was an instructional design introduced by psychologist Charles Ferster in the 1950s as part of his studies on what would make the most effective learning environment. B. F. Skinner was also influential in developing the techni ...
methods for education and memory rehabilitation. In the study of
neural networks A neural network is a group of interconnected units called neurons that send signals to one another. Neurons can be either Cell (biology), biological cells or signal pathways. While individual neurons are simple, many of them together in a netwo ...
in cognitive function, it is often regarded as the neuronal basis of
unsupervised learning Unsupervised learning is a framework in machine learning where, in contrast to supervised learning, algorithms learn patterns exclusively from unlabeled data. Other frameworks in the spectrum of supervisions include weak- or semi-supervision, wh ...
.


Engrams, cell assembly theory, and learning

Hebbian theory provides an explanation for how neurons might connect to become engrams, which may be stored in overlapping cell assemblies, or groups of neurons that encode specific information. Initially created as a way to explain recurrent activity in specific groups of cortical neurons, Hebb's theories on the form and function of cell assemblies can be understood from the following:
The general idea is an old one, that any two cells or systems of cells that are repeatedly active at the same time will tend to become 'associated' so that activity in one facilitates activity in the other.
Hebb also wrote:
When one cell repeatedly assists in firing another, the axon of the first cell develops synaptic knobs (or enlarges them if they already exist) in contact with the soma of the second cell.
D. Alan Allport posits additional ideas regarding cell assembly theory and its role in forming engrams using the concept of auto-association, or the brain's ability to retrieve information based on a partial cue, described as follows:
If the inputs to a system cause the same pattern of activity to occur repeatedly, the set of active elements constituting that pattern will become increasingly strongly inter-associated. That is, each element will tend to turn on every other element and (with negative weights) to turn off the elements that do not form part of the pattern. To put it another way, the pattern as a whole will become 'auto-associated'. We may call a learned (auto-associated) pattern an engram.
Research conducted in the laboratory of Nobel laureate
Eric Kandel Eric Richard Kandel (; born Erich Richard Kandel, November 7, 1929) is an Austrian-born American medical doctor who specialized in psychiatry, a neuroscientist and a professor of biochemistry and biophysics at the College of Physicians and Surgeo ...
has provided evidence supporting the role of Hebbian learning mechanisms at synapses in the marine
gastropod Gastropods (), commonly known as slugs and snails, belong to a large Taxonomy (biology), taxonomic class of invertebrates within the phylum Mollusca called Gastropoda (). This class comprises snails and slugs from saltwater, freshwater, and fro ...
'' Aplysia californica''. Because synapses in the
peripheral nervous system The peripheral nervous system (PNS) is one of two components that make up the nervous system of Bilateria, bilateral animals, with the other part being the central nervous system (CNS). The PNS consists of nerves and ganglia, which lie outside t ...
of marine invertebrates are much easier to control in experiments, Kandel's research found that Hebbian
long-term potentiation In neuroscience, long-term potentiation (LTP) is a persistent strengthening of synapses based on recent patterns of activity. These are patterns of synaptic activity that produce a long-lasting increase in signal transmission between two neuron ...
along with activity-dependent presynaptic facilitation are both necessary for
synaptic plasticity In neuroscience, synaptic plasticity is the ability of synapses to Chemical synapse#Synaptic strength, strengthen or weaken over time, in response to increases or decreases in their activity. Since memory, memories are postulated to be represent ...
and
classical conditioning Classical conditioning (also respondent conditioning and Pavlovian conditioning) is a behavioral procedure in which a biologically potent Stimulus (physiology), stimulus (e.g. food, a puff of air on the eye, a potential rival) is paired with a n ...
in ''Aplysia californica''. While research on invertebrates has established fundamental mechanisms of learning and memory, much of the work on long-lasting synaptic changes between vertebrate neurons involves the use of non-physiological experimental stimulation of brain cells. However, some of the physiologically relevant synapse modification mechanisms that have been studied in vertebrate brains do seem to be examples of Hebbian processes. One such review indicates that long-lasting changes in synaptic strengths can be induced by physiologically relevant synaptic activity using both Hebbian and non-Hebbian mechanisms.


Principles

In
artificial neuron An artificial neuron is a mathematical function conceived as a model of a biological neuron in a neural network. The artificial neuron is the elementary unit of an ''artificial neural network''. The design of the artificial neuron was inspired ...
s and
artificial neural network In machine learning, a neural network (also artificial neural network or neural net, abbreviated ANN or NN) is a computational model inspired by the structure and functions of biological neural networks. A neural network consists of connected ...
s, Hebb's principle can be described as a method of determining how to alter the weights between model neurons. The weight between two neurons increases if the two neurons activate simultaneously, and reduces if they activate separately. Nodes that tend to be either both positive or both negative at the same time have strong positive weights, while those that tend to be opposite have strong negative weights. The following is a formulaic description of Hebbian learning (many other descriptions are possible): :\,w_=x_ix_j, where w_ is the weight of the connection from neuron j to neuron i , and x_i is the input for neuron i . This is an example of pattern learning, where weights are updated after every training example. In a Hopfield network, connections w_ are set to zero if i=j (no reflexive connections allowed). With binary neurons (activations either 0 or 1), connections would be set to 1 if the connected neurons have the same activation for a pattern. When several training patterns are used, the expression becomes an average of the individuals: :w_ = \frac \sum_^p x_i^k x_j^k, where w_ is the weight of the connection from neuron j to neuron i , p is the number of training patterns and x_^k the k -th input for neuron i . This is learning by epoch, with weights updated after all the training examples are presented and is last term applicable to both discrete and continuous training sets. Again, in a Hopfield network, connections w_ are set to zero if i=j (no reflexive connections). A variation of Hebbian learning that takes into account phenomena such as blocking and other neural learning phenomena is the mathematical model of Harry Klopf. Klopf's model assumes that parts of a system with simple adaptive mechanisms can underlie more complex systems with more advanced adaptive behavior, such as neural networks.


Relationship to unsupervised learning, stability, and generalization

Because of the simple nature of Hebbian learning, based only on the coincidence of pre- and post-synaptic activity, it may not be intuitively clear why this form of plasticity leads to meaningful learning. However, it can be shown that Hebbian plasticity does pick up the statistical properties of the input in a way that can be categorized as unsupervised learning. This can be mathematically shown in a simplified example. Let us work under the simplifying assumption of a single rate-based neuron of rate y(t), whose inputs have rates x_1(t) ... x_N(t). The response of the neuron y(t) is usually described as a linear combination of its input, \sum_i w_ix_i, followed by a response function f: :y = f\left(\sum_^N w_i x_i \right). As defined in the previous sections, Hebbian plasticity describes the evolution in time of the synaptic weight w: :\frac = \eta x_i y. Assuming, for simplicity, an identity response function f(a)=a, we can write :\frac = \eta x_i \sum_^N w_j x_j or in
matrix Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the m ...
form: :\frac = \eta \mathbf\mathbf^T\mathbf. As in the previous chapter, if training by epoch is done an average \langle \dots \rangle over discrete or continuous (time) training set of \mathbf can be done:\frac = \langle \eta \mathbf\mathbf^T\mathbf \rangle = \eta \langle \mathbf\mathbf^T\rangle\mathbf = \eta C \mathbf.where C = \langle\, \mathbf\mathbf^T \rangle is the
correlation matrix In statistics, correlation or dependence is any statistical relationship, whether causal or not, between two random variables or bivariate data. Although in the broadest sense, "correlation" may indicate any type of association, in statistics ...
of the input under the additional assumption that \langle\mathbf\rangle = 0 (i.e. the average of the inputs is zero). This is a system of N coupled linear differential equations. Since C is
symmetric Symmetry () in everyday life refers to a sense of harmonious and beautiful proportion and balance. In mathematics, the term has a more precise definition and is usually used to refer to an object that is invariant under some transformations ...
, it is also diagonalizable, and the solution can be found, by working in its eigenvectors basis, to be of the form :\mathbf(t) = k_1e^\mathbf_1 + k_2e^\mathbf_2 + ... + k_Ne^\mathbf_N where k_i are arbitrary constants, \mathbf_i are the eigenvectors of C and \alpha_i their corresponding eigen values. Since a correlation matrix is always a
positive-definite matrix In mathematics, a symmetric matrix M with real entries is positive-definite if the real number \mathbf^\mathsf M \mathbf is positive for every nonzero real column vector \mathbf, where \mathbf^\mathsf is the row vector transpose of \mathbf. Mo ...
, the eigenvalues are all positive, and one can easily see how the above solution is always exponentially divergent in time. This is an intrinsic problem due to this version of Hebb's rule being unstable, as in any network with a dominant signal the synaptic weights will increase or decrease exponentially. Intuitively, this is because whenever the presynaptic neuron excites the postsynaptic neuron, the weight between them is reinforced, causing an even stronger excitation in the future, and so forth, in a self-reinforcing way. One may think a solution is to limit the firing rate of the postsynaptic neuron by adding a non-linear, saturating response function f, but in fact, it can be shown that for ''any'' neuron model, Hebb's rule is unstable. Therefore, network models of neurons usually employ other learning theories such as
BCM theory Bienenstock–Cooper–Munro (BCM) theory, BCM synaptic modification, or the BCM rule, named after Elie Bienenstock, Leon Cooper, and Paul Munro, is a physical theory of learning in the visual cortex developed in 1981. The BCM model proposes a sli ...
, Oja's rule, or the generalized Hebbian algorithm. Regardless, even for the unstable solution above, one can see that, when sufficient time has passed, one of the terms dominates over the others, and :\mathbf(t) \approx e^\mathbf^* where \alpha^* is the ''largest'' eigenvalue of C. At this time, the postsynaptic neuron performs the following operation: :y \approx e^\mathbf^* \mathbf Because, again, \mathbf^* is the eigenvector corresponding to the largest eigenvalue of the correlation matrix between the x_is, this corresponds exactly to computing the first principal component of the input. This mechanism can be extended to performing a full PCA (
principal component analysis Principal component analysis (PCA) is a linear dimensionality reduction technique with applications in exploratory data analysis, visualization and data preprocessing. The data is linearly transformed onto a new coordinate system such that th ...
) of the input by adding further postsynaptic neurons, provided the postsynaptic neurons are prevented from all picking up the same principal component, for example by adding
lateral inhibition In neurobiology, lateral inhibition is the capacity of an excited neuron to reduce the activity of its neighbors. Lateral inhibition disables the spreading of action potentials An action potential (also known as a nerve impulse or "spike" w ...
in the postsynaptic layer. We have thus connected Hebbian learning to PCA, which is an elementary form of unsupervised learning, in the sense that the network can pick up useful statistical aspects of the input, and "describe" them in a distilled way in its output.


Hebbian learning and mirror neurons

Hebbian learning and spike-timing-dependent plasticity have been used in an influential theory of how mirror neurons emerge.Keysers, C. (2011). ''The Empathic Brain''. Mirror neurons are neurons that fire both when an individual performs an action and when the individual sees or hears another perform a similar action. The discovery of these neurons has been very influential in explaining how individuals make sense of the actions of others, since when a person perceives the actions of others, motor programs in the person's brain which they would use to perform similar actions are activated, which add information to the perception and help to predict what the person will do next based on the perceiver's own motor program. One limitation of this idea of mirror neuron functions is explaining how individuals develop neurons that respond both while performing an action and while hearing or seeing another perform similar actions. Neuroscientist Christian Keysers and psychologist David Perrett suggested that observing or hearing an individual perform an action activates brain regions as if performing the action oneself. These re-afferent sensory signals trigger activity in neurons responding to the sight, sound, and feel of the action. Because the activity of these sensory neurons will consistently overlap in time with those of the motor neurons that caused the action, Hebbian learning predicts that the synapses connecting neurons responding to the sight, sound, and feel of an action and those of the neurons triggering the action should be potentiated. The same is true while people look at themselves in the mirror, hear themselves babble, or are imitated by others. After repeated occurrences of this re-afference, the synapses connecting the sensory and motor representations of an action are so strong that the motor neurons start firing to the sound or the vision of the action, and a mirror neuron is created. Numerous experiments provide evidence for the idea that Hebbian learning is crucial to the formation of mirror neurons. Evidence reveals that motor programs can be triggered by novel auditory or visual stimuli after repeated pairing of the stimulus with the execution of the motor program. For instance, people who have never played the piano do not activate brain regions involved in playing the piano when listening to piano music. Five hours of piano lessons, in which the participant is exposed to the sound of the piano each time they press a key is proven sufficient to trigger activity in motor regions of the brain upon listening to piano music when heard at a later time. Consistent with the fact that spike-timing-dependent plasticity occurs only if the presynaptic neuron's firing predicts the post-synaptic neuron's firing, the link between sensory stimuli and motor programs also only seem to be potentiated if the stimulus is contingent on the motor program.


Hebbian theory and cognitive neuroscience

Hebbian learning is linked to cognitive processes like decision-making and social learning. The field of cognitive neuroscience has started to explore the intersection of Hebbian theory with brain regions responsible for reward processing and social cognition, such as the striatum and prefrontal cortex. In particular, striatal projections exposed to Hebbian models exhibit long-term potentiation and long-term depression ''in vivo''. Additionally, models of the prefrontal cortex to stimuli ("mixed selectivity") are not entirely explained by random connectivity, but when a Hebbian paradigm is incorporated, the levels of mixed selectivity in the model are reached. It is hypothesized that Hebbian plasticity in these areas may underlie behaviors like habit formation, reinforcement learning, and even the development of social bonds.


Limitations

Despite the common use of Hebbian models for long-term potentiation, Hebbian theory does not cover all forms of long-term synaptic plasticity. Hebb did not propose any rules for inhibitory synapses or predictions for anti-causal spike sequences (where the presynaptic neuron fires ''after'' the postsynaptic neuron). Synaptic modification may not simply occur only between activated neurons A and B, but at neighboring synapses as well. Therefore, all forms of heterosynaptic plasticity and homeostatic plasticity are considered non-Hebbian. One example is retrograde signaling to presynaptic terminals. The compound most frequently recognized as a retrograde transmitter is
nitric oxide Nitric oxide (nitrogen oxide, nitrogen monooxide, or nitrogen monoxide) is a colorless gas with the formula . It is one of the principal oxides of nitrogen. Nitric oxide is a free radical: it has an unpaired electron, which is sometimes den ...
, which, due to its high solubility and diffusivity, often exerts effects on nearby neurons. This type of diffuse synaptic modification, known as volume learning, is not included in the traditional Hebbian model.


Contemporary developments, artificial intelligence, and computational advancements

Modern research has expanded upon Hebb's original ideas. Spike-timing-dependent plasticity (STDP), for example, refines Hebbian principles by incorporating the precise timing of neuronal spikes to Hebbian theory. Experimental advancements have also linked Hebbian learning to complex behaviors, such as decision-making and emotional regulation. Current studies in
artificial intelligence Artificial intelligence (AI) is the capability of computer, computational systems to perform tasks typically associated with human intelligence, such as learning, reasoning, problem-solving, perception, and decision-making. It is a field of re ...
(AI) and quantum computing continue to leverage Hebbian concepts for developing adaptive algorithms and improving machine learning models. In AI, Hebbian learning has seen applications beyond traditional neural networks. One significant advancement is in reinforcement learning algorithms, where Hebbian-like learning is used to update the weights based on the timing and strength of stimuli during training phases. Some researchers have adapted Hebbian principles to develop more biologically plausible models for learning in artificial systems, which may improve model efficiency and convergence in AI applications. A growing area of interest is the application of Hebbian learning in quantum computing. While classical neural networks are the primary area of application for Hebbian theory, recent studies have begun exploring the potential for quantum-inspired algorithms. These algorithms leverage the principles of quantum superposition and entanglement to enhance learning processes in quantum systems.Current research is exploring how Hebbian principles could inform the development of more efficient quantum machine learning models. New computational models have emerged that refine or extend Hebbian learning. For example, some models now account for the precise timing of neural spikes (as in spike-timing-dependent plasticity), while others have integrated aspects of neuromodulation to account for how neurotransmitters like dopamine affect the strength of synaptic connections. These advanced models provide a more nuanced understanding of how Hebbian learning operates in the brain and are contributing to the development of more realistic computational models. Recent research on Hebbian learning has focused on the role of inhibitory neurons, which are often overlooked in traditional Hebbian models. While classic Hebbian theory primarily focuses on excitatory neurons, more comprehensive models of neural learning now consider the balanced interaction between excitatory and inhibitory synapses. Studies suggest that inhibitory neurons can provide critical regulation for maintaining stability in neural circuits and might prevent runaway positive feedback in Hebbian learning.Cohen, M. R., & Kohn, A. (2011). Measuring and interpreting neuronal correlations. *Nature Neuroscience*, 14(7), 811-819.


See also

* Dale's principle *
Coincidence detection in neurobiology Coincidence detection is a neuronal process in which a neural circuit encodes information by detecting the occurrence of temporally close but spatially distributed input signals. Coincidence detectors influence neuronal information processing b ...
* Leabra * Metaplasticity * Tetanic stimulation * Synaptotropic hypothesis *
Neuroplasticity Neuroplasticity, also known as neural plasticity or just plasticity, is the ability of neural networks in the brain to change through neurogenesis, growth and reorganization. Neuroplasticity refers to the brain's ability to reorganize and rewir ...
*
Behaviorism Behaviorism is a systematic approach to understand the behavior of humans and other animals. It assumes that behavior is either a reflex elicited by the pairing of certain antecedent stimuli in the environment, or a consequence of that indivi ...
* Three-factor learning


References


Further reading

* * * *


External links


Overview
* Hebbian Learning tutorial
Part 1: Novelty FilteringPart 2: PCA
{{DEFAULTSORT:Hebbian Theory Unsupervised learning Neuroplasticity *