Download A computational account for the ontogeny of mirror neurons via

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Holonomic brain theory wikipedia , lookup

Eyeblink conditioning wikipedia , lookup

Synaptogenesis wikipedia , lookup

Neuroethology wikipedia , lookup

Catastrophic interference wikipedia , lookup

Artificial general intelligence wikipedia , lookup

Bird vocalization wikipedia , lookup

Multielectrode array wikipedia , lookup

Neuroeconomics wikipedia , lookup

Neural modeling fields wikipedia , lookup

Embodied language processing wikipedia , lookup

Clinical neurochemistry wikipedia , lookup

Axon wikipedia , lookup

Convolutional neural network wikipedia , lookup

Molecular neuroscience wikipedia , lookup

Activity-dependent plasticity wikipedia , lookup

Single-unit recording wikipedia , lookup

Circumventricular organs wikipedia , lookup

Recurrent neural network wikipedia , lookup

Neuroanatomy wikipedia , lookup

Caridoid escape reaction wikipedia , lookup

Development of the nervous system wikipedia , lookup

Neurotransmitter wikipedia , lookup

Neural oscillation wikipedia , lookup

Stimulus (physiology) wikipedia , lookup

Metastability in the brain wikipedia , lookup

Central pattern generator wikipedia , lookup

Neural coding wikipedia , lookup

Nonsynaptic plasticity wikipedia , lookup

Types of artificial neural networks wikipedia , lookup

Biological neuron model wikipedia , lookup

Donald O. Hebb wikipedia , lookup

Optogenetics wikipedia , lookup

Premovement neuronal activity wikipedia , lookup

Feature detection (nervous system) wikipedia , lookup

Chemical synapse wikipedia , lookup

Pre-Bötzinger complex wikipedia , lookup

Neuropsychopharmacology wikipedia , lookup

Channelrhodopsin wikipedia , lookup

Mirror neuron wikipedia , lookup

Nervous system network models wikipedia , lookup

Synaptic gating wikipedia , lookup

Transcript
A computational account for the ontogeny of
mirror neurons via Hebbian learning
Graduation Project Bachelor Artificial Intelligence
Credits: 18 EC
Author
Lotte Weerts
10423303
University of Amsterdam
Faculty of Science
Science Park 904
1098 XH Amsterdam
Supervisors
dr. Sander Bohté
Centrum Wiskunde & Informatica
Life Sciences Group
Science Park 123
1098 XG Amsterdam
dr. Rajat Mani Thomas
Netherlands Institute for Neuroscience
Social Brain Lab
Meibergdreef 47
1105 BA Amsterdam
prof. dr. Max Welling
University of Amsterdam
Faculty of Science
Science Park 904
1098 XH Amsterdam
June 26, 2015
Abstract
It has been proposed that Hebbian learning could be responsible for the ontogeny
of predictive mirror neurons (Keysers and Gazzola, 2014). Predictive mirror neurons
fire both when an individual performs a certain action and when the action they
encode is most likely to follow a concurrently observed action. Here, we show that
a variation of Oja’s rule (an implementation of Hebbian learning) is sufficient for
the emergence of mirror neurons. An artificial neural network has been created
that simulates the interactions between the premotor cortex (PM) and the superior
temporal sulcus (STS). The PM cortex coordinates self-performed actions, whereas
the STS is a region known to respond to the sight of body movements and the
sound of actions. A quantitative method for analyzing (predictive) mirror neuron
activity in the artificial neural network is presented. A parameter space analysis
shows that a variation of Oja’s rule that uses a threshold produces mirror neuron
behavior, given that the bounds for the weights of inhibitory synapses are lower than
those of excitatory synapses. By extension, this work provides positive evidence
for the association hypothesis, which states that mirror-like behavior of neurons in
the motorcortex arises due to associations established between sensory- and motor
stimuli.
Acknowledgements
My sincere thanks goes to Sander Bohté, for sharing his expertise and
for providing me with excellent guidance and encouragement
throughout the course of this project.
Furthermore, I am extremely grateful to Rajat Thomas, for the
illuminating insight and expertise he has shared with me with great
enthusiasm.
I would also like to thank Max Welling, who took the effort to be my
university supervisor regardless of his busy schedule.
Finally, I want to thank my parents and sisters for their continuous
source of support and motivation.
Contents
1 Introduction
2 Theoretical foundation
2.1 Interpretations of associative learning .
2.1.1 Hebbian learning . . . . . . . .
2.1.2 Rescorla Wagner . . . . . . . .
2.2 Hebbian learning rules . . . . . . . . .
2.2.1 Basic Hebb Rule . . . . . . . .
2.2.2 Covariance rule . . . . . . . . .
2.2.3 BCM . . . . . . . . . . . . . . .
2.2.4 Oja’s rule . . . . . . . . . . . .
2.3 Summary . . . . . . . . . . . . . . . .
1
.
.
.
.
.
.
.
.
.
3
3
3
4
4
5
5
6
6
7
.
.
.
.
.
8
8
9
10
11
12
4 Analysis procedure
4.1 A visualization of mirror neuron behavior . . . . . . . . . . . . . . . . . .
4.2 Analysis procedure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.3 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
13
13
13
16
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
3 Computational model
3.1 A motivation for the structure of the model
3.2 Modeling a single neuron . . . . . . . . . . .
3.3 Modeling action execution and observation .
3.4 Learning rules . . . . . . . . . . . . . . . . .
3.5 Summary . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
5 Results
17
5.1 Comparison of learning rules . . . . . . . . . . . . . . . . . . . . . . . . . . 17
5.2 Parameter space analysis of Oja’s rule . . . . . . . . . . . . . . . . . . . . 20
5.3 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
6 Conclusion
22
7 Discussion and future work
23
7.1 Limitations of this study . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
7.2 Future work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
1
Introduction
In the early 1990s, mirror neurons were discovered in the ventral premotor cortex of
the macaque monkey (Di Pellegrino et al., 1992). These neurons fired both when the
monkeys grabbed an object and when they watched another primate grab that same
object. Mirror neuron-like activity has been observed in monkeys, birds and humans
(Catmur, 2013). More recently, evidence has started to emerge that suggests the existence
of mirror neurons with predictive properties (Keysers and Gazzola, 2014). These neurons
fire when the action they encode is the action most likely to happen next, rather than the
action that is concurrently being observed. A critical question concerns how (predictive)
mirror neurons have developed to behave the way they do. In other words: what is the
ontogeny of mirror neurons?
Two distinct views have emerged with regards to the origin of mirror neurons: the
adaptation hypothesis and the association hypothesis (Heyes, 2010). The adaptation hypothesis refers to the idea that mirror neurons are an adaptation for action understanding
and thus a result of evolution. According to this hypothesis, humans are born with mirror
neurons and experience plays a relatively minor role in their development. Alternatively,
one can view mirror neurons as a product of associative learning, which suggests that
these neurons emerge through a correlated experience of observing and executing the
same action. The mechanism responsible for associative learning itself is a product of
evolution, but associative learning did not evolve for the ’purpose’ of producing mirror
neurons. The important difference between these two theories is the role of sensorimotor
experience, which is crucial for the associative hypothesis but has a less prevalent role
in the adaptation hypothesis. In this work, the focus will lie on further developing the
association hypothesis.
While both the association- and adaptation hypotheses can account for the ontogeny
of mirror neurons, the association hypothesis has several advantages. First, it is consistent
with evidence that indicates that mirror neurons contribute to a range of social cognitive
functions that do not play a role in action understanding, such as emotions (Heyes, 2010).
Second, research suggests that the perceptual motor system is optimized to provide the
brain with input for associative learning (Giudice et al., 2009). For example, infants have
a marked visual preference for hands. Such a preference could provide a rich set of input
stimuli for mirror neurons that are active in manual actions.
Keysers and Gazzola (2014) have proposed a mechanistic perspective on how mirror
neurons in the premotor cortex could emerge due to associative learning, or more precisely, Hebbian learning. Hebbian learning was first proposed by Donald Hebb in 1949 as
an explanation for synaptic plasticity, which refers to the ability of the connections (or
synapses) between neurons to become either weaker or stronger (Cooper et al., 2013b).
Thus, Hebbian learning explains learning on a local level between single neurons. Hebb
postulated the principle as follows: "when an axon of cell A is near enough to excite
a cell B and repeatedly or persistently takes part in firing it, some growth process or
metabolic change takes place in one or both cells such that A’s efficiency, as one of the
cells firing B, is increased".
1
To investigate whether or not Hebbian learning is sufficient to lead to the emergence
of mirror neurons, we present a computational approach that implements the mechanics
described by Keysers and Gazzola (2014). This involves the usage of an artificial neural
network (ANN) to simulate activity in the premotor cortex (PM) and the superior temporal sulcus (STS). The PM cortex coordinates self-performed actions, whereas the STS
is a region known to respond to the sight of body movements and the sound of actions.
Action execution and observation will be simulated by exciting neurons in these two areas. The aim of this simulation is to find out whether or not associative learning - in
particular Hebbian learning - is sufficient to lead to the development of mirror neurons.
Specifically, the problem investigated in this work can be defined as follows:
Can artificial neural networks that evolve via a local Hebbian-like learning rule, when
exposed to action execution and observation, be sufficient to lead to the emergence of
predictive mirror neurons?
Here, we show that Oja’s rule, an implementation of Hebbian learning, is sufficient
to impose predictive mirror-like behavior. First, the proposed research question is considered in relation to other studies on the ontogeny of mirror neurons (chapter 2). This
includes a review of four implementations of Hebbian learning: basic Hebb, the covariance learning rule, BCM, and Oja’s rule. Subsequently, the computational model that
describes the artificial neural network is introduced in chapter 3. Additionally, a variation of Oja’s rule is presented. The procedure for the quantitative analysis of mirror
neuron-like behavior, as explained in chapter 4, has been applied to the recorded activity
of ANNs that evolve via several different learning rules. The results of this analysis are
described in chapter 5. Finally, it is argued in chapter 6 that an artificial neural network
that evolves via a variation of Oja’s rule is sufficient to explain the emergence of mirror
neurons. Related problems that should be considered in future work are dealt with in
chapter 7, the final chapter.
2
2
Theoretical foundation
In this chapter, research that has preceded the current work will be discussed. First, two
computational accounts for the ontogony of mirror neurons via the association hypothesis
will be discussed. Subsequently, an elaboration of several implementations of Hebbian
learning that have been proposed in the literature will be given.
2.1
Interpretations of associative learning
The association hypothesis states that mirror neurons emerge due to associative learning.
This term refers to any learning process in which a new response becomes associated with
a particular stimulus (Terry, 2006). As a refinement of the association hypothesis, two
different implementations of associative learning have been proposed, namely Hebbian
learning (Keysers and Gazzola, 2014) and Rescorla-Wagner (Cooper et al., 2013b).
2.1.1
Hebbian learning
Keysers and Gazzola (2014) have proposed a model in which Hebbian learning could
account for the emergence of mirror neurons. Whereas there are several brain regions
that contain mirror neurons, they illustrate their idea with connections between the
premotor cortex (PM), the inferior posterior parietal cortex (area PF/PFG) and superior
temporal sulcus (STS). Both the PM and PF/PFG play a role in action coordination,
whereas the STS is a region known to respond to the sight of body movements and
the sound of actions. More recently, evidence has started to emerge that some mirror
neurons in the PM have predictive properties (Keysers and Gazzola, 2014). When an
action is observed, these neurons behave as predictors for the action that is most likely
to follow next. Consider the sequence of actions relevant to grabbing a cup of coffee,
which consists of a sequence of reaching, grabbing the cup and bringing the cup to the
mouth. If one observes another person reaching his or her arm, this induces activity in
the STS. Additionally, PM neurons that encode ’grabbing’ - the action that is likely to
be performed next - fire. Such predictive properties are of great importance because it
allows us to anticipate the actions of others. The proposed model explains the emergence
of such neurons as a result of sensorimotor input of self-performed actions. When one
performs a certain action, this has been coordinated by activity in the PM. Subsequently,
one observes him- or herself perform that particular action. This is called reafference.
Once the observation of the action has reached the STS, neurons in the PM that encode
for the next action have already become active. If synaptic plasticity caused by Hebbian
learning is assumed, the simultaneous activity in the STS of the observed action and
activity in the PM for the next action causes an association between neurons in both
regions. Keysers and Gazzola (2014) suggest that such an increase in synaptic strength
could account for the emergence of predictive mirror neurons.
3
2.1.2
Rescorla Wagner
Cooper et al. (2013b) used a computational model to investigate the development of
mirror neurons. Two implementations of associative learning were considered; RescorlaWagner - a supervised learning rule that relies on reducing a prediction error - and an
implementation of Hebbian learning and . Their computational model simulates results of
an experimental study (Cook et al., 2010) that examined automatic imitation. Automatic
imitation is a stimulus-response effect where the physical representation of features that
are task-irrelevant facilitate similar responses but interfere with dissimilar responses. In
this study participants were asked to open and close their hands in response to a color
stimulus (the ’task-relevant’ stimulus) whilst watching a video of an opening or closing
hand (the ’task-irrelevant’ stimulus). Research has shown that magnetic stimulation of
the inferior frontal gyrus (an area where mirror neurons have been found) selectively
disrupts automatic imitation. Therefore, automatic imitation is thought to be an index
for mirror neuron activity. Cook et al. (2010) showed that automatic imitation is reduced
more by contingent and signalled training than by non-contingent sensorimotor training.
The results of this study have been successfully simulated by Cooper et al. (2013b)
using a model based on the Rescorla-Wagner learning rule. The Hebbian learning rule,
however, could not explain the reported data of Cook et al. (2010). Therefore, Cooper
et al. (2013b) argue that prediction error-based learning rules, not Hebbian learning, can
account for the development of mirror neurons.
The study of Cooper et al. (2013b) suggests that Hebbian learning is not sufficient to
impose mirror neuron behavior. However, the used implementation of Hebbian learning
suffers from a computational problem. The weights of the connections between neurons
imposed by this learning rule are in principle unbounded, which makes the learning rule
unstable. More advanced types of Hebbian learning that adhere to such weight constraints, for example Oja’s rule (Oja, 1982), were not considered. Thus, even though
Cooper et al. (2013b) have shown that supervised learning is sufficient to explain automatic imitation data, Hebbian learning should not be dismissed fully as a possible
explanation for the emergence of mirror neurons.
2.2
Hebbian learning rules
At the time Hebbian learning was introduced by Donald Hebb (1949), the field of neuroscience, let alone the field of neuroinformatics, was still in its infancy. Hebb’s postulate
was introduced in a general, qualitative manner. It implies that learning occurs due to
changes in synaptic strength, and that this is caused by a correlation between the input
of a pre- and postsynaptic neuron. These changes have later been defined as long term
potentiation (LTP) and long term depression (LTD), which refer to an increase and decrease of synaptic efficacy respectively (Kandel et al., 2000). However, the qualitative
description of Hebb leaves much room for interpretation. Over time many computational implementations of Hebb’s postulate have been proposed. Here, four of these will
be discussed: the basic Hebb rule, the covariance rule, BCM and Oja’s rule.
4
2.2.1
Basic Hebb Rule
A simple rule that uses the presynaptic activity ai and postsynaptic activity aj to change
the synaptic strength of the connection wj,i is the basic Hebb Rule (Hebb, 1949):
∆wj,i = αhai · aj i
(1)
Here, α is the learning rate, which determines how quickly the synaptic weights
change. The brackets h and i denote that ∆wj,i is the average of all synaptic changes
over a particular time. This articulates the idea that synaptic modification is a slow
process. The rationale behind the formula is that if both neurons are relatively active,
their synaptic connection will increase substantially. However, if one of them or both are
less active, the synaptic strength will increase less.
The major disadvantage of the basic Hebb rule is that the weights can only increase.
Consequently, the basic Hebbian rule accounts for LTP but does not allow LTD. Since
LTD has been recorded in neurons, a rule that fails to account for this phenomena is
biologically implausible. Intuitively, it prevents the system from being able to ’unlearn’
what it has learned.
Moreover, the basic Hebb rule is inherently unbounded, since weights can increase
indefinitely. This is an effect that does not occur in biological systems due to the physical
constraints to the growth of synapses. Aditionally, Hebb’s rule does not take baseline
activity into account. In general, neurons always have some baseline of noisy activity.
Since the simple Hebb rule will increase the weights even if the recorded activity is simply
baseline activity, the weight updates can easily get distorted .
2.2.2
Covariance rule
The covariance rule is an alternative to the basic Hebb rule that implements both LTP
and LTD (Dayan and Abbott, 2001). It is defined as follows:
cov(x, y) =
1
n−1
∆wj,i = α · cov(ai , aj )
n
X
(xi − hxi)(yi − hyi)
(2)
(3)
i=0
As in the basic Hebb rule, α refers to the learning rate. As suggested by its name,
the covariance rule calculates the covariance of the activities of the pre- and postsynaptic
neurons and uses this to determine the weight change. The covariance of two sequences
is negative if they are uncorrelated, and positive if they are correlated. Therefore, the
covariance rule allows for negative weight updates and thus accounts for both LTP and
LTD. By substracting the average activity over a certain time period, the covariance rule
naturally takes the baseline activity into account.
However, similar as the basic Hebb rule, the covariance rule is still unbounded. In
5
theory, the weights can grow infinitely if the activation sequences keep being correlated.
In biology this is not a problem because the synapses are naturally constrained by their
physical properties to grow infinitely. In a computational model, however, such constraints have to be explicitly incorporated in the model, which is not the case for neither
the covariance rule nor the basic Hebb rule.
Another problem with the covariance rule is that ∆wj,i will always be the same as
∆wi,j , since the formula does not differentiate between the pre- and postsynaptic neuron.
Separating the development of two reciprocal connections would make the learning rule
more biologically plausible.
2.2.3
BCM
BCM is a modifcation on the covariance rule that imposes a bound on the synaptic
strengths. The rule, named after its introducers Bienenstock, Cooper and Munro (1982),
is defined as follows:
∆wj,i = hai · aj (aj − θv )i
(4)
θv acts as a threshold that determines whether synapses are strengthened or weakened. To stabilizes synaptic plasticity, BCM reduces the synaptic weights of the postsynaptic neuron if it becomes too active. This can be achieved by implementing θv as a
sliding threshold:
∆θv = ha2j − θv i
(5)
The sliding threshold induces competition between the synapses, because strengthening some synapses increases aj and thus increases θv , which in effect makes it more
difficult for other synapses to increase afterwards. However, one disadvantage of BCM
is that it does not take the baseline activity of the presynaptic neurons into account.
2.2.4
Oja’s rule
Oja’s rule, introduced by Oja (1982), is an alternative learning rule that imposes bounds
on the weights:
∆w[i,j] = hai · aj − α(a2j )w[i,j] i
(6)
Oja’s rule imposes an online re-normalization by constraining the sum of squares of
the synaptic weights. More precisely, the bound of the weights is inversely proportional
to α, that is, |w|2 will relax to α1 (Oja, 1982).
By constraining the weight vector to be of constant length, Oja’s rule imposes competition between different synapses. However, similar to BCM, Oja’s rule does not take
6
the baseline activity of the presynaptic neuron into account.
2.3
Summary
In sum, the field of neuroscience is currently evidencing a debate about the ontogeny of
mirror neurons. The associative hypothesis, which assumes that mirror neurons emerge
due to associative learning, seems to be a promising explanation. Several interpretations of associative learning have been proposed, namely Rescorla-Wagner and Hebbian
learning. Research has shown that supervised learning rules such as Rescorla-Wagner are
sufficient to explain results found in studies on automatic imitation. A different model
that explains the emergence of predictive mirror neurons by means of Hebbian learning
has been proposed, but a proof of concept has yet to be created. The chapter is concluded
by a review of several implementations of Hebbian learning, including basic Hebb, the
covariance rule, BCM and Oja’s rule.
7
3
Computational model
In this chapter we will introduce the computational model that was used to simulate the
emergence of mirror neurons. First, a motivation for the structure of the artificial neuron
network will be presented. Then, the model for a single neuron is presented. Finally, it
will be shown how this model can simulate action executions and observations.
3.1
A motivation for the structure of the model
There are several regions in the brains of monkeys where mirror neurons have been
found. Here, similar as in Keysers and Gazzola (2014), the focus lies on a set of three
interconnected regions: the premotor cortex (PM, area F5), the inferior posterior parietal
cortex (PF/PFG) and the superior temporal sulcus (STS). Both the PM and the PF/PFG
are known to contain mirror neurons. Their main function is the coordination of actions,
whereas the STS is an area in the temporal cortex that responds to the sight of body
movements and the sound of actions. Neurons in the PM and PF/PFG are reciprocally
connected, which means that the connectivity goes both from PM to PF/PFG as the
other way around. Similarly, there are reciprocal connections between the STS and
PF/PFG. Figure 1 shows the structure of the model used to simulate these areas.
In order to make a simulation of the dynamics between these different brain regions
feasible, three simplifications have been introduced. First, the intermediate step via the
PF/PFG is omitted. This results in a model that has only reciprocal connections between
the STS and PM. The artificial neural network will thus consist of two layers of neurons.
Second, it is assumed that the PM and STS have similar encodings for the same
action phases. For example, the action ‘drink a cup of coffee’ could consist of an action
phase sequence of ‘grabbing’, ‘reaching’ and ‘bringing hand to mouth’. In this example,
the PM and STS will each have three neurons that encode these three action phases. It
is not claimed that a single action phase is actually encoded by a single neuron in the
brain. In contrast, the ‘neurons’ of the artificial neural network should not be considered
simulating single neurons. As will be described in the next section, neurons are modeled
by activities that can take values between 0 and 1. In reality, however, a single neuron
can only send a binary signal - the action potential - to postsynaptic neurons. Therefore,
the artificial neurons presented here represent the average spiking rate of a group of
neurons rather than the activity of a single neuron.
Third, the connections from the PM to the STS will be restricted to inhibitory connections between neurons that encode the same action. These connections are inhibitory,
because the activity from PM to STS is thought to be net inhibitory (Hietanen and Perrett, 1993). However, it is biologically implausible to have as many inhibitory connections
as excitatory connections, since inhibitory synapses only comprise a small fraction of the
total number of synapses in the brain (Kandel et al., 2000). Computationally, too many
inhibitory connections can destabilize the network since they allow the PM to completely
suppress all STS activity. For these reasons, only inhibitory connections from the PM to
STS that encode the same action are modeled.
8
Figure 1: The setup used for the simulation. It consists of a two-layered neural network that represents
the PM (red nodes) and STS (blue nodes) cortices. Each action phase, e.g. "reach" or "grasp", is
encoded in one neuron in both the PM and STS. All STS neurons are connected to all PM neurons via
excitatory connections (blue arrows), whereas PM neurons are connected via inhibitory synapses (red
arrows) to the STS neurons that encode the same action. The weight matrices store the strengths of the
connections. In the training phase, PM neurons are sequentially activated to simulate the execution of
an action sequence, "reach, grasp, bring to mouth". The order of the excitation of the PM neurons is
determined by a transition matrix, which contains the probabilities that certain actions are followed by
others. Here, the sequence "reach (A) , grasp (B), bring to mouth (C)" is encoded, since the probabilities
to transition from A → B and B → C are both 1. With a delay of 200 ms, the STS is excited in the same
order. During the testing phase action observation is modeled. Therefore, only the STS is activated
whilst still adhering to the probabilities encoded in the transition matrix.
3.2
Modeling a single neuron
The activity of a single neuron is represented by αi , which can take a value between 0
and 1. The activity, similar as in Cooper et al. (2013a), is calculated as follows:
ai (t + 1) = ρai (t) + (1 − ρ)σ(Ii (t))
X
Ij (t) =
(wj,i ai (t − 1)) + Ej + Bj + N (0, µ2 )
(7)
(8)
i
Here, ρ is a value between 0 and 1 that determines to which extent previous activity
persists over time. Ij (t) refers to the input to the neuron, which is transformed by a value
between 0 and 1 by the sigmoid function σ. The value of Ij (t) depends among others
9
Parameter
Persistance
External excitation
Bias
Gaussian noise
ρ
Ej
Bj
µ
Value
0.990
4.0
2.0
2.0
A
B
C
D
E
F
A
0
0
0
0
0.5
0.5
B
0
0
0
0
0.5
0.5
C
1
0
0
0
0
0
D
0
1
0
0
0
0
E
0
0
1
0
0
0
F
0
0
0
1
0
0
(a) The parameters used in modeling neurons, (b) Example of a transition matrix that encodes for the actions A-C-E and B-D-F. The
drawn from Cooper et al. (2013b)
matrix contains the probability that one action
phase (row) is followed by another (column).
Figure 2: Parameters used in the model (a) and an example of a transition matrix (b)
on the weighted sum of the activities of all neurons. The weight matrix w contains
the weights of the connections. When the artificial network is trained, w is altered.
Additionally, the value of Ij (t) depends on three factors: Ej , Bj and µ. Ej refers to the
amount of stimulation caused by other sources than neurons within the network and can
be either on or off. External stimulation of PM neurons represents choosing the next
action that will be executed. In the case of STS neurons, Ej represents the observation
of a certain action in the real world. The input signal further consist of the activation
bias Bj , which represents baseline activity of the neuron, and Gaussian noise determined
by µ. Each time step t represents 1 ms of activity.
The parameters ρ, Ej , Bj and µ are constant in this model. The values used in our
simulations (figure 2a) were derived from previous work (Cooper et al., 2013b). All weight
matrix entries are initialized to 1 before training commences. Additionally, the network
will first be subjected to a burn-in phase. Here, the weight matrix is held constant and
the neurons are allowed to reach a certain steady-state activity.
3.3
Modeling action execution and observation
Keysers and Gazzola (2014) propose that mirror neuron activity emerges due to synaptic
changes that result from simultaneous action execution and reafference (the observation
of self-performed actions). This implies that the simulation of these mechanics should
consist of two separate phases. In the first phase, the training phase, action execution
and reafference are simulated. During this phase, the weights are being updated. The
second phase, the testing phase, consists of the mere observation of actions. If the correct
weight changes have been made during the training phase, the PM neurons should start
behaving as mirror neurons. The remainder of this section explains how the training and
testing phase have been simulated.
In the artificial neural network, PM activity represents action execution whereas STS
activity represents action observation. In essence, the action observation of self-performed
actions is a delayed version of the action execution signal. It takes approximately 100
ms for a signal in the PM to lead to an actual action execution. After another 100 ms
10
t (ms)
0
300
600
PM-A
[1,
[0,
[0,
PM-B
0,
1,
0,
PM-C
0]
0]
1]
t (ms)
200
500
700
(a) An example of an excitation sequence of
PM neurons during the action phase. If 1, Ej
is added to the input Ij (see formula 8). If
0, nothing is added. Note that the external
excitation persists until a transition is made to
a new action phase, which happens after 300
ms.
STS-A
[1,
[0,
[0,
STS-B
0,
1,
0,
STS-C
0]
0]
1]
(b) An example of an excitation sequence of the
STS neurons. Note that this is simply a delayed
version of the action sequence in c, with a delay
of 200 ms.
Figure 3: Examples of trials of the STS and PM during the training phase
the action observation reaches the STS. This gives a total delay of 200 ms.
The transitions from one action phase to the next are coordinated by a motor planner,
which chooses the next action based on a probability distribution that is encoded in a
transition matrix. A simple example of such a transition matrix is shown in figure 2b. It
encodes two actions, a → c → e and b → d → f . The first sequence can be derived from
the transition matrix since the probabilities to go from both a to c (t[0, 1] = 1) and c
to e (t[1, 2] = 1) are 1. It is assumed that the probability of each entire action sequence
is the same. Therefore, each last action phase in a sequence has a probability of n1 to
go to the first action phase of a new sequence, with n being the total number of action
sequences.
Initially, the motor planner will choose one of the first action phases as a starting
point. Once an action phase is selected, the PM neuron that encodes for this particular
action is externally stimulated (value 1 in figure 3a). The complete list of the excitations
at one particular time step is referred to as a trial. It is assumed that the length of one
action phase is 300 ms, similar as in Keysers and Gazzola (2014). After 300 ms, the next
action phase is chosen at random based on the probabilities encoded in the transition
matrix, which results in a new trial. Action observation entails the same trial, only with
a delay of 200 ms (figure 3b). During the training phase, both the PM and STS trials
are active. During the testing phase, only the STS neurons are excited.
3.4
Learning rules
During the training phase, the weight vector w will be updated according to a certain
neuron rule. Each of the Hebbian learning rules presented in chapter 2 has been implemented. However, since Oja’s rule does not naturally handle the baseline activity well,
we propose an alternative to Oja’s rule. The weight updates are the same as in the
original Oja’s rule (formula 6). However, a threshold is imposed on all the activities of
the neurons:
11
(
ak , if ak ≥θ
ak =
0, otherwise
(9)
This threshold prevents the weights to change from unsubstantial activities. Note
that the addition of the threshold cannot make Oja’s rule unbounded, since the bound
of the original rule only depends on α. Moreover, because the threshold introduces zero
correlation for certain periods of time, the weights will actually be lower compared to
weights imposed by the original learning rule.
3.5
Summary
In this chapter the computational model of the artifical network that models connections
between the premotor cortex (PM) and the superior temporal sulcus (STS) has been
presented. The general structure of connectivity in the network is a result of three
simplifying assumptions. First, the intermediate step of the PM/PFG has been omitted.
Second, each neuron only represents one action phase. Consequently, the neural network
consists of two equally sized layers. Third, the inhibitory P M → ST S connections only
exist between PM and STS neurons that encode the same action. Additionally, the
formula to calculate the activity of a single neuron has been presented, together with
an elaboration on how a transition matrix can be used to simulate action execution and
observation. Finally, a variation of Oja’s rule that uses a threshold has been introduced.
12
4
Analysis procedure
In this chapter an analysis procedure that can detect mirror neuron behavior is introduced. In contrast to the work of Cooper et al. (2013b), the simulation presented here
is not a simulation of one particular experiment. Therefore, it cannot be compared with
experimental data in a supervised manner. This requires designing new ways of investigating a quantification of mirror neuron behavior. First, it will be discussed what mirror
neuron behavior would visually look like in the simulation. Subsequently, a procedure
for calculating a quantitative measurement for the level of mirror neuron behavior in a
sequence of neuron activities is proposed.
4.1
A visualization of mirror neuron behavior
Before an evaluation metric can be designed, it must be clear what is considered to be
’mirror-like behavior’. Figure 4a shows activity that we deem to be typical for mirror
neuron behavior. The STS has been excited externally in a particular sequence and the
PM mimics the behavior of the visual neurons. Figure 4b shows predictive mirror neuron
behavior. The activity of STS-A induces activity in PM-B, which is the action that is
the most likely to follow next. Figure 4c shows a combination of the two, in which the
activity of STS-A induces activity of PM-B. In contrast to figure 4b, this activity persists
until STS-B has stopped to be active. The combined activity can explain both normal
and predictive mirror neuron behavior. Note that these kind of behaviors are expected
to emerge only during the testing phase. During the training phase, the PM neurons will
not be caused by STS activity but by external excitations.
4.2
Analysis procedure
A procedure to measure mirror neuron behavior was designed for the simulations presented in this work. All three types of mirror neuron behavior follow approximately the
same sequence of spikes (except that predictive mirror neurons surpass the first action
of a sequence). Therefore, the analysis procedure consists of the creation of a transition
matrix that describes sequences of activity in both the PM and STS cortices. By comparing the transition matrices, it can be determined to what extent the PM and STS
activity is similar, and thus, to what extent mirror neuron behavior has emerged.
The following procedure was used to calculate a transition matrix from recorded
activity. First, a low-pass filter is applied to the activity signals of all neurons (figure 4a)
to smoothen the signal (figure 4b). Then, significant peaks are substracted by setting a
threshold equal to the mean plus one standard deviation (figure 4c). Subsequently, for
each peak the neuron whose peak is the closest follow-up is determined. For a peak at
time t this is done as follows. First, the thresholded peaks of all neurons from time t to
t+3000 are collected (figure 4d). If a signal contains multiple peaks, all but the first peak
are removed from the signal. Then, a cross correlation analysis is performed between all
pairs of the original signal and the other signals. From this the delays that have the
highst cross correlation can be determined (figure 4e). If the delay of a certain neuron is
13
(a) Normal mirror neuron behavior. The activity of the PM
neurons mirrors exactly the activation of the STS.
(b) Predictive mirror neuron
behavior. The activity of PM
neurons is the same as the activity of STS neurons that will
be active next in the action sequence. For example, the green
PM neuron, which usually precedes a yellow neuron, responds
to the yellow STS neuron.
(c) Combination of normal and
predictive mirror neuron behavior. PM neurons respond
both to the same action phase
and to the action phase that
usually precedes the action
phase they encode.
Figure 4: Examples of three types of mirror neuron-like behavior
very low (−25 to 25 ms) the neuron considered to be simultaneous with the peak that is
under investigation. The neuron i which signal has the lowest positive delay di is selected
as the closest successor. In the transition matrix, tj,i is increased by one (figure 4f). If
the delay dn of another neuron n is very close to the closest delay (dn − di < 50 ms)
it is also selected as the closest successor. In this case, the total probability is evenly
distributed over the transition matrix. As a final step, the rows in the transition matrix
are normalized.
The above procedure can be applied both to STS and PM activity which results in
two matrices, tsts and tpm . In order to determine the similarity s of the two matrices,
one can calculate the Frobenius norm of their difference matrix, which gives the error :
n X
n
p X
pm 2
= (
(tsts
j,i − tj,i ) )
(10)
i=1 j=1
A low error would indicate a high level of mirror neuron-like behavior. To determine
whether or not a low retrieved error is significant, it will be compared to the distribution
14
(a) The recorded activity in the neurons
(b) Activity after application of low-pass
filter
(c) All activities lower than the threshold
are set to zero
(d) Activities within 3000 ms after the
start of the first peak of neuron B
A
A
B
C
D
B
C
D
+1
(f) The transition matrix is updated for
the transition from neuron B to the neuron
C, which has the lowest positive delay.
(e) Delays are determined by calculating
the maximal cross correlation.
Figure 5: A visualization of the procedure used to create a transition matrix from recorded activity.
Step D, E and F are repeated for all peaks detected in step A B and C. Afterwards, the rows of the
transition matrix are normalized.
15
of random errors. The distribution of errors is approximated by calculating the difference
between a set of uniformly at random generated stochastic matrices and the STS matrix.
If the retrieved error is lower than 2% of the errors determined by these random matrices,
it will be deemed significant. In the case of predictive mirror neuron behavior (figure 3b),
the transition from one action sequence to the next will always surpass the first action.
Therefore, this methodology works best for normal and combined mirror behavior (figure
3a and 3c). Note that the method presented here cannot distinguish between the different
types of mirror neuron behavior.
4.3
Summary
In this chapter a quantitative measure for the detection of mirror neuron behavior has
been presented. First, three types of mirror neuron behavior (normal, predictive and
combined) have been identified. Second, the analysis procedure to detect mirror neuron
behavior is described. This procedure consists of the generation of a transition matrix of
both the STS and PM neurons. Afterwards, these transition matrices can be compared
using the Frobenius norm of their difference matrix to obtain the error . This analysis
procedure can detect all presented types (normal, predictive and combined) of mirror
neuron behavior, though it will generally return a higher error for predictive mirror
neuron behavior.
16
5
Results
In this chapter two types of results will be discussed. First, a comparative analysis of the
different learning rules that shows how well each can account for mirror neuron behavior
is presented . Second, a parameter space analysis of the thresholded Oja’s rule is used
to show to what extent different parameter settings affect the level of mirror neuron
behavior that can be observed.
5.1
Comparison of learning rules
Figure 6 shows the mean and standard deviation of the error of the transition matrices
for 10 simulations of basic Hebb, covariance rule, BCM, Oja’s rule and the thresholded
Oja’s rule after a training phase of 250.000 ms for one action sequence consisting of four
action phases. Additionally, a hinton graph of the weight matrices is presented in figure
7. Figure 8 shows the recorded activities in the testing phase for all rules.
As visualized in figure 8a, applying the basic Hebb rule does not impose mirror
neuron behavior. The weight matrix (figure 7a) shows that the weights for all neurons
are approximately the same, which explains why the activities in figure 8b are also very
similar. This effect is caused by the baseline activity that is present for all neurons, which
influences the weight changes imposed by Hebb’s rule.
Figure 6: The average and standard deviation of the error of 10 simulations of the five learning rules. Each simulation consisted of 250.000
ms in both the training and testing phase, for a single action sequence of
4 action phases. The blue line indicates the value under which the norm
is deemed to be significant (< 2% of 10.000.000 randomly generated matrices).
The covariance rule, shown in figure 8b, naturally takes the baseline activity into
account. The weight matrix is more diverse than that of the basic Hebb rule (figure
7b). This causes the average error to be less high than that of Basic Hebb, but still not
17
(a) Basic hebb (α = 0.044),
m = 2.7
(d) Oja’s rule (θ = 0.5, αe =
10−1 ,αi = 1−1 ), m = 3.59
(b) Covariance (α = 0.044),
m = 4.8
(c) BCM (θinit = 0), m = 20.3
(e) Oja’s rule (θ = 0.0, αe =
10−1 , αi = 1−1 ), m = 3.79
Figure 7: A hinton graph that shows the relative differences of the weights of the weight matrix of
all rules. The lower left corner shows all weights from STS to PM neurons, whereas the upper right
corner shows all the PM to STS neurons. Blue and red squares denote inhibitory and excitatory neurons
respectively. They are normalized by the maximum value of the the weight matrix, which is denoted
beneath each figure by m.
significant. Moreover, as enpicted by the large variance of the covariance rule in figure
6, the results are unstable.
Figure 8c shows the activities after applying BCM, a modification on the covariance
rule which imposes a bound on the synaptic strengths. STS activity is completely suppressed during the training phase. Theoretically, BCM should impose bounds on the
weights due to competition between the synapses. However, in this case these bounds
are still too high which causes the suppression of STS activity during the training phase
due to strong inhibitory weights from PM neurons. As a consequence, BCM performs
approximately as well as basic hebb, as shown in figure 6.
An alternative to BCM is Oja’s rule. In contrast to BCM, the bounds of the weights
can be directly affected by the α parameter. To account for differences between the
bounds for inhibitory and excitatory neurons that occur in nature, different bounds were
used. Recall that α−1 is proportional to the bounds of |w|2 . Here, αinhibitory was set
to 1−1 and αexcitatory was set to 10−1 . Figure 7d shows that, if applied directly, all the
18
(a) Basic hebb, α = 0.044
(b) Covariance, α = 0.044
(c) BCM, θinit = 0
(d) Oja’s rule, θ = 0, αe = 10−1 , αi−1
Figure 8: A visualisation of the performance of all learning rules. Each
color denotes the same action phase.
The lines above the graph of STS
activity show the time each neuron
was externally excited. These are the
last 4000 ms of a simulation with a
training- and test phase of 250.000 ms
for four neurons that encode one action sequence (a-b-c-d).
19
(e) Oja’s rule, θ = 0.5, αe = 10−1 , αi−1
(a) θ = 0.3
(b) θ = 0.5.
(c) θ = 0.7
Figure 9: Results of quantitative parameter space analysis. Different images depict the results for
different thresholds θ. The x-axis shows the bounds for inhibitory neurons (αi−1 ), whereas the y-axis
shows the bounds for excitatory neurons (αe−1 ). The lower the error (denoted by a color range from dark
red to light yellow) the better the performance was for a parameter setting. The areas delineated by the
blue line are all retrieved errors that are lower than the errors of 98% of 10 million randomly generated
matrices. If the model is set to these parameter settings, mirror neuron behavior emerges.
PM neurons to behave similarly. Since no mirror behavior occurs, the average error is
high. To account for the baseline activity of the neurons, the learning rule was altered
as proposed in section 2.3, and the threshold θ was set to 0.5. As shown in figure 8e, this
results in a combination of normal and predictive mirror neuron behavior. This results
in an error that is approximately 10 times lower than the other learning rules (figure 6).
5.2
Parameter space analysis of Oja’s rule
The thresholded Oja’s rule has three parameters, namely αinhibitory , αexcitatory and
threshold θ. A parameter space analysis was performed to see to what extent different parameter settings change the outcome of applying Oja’s rule. The thresholded
Oja’s rule was applied in 500 simulations with different parameter settings. Both the
training phase and testing phase consisted of 1.000.000 ms. 19 action phases were modeled, divided over 4 action sequences of length 5, giving a total network size of 38 nodes.
Since modeling 4 action sequences of length 5 requires 20 nodes, and we only model 19,
20
two sequences partly overlap. The procedure as described in chapter 4 was applied to
the activities in the testing phase to determine the error of the difference of the PM
and STS transition matrices. Additionally, the error of 10.000.000 random matrices was
calculated. These were used to approximate the p-values at which a low error is deemed
to be significant.
The results of the parameter space analysis are shown in figure 9. The axes show
the inverse of α since this is approximately equal to the bounds of the weight matrix.
Each figure depicts the retrieved error for a different threshold θ value. The blue line
indicates the section in which the error is lower than 2% of 10.000.000 randomly obtained
errors. It shows that if the threshold is relatively low (0.3, figure 9a), Oja’s rule does
not impose mirror neuron behavior at all. Figure 9b shows that a threshold of 0.5 does
imposes mirror behavior if, in general, the bound of excitatory connections is higher than
the bound of inhibitory connections. If both parameters are higher than approximately
15, no mirror behavior emerges. If the threshold is very high (0.7, figure 9c), neuron
mirror behavior is imposed, but to lesser extent than when the threshold is 0.5. Here,
the excitatory bounds much exceed the inhibitory bounds to a larger extent for mirror
neuron behavior to emerge.
5.3
Summary
Five different learning rules have been implemented. An analysis of simulations of one
simple action sequence shows that, of the rules considered, only the thresholded Oja’s
rule is able to produce mirror neuron behavior. More precisely, it can account for both
normal and predictive mirror neuron behavior. Furthermore, a parameter space analysis
shows that if the threshold is set too low, no mirror neuron behavior is imposed. A
threshold of 0.5 shows a large field of values in which mirror neuron behavior is imposed.
In general, the bounds for inhibitory neurons have to be strictly smaller than those of
excitatory neurons to impose mirror neuron behavior. If the threshold becomes too high,
excitatory bounds much be about twice as high as inhibitory bounds.
21
6
Conclusion
In previous work it has been proposed that Hebbian learning could be responsible for the
ontogeny of predictive mirror neurons (Keysers and Gazzola, 2014). Here, we have shown
that a variation of Oja’s rule (an implementation of Hebbian learning) is sufficient to
explain the emergence of mirror neurons. An artificial neural network that simulates the
interactions between the premotor cortex (PM) and the superior temporal sulcus (STS)
has been created. Different implementations of Hebbian learning have been compared in
performance on a simple action sequence. Additionally, a parameter space analysis has
been performed on the proposed thresholded Oja’s rule to determine the sensitivity of
the parameters on its performance.
We have identified that from the learning rules considered, only the thresholded Oja’s
rule is sufficient to impose mirror neuron behavior. The other learning rules (basic Hebb,
covariance, BCM and the original Oja’s rule) are subject to at least one of the following
three limitations. First, both the covariance rule and the basic Hebb rule are unbounded,
which makes them biologically implausible. Second, all learning rules except for the
covariance rule are sensitive to baseline activity. These changes overshadow associations
between significant STS- and PM activity. This prevents mirror neuron behavior from
emerging. Third, if the inhibitory neurons are not bounded strong enough, as in BCM and
the original Oja’s rule, PM inhibition causes a complete suppression of STS activations.
This prevents any association between the PM and STS to occur. A variation of Oja’s
rule that uses a threshold and different bounds for inhibitory and excitatory synapses
has been shown to surpass each of these limitations.
A parameter space analysis has indicated that the parameters of the proposed thresholded Oja’s rule must adhere to two constraints for mirror neuron behavior to emerge.
First, the threshold value must be higher than the baseline activity, but lower than the
highest peaks measured. If the threshold is too low, small variances in baseline activity
impose weight changes that prevent the correct associations form being made. Choosing
the proper threshold value can be viewed as intrinsic homeostatic plasticity. This refers
to the capacity of neurons to regulate their own excitability relative to network activity
(Turrigiano and Nelson, 2004). Second, mirror neuron behavior can only be imposed if
the bounds for the excitatory neurons are higher than those of the inhibitory neurons.
This is necessary because otherwise inihbitory PM neurons will completely suppress STS
activity which prevents associations from being able to be learned.
In conclusion, we have shown that a thresholded Oja’s rule is sufficient to account for
the emergence of mirror neurons in an artificial neural network that simulates interactions
between the PM and STS cortices. Therefore, this work provides positive evidence for
the proposal that Hebbian learning is sufficient to account for the emergence of mirror
neurons. In the broader sense, this work promotes the idea that statistical sensory motor
contingencies suffice in the explanation for the ontogeny of mirror neurons. Therefore, it
can be regarded as positive evidence for the association hypothesis.
22
7
7.1
Discussion and future work
Limitations of this study
One of the assumptions underlying the model presented here is that the STS encodes
observations of actions of others similarly as the observations of self-performed actions.
While it has been shown that the area STS responds to the sight and sound of actions,
it has not been shown indisputable that the encoding for reafference (sensory input from
one’s own action) and others’ actions is the same. Further research should indicate
whether or not this assumption is valid.
Another limitation of the work presented here is that the input data to the ANN is
simplified compared to action execution and observation in the real world. While the time
delays that were used are comparable with what is observed in the action observation
and execution of primates Keysers and Gazzola (2014), the real world is more noisy and
less predictable than the action sequences that were used. For example, once an action
is chosen in the simulation, the entire action is always executed without interventions.
Future research should indicate whether or not the results presented here still hold when
the network is exposed to more complicated and less predictable action sequences.
7.2
Future work
It has been argued that that Rescorla-Wagner, and not Hebbian learning, account for
the emergence of mirror neurons (Cooper et al., 2013b). However, Oja’s rule was not
considered by Cooper et al. (2013b). In this work we have shown that a modified version
of Oja’s rule which is genuinely unsupervised is sufficient for the emergence of mirror
neurons. In further research, Oja’s rule should be applied to the experimental data
presented by Cooper et al. (2013b) to determine if it is sufficient to replicate the results
of the experimental study of Cook et al. (2010).
In the current implementation of the thresholded Oja’s rule, the threshold was determined manually. An extension of this work could be to model homeostatic plasticity, in
which the threshold value is dynamically set.
Keysers and Gazzola (2014) argued that the inhibitory connections from the PM
to the STS could behave as a ’prediction error’. If the PM correctly predicts the next
action, it inhibits STS activity which would indicate that no audiovisual confirmation is
required. However, as shown in this work, this can easily lead to complete surpression of
STS activity. A chain of activities in the PM cortex should then be imposed without an
exchange of information with the STS. In the model presented here, such a sequence is
not possible because there are no lateral connections modelled. In an extension of this
work one could add lateral connections to the model. Ideally, the transition matrix would
be completely omitted and replaced by interactions of neurons via lateral connections.
As a more realistic extension of this work, one would need to model the ’mirror
circuit’. If each cortical area is considered a node of a graph, then these nodes would
have to be simulated using spiking neuronal models with at least three cortical layers;
one for input, one for prediction and one for prediction error. These models will consist of
23
approximately 1000 neurons per layer at every node and will suffice to represent a much
richer repertoire of actions. Detailed work on the anatomy of the connections between
these nodes will inform the model on connection probabilities between layers within and
across nodes. A model of this nature would be instrumental in determining the origin of
mirror neurons.
24
References
Bienenstock, E. L., Cooper, L. N., and Munro, P. W. (1982). Theory for the development
of neuron selectivity: orientation specificity and binocular interaction in visual cortex.
The Journal of Neuroscience, 2(1):32–48.
Catmur, C. (2013). Sensorimotor learning and the ontogeny of the mirror neuron system.
Neuroscience letters, 540:21–27.
Cook, R., Press, C., Dickinson, A., and Heyes, C. (2010). Acquisition of automatic
imitation is sensitive to sensorimotor contingency. Journal of Experimental Psychology:
Human Perception and Performance, 36(4):840.
Cooper, R. P., Catmur, C., and Heyes, C. (2013a). Are automatic imitation and spatial
compatibility mediated by different processes? Cognitive Science, 37(4):605–630.
Cooper, R. P., Cook, R., Dickinson, A., and Heyes, C. M. (2013b). Associative (not
hebbian) learning and the mirror neuron system. Neuroscience letters, 540:28–36.
Dayan, P. and Abbott, L. F. (2001). Theoretical neuroscience. Cambridge, MA: MIT
Press.
Di Pellegrino, G., Fadiga, L., Fogassi, L., Gallese, V., and Rizzolatti, G. (1992). Understanding motor events: a neurophysiological study. Experimental brain research,
91(1):176–180.
Giudice, M. D., Manera, V., and Keysers, C. (2009). Programmed to learn? the ontogeny
of mirror neurons. Developmental science, 12(2):350–363.
Hebb, D. O. (1949). The organization of behavior: A neuropsychological approach. John
Wiley & Sons.
Heyes, C. (2010). Where do mirror neurons come from? Neuroscience Biobehavioral
Reviews, 34(4):575 – 583.
Hietanen, J. and Perrett, D. (1993). Motion sensitive cells in the macaque superior
temporal polysensory area. Experimental brain research, 93(1):117–128.
Kandel, E. R., Schwartz, J. H., Jessell, T. M., et al. (2000). Principles of neural science,
volume 4. McGraw-Hill New York.
Keysers, C. and Gazzola, V. (2014). Hebbian learning and predictive mirror neurons for
actions, sensations and emotions. Philosophical Transactions of the Royal Society B:
Biological Sciences, 369(1644):20130175.
Oja, E. (1982). Simplified neuron model as a principal component analyzer. Journal of
mathematical biology, 15(3):267–273.
Terry, W. S. (2006). Learning and memory: Basic principles, processes, and procedures.
Pearson/Allyn and Bacon Boston, MA.
Turrigiano, G. G. and Nelson, S. B. (2004). Homeostatic plasticity in the developing
nervous system. Nature Reviews Neuroscience, 5(2):97–107.