Download Lecture 6

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Connectome wikipedia , lookup

Neuroeconomics wikipedia , lookup

Multielectrode array wikipedia , lookup

Catastrophic interference wikipedia , lookup

Donald O. Hebb wikipedia , lookup

Environmental enrichment wikipedia , lookup

Central pattern generator wikipedia , lookup

Artificial general intelligence wikipedia , lookup

Artificial neural network wikipedia , lookup

Apical dendrite wikipedia , lookup

Caridoid escape reaction wikipedia , lookup

Neural oscillation wikipedia , lookup

Axon guidance wikipedia , lookup

Clinical neurochemistry wikipedia , lookup

Eyeblink conditioning wikipedia , lookup

Premovement neuronal activity wikipedia , lookup

Mirror neuron wikipedia , lookup

Convolutional neural network wikipedia , lookup

Molecular neuroscience wikipedia , lookup

Optogenetics wikipedia , lookup

Recurrent neural network wikipedia , lookup

Synaptogenesis wikipedia , lookup

Pre-Bötzinger complex wikipedia , lookup

Single-unit recording wikipedia , lookup

Stimulus (physiology) wikipedia , lookup

Neurotransmitter wikipedia , lookup

Metastability in the brain wikipedia , lookup

Axon wikipedia , lookup

Development of the nervous system wikipedia , lookup

Neuroanatomy wikipedia , lookup

Activity-dependent plasticity wikipedia , lookup

Neural modeling fields wikipedia , lookup

Holonomic brain theory wikipedia , lookup

Neuropsychopharmacology wikipedia , lookup

Nonsynaptic plasticity wikipedia , lookup

Feature detection (nervous system) wikipedia , lookup

Channelrhodopsin wikipedia , lookup

Chemical synapse wikipedia , lookup

Types of artificial neural networks wikipedia , lookup

Neural coding wikipedia , lookup

Efficient coding hypothesis wikipedia , lookup

Synaptic gating wikipedia , lookup

Biological neuron model wikipedia , lookup

Nervous system network models wikipedia , lookup

Transcript
Neural Computation
Spiking Neuron Networks
Sources
• “Handbook of Natural Computing,” Editors
Grzegorz Rosenberg, Thomas Back and Joost N.
Kok, Springer 2014.
• “Deep Learning Tutorial”, Yann LeCun, Marc'Aurelio
Ranzato, ICML Atlanta, 2013.
Spiking Neuron Networks (SNNs)
=
Pulse-Coupled Neural Networks (PCNNs)
• Inspired by natural computation in the brain
• Accurate modeling of synaptic interactions between neurons
• Taking into account the time of spike firing
 Fast adaptation & exponential capacity to memorize
• How can we design efficient learning rules that take advantage of this
computational model?
• Can we already show in theory for simple neuron models whether SNNs
can do more than classical computational models?
• Do we need very accurate neuron models to improve learning?
• What design principles (other than the neuron model) are crucial for
efficient learning?
Neocortex: Cortex means bark in Greek, it lies as a bark over the rest of
the brain with a surface of 2000cm^2. At the back is the occipital area
important for visual processing (the later takes up 40% of the brain) 
very high visual resolution (& capability for associative and therefore
creative thinking?). Frontal area important for short term working
memory, and planning & integration of thoughts.
Cerebellum: contains 50% of all neuron, involved in precise timing functions
Hippocampus: assumed to work as an association area which processes and
associates information from different sources (after removal no new memories
are stored)
• Glial cells provide support to neurons: suck up the spilt-over neuro-transmitters
or provide myelin sheets around axons or neurons
• Neuron cells: 10^11 in our brains
Neurons receive input through
synapses on its dendrites; dendritic
trees often receive more than 10,000
synapses
Neurons communicate using spikes,
i.e. brief 1ms excursions from the
neuron’s membrane voltage,
generated in the axon-hillock from
where it propagates along the axon
to the (10,000) axon terminals
Few hard rules: some release both
excitatory and inhibitory
neurotransmitters, some are without
axons, not all neurons spike etc.
Synaptic plasticity: adjustments,
formation, removal synapses
Rate vs Spike coding:
• Poisson-like rate coding cannot explain the rapid information
transmission for sensory processing
• N neurons:
• Counting spikes  log(N+1) bits
• Binary code  O(N) bits
• Timing code  Means alphabet is not binary but is
{0,1,…,T-1}: gives O(N*log(T)) bits
• Rank order coding  order of spikes matters, the time
axis becomes somewhat part of the length of the code:
log(N!) = O(N*log(N)) bits
• Etc.
1mm^2 surface contains about 10^6 cells
Connectivity in the whole cortex is about
10^6km of fibers
Some neuron models:
The above models is able to accurately predict all 20 firing behaviors, just as the more
complex HH model.
• HH: 1200 floating point computations
• Izhikevich: 13 floating point computations
• Other models lose out on predictability
Supervised learning: teach network explicitly to perform task (e.g.
gradient descent algorithm)
Unsupervised learning: learn interesting features “on its own”
• Deep learning from 2006 onwards! Outperforms SVMs!!
Reservoir computing, SVM, other stuff presented in chapter seems to be
overshadowed by deep learning
Restricted Boltzmann Machines
(Smolensky ,1986, called them “harmoniums”)
• We restrict the connectivity to make learning
easier.
• Only one layer of hidden units.
hidden
• We will deal with more layers later
• No connections between hidden units.
• In an RBM, the hidden units are conditionally
independent given the visible states.
• So we can quickly get an unbiased sample
from the posterior distribution when given
a data-vector.
• This is a big advantage over directed belief
nets
j
i
visible
(Deep Belief Network)
(Restricted Boltzman Machine)
Principles:
• Neuron model: sigmodial seems good enough
• Topology: a flat 2D layer with a small vertical (deep) third dimension (as in our
brains)
• Neurons should be capable of adjusting their weighing factors (a form of
plasticity)