Download Introduction to Neural Networks

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Brain–computer interface wikipedia , lookup

Synaptogenesis wikipedia , lookup

Brain wikipedia , lookup

Neuroinformatics wikipedia , lookup

Aging brain wikipedia , lookup

Neuropsychology wikipedia , lookup

Brain Rules wikipedia , lookup

History of neuroimaging wikipedia , lookup

Functional magnetic resonance imaging wikipedia , lookup

Haemodynamic response wikipedia , lookup

Binding problem wikipedia , lookup

Artificial intelligence wikipedia , lookup

Neuroplasticity wikipedia , lookup

Stimulus (physiology) wikipedia , lookup

Donald O. Hebb wikipedia , lookup

Cognitive neuroscience wikipedia , lookup

Premovement neuronal activity wikipedia , lookup

Multielectrode array wikipedia , lookup

Molecular neuroscience wikipedia , lookup

Mirror neuron wikipedia , lookup

Neurophilosophy wikipedia , lookup

Clinical neurochemistry wikipedia , lookup

Nonsynaptic plasticity wikipedia , lookup

Neuroesthetics wikipedia , lookup

Artificial general intelligence wikipedia , lookup

Feature detection (nervous system) wikipedia , lookup

Neural oscillation wikipedia , lookup

Activity-dependent plasticity wikipedia , lookup

Single-unit recording wikipedia , lookup

Connectome wikipedia , lookup

Mind uploading wikipedia , lookup

Pre-Bötzinger complex wikipedia , lookup

Neuroeconomics wikipedia , lookup

Neural correlates of consciousness wikipedia , lookup

Neural coding wikipedia , lookup

Neuroanatomy wikipedia , lookup

Holonomic brain theory wikipedia , lookup

Biological neuron model wikipedia , lookup

Central pattern generator wikipedia , lookup

Neural modeling fields wikipedia , lookup

Optogenetics wikipedia , lookup

Catastrophic interference wikipedia , lookup

Neural engineering wikipedia , lookup

Artificial neural network wikipedia , lookup

Neuropsychopharmacology wikipedia , lookup

Channelrhodopsin wikipedia , lookup

Synaptic gating wikipedia , lookup

Neural binding wikipedia , lookup

Development of the nervous system wikipedia , lookup

Convolutional neural network wikipedia , lookup

Metastability in the brain wikipedia , lookup

Nervous system network models wikipedia , lookup

Recurrent neural network wikipedia , lookup

Types of artificial neural networks wikipedia , lookup

Transcript
Introduction to Neural
Networks
Neural Networks in the Brain
• Human brain “computes” in an entirely different
way from conventional digital computers.
• The brain is highly complex, nonlinear, and
parallel.
• Organization of neurons to perform tasks much
faster than computers. (Typical time taken in
visual recognition tasks is 100–200 ms.)
• Key features of the biological brain: experience
shapes the wiring through plasticity, and hence
learning becomes the central issue in neural
networks.
Neural Networks as an Adaptive
Machine
• A neural network is a massively parallel
distributed processor made up of simple
processing units, which has a natural propensity
for storing experimental knowledge and making
it available for use.
• Neural networks resemble the brain:
– Knowledge is acquired from the environment
through a learning process.
– Iner neuron connection strengths, known as
synaptic weights, are used to store the acquired
knowledge.
• Procedure used for learning: learning algorithm.
Weights, or even the topology can be adjusted.
Benefits of Neural Networks
• Nonlinearity: nonlinear components, distributed
nonlinearity
• Input-output mapping: supervised learning,
nonparametric statistical inference (model-free
estimation, no prior assumptions)
• Adaptivity: either retain or adapt. Can deal with nonstationary environments. Must overcome stabilityplasticity dilemma.
• Evidential response: decision plus confidence of the
decision can be provided.
• Contextual information: Every neuron in the network
potentially influences every other neuron, so
contextual information is dealt with naturally
Benefits of Neural Networks
• Fault tolerance: performance degrades gracefully.
• VLSI implementability: network of simple
components.
• Uniformity of analysis and design: common
components (neurons), sharability of theories
and learning algorithms, and seamless
integration based on modularity.
• Neurobiological analogy: Neural nets motivated
by neurobiology, and neurobiology also turning
to neural networks for insights and tools.
Human Brain
• Stimulus → Receptors ⇔ Neural Net ⇔
Effectors → Response: Arbib (1987)
• Neurons are slow: 10−3 s per operation,
compared to 10−9 s of modern CPUs.
• Huge number of neurons and on
nections: 1010 neurons, 6 × 1013
connections in human brain.
• Highly energy efficient: 10−16J in the
brain vs. 10−6 J in modern computers
The Biological Neuron
•
•
•
Each neuron is a cell that uses biochemical reactions to receive, process
and transmit information.
Each terminal button is connected to other neurons across a small gap
called a synapse.
A neuron's dendritic tree is connected to a thousand neighbouring
neurons. When one of those neurons fire, a positive or negative charge is
received by one of the dendrites. The strengths of all the received
charges are added together through the processes of spatial and
temporal summation
Models of Neurons
Bias
x1
w
b
1
Input
values
x2
w2

xm

wm
weights

Induced
Field
v
Summing
function
Activation
function
 ()
Output
y
Activation Functions
a if v  c
b if v  c
 (v )  
a if v  c

 (v )  b if v  d
a  (( v  c )( b  a ) /( d  c )) oth

 (v )  z 
 (v ) 
1
1  exp(  xv  y )
 1  v   2 
1
exp   
 

2 
 2   
Stochastic Model
• Instead of deterministic activation, stochastic
activation can be done.
• State of neuron 𝑥 ∈ {+1,-1}
• Probability of firing 𝑃 𝑣
+1 with probability 𝑃(𝑣)
𝑥=
−1 with probability 1 − 𝑃(𝑣)
1
1+exp(−𝑣/𝑇)
• Typical choice of 𝑃 𝑣 =
where 𝑇 is a
pseudo temperature.
• When 𝑇 → 0, the neuron becomes deterministic.
• Think of computer simulation
Signal Flow Graph
•
•
•
•
•
Nodes and links
Links: synaptic links and activation links.
Incoming edges: summation
Outgoing edges: replication
Architectural graph simplifies the above and
abstracts out internal neuronal function.
Signal Flow Graph Example
• Turn the above into a signal-flow graph
Feedback
• Feedback gives dynamics (temporal aspect), and it
is found in almost every part of the nervous system
in every animal.
𝑥𝑗′ 𝑛 = 𝑥𝑗 𝑛 + 𝐵𝑦𝑘 (𝑛)
𝑦𝑘 𝑛 = 𝐴𝑥𝑗′ 𝑛
𝐴
𝑦𝑘 𝑛 =
𝑥𝑗 𝑛
1 − 𝐴𝐵
Feedback
𝑤
𝑦𝑘 𝑛 =
𝑥𝑗 𝑛 = 𝑤
−1
1 − 𝑤𝑧
𝑖𝑛𝑓
𝑖𝑛𝑓
𝑤 𝑙 𝑧 −𝑙 𝑥𝑗 𝑛
𝑙=0
𝑤 𝑙+1 𝑥𝑗 𝑛 − 𝑙
𝑦𝑘 𝑛 =
𝑙=0
𝑧 −1 : 1 − step delay operator
Feedback
• 𝑤 < 1:
– converge (infinite memory,
fading)
• 𝑤 =1
– linearly diverge
• 𝑤 >1
– exponentially diverge
Definition of Neural Networks
• An information processing system that has been
developed as a generalization of mathematical
models of human cognition or neurobiology,
based on the assumptions that
– Information processing occurs at many simple
elements called neurons.
– Signals are passed between neurons over
connection links.
– Each connection link has an associated weight,
which typically multiplies the signal transmitted.
– Each neuron applies an activation function (usually
non-linear) to its net input (sum of weighted input
signals) to determine its output signal.
Network Architectures
• The connectivity of a neural network is
intimately linked with the learning algorithm.
– Single-layer feedforward networks: one input layer,
one layer of computing units (output layer), acyclic
connections.
– Multilayer feedforward networks: one input layer,
one (or more)hidden layers, and one output layer.
Recurrent networks: feedback loop exists.
– Recurrent networks: feedback loop exists.
• Layers can be fully connected or partially
connected.
Similarity Measures
• Similar inputs from similar classes should produce similar
representations, leading to classification into the same
category.
– Reciprocal of Euclidean distance
1
𝑑 𝑥𝑖 ,𝑥𝑗
𝑥𝑖 = 𝑥𝑖1 , 𝑥𝑖2 , 𝑥𝑖3 , … , 𝑥𝑖𝑚
𝑑 𝑥𝑖 , 𝑥𝑗 = 𝑥𝑖 − 𝑥𝑗 =
𝑚
𝑘=1
𝑥𝑖𝑘 − 𝑥𝑗𝑘
– Dot product(inner product)
𝑥𝑖 , 𝑥𝑗 = 𝑥𝑖𝑡 𝑥𝑗 = 𝑚
𝑘=1 𝑥𝑖𝑘 𝑥𝑗𝑘 = 𝑥𝑖
•
𝑡
2 1/2
𝑥𝑗 cos𝜃𝑖𝑗
The two are related, when 𝑥𝑖 = 𝑥𝑗 =1
𝑑 2 𝑥𝑖 , 𝑥𝑗 =
𝑚
𝑘=1
2
𝑥𝑖𝑘 − 𝑥𝑗𝑘 = 𝑥𝑖 − 𝑥𝑗
𝑡
𝑥𝑖 − 𝑥𝑗 = 2 − 2𝑥𝑖𝑡 𝑥𝑗
Similarity Measure
• When two vectors 𝑥𝑖 and 𝑥𝑗 are drawn
from two distributions:
– Mean vector: 𝜇 = 𝐸 𝑥𝑖
– Mahalanobis distance:
2
𝑑𝑖𝑗
=
𝑥𝑖 − 𝑥𝑗
𝑡 −1
Σ
𝑥𝑖 − 𝑥𝑗
– Covariance matrix
Σ = 𝐸 𝑥𝑖 − 𝜇 𝑥𝑖 − 𝜇
𝑡