Download 9-Lecture1(updated)

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Human brain wikipedia , lookup

Aging brain wikipedia , lookup

Neurotransmitter wikipedia , lookup

Premovement neuronal activity wikipedia , lookup

Neuroinformatics wikipedia , lookup

Mirror neuron wikipedia , lookup

Haemodynamic response wikipedia , lookup

Subventricular zone wikipedia , lookup

Brain Rules wikipedia , lookup

Activity-dependent plasticity wikipedia , lookup

Brain wikipedia , lookup

History of neuroimaging wikipedia , lookup

Nonsynaptic plasticity wikipedia , lookup

Multielectrode array wikipedia , lookup

Neural modeling fields wikipedia , lookup

Donald O. Hebb wikipedia , lookup

Neuroplasticity wikipedia , lookup

Neuroesthetics wikipedia , lookup

Neuropsychology wikipedia , lookup

Clinical neurochemistry wikipedia , lookup

Neurophilosophy wikipedia , lookup

Neural oscillation wikipedia , lookup

Cognitive neuroscience wikipedia , lookup

Molecular neuroscience wikipedia , lookup

Feature detection (nervous system) wikipedia , lookup

Neural coding wikipedia , lookup

Axon wikipedia , lookup

Neuroeconomics wikipedia , lookup

Connectome wikipedia , lookup

Stimulus (physiology) wikipedia , lookup

Neural correlates of consciousness wikipedia , lookup

Synaptogenesis wikipedia , lookup

Single-unit recording wikipedia , lookup

Artificial general intelligence wikipedia , lookup

Circumventricular organs wikipedia , lookup

Mind uploading wikipedia , lookup

Central pattern generator wikipedia , lookup

Biological neuron model wikipedia , lookup

Optogenetics wikipedia , lookup

Artificial neural network wikipedia , lookup

Neural engineering wikipedia , lookup

Neuroanatomy wikipedia , lookup

Synaptic gating wikipedia , lookup

Holonomic brain theory wikipedia , lookup

Catastrophic interference wikipedia , lookup

Channelrhodopsin wikipedia , lookup

Neuropsychopharmacology wikipedia , lookup

Development of the nervous system wikipedia , lookup

Metastability in the brain wikipedia , lookup

Convolutional neural network wikipedia , lookup

Nervous system network models wikipedia , lookup

Types of artificial neural networks wikipedia , lookup

Recurrent neural network wikipedia , lookup

Transcript
Artificial Intelligence
Neural Networks
(Chapter 9)
Outline of this Chapter
•
•
•
•
•
•
Biological Neurons
Neural networks History
Artificial Neural Network
Perceptrons
Multilayer Neural Network
Applications of neural networks
Definition
Neural Network
A broad class of models that mimic functioning inside the human brain
There are various classes of NN models.
They are different from each other depending on
 Problem types
 Structure of the model
 Model building algorithm
For this discussion we are going to focus on
Feed-forward Back-propagation Neural Network
(used for Prediction and Classification problems)
Biological Neurons
• The brain is made up of neurons (nerve cells) which
have
– dendrites (inputs)
– a cell body (soma)
– an axon (outputs)
– synapse (connections between cells)
• Synapses can be excitatory (potential-increasing activity) or
inhibitory (potential-decreasing), and may change over time
• The synapse releases a chemical transmitter – the sum of which
can cause a threshold to be reached – causing the neuron to fire
(electrical impulse is sent down the axon )
Biology of a neuron
A bit of biology . . .
Most important functional unit in human brain – a class of cells called –
NEURON
Dendrites
Cell Body
Axon
Synapse
Hippocampal Neurons
Source: heart.cbl.utoronto.ca/ ~berj/projects.html
• Dendrites – Receive information
Schematic
• Cell Body – Process information
• Axon – Carries processed information to other neurons
• Synapse – Junction between Axon end and Dendrites of other Neurons
An Artificial Neuron
Dendrites
X1
X2
Xp
w1
w2
..
.
wp
Cell Body
Axon
Direction of flow of Information
I
f
V = f(I)
I = w1X1 + w2X2
+ w3X3 +… + wpXp
• Receives Inputs X1 X2 … Xp from other neurons or environment
• Inputs fed-in through connections with ‘weights’
• Total Input = Weighted sum of inputs from all sources
• Transfer function (Activation function) converts the input to output
• Output goes to other neurons or environment
Biological Neurons (cont.)
• When the inputs reach some threshold an action potential
(electrical pulse) is sent along the axon to the outputs.
• The pulse spreads out along the axon reaching synapse &
releasing transmitters into the bodies of other cells.
• Learning occurs as a result of the synapse’ plasticicity: They
exhibit long-term changes in connection strength.
• There are about 1011 neurons and about 1014 synapses in the
human brain(!)
• A neuron may connect to as many as 100,000 other neurons
Brain structure
• We still do not know exactly how the brain works.
e.g, born with about 100 billion neurons in our brain. Many die as we
progress through life, & are not replaced, but we continue to learn.
But we do know certain things about it.
• Different areas of the brain have different functions
– Some areas seem to have the same function in all humans (e.g., Broca’s
region- speech & language); the overall layout is generally consistent
– Some areas vary in their function; also, the lower-level structure and function
vary greatly
senses
emotions, reasoning,
planning, movement,
& parts of speech.
hearing, memory,
meaning, and language
vision & ability to
recognize objects
Brain structure (cont.)
• We don’t know how different functions are “assigned” or
acquired
– Partly the result of the physical layout / connection to inputs (sensors)
and outputs (effectors)
– Partly the result of experience (learning)
• We really don’t understand how this neural structure/ collection
of simple cells leads to action, consciousness and thought
• Artificial neural networks are not nearly as complex as the actual
brain structure
Comparing brains with computers
Computer
Computational units
Storage units
Cycle time
Bandwidth
Neuron updates/sec
•
•
•
•
•
•
Human Brain
1 CPU, 105 gates
109 Bits RAM, 1011 bits disk
1011 neurons
1011neurons, 1014 synapses
10-8 Sec
109 bits/sec
105
10-3 sec
1014 bits/sec
1014
They are more neurons in human brain than they are bits in computers
Human brain is evolving very slowly---computer memories are growing rapidly.
There are a lot more neurons than we can reasonably model in modern digital computers,
and they all fire in parallel
NN running on a serial computer requires 100 of cycles to decided if a single N will fire--in real brain, all Ns do this in a single step.
e.g. brain recognizes a face in less than a sec--- billion of cycles
Neural networks are designed to be massively parallel
The brain is effectively a billion times faster at what it does
Neural networks History
•
McCulloch & Pitts (1943) are generally recognised as the designers of the
first neural network
•
Many of their ideas still used today (e.g. a neuron has a threshold level
and once that level is reached the neuron fires is still the fundamental way
in which artificial neural networks operate)
•
Hebb (1949) developed the first learning rule (on the premise that if two
neurons were active at the same time the strength between them should
be increased).
•
During the 50’s and 60’s many researchers such as Minsky & Papert,
worked on the perceptron (NNModel)
•
1969 saw the death of neural network research for about 15 years
•
Only in the mid 80’s (Parker and LeCun) NN research revived.
Artificial Neural Network
•
(Artificial) Neural networks are made up of nodes/units connected by
links which have
– inputs edges, each link has a numeric weight
– outputs edges (with weights)
– an activation level (a function of the inputs)
The computation is split into 2 components:
1. Linear component, called input function (ini)-- computes the weighted
sum of the unit’s input values.
2. Non-linear component, called activation function (g)– transforms the
weighted sum into the final value that serves as the unit’s activation value:
ai = g(ini) = g( aj wj,i )
•
Some nodes are inputs (perception), some are outputs (action)
Modeling a Neuron
in i   j Wj , iaj
Each unit does a local computation based on inputs from its neighbours &
compute a new activation level – sends along each of its output links
aj: Activation value of unit j
wj,I: Weight on the link from unit j to unit i
inI: Weighted sum of inputs to unit i
aI: Activation value of unit i
g: Activation function.
Activation Functions
• Different models are obtained by using different mathematical functions for g.
• 3 common choices are:
threshold function
logistic function
Step(x)
=
1 if x >= 0, else 0
Sign(x)
=
+1 if x >= 0, else –1
Sigmoid(x) =
1/(1+e-x)
( in which we try to minimize the error by adjusting the weights of the network,
e : represents error degree)
1  represents the firing of a pulse down the axon, & 0 represents no firing.
t (threshold) represents the min total weighted input needed to cause the neuron to
ANN (Artificial Neural Network) – Feed-forward Network
A collection of neurons form a ‘Layer’
X1
X2
X3
X4
Input Layer
Direction of information flow
- Each neuron gets ONLY
one input, directly from outside
Hidden Layer
- Connects Input and Output
layers
Output Layer
- Output of each neuron
directly goes to outside
y1
y2
ANN (Artificial Neural Network) – Feed-forward Network
Number of hidden layers can be
None
One
More
Implementing logical functions
• McCulloch and Pitts: every Boolean function AND, OR, & NOT can be
represented by units with suitable weights & thresholds.
• We can use these units to build a network to compute any Boolean function
(t = threshold or the value of the Bias weight that determines the threshold to
cause the neuron to fire)
Network Structure
They are 2 main categories of NN structure:
• Feed-forward/acyclic networks:
 allow signals to travel one way only; from input to output. There is no
feedback (loops).
 Tend to be straight forward networks that associate inputs with outputs. (i.e.
pattern recognition.)
 Usually arranged in layers– each unit receives input only from units in
preceding layer, no links between units in the same layer.
– single-layer perceptrons
– multi-layer perceptrons
• Recurrent/cyclic networks:
– Feeds its outputs back into its own inputs.
– recurrent neural nets have directed cycles with delays
– The links can form arbitrary topologies.
• The brain is recurrent network – activation is fed back to the units that caused
it.
Feed-forward example
No direct connection to the
outside world
Simple NN with 2 inputs, 2 hidden units & 1 output unit.
a5 = g(W3,5.a3 +W4,5.a4)
= g(W3,5. g(W1,3.a1 +W2,3.a2) +W4,5 g(W1,4.a1 +W2,4.a2))
By adjusting weights, we change the function that the Network
represents: learning occurs in NN this way!
Perceptron
• Is a network with all inputs connected
directly to the output.
• This is called a single layer NN (Neural
Network) or a Perceptron Network.
• It is a simple form of NN that is used for
classification of linearly separable
patterns. (i.e. If we have 2 results we can
separate them with a line with each group
result on a different side of the line)
Perceptron or a Single-layer NN
 A Feed-Forward NN with no
hidden units.
 Output units all operate
separately--no shared
weights.
 First Studied in the 50’s
 Other networks were known
about but the perceptron was
the only one capable of
learning and thus all research
was concentrated in this area.
 A single weight only affects
one output so we can limit
our study to a model as
shown on the right
 Notation can be simpler, i.e.
O  Step0 j WjIj
Multilayer NN
• Network with 1/more layers of hidden units
• Layers are usually fully connected;
• numbers of hidden units typically chosen by hand
Summary
• Most brains have lots of neurons; each neuron  linear threshold unit (?)
• Perceptrons (one-layer networks) insufficiently expressive
• Multi-layer networks are sufficiently expressive; can be trained by gradient
descent, i.e., error back-propagation
• Many applications: speech, driving, handwriting, fraud detection, etc.
• Engineering, cognitive modelling, and neural system modelling
• sub fields have largely diverged
End of Chapter 9