Download Chapter 1

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Pattern language wikipedia , lookup

Perceptual control theory wikipedia , lookup

History of artificial intelligence wikipedia , lookup

Gene expression programming wikipedia , lookup

Neural modeling fields wikipedia , lookup

Hierarchical temporal memory wikipedia , lookup

Pattern recognition wikipedia , lookup

Catastrophic interference wikipedia , lookup

Convolutional neural network wikipedia , lookup

Transcript
Introduction to Neural
Networks
John Paxton
Montana State University
Summer 2003
Textbook
Fundamentals of Neural Networks:
Architectures, Algorithms, and Applications
Laurene Fausett
Prentice-Hall
1994
Chapter 1: Introduction
• Why Neural Networks?
Training techniques exist.
High speed digital computers.
Specialized hardware.
Better capture biological neural systems.
Who is interested?
• Electrical Engineers – signal processing,
control theory
• Computer Engineers – robotics
• Computer Scientists – artificial
intelligence, pattern recognition
• Mathematicians – modelling tool when
explicit relationships are unknown
Characterizations
• Architecture – a pattern of connections
between neurons
• Learning Algorithm – a method of
determining the connection weights
• Activation Function
Problem Domains
•
•
•
•
•
Storing and recalling patterns
Classifying patterns
Mapping inputs onto outputs
Grouping similar patterns
Finding solutions to constrained
optimization problems
A Simple Neural Network
x1
w1
y
x2
w2
yin = x1w1 + x2w2
Activation is f(yin)
Biological Neuron
• Dendrites receive electrical signals
affected by chemical process
• Soma fires at differing frequencies
dendrite
soma
axon
Observations
• A neuron can receive many inputs
• Inputs may be modified by weights at the
receiving dendrites
• A neuron sums its weighted inputs
• A neuron can transmit an output signal
• The output can go to many other neurons
Features
• Information processing is local
• Memory is distributed (short term =
signals, long term = dendrite weights)
• The dendrite weights learn through
experience
• The weights may be inhibatory or
excitatory
Features
• Neurons can generalize novel input stimuli
• Neurons are fault tolerant and can sustain
damage
Applications
• Signal processing, e.g. suppress noise on
a phone line.
• Control, e.g. backing up a truck with a
trailer.
• Pattern recognition, e.g. handwritten
characters or face sex identification.
• Diagnosis, e.g. aryhthmia classification or
mapping symptoms to a medical case.
Applications
• Speech production, e.g. NET Talk.
Sejnowski and Rosenberg 1986.
• Speech recognition.
• Business, e.g. mortgage underwriting.
Collins et. Al. 1988.
• Unsupervised, e.g. TD-Gammon.
Single Layer Feedforward NN
x1
w11
y1
w1m
wn1
xn
ym
wnm
Multilayer Neural Network
• More powerful
• Harder to train
x1
xn
z1
y1
zp
ym
Setting the Weight
• Supervised
• Unsupervised
• Fixed weight nets
Activation Functions
• Identity
f(x) = x
• Binary step
f(x) = 1 if x >= q
f(x) = 0 otherwise
• Binary sigmoid
f(x) = 1 / (1 + e-sx)
Activation Functions
• Bipolar sigmoid
f(x) = -1 + 2 / (1 + -sx)
• Hyperbolic tangent
f(x) = (ex – e-x) / (ex + e-x)
History
•
•
•
•
1943 McCulloch-Pitts neurons
1949 Hebb’s law
1958 Perceptron (Rosenblatt)
1960 Adaline, better learning rule (Widrow,
Huff)
• 1969 Limitations (Minsky, Papert)
• 1972 Kohonen nets, associative memory
History
•
•
•
•
1977 Brain State in a Box (Anderson)
1982 Hopfield net, constraint satisfaction
1985 ART (Carpenter, Grossfield)
1986 Backpropagation (Rumelhart, Hinton,
McClelland)
• 1988 Neocognitron, character recognition
(Fukushima)
McCulloch-Pitts Neuron
x1
f(yin) = 1 if yin >= q
x2
x3
y
Exercises
•
•
•
•
2 input AND
2 input OR
3 input OR
2 input XOR