Download Slayt 1 - Department of Information Technologies

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Multielectrode array wikipedia , lookup

Optogenetics wikipedia , lookup

Caridoid escape reaction wikipedia , lookup

Artificial neural network wikipedia , lookup

Molecular neuroscience wikipedia , lookup

Development of the nervous system wikipedia , lookup

Feature detection (nervous system) wikipedia , lookup

Neuroanatomy wikipedia , lookup

Mirror neuron wikipedia , lookup

Axon wikipedia , lookup

Neural coding wikipedia , lookup

Central pattern generator wikipedia , lookup

Chemical synapse wikipedia , lookup

Perceptual control theory wikipedia , lookup

Pattern recognition wikipedia , lookup

Channelrhodopsin wikipedia , lookup

Neurotransmitter wikipedia , lookup

Holonomic brain theory wikipedia , lookup

Sparse distributed memory wikipedia , lookup

Nonsynaptic plasticity wikipedia , lookup

Neuropsychopharmacology wikipedia , lookup

Catastrophic interference wikipedia , lookup

Metastability in the brain wikipedia , lookup

Neural modeling fields wikipedia , lookup

Single-unit recording wikipedia , lookup

Stimulus (physiology) wikipedia , lookup

Convolutional neural network wikipedia , lookup

Recurrent neural network wikipedia , lookup

Synaptic gating wikipedia , lookup

Types of artificial neural networks wikipedia , lookup

Biological neuron model wikipedia , lookup

Nervous system network models wikipedia , lookup

Transcript
Introduction
Course Objectives
This course gives a basic neural network architectures and learning rules.
Emphasis is placed on the mathematical analysis of these networks, on
methods of training them and on their application to practical
engineering problems in such areas as pattern recognition, signal
processing and control systems.
What Will Not Be Covered
• Review of all architectures and
learning rules
• Implementation
– VLSI
– Optical
– Parallel Computers
• Biology
• Psychology
Historical Sketch
• Pre-1940: von Hemholtz, Mach, Pavlov, etc.
– General theories of learning, vision, conditioning
– No specific mathematical models of neuron operation
• 1940s: Hebb, McCulloch and Pitts
– Mechanism for learning in biological neurons
– Neural-like networks can compute any arithmetic function
• 1950s: Rosenblatt, Widrow and Hoff
– First practical networks and learning rules
• 1960s: Minsky and Papert
– Demonstrated limitations of existing neural networks, new
learning algorithms are not forthcoming, some research
suspended
• 1970s: Amari, Anderson, Fukushima, Grossberg, Kohonen
– Progress continues, although at a slower pace
• 1980s: Grossberg, Hopfield, Kohonen, Rumelhart, etc.
– Important new developments cause a resurgence in the field
• Aerospace
Applications
– High performance aircraft autopilots, flight path simulations,
aircraft control systems, autopilot enhancements, aircraft
component simulations, aircraft component fault detectors
• Automotive
– Automobile automatic guidance systems, warranty activity
analyzers
• Banking
– Check and other document readers, credit application evaluators
• Defense
– Weapon steering, target tracking, object discrimination, facial
recognition, new kinds of sensors, sonar, radar and image signal
processing including data compression, feature extraction and
noise suppression, signal/image identification
• Electronics
– Code sequence prediction, integrated circuit chip layout, process
control, chip failure analysis, machine vision, voice synthesis,
nonlinear modeling
• Financial
Applications
– Real estate appraisal, loan advisor, mortgage screening, corporate
bond rating, credit line use analysis, portfolio trading program,
corporate financial analysis, currency price prediction
• Manufacturing
– Manufacturing process control, product design and analysis,
process and machine diagnosis, real-time particle identification,
visual quality inspection systems, beer testing, welding quality
analysis, paper quality prediction, computer chip quality analysis,
analysis of grinding operations, chemical product design analysis,
machine maintenance analysis, project bidding, planning and
management, dynamic modeling of chemical process systems
• Medical
– Breast cancer cell analysis, EEG and ECG analysis, prosthesis
design, optimization of transplant times, hospital expense
reduction, hospital quality improvement, emergency room test
advisement
Applications
• Robotics
– Trajectory control, forklift robot, manipulator controllers, vision
systems
• Speech
– Speech recognition, speech compression, vowel classification, text
to speech synthesis
• Securities
– Market analysis, automatic bond rating, stock trading advisory
systems
• Telecommunications
– Image and data compression, automated information services,
real-time translation of spoken language, customer payment
processing systems
• Transportation
– Truck brake diagnosis systems, vehicle scheduling, routing systems
Biology
• Neurons respond slowly
• The brain uses massively parallel computation
– 1011 neurons in the brain
– 104 connections per neuron
Biology
The dendrites are tree-like receptive networks of nerve fibers that carry electrical signals into
the cell body
The cell body effectively sums and thresholds these incoming signals.
The axon is a single long fiber that carries the signal from the cell body out to other neurons.
The point of contact between an axon of one cell and a dendrite of another cell is called a
synapse.
Neuron Model
Neuron Model
the weight “w” corresponds to the strength of a synapse
the cell body is represented by the summation and the transfer
function
the neuron output “a”represents the signal on the axon
Single-Input Neuron Model
The scalar input “p” is multiplied by “w” the scalar weight “w” to form “wp”, one of
the terms that is sent to the summer.
The other input, 1, is multiplied by a bias “b”and then passed to the summer.
The summer output “n”, often referred to as the net input , goes into a transfer
function, which produces the scalar neuron output “a” .
Neuron Model
Transfer Functions
The hard limit transfer function
sets the output of the neuron to 0 if the function argument is less than 0 or
sets the output of the neuron to 1 if its argument is greater than or equal to 0.
Neuron Model
Transfer Functions
The output of a linear transfer function is equal to its input
a=n
Neuron Model
Transfer Functions
This transfer function takes the input and squashes the output into the range 0 to 1,
according to the expression:
Multiple-Input Neuron Model
The neuron has a bias b, which is summed with the weighted inputs to form the net
input n
n = Wp + b
the neuron output can be written as
a = f (Wp + b)
Multiple-Input Neuron Model
The neuron has a bias b, which is summed with the weighted inputs to form the net input n
n = Wp + b
the neuron output can be written as
a = f (Wp + b)
The first index indicates the particular neuron destination for that weight.
The second index indicates the source of the signal fed to the neuron.
Multiple-Input Neuron Model
Abbreviated Notation
Network Architectures
A Layer of Neurons
Network Architectures
A Layer of Neurons
Each of the R inputs is connected to each of the neurons
The layer includes the weight matrix W, the summers, the bias vector b, the
transfer function boxes and the output vector a
Network Architectures
A Layer of Neurons
Abbreviated Notation
Network Architectures
Multiple Layers of Neurons
•Each layer has its own weight matrix , its own bias vector , a net input
vector and an output vector
•The number of the layer as a superscript to the names for each of these
variables
•A layer whose output is the network output is called an output layer. The
other layers are called hidden layers.
Network Architectures
Multiple Layers of Neurons
Abbreviated Notation
Simulation using MATLAB
Neuron output=?
Simulation using MATLAB
To set up this feedforward network
net = newlin([1 3;1 3],1);
Simulation using MATLAB
To set up this feedforward network
net = newlin([1 3;1 3],1);
Assignments
net.IW{1,1} = [1 2];
net.b{1} = 0;
P = [1 2 2 3; 2 1 3 1];
Simulation using MATLAB
To set up this feedforward network
simulate the network
net = newlin([1 3;1 3],1);
Assignments
net.IW{1,1} = [1 2];
net.b{1} = 0;
P = [1 2 2 3; 2 1 3 1];
A = sim(net,P)
A=
5
4
8
5