Download Folie 1

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Optogenetics wikipedia , lookup

Connectome wikipedia , lookup

Neuroeconomics wikipedia , lookup

Artificial general intelligence wikipedia , lookup

Neural engineering wikipedia , lookup

Donald O. Hebb wikipedia , lookup

Neuroanatomy wikipedia , lookup

Cognitive neuroscience wikipedia , lookup

Neuropsychology wikipedia , lookup

Cognitive science wikipedia , lookup

Neuropsychopharmacology wikipedia , lookup

Artificial neural network wikipedia , lookup

Neuroinformatics wikipedia , lookup

Biological neuron model wikipedia , lookup

Time series wikipedia , lookup

AI winter wikipedia , lookup

Embodied cognitive science wikipedia , lookup

History of artificial intelligence wikipedia , lookup

Metastability in the brain wikipedia , lookup

Holonomic brain theory wikipedia , lookup

Mathematical model wikipedia , lookup

Neural modeling fields wikipedia , lookup

Hierarchical temporal memory wikipedia , lookup

Catastrophic interference wikipedia , lookup

Mind uploading wikipedia , lookup

Neurophilosophy wikipedia , lookup

Convolutional neural network wikipedia , lookup

Recurrent neural network wikipedia , lookup

Nervous system network models wikipedia , lookup

Types of artificial neural networks wikipedia , lookup

Transcript
Multilayer Perceptrons
A discussion of
The Algebraic Mind
Chapters 1+2
Jasmin Steinwender ([email protected])
Sebastian Bitzer ([email protected])
Seminar Cognitive Architecture
University of Osnabrueck
April 16th 2003
The General Question
What are the processes and
representations underlying mental
activity?
16.04.2003
Multilayer Perceptrons
2
Connectionism
•
•
•
•
•
vs.
Symbol manipulation
Also referred to as parallel-distributed
processing (PDP) or neural network
models
Hypothesis that cognition is a dynamic
pattern of connections and activations
in a 'neural net.'
Model of the parallel processor and the
relevance to the anatomy and function
of neurons.
Consists of simple neuron- like
processing elements: units
•
•
•
•
Biological plausible?
brain consisting of neurons,
evidence for hebbian learning in
the brain
•
16.04.2003
•
„classical view“
Production rules
Hierarchical binary trees
computer-like application of rules
and manipulation of symbols
Mind as symbol manipulator
(Marcus)
Biological plausible?
Brain circuits as representation of
generalization and rules
Multilayer Perceptrons
3
BUT…
Ambiguity of the term connectionism:
in the huge variety of connectionist models
some will also include symbol-manipulation
16.04.2003
Multilayer Perceptrons
4
Two types of Connectionism
1.
implementational connectionism:
- a form of connectionism that would seek to understand how systems of
neuron-like entities could implement symbols
2.
eliminative connectionism:
- which denies that the mind can be usefully understood in terms of symbolmanipulation
→ „ …eliminative connectionism cannot work(…): eliminativist models (unlike
humans) provably cannot generalize abstractions to novel items that contain
features that did not appear in the training set.”
Gary Marcus:
http://listserv.linguistlist.org/archives/info-childes/infochi/Connectionism/connectionist5.html and
http://listserv.linguistlist.org/archives/info-childes/infochi/Connectionism/connectionism11.html
16.04.2003
Multilayer Perceptrons
5
Symbol manipulation
-3 separable Hypothesis•
1.
2.
3.
Will be explicitly explained in the whole book, now just
mentioned
„The mind represents abstract relationships between
variables“
„The mind has a system of recursively structured
representations“
„ The mind distinguishes between mental representations
of individuals and mental representation of kinds“
If the brain is a symbol-manipulator, then one of this
hypotheses must hold.
16.04.2003
Multilayer Perceptrons
6
Introduction to Multilayer
Perceptrons
• simple perceptron
– local vs. distributed
– linearly separable
• hidden layers
• learning
16.04.2003
Multilayer Perceptrons
7
The Simple Perceptron I
i4
w4
i3
w3
i2
w5
5
o  f act ( in  wn )
n 1
w2
o
i1
16.04.2003
i5
w1
Multilayer Perceptrons
8
Activation functions
16.04.2003
Multilayer Perceptrons
9
The Simple Perceptron II
a single-layer feed-forward mapping network
o1
i1
16.04.2003
o2
i2
o3
i3
Multilayer Perceptrons
i4
10
Local vs. distributed representations
representation of CAT
local
o1
i1
cat
16.04.2003
distributed
o2
i2
o1
i3
o2
i1
i2
i3
furry
fourlegged
whiskered
Multilayer Perceptrons
11
Linear (non-)separable functions I
[Trappenberg]
16.04.2003
Multilayer Perceptrons
12
Linear (non-)separable functions II
boolean functions
n
2
Number of linear
separable functions
14
Number of linear nonseparable functions
2
3
104
151
4
1,882
63654
5
94,572
~4.3109
6
15,028,134
~1.81019
16.04.2003
Multilayer Perceptrons
13
Hidden Layers
o1
o2
o3
h1
i1
16.04.2003
i2
h2
i3
Multilayer Perceptrons
i4
14
Learning
16.04.2003
Multilayer Perceptrons
15
Backpropagation
o1
o2
h1
i1
i2
16.04.2003
• compare actual
output - right o.,
change weights
o3
• based on
comparison from
above change
weights in deeper
layers, too
h2
i3
i4
Multilayer Perceptrons
16
Multilayer Perceptron (MLP)
A type of feedforward neural network that is an extension of the
perceptron in that it has at least one hidden layer of neurons.
Layers are updated by starting at the inputs and ending with the
outputs. Each neuron computes a weighted sum of the incoming
signals, to yield a net input, and passes this value through its
sigmoidal activation function to yield the neuron's activation
value. Unlike the perceptron, an MLP can solve linearly
inseparable problems.
Gary William Flake,
The Computational Beauty of Nature,
MIT Press, 2000
16.04.2003
Multilayer Perceptrons
17
Other network structures
MLPs
16.04.2003
Multilayer Perceptrons
18
Vicky
Andy
The
Family-Tree
Model
Penny
Arthur
distributed
encoding
of patient
(6 nodes)
hidden layer (12 nodes)
distributed
encoding of
relationship
(6 nodes)
distributed
encoding
of agent
(6 nodes)
Vicky
Andy
16.04.2003
others
Penny
others
Multilayer Perceptrons
other
sis
mom
dad
19
The sentence prediction model
16.04.2003
Multilayer Perceptrons
20
The appeal of MLPs
(preliminary considerations)
1. Biological plausibility
– independent nodes
– change of connection weights resembles
synaptic plasticity
– parallel processing
 brain is a network and MLPs are too
16.04.2003
Multilayer Perceptrons
21
Evaluation Of The Preliminaries
1.
Biological plausibility
•
Biological plausibility considerations make no distinction between
eliminative and implementing connectionist models
Multilayered perceptron as „more compatible than symbolic
models“, BUT nodes and their connections only loosely model
neurons and synapses
Back-propagation MLP lacks brain-like structure and requires
varying synapses (inhibitory and excitatory)
Also symbol-manipulation models consist of multiple units and
operate in parallel → brain-like structure
Not yet clear what is biological plausible – biological knowledge
changes over time
•
•
•
•
16.04.2003
Multilayer Perceptrons
22
Remarks on Marcus
difficult to argue against his arguments:
– sometimes addresses comparison between
eliminative and implementational connectionist
models
– sometimes he compares connectionism and
classical symbol-manipulation
16.04.2003
Multilayer Perceptrons
23
Remarks on Marcus
1. Biological plausibility
(comparison MLPs – classical symbol-manipulation)
– MLPs are just an abstraction
– no need to model newest detailed biological
knowledge
– even if not everything is biological plausible,
still MLPs are more likely
16.04.2003
Multilayer Perceptrons
24
Preliminary considerations II
2. Universal function approximators
– “multilayer networks can approximate any
function arbitrarily well” [Trappenberg]
– “information is frequently mapped between
different representations” [Trappenberg]
– mapping of one representation to another can
be seen as a function
16.04.2003
Multilayer Perceptrons
25
Evaluation Of The Preliminaries II
2. Universal function approximators
• MLP cannot capture all functions (f. e. partial recursive
func. – models computational properties of human
language)
• No guarantee of generalization ability from limited data
like humans
• Unrealistic need of infinite resources for universal function
approximation
• Symbol-manipulators could also approximate any function
16.04.2003
Multilayer Perceptrons
26
Preliminary considerations III
3. Little innate structure
– children have relatively little innate structure
 “simulate developmental phenomena in new
and … exciting ways” [Elman et al., 1996]
e.g. model of balance beam problem
[McClelland, 1989] fits data from children
– domain-specific representations from domaingeneral architectures
16.04.2003
Multilayer Perceptrons
27
Evaluation Of The Preliminaries III
3. Little innate structure
• There also exist symbol-manipulating
models with little innate structure
• Possibility to prespecify the connection
weights of MLP
16.04.2003
Multilayer Perceptrons
28
Preliminary considerations IV
4. Graceful degradation
– tolerate noise during processing and in input
– tolerate damage (loss of nodes)
16.04.2003
Multilayer Perceptrons
29
Evaluation Of The Preliminaries IV
4. Learning and graceful degradation
• No unique ability of all MLP
• Symbol-manipulation models which can
also handle degradation
• No yet empirical data that humans recover
from degraded input
16.04.2003
Multilayer Perceptrons
30
Preliminary considerations V
5. Parsimony
– one just has to give the architecture and
examples
– more generally applicable mechanisms
(e.g. inflecting verbs)
16.04.2003
Multilayer Perceptrons
31
Evaluation Of The Preliminaries V
5. Parsimony
• MLP connections interpreted as free
parameters → less parsimonious
• Complexity may be more biological
plausible than parsimony
• Parsimony as criterion only if both models
cover the data adequately
16.04.2003
Multilayer Perceptrons
32
What truly distinguishes MLP from
Symbol -manipulation
• Is not clear, because…
…both can be context independent
…both can be counted as having symbols
…both can be localist or distributed
16.04.2003
Multilayer Perceptrons
33
We are left with the question:
Is the mind a system that represents
• abstract relationships between variables OR
• operations over variables OR
• structured representations
• and distinguishes between mental
representations of individuals and of kinds
We will find out later in the book…
16.04.2003
Multilayer Perceptrons
34
Discussion
“… I agree with Stemberger that connectionism can make
a valuable contribution to cognitive science. The only
place that we differ is that, first, he thinks that the
contribution will be made by providing a way of
*eliminating* symbols, whereas I think that connectionism
will make its greatest contribution by accepting the
importance of symbols, seeking ways of supplementing
symbolic theories and seeking ways of explaining how
symbols could be implemented in the brain. Second,
Stemberger feels that symbols may play no role in
cognition; I think that they do.”
Gary Marcus:
http://listserv.linguistlist.org/archives/info-childes/infochi/Connectionism/connectionist8.html
16.04.2003
Multilayer Perceptrons
35
References
• Marcus, Gary F.: The Algebraic Mind, MIT
Press, 2001
• Trappenberg, Thomas P.: Fundamentals of
Computational Neuroscience, OUP, 2002
• Dennis, Simon & McAuley, Devin:
Introduction to Neural Networks,
http://www2.psy.uq.edu.au/~brainwav/Manual/WhatIs.html
16.04.2003
Multilayer Perceptrons
36