* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Folie 1
Optogenetics wikipedia , lookup
Neuroeconomics wikipedia , lookup
Artificial general intelligence wikipedia , lookup
Neural engineering wikipedia , lookup
Donald O. Hebb wikipedia , lookup
Neuroanatomy wikipedia , lookup
Cognitive neuroscience wikipedia , lookup
Neuropsychology wikipedia , lookup
Cognitive science wikipedia , lookup
Neuropsychopharmacology wikipedia , lookup
Artificial neural network wikipedia , lookup
Neuroinformatics wikipedia , lookup
Biological neuron model wikipedia , lookup
Time series wikipedia , lookup
Embodied cognitive science wikipedia , lookup
History of artificial intelligence wikipedia , lookup
Metastability in the brain wikipedia , lookup
Holonomic brain theory wikipedia , lookup
Mathematical model wikipedia , lookup
Neural modeling fields wikipedia , lookup
Hierarchical temporal memory wikipedia , lookup
Catastrophic interference wikipedia , lookup
Mind uploading wikipedia , lookup
Neurophilosophy wikipedia , lookup
Convolutional neural network wikipedia , lookup
Recurrent neural network wikipedia , lookup
Multilayer Perceptrons A discussion of The Algebraic Mind Chapters 1+2 Jasmin Steinwender ([email protected]) Sebastian Bitzer ([email protected]) Seminar Cognitive Architecture University of Osnabrueck April 16th 2003 The General Question What are the processes and representations underlying mental activity? 16.04.2003 Multilayer Perceptrons 2 Connectionism • • • • • vs. Symbol manipulation Also referred to as parallel-distributed processing (PDP) or neural network models Hypothesis that cognition is a dynamic pattern of connections and activations in a 'neural net.' Model of the parallel processor and the relevance to the anatomy and function of neurons. Consists of simple neuron- like processing elements: units • • • • Biological plausible? brain consisting of neurons, evidence for hebbian learning in the brain • 16.04.2003 • „classical view“ Production rules Hierarchical binary trees computer-like application of rules and manipulation of symbols Mind as symbol manipulator (Marcus) Biological plausible? Brain circuits as representation of generalization and rules Multilayer Perceptrons 3 BUT… Ambiguity of the term connectionism: in the huge variety of connectionist models some will also include symbol-manipulation 16.04.2003 Multilayer Perceptrons 4 Two types of Connectionism 1. implementational connectionism: - a form of connectionism that would seek to understand how systems of neuron-like entities could implement symbols 2. eliminative connectionism: - which denies that the mind can be usefully understood in terms of symbolmanipulation → „ …eliminative connectionism cannot work(…): eliminativist models (unlike humans) provably cannot generalize abstractions to novel items that contain features that did not appear in the training set.” Gary Marcus: http://listserv.linguistlist.org/archives/info-childes/infochi/Connectionism/connectionist5.html and http://listserv.linguistlist.org/archives/info-childes/infochi/Connectionism/connectionism11.html 16.04.2003 Multilayer Perceptrons 5 Symbol manipulation -3 separable Hypothesis• 1. 2. 3. Will be explicitly explained in the whole book, now just mentioned „The mind represents abstract relationships between variables“ „The mind has a system of recursively structured representations“ „ The mind distinguishes between mental representations of individuals and mental representation of kinds“ If the brain is a symbol-manipulator, then one of this hypotheses must hold. 16.04.2003 Multilayer Perceptrons 6 Introduction to Multilayer Perceptrons • simple perceptron – local vs. distributed – linearly separable • hidden layers • learning 16.04.2003 Multilayer Perceptrons 7 The Simple Perceptron I i4 w4 i3 w3 i2 w5 5 o f act ( in wn ) n 1 w2 o i1 16.04.2003 i5 w1 Multilayer Perceptrons 8 Activation functions 16.04.2003 Multilayer Perceptrons 9 The Simple Perceptron II a single-layer feed-forward mapping network o1 i1 16.04.2003 o2 i2 o3 i3 Multilayer Perceptrons i4 10 Local vs. distributed representations representation of CAT local o1 i1 cat 16.04.2003 distributed o2 i2 o1 i3 o2 i1 i2 i3 furry fourlegged whiskered Multilayer Perceptrons 11 Linear (non-)separable functions I [Trappenberg] 16.04.2003 Multilayer Perceptrons 12 Linear (non-)separable functions II boolean functions n 2 Number of linear separable functions 14 Number of linear nonseparable functions 2 3 104 151 4 1,882 63654 5 94,572 ~4.3109 6 15,028,134 ~1.81019 16.04.2003 Multilayer Perceptrons 13 Hidden Layers o1 o2 o3 h1 i1 16.04.2003 i2 h2 i3 Multilayer Perceptrons i4 14 Learning 16.04.2003 Multilayer Perceptrons 15 Backpropagation o1 o2 h1 i1 i2 16.04.2003 • compare actual output - right o., change weights o3 • based on comparison from above change weights in deeper layers, too h2 i3 i4 Multilayer Perceptrons 16 Multilayer Perceptron (MLP) A type of feedforward neural network that is an extension of the perceptron in that it has at least one hidden layer of neurons. Layers are updated by starting at the inputs and ending with the outputs. Each neuron computes a weighted sum of the incoming signals, to yield a net input, and passes this value through its sigmoidal activation function to yield the neuron's activation value. Unlike the perceptron, an MLP can solve linearly inseparable problems. Gary William Flake, The Computational Beauty of Nature, MIT Press, 2000 16.04.2003 Multilayer Perceptrons 17 Other network structures MLPs 16.04.2003 Multilayer Perceptrons 18 Vicky Andy The Family-Tree Model Penny Arthur distributed encoding of patient (6 nodes) hidden layer (12 nodes) distributed encoding of relationship (6 nodes) distributed encoding of agent (6 nodes) Vicky Andy 16.04.2003 others Penny others Multilayer Perceptrons other sis mom dad 19 The sentence prediction model 16.04.2003 Multilayer Perceptrons 20 The appeal of MLPs (preliminary considerations) 1. Biological plausibility – independent nodes – change of connection weights resembles synaptic plasticity – parallel processing brain is a network and MLPs are too 16.04.2003 Multilayer Perceptrons 21 Evaluation Of The Preliminaries 1. Biological plausibility • Biological plausibility considerations make no distinction between eliminative and implementing connectionist models Multilayered perceptron as „more compatible than symbolic models“, BUT nodes and their connections only loosely model neurons and synapses Back-propagation MLP lacks brain-like structure and requires varying synapses (inhibitory and excitatory) Also symbol-manipulation models consist of multiple units and operate in parallel → brain-like structure Not yet clear what is biological plausible – biological knowledge changes over time • • • • 16.04.2003 Multilayer Perceptrons 22 Remarks on Marcus difficult to argue against his arguments: – sometimes addresses comparison between eliminative and implementational connectionist models – sometimes he compares connectionism and classical symbol-manipulation 16.04.2003 Multilayer Perceptrons 23 Remarks on Marcus 1. Biological plausibility (comparison MLPs – classical symbol-manipulation) – MLPs are just an abstraction – no need to model newest detailed biological knowledge – even if not everything is biological plausible, still MLPs are more likely 16.04.2003 Multilayer Perceptrons 24 Preliminary considerations II 2. Universal function approximators – “multilayer networks can approximate any function arbitrarily well” [Trappenberg] – “information is frequently mapped between different representations” [Trappenberg] – mapping of one representation to another can be seen as a function 16.04.2003 Multilayer Perceptrons 25 Evaluation Of The Preliminaries II 2. Universal function approximators • MLP cannot capture all functions (f. e. partial recursive func. – models computational properties of human language) • No guarantee of generalization ability from limited data like humans • Unrealistic need of infinite resources for universal function approximation • Symbol-manipulators could also approximate any function 16.04.2003 Multilayer Perceptrons 26 Preliminary considerations III 3. Little innate structure – children have relatively little innate structure “simulate developmental phenomena in new and … exciting ways” [Elman et al., 1996] e.g. model of balance beam problem [McClelland, 1989] fits data from children – domain-specific representations from domaingeneral architectures 16.04.2003 Multilayer Perceptrons 27 Evaluation Of The Preliminaries III 3. Little innate structure • There also exist symbol-manipulating models with little innate structure • Possibility to prespecify the connection weights of MLP 16.04.2003 Multilayer Perceptrons 28 Preliminary considerations IV 4. Graceful degradation – tolerate noise during processing and in input – tolerate damage (loss of nodes) 16.04.2003 Multilayer Perceptrons 29 Evaluation Of The Preliminaries IV 4. Learning and graceful degradation • No unique ability of all MLP • Symbol-manipulation models which can also handle degradation • No yet empirical data that humans recover from degraded input 16.04.2003 Multilayer Perceptrons 30 Preliminary considerations V 5. Parsimony – one just has to give the architecture and examples – more generally applicable mechanisms (e.g. inflecting verbs) 16.04.2003 Multilayer Perceptrons 31 Evaluation Of The Preliminaries V 5. Parsimony • MLP connections interpreted as free parameters → less parsimonious • Complexity may be more biological plausible than parsimony • Parsimony as criterion only if both models cover the data adequately 16.04.2003 Multilayer Perceptrons 32 What truly distinguishes MLP from Symbol -manipulation • Is not clear, because… …both can be context independent …both can be counted as having symbols …both can be localist or distributed 16.04.2003 Multilayer Perceptrons 33 We are left with the question: Is the mind a system that represents • abstract relationships between variables OR • operations over variables OR • structured representations • and distinguishes between mental representations of individuals and of kinds We will find out later in the book… 16.04.2003 Multilayer Perceptrons 34 Discussion “… I agree with Stemberger that connectionism can make a valuable contribution to cognitive science. The only place that we differ is that, first, he thinks that the contribution will be made by providing a way of *eliminating* symbols, whereas I think that connectionism will make its greatest contribution by accepting the importance of symbols, seeking ways of supplementing symbolic theories and seeking ways of explaining how symbols could be implemented in the brain. Second, Stemberger feels that symbols may play no role in cognition; I think that they do.” Gary Marcus: http://listserv.linguistlist.org/archives/info-childes/infochi/Connectionism/connectionist8.html 16.04.2003 Multilayer Perceptrons 35 References • Marcus, Gary F.: The Algebraic Mind, MIT Press, 2001 • Trappenberg, Thomas P.: Fundamentals of Computational Neuroscience, OUP, 2002 • Dennis, Simon & McAuley, Devin: Introduction to Neural Networks, http://www2.psy.uq.edu.au/~brainwav/Manual/WhatIs.html 16.04.2003 Multilayer Perceptrons 36