* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Neural Networks Coursework
Collaborative information seeking wikipedia , lookup
Visual Turing Test wikipedia , lookup
Affective computing wikipedia , lookup
Computer vision wikipedia , lookup
Incomplete Nature wikipedia , lookup
Self-reconfiguring modular robot wikipedia , lookup
Catastrophic interference wikipedia , lookup
Combining Multiple Modes of Information using Unsupervised Neural Classifiers Khurshid Ahmad, Matthew Casey, Bogdan Vrusias, Panagiotis Saragiotis http://www.computing.surrey.ac.uk/ncg/ Neural Computing Group, Department of Computing, School of Electronics and Physical Sciences, University of Surrey 1 Content • Report on preliminary experiments to: – Attempt to improve classification through combining modalities of information – Use a modular co-operative neural network system combining unsupervised learning techniques • Tested using: – Scene-of-crime images and collateral text – Number magnitude and articulation 2 Background • Consider how we may improve classification through combination: – Combining like classifiers (e.g. ensemble systems) – Combining expert classifiers (e.g. modular systems) • Concentrate on a modular approach to combining modalities of information – For example, Kittler et al (1998): • Personal identity verification using frontal face, face profile and voice inputs 3 Multi-net Systems • Concept of combining neural network systems has been discussed for a number of years – Both ensemble and modular systems – Ensemble more prevalent • Term multi-net systems has been promoted by Sharkey (1999, 2002) who recently advocated the use of modular systems – For example, mixture-of-experts by Jacobs et al 1991 4 Multi-net Systems • Neural network techniques for classification tend to subscribe to the supervised learning paradigm – Ensemble methods – Mixture-of-experts • Exceptions include Lawrence et al (1997) and Ahmad et al (2002) • Unsupervised techniques give rise to problems of interpretation 5 Self-organised Combinations • Our approach is based upon the combination of different Hebbian-like learning systems • Hebb’s neurophysiological postulate (1949) – Strength of connection is increased when both sides of the connection are active 6 Self-organised Combinations • Willshaw & von der Malsburg (1976) – Used Hebbian learning to associate patterns of activity in a 2-d pre-synaptic (input) layer and a 2-d postsynaptic (output) layer – Pre-synaptic neurons become associated with postsynaptic neurons • Kohonen (1997) extended this in his Selforganising Map (SOM) – Statistical approximation of the input space – Topological map showing relatedness of input patterns – Clusters used to show classes 7 Self-organised Combinations • Our architecture builds further on this using the multi-net paradigm • Can be compared to Hebb’s superordinate combination of cell assemblies • Two SOMs linked by Hebbian connections – One SOM learns to classify a primary modality of information – One SOM learns to classify a collateral modality of information – Hebbian connections associate patterns of activity in each SOM 8 Self-organised Combinations . . . Primary Vector . . . Primary SOM Bi-directional Hebbian Network Collateral SOM Collateral Vector • SOMs and Hebbian connections trained synchronously 9 Self-organised Combinations • Hebbian connections associate neighbourhoods of activity – Not just a one-to-one linear association – Each SOM’s output is formed by a pattern of activity centred on the winning neuron for the primary and collateral input • Training complete when both SOM classifiers have learned to classify their respective inputs 10 Classifying Images and Text Class Body Single objects (close-up) Primary Image Collateral Text Full length shot of body Nine millimetre browning high power self-loading pistol 11 Classifying Images and Text • Classify images based upon images and texts • Primary modality of information: – 66 images from the scene-of-crime domain – 112-d vector based upon colour, edges and texture • Collateral modality of information: – 66 texts describing image content – 50-d binary vector term frequency analysis • 8 expert defined classes • 58 vector pairs used for training, 8 for testing 12 Training • • • • Image SOM: 15 by 15 neurons Text SOM: 15 by 15 neurons Initial random weights Gaussian neighbourhood function with initial radius 8 neurons, reducing to 1 neuron • Exponentially decreasing learning rate, initially 0.9, reducing to 0.1 • Hebbian connection weights normalised • Trained for 1000 epochs 13 Testing • Tested with 8 image and text vectors – Successful classification if test vector’s winner corresponds with identified cluster for class • Image SOM: – Correctly classified 4 images • Text SOM: – Correctly classified 5 texts 14 Testing • For misclassified images – Text classification was determined – Translated into image classification via Hebbian activation • Similarly for misclassified texts • Image SOM: – Further 3 images classified out of 4 (total 7 out of 8) • Text SOM: – Further 2 texts classified out of 3 (total 7 out of 8) 15 Comparison • Contrast with single modality of classification in image or text SOM • Compared with a single SOM classifier – 15 by 15 neurons – Trained on combined image and text vectors (162-d vectors) – 3 out of 8 test vectors correctly classified 16 Classifying Number • Classify numbers based upon (normalised) image or articulation? • Primary modality of information: – Magnitude representation of the numbers 1 to 22 – 66-d binary vector with 3 bits per magnitude • Collateral modality of information: – Articulation representation of the numbers 1 to 22 – 16-d vector representing phonemes • 22 different numbers to classify • 16 vector pairs used for training, 6 testing 17 Training • • • • Magnitude SOM: 66 by 1 neurons Articulation SOM: 16 by 16 neurons Initial random weights Gaussian neighbourhood function with initial radius 33 (primary) and 8 (collateral) neurons, reducing to 1 neuron • Exponentially decreasing learning rate, initially 0.5 • Hebbian connection weights normalised • Trained for 1000 epochs 18 Testing • Tested with 6 magnitude and articulation vectors – Successful classification if test vector’s winner corresponds with identified cluster for class • Magnitude SOM: – Correctly classified 6 magnitudes – Magnitudes arranged in a ‘number line’ • Articulation SOM: – Similar phonetic responses, but essentially misclassified all 6 articulations 19 Testing • For misclassified articulation vectors – Magnitude classification was determined – Translated into articulation classification via Hebbian activation • Articulation SOM: – 3 articulation vectors classified out of 6 – Remaining 3 demonstrate that Hebbian association not sufficient to give rise to better classification 20 Comparison • Contrast with single modality of classification in magnitude or articulation SOM • Compared with a single SOM classifier – 16 by 16 neurons – Trained on combined magnitude and articulation vectors (82-d vectors) – Misclassified all 6 articulation vectors – SOM shows test numbers are similar in ‘sound’ to numbers in the training set – Combined SOM does not demonstrate ‘number line’ and cannot capitalise upon it 21 Summary • Preliminary results show that: – Modular co-operative multi-net system using unsupervised learning techniques can improve classification with multiple modalities – Hebb’s superordinate combination of cell assemblies? • Future work: – Evaluate against larger sets of data – Further understanding of clustering and classification in SOMs – Further explore linkage of neighbourhoods, more than just a one-to-one mapping, and theory underlying model 22 Acknowledgements • Supported by the EPSRC Scene of Crime Information System project (Grant No.GR/M89041) – University of Sheffield – University of Surrey – Five UK police forces • Images supplied by the UK Police Training College at Hendon, with text transcribed by Chris Handy 23 References Ahmad, K., Casey, M.C. & Bale, T. (2002). Connectionist Simulation of Quantification Skills. Connection Science, vol. 14(3), pp. 165-201. Jacobs, R.A., Jordan, M.I. & Barto, A.G. (1991). Task Decomposition through Competition in a Modular Connectionist Architecture: The What and Where Vision Tasks. Cognitive Science, vol. 15, pp. 219-250. Hebb, D.O. (1949). The Organization of Behavior: A Neuropsychological Theory. New York: John Wiley & Sons. Kittler, J., Hatef, M., Duin, R.P.W. & Matas, J. (1998). On Combining Classifiers. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 20(3), pp. 226-239. Kohonen, T. (1997). Self-Organizing Maps, 2nd Ed. Berlin, Heidelberg, New York: SpringerVerlag. Lawrence, S., Giles, C.L., Ah Chung Tsoi & Back, A.D. (1997). Face Recognition: A Convolutional Neural Network Approach. IEEE Transactions on Neural Networks, vol. 8(1), pp. 98-113. Sharkey, A.J.C. (1999). Combining Artificial Neural Nets: Ensemble and Modular Multi-Net Systems. Berlin, Heidelberg, New York: Springer-Verlag. Sharkey, A.J.C. (2002). Types of Multinet System. In Roli, F. & Kittler, J. (Ed), Proceedings of the Third International Workshop on Multiple Classifier Systems (MCS 2002), pp. 108-117. Berlin, Heidelberg, New York: Springer-Verlag. Willshaw, D.J. & von der Malsburg, C. (1976). How Patterned Neural Connections can be set up by Self-Organization. Proceedings of the Royal Society, Series B, vol. 194, pp. 431-445. 24 Combining Multiple Modes of Information using Unsupervised Neural Classifiers Khurshid Ahmad, Matthew Casey, Bogdan Vrusias, Panagiotis Saragiotis http://www.computing.surrey.ac.uk/ncg/ Neural Computing Group, Department of Computing, School of Electronics and Physical Sciences, University of Surrey 25 Multi-net Systems Combination Decision Top-down Static Multi-net Systems Bottom-up Dynamic Co-operative Combination Mechanism Competitive Bottom-up Combination Method Components Ensemble Hybrid Modular (Fusion) Sharkey (2002) – Types of Multi-net System 26