* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download AMAM Conference 2005
Embodied cognition wikipedia , lookup
Ecological interface design wikipedia , lookup
Incomplete Nature wikipedia , lookup
Pattern recognition wikipedia , lookup
Kevin Warwick wikipedia , lookup
Embodied language processing wikipedia , lookup
Philosophy of artificial intelligence wikipedia , lookup
Machine learning wikipedia , lookup
Concept learning wikipedia , lookup
Existential risk from artificial general intelligence wikipedia , lookup
Perceptual control theory wikipedia , lookup
History of artificial intelligence wikipedia , lookup
Index of robotics articles wikipedia , lookup
Visual servoing wikipedia , lookup
List of Doctor Who robots wikipedia , lookup
Self-reconfiguring modular robot wikipedia , lookup
Robotic automation software wikipedia , lookup
Adaptive collaborative control wikipedia , lookup
AMAM Conference 2005 Adaptive Motion in Animals and Machines Outline of the talk Short AMAM conference overview Introduction to Embodied Artificial Intelligence (keynotes, R. Pfeifer) More detailed look at: Sensory Motor Coordination Value-Systems AMAM: Conference Overview Motivation of studying Biology Source of inspiration for robotics Model features of rather simple animals (insects…) Robots and animals have to solve the same physical problems Robots are useful tools for computational neuroscience Testing Neural Models within a complete sensing-acting loop Biorobotics Bio-inspired technologies New sensors: Whiskers and Antennas Muscle-Like (flexible) actuators Flexible robotic arms and hands Biped and humanoid robots Numerical Models of animal and human locomotion Central Pattern Generator based and other control methods Some robots for illustratoin: AMAM: robots Scorpion [Kirchner05] 8 legged robot BigDog [Buehler, Boston Dynamics] AMAM: Robots Fish Robot Iida Stumpy „Special“ robot to investigate cheap design locomotion (Iida) AMAM Conference: Robots ZAR 4 [boblan05] Bionic robot arm driven by artificial muscels And many more: Insects : Coackroaches[ritzmann 05] Worm [menciassi05] Amoebic Robots [ishiguro05] Bisam Rat [albiez05] Embodied Artificial Intelligence Not interested in the control aspects of robots alone, but rather in designing entire systems Morphology, Materials + Control Synthetic Methology: Understanding intelligent behavior by building Concentrate on complete autonomous robots [Pfeifer99, Iida03] Self-Sufficient: Sustain itself over a extended period of time Situatedness: acquires all information about the environment from its own sensory system „Lives“ in a specified ecological niche: no need for universal robots Embodiment: real physical agents Adaptivity „Why do plants have no brain? They do not move.“ [Brooks] Often aspects of only simple animals are modeled by robots (locomotion of insects…) It took evolution 3 billion years to evolve insects/legged locomotion, but only 500 million more years to develop humans => locomotion must be a hard problem Embodied AI: Principles Emergence: Emergent Behaviours: „emerge“ by the interaction of the robot with the environment Not preprogrammed Agent is the result of its history Exploit the dynamics of the system More adaptive : developmental mechanisms Diversity Compliance: Exploiting ecologicol niche / behavioral diversity Exploration/Exploitation trade off Embodied AI: Principles Parallel, loosely coupled processes Intelligence emerge from a lager number of parallel processes Processes are connected to the agent‘s sensory-motor aparatus Coupling through embodiment or coordination No functional decompositon/hierarchical control like in traditional robotic Supsumption architecture [brooks86] Sensory-Motor Coordination Structuring sensory input Generation of good sensory-motor patterns: Correlated Stationarity Can simplify learning Dimensionality Reduction of sensory-motor space [lungeralla05, boekhorst03] Embodied AI: Principles „Morphological“ Computation Parts of the control can be „computed“ by the morphology Springs and flexible material Exploit system dynamics for control Facets in flies, motion paralax E.g. Exploit gravity and flexible actuators Can simplify control considerably Increase learning speed by morphology „Extreme“ Example: Passive dynamic walker Cheap Design: Exploit physics and constraints of ecological niche Use the most simple architecture for a given task Embodied AI: Principles Redundancy: Overlap of functionality in the subsystems Required for diversity and adaptivity Ecological Balance: Sensory system, Motor system Complexity of the sensory, motor and neural system has to match for a given task Balance between morphology, materials and control [Ishiguro03] Value Principle Motivation of the robot to do something (should be more general than RL) Essential for every complete autonomous agent No generally accepted solution exists 2 approaches will be discussed in more detail Traditional Robotics / AI In difference to traditional robotics Limited numbers of degrees of freedom (e.g. wheels) Stiff structure and joints (servo motors) Limited natural dynamics Centralized rule-based control Easy to control All Computation has to be done by the control system Functional decomposition „Sense-think-act“ cycle Problems: Frame problem Symbol grounding problem Sensory-Motor coordination (SMC) [Pfeifer99, Lungarella05] Used for categorization Traditional approach: Sensory-input to category mapping Prototype or example matching Difficulties: Often this mapping is not learnable Noise and Inaccuracies in Sensors Ambigious sensory input (Type 2 problems) Categorization: Example [Nolfi97] Learn 2 categories (Wall, Cylinder) with IR sensors Data for: Learn with neural network 180 orientations, 20 distances Just linear output units 4 resp 8 hidden neurons Very bad results: 35 % correct categorization Back dots: correct categoritization SMC: Categorization Approach the problem through interacting with the environment Object related actions to structure the input Simplifies the problem of categorization No real internal category representation Just different behaviors for different categories Empirical studies about Dimensionality Reduction [lungarella05] Example in infants: Look at object from different directions in the same distance SMC: Example Learning optimal categorization strategy through a genetic algorithm Nolfi‘s experiment: Fitness: Time the robot is near the cylinder Evolved Behavior: Robot never stops in front of target: Move back/forth and left/right hand side SMC: Example Learning to distinguish circles and diamonds [Beer96] Catching circles, avoiding diamonds Agent can only move horizontally Again evolved controller SMC: Example Results: Not merely centering and statically pattern matching Dynamic strategy, with active scanning Both policies evolve sensorymotor coordination strategies Examples show quite good the idea of sensory-motor coordination Other examples: Catching Circle Darwin II [Reeke89] Garbage Collector [Pfeifer97, Schleier96] Avoiding Diamond SMC: Conclusion Nice new ideas for categorization tasks and robotics in generell Simple examples that illustrate the use of SMC for categorization Examples are „well-suited“ for SMC No complex categorization problem (e.g for visual object recognition) found in the literature Only numerical results which proofs dimensionality reduction How to use them? Critic: Humans are also able to do categorization very well without sensory-motor interaction The emphasis of SMC is a bit overstressed by the authors Value Systems & Developmental Learning [oudeyer04/05, steels03] Intrinsic Motivation of the Agent: learn more about the environment Ideal case: open-end learning Many different behaviors may emerge 2 approaches to this problem discussed in more detail Very adaptive Intelligent Adaptive Curiosity (IAC) [oudeyer04] Autotelic Principle [steels03] Still in the beginning, only for toy examples Other approaches comming from RL Intrinsically motivated RL [singh04] Self Motivated Development [schmidhuber05] IAC: Motivation Push agent towards situations in which it maximizes learning progress Balance between the „unknown“ and the „predictable“ Goal: Improve prediction machine P( A(t ), SM (t )) S (t 1) A(t) … action SM(t)… sensory-motor context S(t+1)… prediction IAC: framework Prediction error E (t ) || S (t 1) Sa(t 1) || => Decrease E(t) First naive approach Learning Progress LP(t ) ( Em(t ) Em(t DELAY )) Em(t)… mean Error at time t Do not reward high error values, reward high LP Meta Learning Machine (predicts error) MP ( A(t ), SM (t )) Ep(t 1) Choose action which maximizes Learning Progress Problem ? IAC: Problem of naive approach: Transition from complex, not predictable situations to simple situations is considered as learning progress Solution: Instead of comparing the LP succesive in time, compare the LP succesive in state space IAC: algorithm Prediction machine P Consists of a set of local experts. Each expert consists of training examples Simple NN algorithm is used for prediction Build kd-tree incrementally : experts in the leaves Each expert stores prediction errors and the mean Calculate local learning progress LPi(t) = -(Empi(t) – Empi(t – DELAY) Used for action selection Very simple algorithms used More sophisticated algorithms have a good chance to improve performance IAC: experiments Toy example: 2 wheeled robot, can produce sound Toy: position depends on sound frequency intervall f1 : moves randomly f2 : stops moving f3 : toy jumps to robot Predictor: predict relative position of the toy IAC: experiments Results: Basically 3 experts First explores intervall f3, then intervall f2 f1 is not explored : not predictable IAC: experiments Playground experiment AIBO robot on a baby play mat Various toys: can be bitten, bashed or simply detected IAC: Playground Experiment Motor Control: Turning head (2 DoF, pan + tilt) Bashing (2 DoF, strength + angle) Crouch + Bite (1 DoF, crouches given distance in direction it is looking at) Perception: 3 High level sensors (just binary values) Visual object detection Biting Sensor Infra-red distance sensor Bashing + Biting only produce visible results if applied in front of an appropriate object Agent knows nothing about sensorimotor affordances IAC: Results Different stages evolves Stage 1: random exploration + body babbling Stage 2: Most of the time looking around (no biting + bashing) Stage 3: biting and bashing Stage 4: Starts to look at objects Sometimes produces something, robot still not oriented to objects Learns precise location of the object Stage 5: Trying bite biteable object, trying to bash bashable object The Autotelic principle [steels03] Autotelic activities: no real reward Motivational driving signal comes from the individual itself Balance between high challenge and required skill Climbing, painting… too high: withdrawal too low: boredom Operational description given in [steels03], no real experiments found Autotelic Principle: Operational Descripion Agent: Organised in number of sub-agencies (components) Establish input/output mapping based on knowledge Each component must be parameterized to self adjust challenge levels Precision of movement, weights of objects… Parameter vector pi for each component Goal: not to reach a stable state, keep exploring parameter landscape Each component has also an associated skill vector Autotelic Principle: Operational Descripion Self Regulation: Operation phase: Clamp challenge parameters, learn skills through learning Shake-Up phase: Increase challenge: skill level already too high Decrease challenge: performance could not be reached Conclusion: Value Systems Both approaches try to create open-ended learner Interesting ideas Only very simple algorithms used, or not even implemented Can help to structure learning progress in complex environments Open for improvement Complete autonomous agents will need some sort of developmental value system No complex real-world experiments found Scalable? Conclusion: Embodied Intelligence Provides new ways of thinking about robotic / intelligence in general Provides a better understanding of intelligent behavior by modelling the behavior. Good principles to design an agent Claims to solve many problems of traditionial AI Good and promising ideas Somehow the algorithmic solutions for more complex systems are missing Actually: same problems as for traditional AI Works for small problems Hard to scale up The End Thank you! Literature [pfeifer99] R. Pfeifer and C. Schleier, Understanding Intelligence, MIT Press [iida03] F. Iida and R. Pfeifer, Embodied Artificial Intelligence [kirchner05] D. Spenneberg, F. Kirchner, Embodied Categorization of spatial environments on the Basis of Proprioceptive Data, AMAM 2005 [ritzmann05] R. Ritzmann, R. Quinn, Convergent Evolution and locomotion through complex terrain by insects, vertebrates and robots, AMAM 2005 [menciassi05] A. Menciassi, S. Spina, Bioinspired robotic worms for locomotion in unstructered environments, AMAM2005 [ishiguro05] A. Ishiguro, M. Shimizu, Slimebot: A Modular robot that exhibits amoebic locomotion, AMAM2005 [albiez05] J. Albiez, T. Hinkel, Reactive Foot-control for quadruped walking, AMAM2005 [boblan05] I. Boblan, R. Bannasch, A Humanlike Robot Arm and Hand with fluidic muscles: The human muscle and the control of technical realization, AMAM 2005 [lungeralla05] M. Lungarella, O. Sporns, Information Self-Structuring: Key Principle for Learning and Development [broekhorst03] R. Broekhorst, M. Lungarella, Dimensionality Reduction through sensory motor-coordination Literature [ishiguro03] A. Ishiguro, T. Kawakatsu, How should control and body systems be coupled? A robotic case study, Embodied artificial intellingence 2003 [nolfi97] S. Nolfi, Evolving non-trivial behavior on autonomous robots: Adaptation is more powerful than decompositionand integration [beer96] R. Beer, Toward the Evolution of Dynamical Neural Networks for Minimally Cognitive Behavior [reeke89] G. Reeke, O. Sporns, Synthetic neural modeling: A multilevel approach to analysis of brain complexity [pfeifer97] R. Pfeifer, C. Schleier, Sensory-motor coordination: The metaphor and beyond: Practice and future of autonmous robots [schleier96] C. Schleier, D. Lambrinos, Categorization in a real world agent using haptic exploration and active perception [oudeyer04] P. Oudeyer, F. Kaplan, Intelligent Adaptive Curiosity: a source of SelfDevelopment [oudeyer05] P. Oudeyer, F. Kaplan, The Playground Experiment: Task independent development of a curious robot. [steels03] L. Steels, The Autotelic Principle [singh04] S. Singh, A. Barto, Intrinsically Motivated Learning of Hierarical Collections of Skills [schmidhuber05] J. Schmidhuber, Self-Motivated Development Through Rewards for Predictor Errors/Improvements Measure influence of SMC [lungeralla05, broekhorst03] New experiments with SMC Measure the effect of SMC with information processing quantities Experiments of Broekhorst: Robot: 5 different Experiments: Wheeled CCD camera (compressed to 10 x 10 pixels) IR sensors (12) Measure angular velocity Control setup: Move forward Moving object Wiggling : Move forward in oscillatory movement Tracking 1: Move forward + track object Tracking 2: Move forward + track moving object Preprogrammed control Measure Influence of SMC [broekhorst03] Quantify dimension of the sensory information Measure Correlation on most significant principal components from the different modalities (R*) 3 different information quantities N H p(i ) log p(i ) Shannon entropy i …Eigenvalue of R* Dominance of the highest eigenvector Number of PC‘s that explain 95% of variance i 1 Results: Difference: Variance in the experiments SMC experiments have higher variance SMC experiments and non SMC experiments can be distinguished No further straithforward results Measure Influence of SMC [lungarella05] Experimental Setup: Active Vision: (compressed 55 x 75 pixels) looking at screen 2 behaviors: Foveation: „follow red area“ Random: Same motion structure, not coordinated 2 scenarios Artificial Scene: Random Data with moving red block Natural Images Measure Influence of SMC [lungarella05] Quantify sensory information Entropy Joint-Entropy Mutual Information Integration : Multivariate Mutual Information I ( X ) H ( xi ) H ( X ) i Complexity : C ( X ) H ( X ) H ( xi | X xi ) i Quantify Dimensionality Reduction PCA Isomap ([tenenbaum01], also recognizes non-linear dimensions) Results for foveation behavior Entropy in central regions decreased Mutual information increased Results for foveation behavior Integration and Complexity where much larger in the center Results for foveation behavior Reduced dimensionality (isomap) Mutual information between center and motor actions also increased