Integrating Language and Motor Function on a Humanoid Robot

Similar documents
A Case-Based Approach To Imitation Learning in Robotic Agents

Seminar - Organic Computing

Evolution of Symbolisation in Chimpanzees and Neural Nets

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Software Maintenance

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Evolutive Neural Net Fuzzy Filtering: Basic Description

Accelerated Learning Course Outline

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Accelerated Learning Online. Course Outline

Python Machine Learning

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Learning Methods for Fuzzy Systems

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Word Segmentation of Off-line Handwritten Documents

AQUA: An Ontology-Driven Question Answering System

Introduction to Simulation

DIGITAL GAMING & INTERACTIVE MEDIA BACHELOR S DEGREE. Junior Year. Summer (Bridge Quarter) Fall Winter Spring GAME Credits.

INPE São José dos Campos

GACE Computer Science Assessment Test at a Glance

Quantitative Evaluation of an Intuitive Teaching Method for Industrial Robot Using a Force / Moment Direction Sensor

Degree Qualification Profiles Intellectual Skills

Human Emotion Recognition From Speech

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Probabilistic Latent Semantic Analysis

Speech Recognition at ICSI: Broadcast News and beyond

Modeling function word errors in DNN-HMM based LVCSR systems

On the Formation of Phoneme Categories in DNN Acoustic Models

Eli Yamamoto, Satoshi Nakamura, Kiyohiro Shikano. Graduate School of Information Science, Nara Institute of Science & Technology

Continual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Abstractions and the Brain

Robot manipulations and development of spatial imagery

SARDNET: A Self-Organizing Feature Map for Sequences

Data Fusion Models in WSNs: Comparison and Analysis

Reinforcement Learning by Comparing Immediate Reward

Automating the E-learning Personalization

MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE

Circuit Simulators: A Revolutionary E-Learning Platform

On the Combined Behavior of Autonomous Resource Management Agents

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

THE DEPARTMENT OF DEFENSE HIGH LEVEL ARCHITECTURE. Richard M. Fujimoto

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

BODY LANGUAGE ANIMATION SYNTHESIS FROM PROSODY AN HONORS THESIS SUBMITTED TO THE DEPARTMENT OF COMPUTER SCIENCE OF STANFORD UNIVERSITY

Time series prediction

WHEN THERE IS A mismatch between the acoustic

Designing a Rubric to Assess the Modelling Phase of Student Design Projects in Upper Year Engineering Courses

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

Corpus Linguistics (L615)

A Pipelined Approach for Iterative Software Process Model

Modeling function word errors in DNN-HMM based LVCSR systems

The Evolution of Random Phenomena

A Bayesian Model of Imitation in Infants and Robots

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass

Axiom 2013 Team Description Paper

Knowledge Transfer in Deep Convolutional Neural Nets

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

9.85 Cognition in Infancy and Early Childhood. Lecture 7: Number

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

CS Machine Learning

Physics 270: Experimental Physics

Rajesh P. N. Rao, Aaron P. Shon and Andrew N. Meltzoff

On-Line Data Analytics

Knowledge-Based - Systems

NCEO Technical Report 27

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

The Learning Tree Workshop: Organizing Actions and Ideas, Pt I

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

What s in a Step? Toward General, Abstract Representations of Tutoring System Log Data

Statewide Framework Document for:

Introduction to Modeling and Simulation. Conceptual Modeling. OSMAN BALCI Professor

Natural Language Processing. George Konidaris

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)

Artificial Neural Networks

Integrating simulation into the engineering curriculum: a case study

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

arxiv: v2 [cs.ro] 3 Mar 2017

Linking Task: Identifying authors and book titles in verbose queries

Lecture 1: Machine Learning Basics

Test Effort Estimation Using Neural Network

B. How to write a research paper

XXII BrainStorming Day

LEGO MINDSTORMS Education EV3 Coding Activities

While you are waiting... socrative.com, room number SIMLANG2016

Designing e-learning materials with learning objects

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Visual CP Representation of Knowledge

Dyslexia/dyslexic, 3, 9, 24, 97, 187, 189, 206, 217, , , 367, , , 397,

Concept Acquisition Without Representation William Dylan Sabo

An Embodied Model for Sensorimotor Grounding and Grounding Transfer: Experiments With Epigenetic Robots

Missouri Mathematics Grade-Level Expectations

SNAP, CRACKLE AND POP! INFUSING MULTI-SENSORY ACTIVITIES INTO THE EARLY CHILDHOOD CLASSROOM SUE SCHNARS, M.ED. AND ELISHA GROSSENBACHER JUNE 27,2014

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

Self-Supervised Acquisition of Vowels in American English

THE ROLE OF TOOL AND TEACHER MEDIATIONS IN THE CONSTRUCTION OF MEANINGS FOR REFLECTION

Transcription:

1 Integrating Language and Motor Function on a Humanoid Robot L. Majure, L. Niehaus, A. Duda, A. Silver, L. Wendt, S. Levinson Abstract In order to design an artificial entity that is able to use language naturally, a computer program that simply manipulates lexical symbols is not good enough. Using these symbols without understanding the meaning attached to them is pointless. Not only does use of natural language require a framework for computational function (the brain), but it requires a physical embodiment. Our lab s focus is therefore to create such a framework using mathematical approaches in statistical pattern recognition, as well as dynamical systems models. While these aim to emulate the brain, the icub humanoid robot serves as our embodied platform. Through the fusion of a rich sensorimotor periphery, we have begun to develop a cognitive architecture capable of autonomous language acquisition. I. INTRODUCTION The essence of cognition is learning and recalling a model of the world. This model is built by forming associations among stimuli received from the array of senses and motor experience. Multisensory information allows for a robust model which can make correct inferences and predictions even when observations are noisy or incomplete. One reasons that artificial intelligence techniques rarely match up to the sophistication and flexibility of human thought is their unimodality. The best way to replicate the ability of humans to integrate complex multisensory information and act on the world is to implement proposed algorithms on an embodied, autonomous platform. The Language Acquisition and Robotics Lab began using modified hobby robots, which explored the world with their stereo vision, ears, wheels, and gripper. Cascaded hidden Markov models were used to form associative memories, allowing the robot to learn words for objects in its environment. [1]. Simple syntax was also learned from acting on objects and associating verbs with the actions [2]. The Language Acquisition and Robotics Lab obtained an icub humanoid robot from the RobotCub Consortium in the spring of 2010. This robot is capable of more complex interactions with the world than the previous non-humanoids. The icub allows the lab to study motor control and learning in further depth than has been possible before. Especially of interest is the connection between motor experience and language II. MOTOR LEARNING AND CONTROL A. Motor representations for learning and association Motor representations need to be chosen in such a way that they are adaptive, computationally reasonable, and available for association with other data streams. The analogous function in the human brain is located in the parietal lobe, where spatial Fig. 1. Bert, our icub robot information from different coordinate systems is integrated and motions are planned. Robot motor control requires a mapping between spatial coordinates and the joint angles needed to place the body in that position. In many robotics applications it is sufficient to hard code the kinematics of the robot, allowing classical control techniques to be used. However, in an adaptive cognitive framework, a learned model is desired. This is both for studying possible mechanisms of human motor control and so that the robot can adapt to changes in its dynamic properties. In addition this learned model should be able to emit a reduced-dimensionality signal for language integration. There are two general methods for motor representation used by the lab, each with their own distinct advantages and disadvantages. The self-organizing map is a model which has been demonstrated to learn correspondences between different representations of space. It has been used for planning reaching tasks by mapping joint angles to hand position [3]. SOMs can be used for motor babbling, or random kinematic exploration, the technique infants use to learn about their body. The outputs of the SOMs, the neuron activations, form a topologically smooth representation of the robot s kinematic space. These can be used as inputs to a sequence classifier, or an associative language engine. Another model which has been successfully applied to motor representation is the Hidden Markov Model[4]. The HMM is used for many pattern and sequence recognition systems, and forms the basis of our associative language engine. The primary advantage of the HMM is its ability to encode sequence information. HMMs have been applied to

2 gesture recognition[5], and are able to capture and reproduce short atomic gestures quite reliably. The main drawback to the HMM in this usage however is the direct use of joint angles as input features to the model. This high dimensional input introduces noise issues as well as the need for increased model complexity and training time. This problem is often solved through the use of Principle Component Analysis, which aligns the data along the axes of greatest variation. This creates a reduced dimensionality input, with dimensions that were primarily noise removed, making the classification problem easier. An even better solution is to combine the SOM and HMM, which is the approach being actively pursued in our lab. This provides the robot with the crucial ability to perform pattern and sequence recognition on features which are directly related to its internal kinematic model. The SOM can be used to emit a discrete observation which is directly mappable back to a given motor pose in its kinematic space. The bank of HMMs can classify basic learned sequences of these poses (gestures or motor words ), and themselves emit a discrete representation of atomic actions. These quantized gestures are the crux of the imitation and language grounding problems which will be discussed in the following sections. B. Imitation of human motions Much of the motor task learning of children is driven by imitation. A humanoid robot is uniquely suited to studying this cognitive skill, due to its similar body configuration to humans. Imitation requires awareness of body layout and an ability to map other bodies onto ones own. This is an active area of research for the lab. An important reason to tie imitation to language is segmentation and generalization of motions. Much as linguistic labels allow visual objects to be described generally, they can be used similarly for motor objects. Imitation is essentially a hidden variable problem, in that the sequence of motor signals used to generate the motion must be inferred by the observing robot. The model of the motion being imitated is updated by the learning robot observing its own performance and comparing to the trained action. In a sense, this view of imitation is similar to the Motor Theory of Language [6]. The learned action has a symbolic or linguistic meaning, which can only be inferred and reproduced by understanding the sequence of motor gestures used to produce it. C. Learning fine motor control for precision tasks Aside from the problem of learning labels and effects of motions, eventually fine motor control needs to be implemented if the robot is to accomplish certain physical tasks. This encompasses several behavioral goals for the icub, including manipulation of small objects and walking. The short-term project addressing this problem is getting the robot to balance an inverted pendulum on its hand. It is the intention of this project that the results can be extended to smooth control of motion, balance, precision, and timing. This area of research mirrors the function of the cerebellum in humans. Fig. 2. Sensorimotor integration architecture III. LANGUAGE Up to this point we have discussed the methods our lab uses to create biologically feasible internal models of motor space as well as ways of creating a repertoire of motor abilities through imitation learning, and practice of fine motor abilities. However the overarching goal of much of this research is to study the extent to which motor function dictates and is a necessity for language. In the beginning, we set out a plan based on the fundamental idea, shared by many[7], that this language-action interaction is both compositional and hierarchical in nature. Many architectures have already been generated, with structure that facilitates such learning. These structures have been both directed[8] and and emergent[9] in nature. Both of these experiments were aimed at producing systems that were able to identify reusable primitives and higher level programs which could exploit the modularity and reusability of said primitives. In addition, both of these experiments we carried out with a focus on motor and visual modalities. Our experiments however focus on a similar idea of reuse and composition of low-level primitives, but with the end goal being the study of the acquisition of language. Fortunately, the motor learning and language learning problems share similar methods and a fair amount of history. The Motor Theory of Speech Perception[6], has been posited as an explanation for humans ability to so accurately recognize speech, which automatic speech recognition systems fail in many scenarios. While not unanimously accepted, there has been mounting evidence in recent years, that the motor cortex is integral in both the basic recognition of phonemes and of higher level concepts such as action words[10][11]. Therefore, one of the primary areas of research in the lab is the creation of an architecture, that is the same horizontally (i.e. across modalities), and vertically (integrating concepts to higher levels of abstraction). Figure 2 shows a schematic depiction of such an architecture. Our lab has already had success integrating across modalities[2], in our experiments with a wheeled robot. In this case, the two modalities were vision and auditory senses. The robot was successful in learning both acoustic patterns on word and phoneme level, as well as visual patterns representing various objects. In all of these cases, an online-training

3 Fig. 3. Illustrative example of a concept grounded over multiple sensory modalities version of the HMM[12] was the fundamental model used for recognition. The test procedure consisted of remembering and then recalling some fixed number of objects present in the robot s playpen. The vertical aspect to the architecture in Figure 2, is twofold. At higher levels of the model, the robot s brain integrates both over modalities and over sequences. Integration over modalities attempts to solve the symbol grounding problem of language. These concept HMMs create connections between observations in different senses to form a mental model of an object or idea in the real world (the word apple, a picture of an apple, the feel of an apple, as shown in figure 3). Integration over sequences provides for a hierarchical or compositional use of atomic sequences to achieve sufficient representation of a concept. Examples of audio hierarchy are phonemes, morphemes, words and sentences. Action analogs of these might be poses, gestures, or actions. Current goals of our lab with respect to the language engine include the introduction of an efficient motor representation to the associative memory, as well as developing a repeatable, self-organizing system (not necessarily based in traditional statistical pattern recognition), for the discovery of modules at differing timescales. A self-similar structure would ideally facilitate such concept discovery, and provide a compact, expandable internal world representation that could be directly manipulated by a behavioral system. IV. MULTI-SCALE MODEL OF ASSOCIATIVE MEMORY To this end, we are currently designing a new multi-scale model based on dynamical systems and neural networks that will serve as the foundation for an associative memory to be implemented within the icub. In the sections below we outline the details at each scale. Furthermore, we mention future work that we plan to complete in the near-term. A. Scale 0: Hodgkin-Huxley Neuron Model The model begins with the classic Hodgkin-Huxley (HH) model of a single neuron [13]. What makes the HH neuron model most useful is the wide range of nonlinear behaviors observed from various inputs. For a detailed summary of these behaviors (tonic spiking, resonator, integrator, etc.) and a comparison of the most widely-used neuron models (including the integrate-and-fire, resonate-and-fire, Izhikevich, and FitzHugh- Nagumo), which justifies its use as a biophysically meaningful neuron model, see [14]. B. Scale 1: Components Components will consist of large populations of HH neurons. Adjacent neurons will obey a Hebbian plasticity rule [15], [16] based on causal synchronous spiking. Specifically, adjacent neurons that fire within a short and biophysically meaningful period of time, τ, will have their connection strengthened, while those neurons that are adjacent to a spiking neuron and do not spike within τ will have their connection weakened (with appropriate directionality considered). The overall state of a component will be determined by a neuronal population coding scheme based on their phase synchrony. For example, suppose that a given population of neurons evolve to have a Scale-Free network structure, one in which the probability that a given vertex has degree k follows a power-law distribution [17], which recent work has suggested is reasonable [18], [17], [19]. Furthermore, suppose that specifically, we have a canonical example- the Barabasi Albert Scale-Free Network Model with 50 vertices. Fig. 4. Barabasi Albert Scale-Free Network Model [17] with N=50 Upon this structure neurons are placed, which for the purpose of illustration (and simplicity) have been idealized as sinusoids. Initially the neurons begin with phase angles randomly sampled from the unit circle (see Figure 2, t=0.01 s). However, over time the neurons become more synchronized. After 1.69 s, a tight angular region defines the boundaries for the phases (see Figure 2, t=1.69 s). This can be accounted for by the structural property that in such networks, certain neurons emerge with very high degree, which means that those within their neighborhood will tend to synchronize. A number of such high degree neurons leads to the range of phase angles observed. The state of the component can be characterized by the variance of the population s phase synchrony. Using this metric we are able to take multi-dimensional, parallel input streams, feed them into a given population and map these to a 2D representation. The state of the different components can be used to uniquely determine a trajectory, which is explained in the next section. C. Scale 2: Memory Encoding Let us assume that we have multiple components. There will be neural pathways, represented by edges, that connect

4 determine the general characteristics of the expected trajectory by examining the cycles in the graph of Figure 10. There is a positive 3-cycle (the product of edge labels is positive), (x 1 x 2 x 3 ), which will ensure multistationarity. There is a negative 2-cycle (the product of edge labels is negative), (x 1 x 3 ), which will ensure stable periodicity. The edge labels for the system are defined as follows: a 11 = x1.5 a 21 = x2 a 31 = x3 = 3x 2 1 a 12 = x1 a 22 = x2 a 32 = x3 = 1 a 13 = x1 a 23 = x2 a 33 = x3 As a result, the system may be captured with the following set of equations: x 1.5x 1 x 3 (1) x 2 = x 1 x 2 (2) x 3 = x 3 1 + x 2 (3) In the traditional way, the steady states are found to be (0, 0, 0), ( 1, 1, 0.5), and (1, 1, 0.5). Using these pieces of information reveals the primary characteristics of the given trajectory, which are confirmed upon simulation (see Figure 4). We provide the graphical results of simulation to show the accuracy of the analytical predictions. Fig. 5. Phase Synchrony at t=0.01, 1.69 s for Barabasi Albert Scale-Free Network Model with N=50 (Note: Radial distance only exists to distinguish neurons.) said components. Each edge will be assigned a value. Loops will be assigned a value that results from a function of the phase synchrony of the population within the component. Edges incident to two different components will be assigned a value that results from a function that compares the relative phase synchrony of the two adjacent components. For the sake of clarity, let us consider a Toy example that consists of 3 components (shown in Figure 3). Before simulating the system, we may determine the qualitative behavior of the system by examining the cycles within Figure 3. Fig. 7. Example memory state for system, t (1, 1000) Such a trajectory we consider a memory state ; it may be viewed as the internal representation of the corresponding external stimulus. The system may be trained such that only those stimuli which have been encoded represent stable states. Sequences of such trajectories may be used to represent more complex memories. In the near-term we hope to fully implement the model using 50,000+ HH neurons per component and at least 3 components (we will restrict the component number in order to aid the dimensionality reduction). This computationally intensive work will be carried out on the Turing Cluster or Blue Waters Supercomputer (once it goes online). Fig. 6. Toy example: 3 components with simple edge weights We may use methods described in [20] that allow one to V. CONCLUSION This paper has given a brief overview of the design philosophy, methods and specific models used in order to create an intelligent agent which is able to manipulate language in the way that we do. At its core is the need for an associative

5 memory to ground semantic concepts in physical observations and experiences. Our lab has recently experienced a great leap forward with our recent acquisition of an icub platform, and now has the opportunity to expand our understanding of what methods can be used to effectively model the function of the human brain with respect to cognition and language. However, with the great increase in platform complexity comes a need for new techniques to handle the range of inputs now possible. Our three main areas of focus now have become developing human-like fine motor abilities, developing a way to represent these abilities internally, and developing new associative memories which will be able to make sense of the large amount of data generated by these new skills. REFERENCES [1] M. McClain, Semantic based learning of syntax in an autonomous robot, Ph.D. dissertation, University of Illinois at Urbana-Champaign, 2006. [2] K. Squire, Hmm-based semantic learning for a mobile robot, Ph.D. dissertation, University of Illinois at Urbana-Champaign, 2004. [3] C. Gaskett and G. Cheng, Online learning of a motor map for humanoid robot reaching, in Proc. of the 2nd Intl. Conf. on Computational Intelligence, Robotics and Autonomous Systems, 2003. [4] L. Rabiner, A tutorial on hidden markov models and selected applications in speech recognition, Proceedings of the IEEE, vol. 77, pp. 257 285, 1989. [5] S. Calinon and A. Billard, Incremental learning of gestures by imitation in a humanoid robot, Proc. of the ACM/IEEE intl. conf. on Human-robot interaction, pp. 255 262, 2007. [6] A. Liberman and I. G. Mattingly, The motor theory of speech perception revised, Cognition, vol. 21, pp. 1 36, 1985. [7] e. a. Cangelosi A., Metta G., Integration of action and language knowledge: A roadmap for developmental robotics, IEEE Transactions on Autonomous Mental Development., (in press). [8] A. Sadeghipour and S. Kopp, A probabilistic model of motor resonance for embodied gesture perception, Proceedings of the 9th International Conference on Intelligent Virtual Agents, 2009. [9] Y. Yamashita and J. Tani, Emergence of functional hierarchy in a multiple timescale neural network model: A humanoid robot experiment, PLoS Computational Biology, vol. 4, 2008. [10] F. Pulvermueller, Brain mechanisms linking language and action, Nature Reviews Neuroscience, vol. 6, p. 576âĂŞ582, 2005. [11] T. Nazir, The role of sensory-motor systems for language understanding, Journal of Psychology - Paris, vol. 102, pp. 1 3, 2008. [12] V. Krishnamurthy and G. Yin, Recursive algorithms for estimation of hidden markov models and autoregressive models with markov regime, IEEE Trans. on Information Theory, vol. 48, pp. 458 476, 2002. [13] A. Hodgkin and A. Huxley, A quantitative description of membrane current and its application to conduction and excitation in nerve, J. Physiology, vol. 117, pp. 500 544, 1952. [14] E. M. Izhikevich, Which model to use for spiking neurons? IEEE Trans. on Neural Networks, vol. 15, no. 5, 2004. [15] D. O. Hebb, The organization of behavior. New York, NY: Wiley, 1949. [16] E. R. Kandel, Small systems of neurons, Sci. Am., vol. 241, no. 3, pp. 66 76, 1979. [17] A.-L. Barabasi and R. Albert, Emergence of scaling in random networks, Science, vol. 286, pp. 509 512, 1999. [18] L. A. et. al., Classes of behavior of small-world networks, Proc. Natl. Acad. Sci., vol. 97, pp. 11 149 11 152, 2000. [19] A. Barrat and M. Weigt, On the properties of small-world network models, Eur. Phys. J., vol. 13, pp. 547 560, 2000. [20] R. Thomas and M. Kaufman, Conceptual tools for the integration of data, C. R. Bio., vol. 325, pp. 505 514, 2002.