MIRROR NEURONS AND THE ACTION THEORY OF LANGUAGE ORIGINS

Similar documents
AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Evolution of Symbolisation in Chimpanzees and Neural Nets

DIGITAL GAMING & INTERACTIVE MEDIA BACHELOR S DEGREE. Junior Year. Summer (Bridge Quarter) Fall Winter Spring GAME Credits.

Seminar - Organic Computing

Psychology and Language

The Thinking Hand: Embodiment of Tool Use, Social Cognition and Metaphorical Thinking and Implications for Learning Design

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

BUILD-IT: Intuitive plant layout mediated by natural interaction

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

The Mirror System, Imitation, and the Evolution of Language DRAFT: December 10, 1999

Saliency in Human-Computer Interaction *

Automating the E-learning Personalization

Emotional Variation in Speech-Based Natural Language Generation

A Case-Based Approach To Imitation Learning in Robotic Agents

What s in a Step? Toward General, Abstract Representations of Tutoring System Log Data

EUROPEAN UNIVERSITIES LOOKING FORWARD WITH CONFIDENCE PRAGUE DECLARATION 2009

DICTE PLATFORM: AN INPUT TO COLLABORATION AND KNOWLEDGE SHARING

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance

Grounding Language for Interactive Task Learning

Knowledge-Based - Systems

Rover Races Grades: 3-5 Prep Time: ~45 Minutes Lesson Time: ~105 minutes

Conversation Starters: Using Spatial Context to Initiate Dialogue in First Person Perspective Games

MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE

Accelerated Learning Course Outline

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

An Embodied Model for Sensorimotor Grounding and Grounding Transfer: Experiments With Epigenetic Robots

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Some Principles of Automated Natural Language Information Extraction

Accelerated Learning Online. Course Outline

Emergent Narrative As A Novel Framework For Massively Collaborative Authoring

While you are waiting... socrative.com, room number SIMLANG2016

Circuit Simulators: A Revolutionary E-Learning Platform

English Language and Applied Linguistics. Module Descriptions 2017/18

A THEORETICAL FRAMEWORK FORA TASK-BASED SYLLABUS FOR PRIMARY SCHOOLS IN SOUTH AFRICA

RWTH Aachen University

Learning Prospective Robot Behavior

SPATIAL SENSE : TRANSLATING CURRICULUM INNOVATION INTO CLASSROOM PRACTICE

AQUA: An Ontology-Driven Question Answering System

Designing Autonomous Robot Systems - Evaluation of the R3-COP Decision Support System Approach

Laboratorio di Intelligenza Artificiale e Robotica

Jacqueline C. Kowtko, Patti J. Price Speech Research Program, SRI International, Menlo Park, CA 94025

DISTANCE LEARNING OF ENGINEERING BASED SUBJECTS: A CASE STUDY. Felicia L.C. Ong (author and presenter) University of Bradford, United Kingdom

Ben Kokkeler University of Twente 10 th September 2015 HEIR Network Conference University of the West of Scotland, Paisley

Building Student Understanding and Interest in Science through Embodied Experiences with LEGO Robotics

NAME: East Carolina University PSYC Developmental Psychology Dr. Eppler & Dr. Ironsmith

Axiom 2013 Team Description Paper

A MULTI-AGENT SYSTEM FOR A DISTANCE SUPPORT IN EDUCATIONAL ROBOTICS

Psychology of Speech Production and Speech Perception

Integrating Meta-Level and Domain-Level Knowledge for Task-Oriented Dialogue

UC Merced Proceedings of the Annual Meeting of the Cognitive Science Society

ECE-492 SENIOR ADVANCED DESIGN PROJECT

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Laboratorio di Intelligenza Artificiale e Robotica

An Interactive Intelligent Language Tutor Over The Internet

Artificial Neural Networks

Litterature review of Soft Systems Methodology

Modeling user preferences and norms in context-aware systems

Organizing Comprehensive Literacy Assessment: How to Get Started

Developing a TT-MCTAG for German with an RCG-based Parser

Candidates must achieve a grade of at least C2 level in each examination in order to achieve the overall qualification at C2 Level.

Learning Methods for Fuzzy Systems

AC : DESIGNING AN UNDERGRADUATE ROBOTICS ENGINEERING CURRICULUM: UNIFIED ROBOTICS I AND II

Artificial Neural Networks written examination

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

Visual CP Representation of Knowledge

Understanding and Supporting Dyslexia Godstone Village School. January 2017

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Multidisciplinary Engineering Systems 2 nd and 3rd Year College-Wide Courses

Document number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering

A basic cognitive system for interactive continuous learning of visual concepts

Think A F R I C A when assessing speaking. C.E.F.R. Oral Assessment Criteria. Think A F R I C A - 1 -

File # for photo

Agent-Based Software Engineering

ASTEN Fellowship report Priscilla Gaff Program Coordinator Life Science

SAM - Sensors, Actuators and Microcontrollers in Mobile Robots

INSPIRE A NEW GENERATION OF LIFELONG LEARNERS

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

CS 598 Natural Language Processing

THE DEPARTMENT OF DEFENSE HIGH LEVEL ARCHITECTURE. Richard M. Fujimoto

City University of Hong Kong Course Syllabus. offered by Department of Architecture and Civil Engineering with effect from Semester A 2017/18

USER ADAPTATION IN E-LEARNING ENVIRONMENTS

arxiv: v2 [cs.ro] 3 Mar 2017

TC The Power of Non Formal Education 2014

Unit 7 Data analysis and design

Building A Baby. Paul R. Cohen, Tim Oates, Marc S. Atkin Department of Computer Science

Initial English Language Training for Controllers and Pilots. Mr. John Kennedy École Nationale de L Aviation Civile (ENAC) Toulouse, France.

PRODUCT COMPLEXITY: A NEW MODELLING COURSE IN THE INDUSTRIAL DESIGN PROGRAM AT THE UNIVERSITY OF TWENTE

EOSC Governance Development Forum 4 May 2017 Per Öster

Applications of memory-based natural language processing

COMPUTER-AIDED DESIGN TOOLS THAT ADAPT

Emma Kushtina ODL organisation system analysis. Szczecin University of Technology

Spinal Cord. Student Pages. Classroom Ac tivities

ARCHITECTURES FOR COSY ROBOTS

Organizational Knowledge Distribution: An Experimental Evaluation

Word Segmentation of Off-line Handwritten Documents

Strategy Study on Primary School English Game Teaching

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Learning or lurking? Tracking the invisible online student

Concept Acquisition Without Representation William Dylan Sabo

Transcription:

MIRROR NEURONS AND THE ACTION THEORY OF LANGUAGE ORIGINS Luc Steels Vrije Universiteit Brussel (AI Lab) and Sony Computer Science Laboratory Paris 6 Rue Amyot tel: 33-1-44 08 05 05 fax: 33-1-45-87-87-50 email: steels@arti.vub.ac.be 75005 Paris a. Introduction and Goal The research reported here attempts to understand how language may have originated from sensori-motor competences. Recently the observation of mirror neurons [1] has lead to the suggestion that there is not only a rich representation of motor action but also that this representation is used for multiple purposes: action execution, action planning, action imaging, and action recognition. Of particular importance is the observation that one agent can recognise an action plan of another one and that the same neurons are involved. The relevance of this for the origins of language has been pointed out by Rizzolatti and Arbib [2]. Here we go a step further, arguing that the meaning of a language utterance in general is a series of physical or mental actions that the speaker wants the hearer to perform, rather than a declarative statement to be stored whose only relevance are its truth conditions. For example, when a speaker says "Can you give me the black box on the table?", he wants the hearer to hand over an object (which means to grasp it and move it in the direction of the speaker). To know which object is involved, the speaker wants the hearer to direct his or her attention to a table in the shared context, to identify the objects which can be compared to the prototype of a box, and then focus on the one box which has a black colour. These mental actions are as situated and grounded as motor actions like grasping. From this action-oriented view of language semantics, language understanding amounts to the recognition of the plan intended by the hearer and the utterance is seen as giving hints about which plan is intended. The production of an utterance can also be seen as involving the contruction of an action plan and thus parsing amounts to the recognition of which production plans have been used by the speaker. So the production of an utterance, both the conceptualisation of what to say and the decision on how to say it, can be viewed as the planning of a series of actions, and the interpretation of an utterance can be seen

as the recognition of these action patterns and their subsequent execution. Taking this point of view has two important implications: (1) It helps to understand how language might have originated. If the mechanisms required for language are essentially the same as those required for motor planning, execution, and recognition, then it is less a mystery how homo sapiens could have started to evolve language. We no longer need a scenario based on genetic mutations (as in [5]) but can assume a pre-adaptation scenario, in which existing brain structures and processes became used for language communication. (2) It leads to a greater overall economy of the human cognitive system because fewer special-purpose components (like a dedicated language organ) are needed. To demonstrate the theoretical viability of this thesis we have to show that the same representational framework is adequate for sensorimotor behavior and both for the conceptualisation and interpretation of utterances and for the verbal behavior itself (the production and recognition of utterances). We also have to show that the same learning mechanisms are involved. This is obviously a very non-trivial exercise given the complexity involved. b. Materials and methods So far, we have been developing formal models and have conducted computer simulations and experiments with physical robots to test them. The robots have a sensori-motor layer for executing autonomous behaviors, and a fully integrated cognitive layer for planning, memory, and communication. The robot bodies in our experiments range in complexity from steerable cameras [3] to small mobile robots, animallike robots (specifically the dog-shaped SONY AIBO), and humanoid torsos. The robots play language games, either among themselves or with a human player. Each language game is a situated interaction between at least two agents about something in their shared environment. It involves perception, conceptualisation, communication, interpretation, and action. An example game that we have used extensively is the guessing game, in which the speaker draws the attention of the hearer to an object in the shared reality by verbal means [3]. In one large-scale experiment, a growing population of close to 3000 (virtual) agents was employed which

used the (real) robot bodies to engage in guessing games about scenes consisting of geometrical figures on a white board in front of them. Another example game that we have used extensively is the "Where-Is- It?" game, in which agents locate objects based on a spatial map acquired by exploring and remembering the environment and verbal suggestions of a path to follow. The first step in our research has been to operationalise a representational framework of actions and action plans in the form of schemata. Each schema has a number of slots, constraints on each slot, and an action plan in the form of augmented finite state machines. The automaton schedules and de-schedules sensori-motor behaviors and moves from one state to another based on success or failure in behavior execution. The constraints are maintained by propagating information as fast as it becomes available using data-flow computation. A schema may itself be a specialisation of a more abstract schema and may call upon other schemas. This representational framework was demonstrated to be adequate for the actual high level control of grounded robotic behaviors. We have also developed a learning system capable to acquire new motor schemata by the exploration of a search space of possible concatentations of the primitive actions and by a chunking of successful paths. Our second step has been to use the same framework to plan the meaning of natural language utterances as needed for language games. The primitive actions in this case are operations over cognitive spaces, such as filtering a set into a subset, shifting the focus of attention from one object to another, or ordering the members of a set into a sequence and retrieving the first member. These conceptual schemata are tightly coupled with the sensori-motor layer in the sense that the information items and facts used by them have all been deposited in memory by sensori-motor behaviors and are continuously upgraded by them. The third step has been to use the same framework for the execution and the recognition of the utterances themselves. The primitive actions of verbal schemata center on the production (or recognition) of parts of utterances in a specific order and on the realisation of suprasegmental modulations such as prosody and stress patterns. The planning of verbal behavior is itself a highly complex process and known to be distinct from the actual execution of the plan. We are interested in natural dialogues which are highly situated in the specific interaction context of speaker and hearer, with many false starts, hesitations, irrelevant words, etc. This makes verbal behavior much closer to sensori-motor

behavior than is usually assumed, particularly by linguistic theories that exclusively lok at "clean" written language. Finally we have developed a two-way associative memory that is mapping conceptual schemata to verbal schemata. While parsing an utterance the hearer must recognise which verbal schemata wre involved and map them to the conceptual schemata that could have been intended by the speaker. While producing an utterance the speaker must conceptualise what he wants to say in terms of conceptual schemata and map them onto verbal schemata that constitute a plan for how to express the meaning. We have been experimenting with memory-based learning techniques to gradually build up the repertoire of form-meaning mappings [4]. c. Results At this point we have been able to demonstrate the complete architecture on autonomous robotic agents. For example, in the largescale experiment alluded to earlier [3], we have observed that a stable communication system based on a vocabulary of a few thousand words indeed emerges and is maintained in the population even if new members continuously enter or leave the system. A self-organising semiotic dynamics has been observed damping synonymy and polysemy due to a positive feedback loop between use and success [5]. These grammatical forms express the conceptual plans made by the speaker and recognised by the hearer. Even though a vast amount of work is still required to enrich the schema repertoires by the addition of more primitive actions and by integrating more complex learning mechanisms, we can say that based on the results so far the original thesis has gained in plausibility. The planning and plan execution mechanisms required for sensori-motor behavior can form the basis of language. d. Conclusions The Action Theory of the origins of language argues that there is a very tight analogy between the ability to plan and recognise a motor action and the ability to plan and recognise an utterance, both its content (what to say) and its form (how to say it). Our research is developing in full detail this analogy by operationalising it on physical robots. We believe that such experimentation is complementary to neurobiological observation and a potentially rich source for detailed models of human verbal behavior.

e. References [1] Gallese, V., L. Fadiga, L. Fogassi, G. Rizzolatti (1996) Action recognition in the premotor cortex. Brain 119:593-609. [2] Rizzolatti, G. and M. Arbib (1998) Language within our grasp. Trends Neuroscience. 21:188-194. [3] Steels, L. (1998) The origins of syntax in visually grounded robotic agents. Artificial Intelligence 103 (1-2), 133-156. [4] Steels, L. (2000) The Emergence of Grammar in Communicating Robots. In: Proceedings of the European Conference on AI, Berlin. IOS Press, Amsterdam. [5] Pinker, S. and P. Bloom (1990) Natural Language and Natural Selection. Behavioral and Brain Sciences, 13, 707-784.