• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Multiple paired forward and inverse models for motor control. Neural Networks (1998)

by D M Wolpert, M Kawato
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 394
Next 10 →

The neural basis of human error processing: Reinforcement learning, dopamine, and the error-related negativity

by Clay B. Holroyd, Michael G. H. Coles - PSYCHOLOGICAL REVIEW 109:679–709 , 2002
"... The authors present a unified account of 2 neural systems concerned with the development and expression of adaptive behaviors: a mesencephalic dopamine system for reinforcement learning and a “generic ” error-processing system associated with the anterior cingulate cortex. The existence of the error ..."
Abstract - Cited by 430 (20 self) - Add to MetaCart
The authors present a unified account of 2 neural systems concerned with the development and expression of adaptive behaviors: a mesencephalic dopamine system for reinforcement learning and a “generic ” error-processing system associated with the anterior cingulate cortex. The existence of the error-processing system has been inferred from the error-related negativity (ERN), a component of the event-related brain potential elicited when human participants commit errors in reaction-time tasks. The authors propose that the ERN is generated when a negative reinforcement learning signal is conveyed to the anterior cingulate cortex via the mesencephalic dopamine system and that this signal is used by the anterior cingulate cortex to modify performance on the task at hand. They provide support for this proposal using both computational modeling and psychophysiological experimentation. Human beings learn from the consequences of their actions. Thorndike (1911/1970) originally described this phenomenon with his law of effect, which made explicit the commonsense notion that actions that are followed by feelings of satisfaction are more likely to be generated again in the future, whereas actions that are followed by negative outcomes are less likely to reoccur. This
(Show Context)

Citation Context

...n which a single motor controller is given responsibility over the motor system. Instead, more sophisticated algorithms exist for blending the output of several modules (e.g., Whitehead et al., 1993; =-=Wolpert & Kawato, 1998-=-). A future simulation could adopt such an algorithm, with the anterior cingulate cortex combining the output of several motor controllers to guide the response selection process. Moreover, 2 the “tem...

Learning to Perceive the World as Articulated: An Approach for Hierarchical Learning in Sensory-Motor Systems

by Jun Tani, Stefano Nolfi - NEURAL NETWORKS , 1999
"... This paper describes how agents can learn an internal model of the world structurally by focusing on the problem of behavior-based articulation. We develop an on-line learning scheme -- the so-called mixture of recurrent neural net (RNN) experts -- in which a set of RNN modules becomes self-organ ..."
Abstract - Cited by 141 (31 self) - Add to MetaCart
This paper describes how agents can learn an internal model of the world structurally by focusing on the problem of behavior-based articulation. We develop an on-line learning scheme -- the so-called mixture of recurrent neural net (RNN) experts -- in which a set of RNN modules becomes self-organized as experts on multiple levels in order to account for the different categories of sensory-motor flow which the robot experiences. Autonomous switching of activated modules in the lower level actually represents the articulation of the sensory-motor flow. In the meanwhile, a set of RNNs in the higher level competes to learn the sequences of module switching in the lower level, by which articulation at a further more abstract level can be achieved. The proposed scheme was examined through simulation experiments involving the navigation learning problem. Our dynamical systems analysis clarified the mechanism of the articulation; the possible correspondence between the articulation...

The case for motor involvement in perceiving conspecifics

by Margaret Wilson, Günther Knoblich - Psychological Bulletin , 2005
"... Perceiving other people’s behaviors activates imitative motor plans in the perceiver, but there is disagreement as to the function of this activation. In contrast to other recent proposals (e.g., that it subserves overt imitation, identification and understanding of actions, or working memory), here ..."
Abstract - Cited by 137 (2 self) - Add to MetaCart
Perceiving other people’s behaviors activates imitative motor plans in the perceiver, but there is disagreement as to the function of this activation. In contrast to other recent proposals (e.g., that it subserves overt imitation, identification and understanding of actions, or working memory), here it is argued that imitative motor activation feeds back into the perceptual processing of conspecifics’ behaviors, generating top-down expectations and predictions of the unfolding action. Furthermore, this account incorporates recent ideas about emulators in the brain—mental simulations that run in parallel to the external events they simulate—to provide a mechanism by which motoric involvement could contribute to perception. Evidence from a variety of literatures is brought to bear to support this account of perceiving human body movement.
(Show Context)

Citation Context

... real time. Note that this account presupposes that motor knowledge is used simultaneously for emulating and planning. The highly parallel and modular architecture of current models of motor control (=-=Wolpert & Kawato, 1998-=-) makes such an assumption plausible. Such an account would predict that interference occurs only if emulating and planning access the same internal models of a given movement (cf. Hamilton et al., 20...

Imitation as a dual-route process featuring predictive and learning components: a biologically plausible computational model

by Yiannis Demiris, Gillian Hayes , 2002
"... ..."
Abstract - Cited by 137 (7 self) - Add to MetaCart
Abstract not found

Abnormalities in the Awareness and Control of Action

by Christopher D. Frith, Sarah-jayne Blakemore, Daniel M. Wolpert - Philos Trans R Soc Lond B Biol Sci , 2000
"... this paper is to consider the extent to which we are aware of the functioning of some ..."
Abstract - Cited by 111 (2 self) - Add to MetaCart
this paper is to consider the extent to which we are aware of the functioning of some
(Show Context)

Citation Context

... in a variety of di¡erent environments relies on a `divide-and-conquer’ strategy. Complex tasks are decomposed into simpler subtasks, each learned by a separate controller (Ghahramani & Wolpert 1997; =-=Wolpert & Kawato 1998-=-; Blakemore et al. 1998a). Therefore, rather than having a single controller, multiple controllers develop, each tuned to a particular sensorimotor context. At any given time, one or a subset of these...

Emergence of functional hierarchy in a multiple timescale neural network model: A humanoid robot experiment”, PLoS

by Yuichi Yamashita, Jun Tani - Computational Biology
"... It is generally thought that skilled behavior in human beings results from a functional hierarchy of the motor control system, within which reusable motor primitives are flexibly integrated into various sensori-motor sequence patterns. The underlying neural mechanisms governing the way in which cont ..."
Abstract - Cited by 93 (15 self) - Add to MetaCart
It is generally thought that skilled behavior in human beings results from a functional hierarchy of the motor control system, within which reusable motor primitives are flexibly integrated into various sensori-motor sequence patterns. The underlying neural mechanisms governing the way in which continuous sensori-motor flows are segmented into primitives and the way in which series of primitives are integrated into various behavior sequences have, however, not yet been clarified. In earlier studies, this functional hierarchy has been realized through the use of explicit hierarchical structure, with local modules representing motor primitives in the lower level and a higher module representing sequences of primitives switched via additional mechanisms such as gate-selecting. When sequences contain similarities and overlap, however, a conflict arises in such earlier models between generalization and segmentation, induced by this separated modular structure. To address this issue, we propose a different type of neural network model. The current model neither makes use of separate local modules to represent primitives nor introduces explicit hierarchical structure. Rather than forcing architectural hierarchy onto the system, functional hierarchy emerges through a form of self-organization that is based on two distinct types of neurons, each with different time properties (‘‘multiple timescales’’). Through the introduction of multiple timescales, continuous sequences of behavior are segmented into reusable primitives, and the primitives, in turn, are flexibly integrated into novel sequences. In experiments, the proposed network model, coordinating the physical body of a humanoid robot through
(Show Context)

Citation Context

...preservesthe internal state, which enables it to reproduce complex dynamics. It is thought thatsthis type of prediction mechanism is realized in animal and human brains through assimple forward model =-=[24, 25]-=-.sIn the CTRNN, proprioception and vision inputs were sparsely encoded with thespreserving topology of the input space (see Method for details). This topology preservingssparse encoding of sensori-mot...

Multiple model-based reinforcement learning

by Kenji Doya, Kazuyuki Samejima - Neural Computation , 2002
"... We propose a modular reinforcement learning architecture for non-linear, non-stationary control tasks, which we call multiple model-based reinforcement learn-ing (MMRL). The basic idea is to decompose a complex task into multiple domains in space and time based on the predictability of the environme ..."
Abstract - Cited by 85 (5 self) - Add to MetaCart
We propose a modular reinforcement learning architecture for non-linear, non-stationary control tasks, which we call multiple model-based reinforcement learn-ing (MMRL). The basic idea is to decompose a complex task into multiple domains in space and time based on the predictability of the environmental dynamics. The 1 system is composed of multiple modules, each of which consists of a state predic-tion model and a reinforcement learning controller. The “responsibility signal,” which is given by the softmax function of the prediction errors, is used to weight the outputs of multiple modules as well as to gate the learning of the predic-tion models and the reinforcement learning controllers. We formulate MMRL for both discrete-time, finite state case and continuous-time, continuous state case. The performance of MMRL was demonstrated for discrete case in a non-stationary hunting task in a grid world and for continuous case in a non-linear, non-stationary control task of swinging up a pendulum with variable physical parameters. 1
(Show Context)

Citation Context

...uence prediction. The use of the softmax function for module selection and combination was originally proposed for a tracking control paradigm as the “Multiple Paired Forward-Inverse Models (MPFIM)” (=-=Wolpert and Kawato 1998-=-; Wolpert et al. 1998; Haruno et al. 1999). It was recently reformulated as “MOdular Selection and Identification for Control (MOSAIC)” (Wolpert and Ghahramani 2000). In this paper, we apply the idea ...

Hierarchical attentive multiple models for execution and recognition of actions

by Yiannis Demiris, Bassam Khadhouri - ROBOTICS AND AUTONOMOUS SYSTEMS , 2005
"... According to the motor theories of perception, the motor systems of an observer are actively involved in the perception of actions when these are performed by a demonstrator. In this paper we review our computational architecture, HAMMER (Hierarchical Attentive Multiple Models for Execution and Reco ..."
Abstract - Cited by 84 (20 self) - Add to MetaCart
According to the motor theories of perception, the motor systems of an observer are actively involved in the perception of actions when these are performed by a demonstrator. In this paper we review our computational architecture, HAMMER (Hierarchical Attentive Multiple Models for Execution and Recognition), where the motor control systems of a robot are organised in a hierarchical, distributed manner, and can be used in the dual role of (a) competitively selecting and executing an action, and (b) perceiving it when perfomed by a demonstrator. We subsequently demonstrate that such arrangement can provide a principled method for the top-down control of attention during action perception, resulting in significant performance gains. We assess these performance gains under a variety of resource allocation strategies.

Learning from Observation Using Primitives

by Darrin C. Bentivegna, Christopher G. Atkeson , 2004
"... This paper describes the use of task primitives in robot learning from observation. A framework has been developed that uses observed data to initially learn a task and then the agent goes on to increase its performance through repeated task performance (learning from practice). Data that is collect ..."
Abstract - Cited by 70 (5 self) - Add to MetaCart
This paper describes the use of task primitives in robot learning from observation. A framework has been developed that uses observed data to initially learn a task and then the agent goes on to increase its performance through repeated task performance (learning from practice). Data that is collected while a human performs a task is parsed into small parts of the task called primitives. Modules are created for each primitive that encode the movements required during the performance of the primitive, and when and where the primitives are performed. The feasibility of this method is currently being tested with agents that learn to play a virtual and an actual air hockey game. 1
(Show Context)

Citation Context

...ing these robots using learning from observation methods and/or using primitives [13, 34, 37, 69, 112]. Primitives in these robotic systems are mostly being explored at two levels; task [1] and motor =-=[133, 114]-=-. Task-level primitives are specialized to the task that needs to be performed and the underlying hardware may or may not be taken into account. Motor-level primitives provide the robot with basic mov...

Challenge Point: A Framework for Conceptualizing the Effects of Various Practice Conditions in Motor Learning

by Mark A. Guadagnoli, Timothy D. Lee
"... ABSTRACT. The authors describe the effects of practice conditions in motor learning (e.g., contextual interference, knowledge of results) within the constraints of 2 experimental variables: skill level and task difficulty. They use a research framework to conceptualize the interaction of those varia ..."
Abstract - Cited by 70 (0 self) - Add to MetaCart
ABSTRACT. The authors describe the effects of practice conditions in motor learning (e.g., contextual interference, knowledge of results) within the constraints of 2 experimental variables: skill level and task difficulty. They use a research framework to conceptualize the interaction of those variables on the basis of concepts from information theory and information processing. The fundamental idea is that motor tasks represent different challenges for performers of different abilities. The authors propose that learning is related to the information arising from performance, which should be optimized along functions relating the difficulty of the task to the skill level of the performer. Specific testable hypotheses arising from the framework are also described. Key words: augmented feedback, contextual interference, motor learning P ractice is generally considered to be the single most important factor responsible for the permanent improvement in the ability to perform a motor skill (i.e.,
(Show Context)

Citation Context

...s for the anticipation of feedback sensations. The action plan works in a manner suggested by inverse and forward internal models of motor control and learning (Wolpert, Ghahramani, & Flanagan, 2001; =-=Wolpert & Kawato, 1998-=-). According to that theory, the processing system uses pairs of inverse (motor predictor) and forward (sensory predictor) internal models to generate output signals, and it uses sensory feedback as t...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University