TY - JOUR
T1 - Acquisition of viewpoint transformation and action mappings via sequence to sequence imitative learning by deep neural networks
AU - Nakajo, Ryoichi
AU - Murata, Shingo
AU - Arie, Hiroaki
AU - Ogata, Tetsuya
N1 - Funding Information:
This work was supported by the JST, CREST Grant Number JPMJCR15E3, and JSPS KAKENHI Grant Numbers 15H01710 and 16H05878.
Publisher Copyright:
Copyright © 2018 Nakajo, Murata, Arie and Ogata.
PY - 2018
Y1 - 2018
N2 - We propose an imitative learning model that allows a robot to acquire positional relations between the demonstrator and the robot, and to transform observed actions into robotic actions. Providing robots with imitative capabilities allows us to teach novel actions to them without resorting to trial-and-error approaches. Existing methods for imitative robotic learning require mathematical formulations or conversion modules to translate positional relations between demonstrators and robots. The proposed model uses two neural networks, a convolutional autoencoder (CAE) and a multiple timescale recurrent neural network (MTRNN). The CAE is trained to extract visual features from raw images captured by a camera. The MTRNN is trained to integrate sensory-motor information and to predict next states. We implement this model on a robot and conducted sequence to sequence learning that allows the robot to transform demonstrator actions into robot actions. Through training of the proposedmodel, representations of actions, manipulated objects, and positional relations are formed in the hierarchical structure of the MTRNN. After training, we confirm capability for generating unlearned imitative patterns.
AB - We propose an imitative learning model that allows a robot to acquire positional relations between the demonstrator and the robot, and to transform observed actions into robotic actions. Providing robots with imitative capabilities allows us to teach novel actions to them without resorting to trial-and-error approaches. Existing methods for imitative robotic learning require mathematical formulations or conversion modules to translate positional relations between demonstrators and robots. The proposed model uses two neural networks, a convolutional autoencoder (CAE) and a multiple timescale recurrent neural network (MTRNN). The CAE is trained to extract visual features from raw images captured by a camera. The MTRNN is trained to integrate sensory-motor information and to predict next states. We implement this model on a robot and conducted sequence to sequence learning that allows the robot to transform demonstrator actions into robot actions. Through training of the proposedmodel, representations of actions, manipulated objects, and positional relations are formed in the hierarchical structure of the MTRNN. After training, we confirm capability for generating unlearned imitative patterns.
KW - Deep neural networks
KW - Human-robot interaction
KW - Imitative learning
KW - Recurrent neural networks
KW - Sequence to sequence learning
UR - http://www.scopus.com/inward/record.url?scp=85074134808&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85074134808&partnerID=8YFLogxK
U2 - 10.3389/fnbot.2018.00046
DO - 10.3389/fnbot.2018.00046
M3 - Article
AN - SCOPUS:85074134808
SN - 1662-5218
VL - 12
JO - Frontiers in Neurorobotics
JF - Frontiers in Neurorobotics
M1 - 46
ER -