Intrinsic motivation and mental replay enable efficient online adaptation in stochastic recurrent networks

Research output: Contribution to journalArticleResearchpeer-review

Standard

Intrinsic motivation and mental replay enable efficient online adaptation in stochastic recurrent networks. / Tanneberg, Daniel; Peters, Jan; Rueckert, Elmar.
In: Neural networks, Vol. 109.2019, No. January, 01.2019, p. 67-80.

Research output: Contribution to journalArticleResearchpeer-review

Vancouver

Tanneberg D, Peters J, Rueckert E. Intrinsic motivation and mental replay enable efficient online adaptation in stochastic recurrent networks. Neural networks. 2019 Jan;109.2019(January):67-80. Epub 2018 Oct 22. doi: 10.1016/j.neunet.2018.10.005

Author

Bibtex - Download

@article{fbcf6f8f9a4b4b31a86fbd2c07494c46,
title = "Intrinsic motivation and mental replay enable efficient online adaptation in stochastic recurrent networks",
abstract = "Autonomous robots need to interact with unknown, unstructured and changing environments, constantly facing novel challenges. Therefore, continuous online adaptation for lifelong-learning and the need of sample-efficient mechanisms to adapt to changes in the environment, the constraints, the tasks, or the robot itself are crucial. In this work, we propose a novel framework for probabilistic online motion planning with online adaptation based on a bio-inspired stochastic recurrent neural network. By using learning signals which mimic the intrinsic motivation signal cognitive dissonance in addition with a mental replay strategy to intensify experiences, the stochastic recurrent network can learn from few physical interactions and adapts to novel environments in seconds. We evaluate our online planning and adaptation framework on an anthropomorphic KUKA LWR arm. The rapid online adaptation is shown by learning unknown workspace constraints sample-efficiently from few physical interactions while following given way points.",
keywords = "Autonomous robots, Experience replay, Intrinsic motivation, Neural sampling, Online learning, Spiking recurrent networks",
author = "Daniel Tanneberg and Jan Peters and Elmar Rueckert",
note = "Publisher Copyright: {\textcopyright} 2018 Elsevier Ltd",
year = "2019",
month = jan,
doi = "10.1016/j.neunet.2018.10.005",
language = "English",
volume = "109.2019",
pages = "67--80",
journal = "Neural networks",
issn = "0893-6080",
publisher = "Elsevier",
number = "January",

}

RIS (suitable for import to EndNote) - Download

TY - JOUR

T1 - Intrinsic motivation and mental replay enable efficient online adaptation in stochastic recurrent networks

AU - Tanneberg, Daniel

AU - Peters, Jan

AU - Rueckert, Elmar

N1 - Publisher Copyright: © 2018 Elsevier Ltd

PY - 2019/1

Y1 - 2019/1

N2 - Autonomous robots need to interact with unknown, unstructured and changing environments, constantly facing novel challenges. Therefore, continuous online adaptation for lifelong-learning and the need of sample-efficient mechanisms to adapt to changes in the environment, the constraints, the tasks, or the robot itself are crucial. In this work, we propose a novel framework for probabilistic online motion planning with online adaptation based on a bio-inspired stochastic recurrent neural network. By using learning signals which mimic the intrinsic motivation signal cognitive dissonance in addition with a mental replay strategy to intensify experiences, the stochastic recurrent network can learn from few physical interactions and adapts to novel environments in seconds. We evaluate our online planning and adaptation framework on an anthropomorphic KUKA LWR arm. The rapid online adaptation is shown by learning unknown workspace constraints sample-efficiently from few physical interactions while following given way points.

AB - Autonomous robots need to interact with unknown, unstructured and changing environments, constantly facing novel challenges. Therefore, continuous online adaptation for lifelong-learning and the need of sample-efficient mechanisms to adapt to changes in the environment, the constraints, the tasks, or the robot itself are crucial. In this work, we propose a novel framework for probabilistic online motion planning with online adaptation based on a bio-inspired stochastic recurrent neural network. By using learning signals which mimic the intrinsic motivation signal cognitive dissonance in addition with a mental replay strategy to intensify experiences, the stochastic recurrent network can learn from few physical interactions and adapts to novel environments in seconds. We evaluate our online planning and adaptation framework on an anthropomorphic KUKA LWR arm. The rapid online adaptation is shown by learning unknown workspace constraints sample-efficiently from few physical interactions while following given way points.

KW - Autonomous robots

KW - Experience replay

KW - Intrinsic motivation

KW - Neural sampling

KW - Online learning

KW - Spiking recurrent networks

UR - http://www.scopus.com/inward/record.url?scp=85056187742&partnerID=8YFLogxK

UR - https://cps.unileoben.ac.at/wp/NeuralNetworks2018Tanneberg.pdf

U2 - 10.1016/j.neunet.2018.10.005

DO - 10.1016/j.neunet.2018.10.005

M3 - Article

C2 - 30408695

AN - SCOPUS:85056187742

VL - 109.2019

SP - 67

EP - 80

JO - Neural networks

JF - Neural networks

SN - 0893-6080

IS - January

ER -