Autonomy is fundamental for artificial agents acting in complex real-world scenarios. The acquisition of many differentskills is pivotal to foster versatile autonomous behaviour and thus a main objective for robotics and machine learning.Intrinsic motivations have proven to properly generate a task-agnostic signal to drive the autonomous acquisition ofmultiple policies in settings requiring the learning of multiple tasks. However, in real-world scenarios tasks may beinterdependent so that some of them may constitute the precondition for learning other ones. Despite different strategieshave been used to tackle the acquisition of interdependent/hierarchical tasks, fully autonomous open-ended learning inthese scenarios is still an open question. Building on previous research within the framework of intrinsically-motivated open-ended learning, we propose an architecture for robot control that tackles this problem from the point of view ofdecision making, i.e. treating the selection of tasks as a Markov Decision Process where the system selects the policiesto be trained in order to maximise its competence over all the tasks. The system is then tested with a humanoid robotsolving interdependent multiple reaching tasks.

Autonomous Open-Ended Learning of Interdependent Tasks

Vieri Giuliano Santucci;Emilio Cartoni;Gianluca Baldassarre
2019

Abstract

Autonomy is fundamental for artificial agents acting in complex real-world scenarios. The acquisition of many differentskills is pivotal to foster versatile autonomous behaviour and thus a main objective for robotics and machine learning.Intrinsic motivations have proven to properly generate a task-agnostic signal to drive the autonomous acquisition ofmultiple policies in settings requiring the learning of multiple tasks. However, in real-world scenarios tasks may beinterdependent so that some of them may constitute the precondition for learning other ones. Despite different strategieshave been used to tackle the acquisition of interdependent/hierarchical tasks, fully autonomous open-ended learning inthese scenarios is still an open question. Building on previous research within the framework of intrinsically-motivated open-ended learning, we propose an architecture for robot control that tackles this problem from the point of view ofdecision making, i.e. treating the selection of tasks as a Markov Decision Process where the system selects the policiesto be trained in order to maximise its competence over all the tasks. The system is then tested with a humanoid robotsolving interdependent multiple reaching tasks.
2019
Istituto di Scienze e Tecnologie della Cognizione - ISTC
Interdependent Tasks
Hierarchical Skill Learning
Reinforcement Learning
Autonomous Robotics
Intrinsic Motivations
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.14243/366818
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact