This paper proposes a new open-ended learning framework which aims at implementing an autonomous agent using intrinsic motivations (IM) at two different levels. At the first level, the IM paradigm is exploited by the agent to learn new operational skills, described in terms of sub-symbolic options. After discovering the options, the agent iteratively: (1) executes them to explore the world, collecting the necessary data and (2) automatically abstracts the collected data into a high-level representation of the domain, expressed in PPDDL language. At the second level, the IM paradigm is used to exploit the abstracted representation of the domain by identifying particular symbolic states deemed promising according to a specific criterium, which in the present work is the farthest distance covered by the agent (i.e., the most promising states are those that rest at the frontier of the visited space). Once these states are identified, they can be successively reached through an internally generated high-level plan and used as promising starting points for discovering new knowledge. The presented framework is tested in the so-called Treasure Game domain described in the recent literature. The tests we have performed show that the proposed idea of implementing intrinsic motivations at two different levels of abstraction facilitates the discovery of new knowledge, compared to a previous approach proposed in the literature.

Intrinsically Motivated High-Level Planning for Agent Exploration

Oddi, Angelo;Rasconi, Riccardo;Santucci, Vieri Giuliano
2023

Abstract

This paper proposes a new open-ended learning framework which aims at implementing an autonomous agent using intrinsic motivations (IM) at two different levels. At the first level, the IM paradigm is exploited by the agent to learn new operational skills, described in terms of sub-symbolic options. After discovering the options, the agent iteratively: (1) executes them to explore the world, collecting the necessary data and (2) automatically abstracts the collected data into a high-level representation of the domain, expressed in PPDDL language. At the second level, the IM paradigm is used to exploit the abstracted representation of the domain by identifying particular symbolic states deemed promising according to a specific criterium, which in the present work is the farthest distance covered by the agent (i.e., the most promising states are those that rest at the frontier of the visited space). Once these states are identified, they can be successively reached through an internally generated high-level plan and used as promising starting points for discovering new knowledge. The presented framework is tested in the so-called Treasure Game domain described in the recent literature. The tests we have performed show that the proposed idea of implementing intrinsic motivations at two different levels of abstraction facilitates the discovery of new knowledge, compared to a previous approach proposed in the literature.
2023
Istituto di Scienze e Tecnologie della Cognizione - ISTC
9783031475450
9783031475467
Intrinsic Motivations
Open-ended learning
Planning
File in questo prodotto:
File Dimensione Formato  
Intrinsically Motivated High-Level Planning for Agent Exploration.pdf

solo utenti autorizzati

Descrizione: Sartor, G., Oddi, A., Rasconi, R., Santucci, V.G. (2023). Intrinsically Motivated High-Level Planning for Agent Exploration. In: Basili, R., Lembo, D., Limongelli, C., Orlandini, A. (eds) AIxIA 2023 – Advances in Artificial Intelligence. AIxIA 2023. Lecture Notes in Computer Science(), vol 14318. Springer, Cham. https://doi.org/10.1007/978-3-031-47546-7_9
Tipologia: Versione Editoriale (PDF)
Licenza: NON PUBBLICO - Accesso privato/ristretto
Dimensione 1.71 MB
Formato Adobe PDF
1.71 MB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.14243/522717
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? ND
social impact