Dexterous manipulation, often facilitated by multi-fingered robotic hands, holds solid impact for real-world ap-plications. Soft robotic hands, due to their compliant nature, offer flexibility and adaptability during object grasping and manipulation. Yet, benefits come with challenges, particularly in the control development for finger coordination. Reinforce-ment Learning (RL) can be employed to train object-specific in-hand manipulation policies, but limiting adaptability and generalizability. We introduce a Continual Policy Distillation (CPD) framework to acquire a versatile controller for in-hand manipulation, to rotate different objects in shape and size within a four-fingered soft gripper. The framework leverages Policy Distillation (PD) to transfer knowledge from expert policies to a continually evolving student policy network. Exemplar-based rehearsal methods are then integrated to mitigate catastrophic forgetting and enhance generalization. The performance of the CPD framework over various replay strategies demonstrates its effectiveness in consolidating knowledge from multiple experts and achieving versatile and adaptive behaviours for in-hand manipulation tasks.

Continual policy distillation of reinforcement learning-based controllers for soft robotic in-hand manipulation

Li L.;
2024

Abstract

Dexterous manipulation, often facilitated by multi-fingered robotic hands, holds solid impact for real-world ap-plications. Soft robotic hands, due to their compliant nature, offer flexibility and adaptability during object grasping and manipulation. Yet, benefits come with challenges, particularly in the control development for finger coordination. Reinforce-ment Learning (RL) can be employed to train object-specific in-hand manipulation policies, but limiting adaptability and generalizability. We introduce a Continual Policy Distillation (CPD) framework to acquire a versatile controller for in-hand manipulation, to rotate different objects in shape and size within a four-fingered soft gripper. The framework leverages Policy Distillation (PD) to transfer knowledge from expert policies to a continually evolving student policy network. Exemplar-based rehearsal methods are then integrated to mitigate catastrophic forgetting and enhance generalization. The performance of the CPD framework over various replay strategies demonstrates its effectiveness in consolidating knowledge from multiple experts and achieving versatile and adaptive behaviours for in-hand manipulation tasks.
2024
Istituto di Scienza e Tecnologie dell'Informazione "Alessandro Faedo" - ISTI
979-8-3503-8182-5
Continual Learning
In-hand Manipulation
Policy Distillation
Reinforcement Learning
Soft Robots
File in questo prodotto:
File Dimensione Formato  
Continual_Policy_Distillation_of_Reinforcement_Learning-based_Controllers_for_Soft_Robotic_In-Hand_Manipulation.pdf

solo utenti autorizzati

Descrizione: Continual policy distillation of reinforcement learning-based controllers for soft robotic in-hand manipulation
Tipologia: Versione Editoriale (PDF)
Licenza: NON PUBBLICO - Accesso privato/ristretto
Dimensione 2.7 MB
Formato Adobe PDF
2.7 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
Lanpei et al_Continual_preprintRoboSoft_2024.pdf

accesso aperto

Descrizione: Continual policy distillation of reinforcement learning-based controllers for soft robotic in-hand manipulation
Tipologia: Documento in Pre-print
Licenza: Creative commons
Dimensione 2.63 MB
Formato Adobe PDF
2.63 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.14243/559736
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 10
  • ???jsp.display-item.citation.isi??? 5
social impact