In the context of socially assistive robotics, there is a growing need for interaction strategies that can adapt to users' emotional states in real time, as fixed or generic communication styles often fail to sustain user engagement or meet individual motivational needs, especially in long-term human-robot interaction. To address this challenge, this paper presents a novel framework for adaptive interaction style modulation in socially assistive agents, combining large language models (LLMs) with reinforcement learning based on real-time emotion recognition. The proposed architecture leverages multimodal sensing to monitor the user's affective state and dynamically selects among predefined communicative styles using Thompson Sampling. At each interaction turn, the user's emotional feedback is converted into a scalar reward, allowing the system to reinforce styles that yield more positive affective outcomes. Style conditioning is operationalized through prompting strategies that guide the LLM to generate responses aligned with the selected tone. A preliminary evaluation using VADER sentiment analysis demonstrates that stylistic prompts successfully induce measurable differences in sentiment polarity, neutrality, and verbosity. These findings suggest the viability of our approach to style-aware dialogue generation and support its potential for long-term adaptation in personalized human-agent interaction.

Adaptive Interaction Style Modulation via Reinforcement Learning and Prompted Language Generation

Tamantini C.
Primo
;
Beraldo G.;Umbrico A.;Orlandini A.
Ultimo
2025

Abstract

In the context of socially assistive robotics, there is a growing need for interaction strategies that can adapt to users' emotional states in real time, as fixed or generic communication styles often fail to sustain user engagement or meet individual motivational needs, especially in long-term human-robot interaction. To address this challenge, this paper presents a novel framework for adaptive interaction style modulation in socially assistive agents, combining large language models (LLMs) with reinforcement learning based on real-time emotion recognition. The proposed architecture leverages multimodal sensing to monitor the user's affective state and dynamically selects among predefined communicative styles using Thompson Sampling. At each interaction turn, the user's emotional feedback is converted into a scalar reward, allowing the system to reinforce styles that yield more positive affective outcomes. Style conditioning is operationalized through prompting strategies that guide the LLM to generate responses aligned with the selected tone. A preliminary evaluation using VADER sentiment analysis demonstrates that stylistic prompts successfully induce measurable differences in sentiment polarity, neutrality, and verbosity. These findings suggest the viability of our approach to style-aware dialogue generation and support its potential for long-term adaptation in personalized human-agent interaction.
2025
Istituto di Scienze e Tecnologie della Cognizione - ISTC
Human-Robot Interaction
Large Language Models
Prompt Engineering
Reinforcement Learning
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.14243/562025
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ente

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact