Advances in Natural Language Processing led to the introduction of Large Language Models (LLMs), that have been found endowed of enriched capabilities and improved performance results when increased in size. Researcher from various disciples argue about whether or not, among the capabilities of LLMs, there is the one of using knowledge about knowledge - usually considered one of the antechambers of meta-cognition in cognitive agents - about a particular task in order to improve or self-correct previous errors. In this work we propose a novel fine-tuning approach, named EXAR, based on a multi-stage LLM fine-tuning leveraging past predictions from an early version of the same, aimed to inject metacognitive features for the task of Question-Answering. The conducted experiments on LLAMA-2-7B-CHAT showed promising improvements on the quality of the outcomes, due to LLM acquired ability to detect its own wrong predictions and forcing itself to repeat submissions, thorough a prompt designed to fix inadmissible predictions whenever are detected. Such detection is achieved by enquiring the same LLM acting as meta-validator, through another prompt specifically designed for such purpose.

Eliciting Metaknowledge in Large Language Models

Carmelo Fabio Longo
Methodology
;
Luana Bulla
Writing – Review & Editing
;
Antonio Lieto
Supervision
2024

Abstract

Advances in Natural Language Processing led to the introduction of Large Language Models (LLMs), that have been found endowed of enriched capabilities and improved performance results when increased in size. Researcher from various disciples argue about whether or not, among the capabilities of LLMs, there is the one of using knowledge about knowledge - usually considered one of the antechambers of meta-cognition in cognitive agents - about a particular task in order to improve or self-correct previous errors. In this work we propose a novel fine-tuning approach, named EXAR, based on a multi-stage LLM fine-tuning leveraging past predictions from an early version of the same, aimed to inject metacognitive features for the task of Question-Answering. The conducted experiments on LLAMA-2-7B-CHAT showed promising improvements on the quality of the outcomes, due to LLM acquired ability to detect its own wrong predictions and forcing itself to repeat submissions, thorough a prompt designed to fix inadmissible predictions whenever are detected. Such detection is achieved by enquiring the same LLM acting as meta-validator, through another prompt specifically designed for such purpose.
2024
Istituto di Scienze e Tecnologie della Cognizione - ISTC - Sede Secondaria Catania
inglese
File in questo prodotto:
File Dimensione Formato  
POSTER_ACS.pdf

accesso aperto

Licenza: Dominio pubblico
Dimensione 2.65 MB
Formato Adobe PDF
2.65 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.14243/514389
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact