SUGAR: Leveraging Contextual Confidence for Smarter Retrieval

  • Hanna Zubkova
  • , Ji Hoon Park
  • , Seong Whan Lee*
  • *Corresponding author for this work

Research output: Contribution to journalConference articlepeer-review

Abstract

Bearing in mind the limited parametric knowledge of Large Language Models (LLMs), retrieval-augmented generation (RAG) which supplies them with the relevant external knowledge has served as an approach to mitigate the issue of hallucinations to a certain extent. However, uniformly retrieving supporting context makes response generation source-inefficient, as triggering the retriever is not always necessary, or even inaccurate, when a model gets distracted by noisy retrieved content and produces an unhelpful answer. Motivated by these issues, we introduce Semantic Uncertainty Guided Adaptive Retrieval (SUGAR), where we leverage context-based entropy to actively decide whether to retrieve and to further determine between single-step and multi-step retrieval. Our empirical results show that selective retrieval guided by semantic uncertainty estimation improves the performance across diverse question answering tasks, as well as achieves a more efficient inference.

Bibliographical note

Publisher Copyright:
© 2025 IEEE.

Keywords

  • large language models
  • question answering
  • retrieval augmented generation
  • uncertainty estimation

ASJC Scopus subject areas

  • Software
  • Signal Processing
  • Electrical and Electronic Engineering

Fingerprint

Dive into the research topics of 'SUGAR: Leveraging Contextual Confidence for Smarter Retrieval'. Together they form a unique fingerprint.

Cite this