Do Not Design, Learn: A Trainable Scoring Function for Uncertainty Estimation in Generative LLMs
Uncertainty estimation (UE) of generative large language models (LLMs) is crucial for evaluating the reliability of generated sequences. A significant subset of UE methods utilize token probabilities to assess uncertainty, aggregating multiple token probabilities into a single UE score using a scori...
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Uncertainty estimation (UE) of generative large language models (LLMs) is
crucial for evaluating the reliability of generated sequences. A significant
subset of UE methods utilize token probabilities to assess uncertainty,
aggregating multiple token probabilities into a single UE score using a scoring
function. Existing scoring functions for probability-based UE, such as
length-normalized scoring and semantic contribution-based weighting, are
designed to solve certain aspects of the problem but exhibit limitations,
including the inability to handle biased probabilities and complex semantic
dependencies between tokens. To address these issues, in this work, we propose
Learnable Response Scoring (LARS) function, a novel scoring function that
leverages supervised data to capture complex dependencies between tokens and
probabilities, thereby producing more reliable and calibrated response scores
in computing the uncertainty of LLM generations. Our comprehensive experiments
across question-answering and arithmetical reasoning tasks with various
datasets demonstrate that LARS significantly outperforms existing scoring
functions, achieving improvements of up to 16\% AUROC score. |
---|---|
DOI: | 10.48550/arxiv.2406.11278 |