ICT: Image-Object Cross-Level Trusted Intervention for Mitigating Object Hallucination in Large Vision-Language Models
Despite the recent breakthroughs achieved by Large Vision Language Models (LVLMs) in understanding and responding to complex visual-textual contexts, their inherent hallucination tendencies limit their practical application in real-world scenarios that demand high levels of precision. Existing metho...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Despite the recent breakthroughs achieved by Large Vision Language Models
(LVLMs) in understanding and responding to complex visual-textual contexts,
their inherent hallucination tendencies limit their practical application in
real-world scenarios that demand high levels of precision. Existing methods
typically either fine-tune the LVLMs using additional data, which incurs extra
costs in manual annotation and computational resources or perform comparisons
at the decoding stage, which may eliminate useful language priors for reasoning
while introducing inference time overhead. Therefore, we propose ICT, a
lightweight, training-free method that calculates an intervention direction to
shift the model's focus towards different levels of visual information,
enhancing its attention to high-level and fine-grained visual details. During
the forward pass stage, the intervention is applied to the attention heads that
encode the overall image information and the fine-grained object details,
effectively mitigating the phenomenon of overly language priors, and thereby
alleviating hallucinations. Extensive experiments demonstrate that ICT achieves
strong performance with a small amount of data and generalizes well across
different datasets and models. Our code will be public. |
---|---|
DOI: | 10.48550/arxiv.2411.15268 |