Enhancing Uncertainty Modeling with Semantic Graph for Hallucination Detection

Large Language Models (LLMs) are prone to hallucination with non-factual or unfaithful statements, which undermines the applications in real-world scenarios. Recent researches focus on uncertainty-based hallucination detection, which utilizes the output probability of LLMs for uncertainty calculatio...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Chen, Kedi, Chen, Qin, Zhou, Jie, Tao, Xinqi, Ding, Bowen, Xie, Jingwen, Xie, Mingchen, Li, Peilong, Zheng, Feng, He, Liang
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Chen, Kedi
Chen, Qin
Zhou, Jie
Tao, Xinqi
Ding, Bowen
Xie, Jingwen
Xie, Mingchen
Li, Peilong
Zheng, Feng
He, Liang
description Large Language Models (LLMs) are prone to hallucination with non-factual or unfaithful statements, which undermines the applications in real-world scenarios. Recent researches focus on uncertainty-based hallucination detection, which utilizes the output probability of LLMs for uncertainty calculation and does not rely on external knowledge or frequent sampling from LLMs. Whereas, most approaches merely consider the uncertainty of each independent token, while the intricate semantic relations among tokens and sentences are not well studied, which limits the detection of hallucination that spans over multiple tokens and sentences in the passage. In this paper, we propose a method to enhance uncertainty modeling with semantic graph for hallucination detection. Specifically, we first construct a semantic graph that well captures the relations among entity tokens and sentences. Then, we incorporate the relations between two entities for uncertainty propagation to enhance sentence-level hallucination detection. Given that hallucination occurs due to the conflict between sentences, we further present a graph-based uncertainty calibration method that integrates the contradiction probability of the sentence with its neighbors in the semantic graph for uncertainty calculation. Extensive experiments on two datasets show the great advantages of our proposed approach. In particular, we obtain substantial improvements with 19.78% in passage-level hallucination detection.
doi_str_mv 10.48550/arxiv.2501.02020
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2501_02020</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2501_02020</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2501_020203</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjUw1DMwAkJOBj_XvIzEvOTMvHSF0Lzk1KKSxMy8kkoF3_yU1ByQYHlmSYZCcGpuYl5JZrKCe1FiQYZCWn6RgkdiTk4pUFtiSWZ-noJLaklqMojFw8CalphTnMoLpbkZ5N1cQ5w9dME2xxcUZeYmFlXGg1wQD3aBMWEVAJURPAk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Enhancing Uncertainty Modeling with Semantic Graph for Hallucination Detection</title><source>arXiv.org</source><creator>Chen, Kedi ; Chen, Qin ; Zhou, Jie ; Tao, Xinqi ; Ding, Bowen ; Xie, Jingwen ; Xie, Mingchen ; Li, Peilong ; Zheng, Feng ; He, Liang</creator><creatorcontrib>Chen, Kedi ; Chen, Qin ; Zhou, Jie ; Tao, Xinqi ; Ding, Bowen ; Xie, Jingwen ; Xie, Mingchen ; Li, Peilong ; Zheng, Feng ; He, Liang</creatorcontrib><description>Large Language Models (LLMs) are prone to hallucination with non-factual or unfaithful statements, which undermines the applications in real-world scenarios. Recent researches focus on uncertainty-based hallucination detection, which utilizes the output probability of LLMs for uncertainty calculation and does not rely on external knowledge or frequent sampling from LLMs. Whereas, most approaches merely consider the uncertainty of each independent token, while the intricate semantic relations among tokens and sentences are not well studied, which limits the detection of hallucination that spans over multiple tokens and sentences in the passage. In this paper, we propose a method to enhance uncertainty modeling with semantic graph for hallucination detection. Specifically, we first construct a semantic graph that well captures the relations among entity tokens and sentences. Then, we incorporate the relations between two entities for uncertainty propagation to enhance sentence-level hallucination detection. Given that hallucination occurs due to the conflict between sentences, we further present a graph-based uncertainty calibration method that integrates the contradiction probability of the sentence with its neighbors in the semantic graph for uncertainty calculation. Extensive experiments on two datasets show the great advantages of our proposed approach. In particular, we obtain substantial improvements with 19.78% in passage-level hallucination detection.</description><identifier>DOI: 10.48550/arxiv.2501.02020</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language</subject><creationdate>2025-01</creationdate><rights>http://creativecommons.org/licenses/by-nc-sa/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2501.02020$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2501.02020$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Chen, Kedi</creatorcontrib><creatorcontrib>Chen, Qin</creatorcontrib><creatorcontrib>Zhou, Jie</creatorcontrib><creatorcontrib>Tao, Xinqi</creatorcontrib><creatorcontrib>Ding, Bowen</creatorcontrib><creatorcontrib>Xie, Jingwen</creatorcontrib><creatorcontrib>Xie, Mingchen</creatorcontrib><creatorcontrib>Li, Peilong</creatorcontrib><creatorcontrib>Zheng, Feng</creatorcontrib><creatorcontrib>He, Liang</creatorcontrib><title>Enhancing Uncertainty Modeling with Semantic Graph for Hallucination Detection</title><description>Large Language Models (LLMs) are prone to hallucination with non-factual or unfaithful statements, which undermines the applications in real-world scenarios. Recent researches focus on uncertainty-based hallucination detection, which utilizes the output probability of LLMs for uncertainty calculation and does not rely on external knowledge or frequent sampling from LLMs. Whereas, most approaches merely consider the uncertainty of each independent token, while the intricate semantic relations among tokens and sentences are not well studied, which limits the detection of hallucination that spans over multiple tokens and sentences in the passage. In this paper, we propose a method to enhance uncertainty modeling with semantic graph for hallucination detection. Specifically, we first construct a semantic graph that well captures the relations among entity tokens and sentences. Then, we incorporate the relations between two entities for uncertainty propagation to enhance sentence-level hallucination detection. Given that hallucination occurs due to the conflict between sentences, we further present a graph-based uncertainty calibration method that integrates the contradiction probability of the sentence with its neighbors in the semantic graph for uncertainty calculation. Extensive experiments on two datasets show the great advantages of our proposed approach. In particular, we obtain substantial improvements with 19.78% in passage-level hallucination detection.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2025</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjUw1DMwAkJOBj_XvIzEvOTMvHSF0Lzk1KKSxMy8kkoF3_yU1ByQYHlmSYZCcGpuYl5JZrKCe1FiQYZCWn6RgkdiTk4pUFtiSWZ-noJLaklqMojFw8CalphTnMoLpbkZ5N1cQ5w9dME2xxcUZeYmFlXGg1wQD3aBMWEVAJURPAk</recordid><startdate>20250102</startdate><enddate>20250102</enddate><creator>Chen, Kedi</creator><creator>Chen, Qin</creator><creator>Zhou, Jie</creator><creator>Tao, Xinqi</creator><creator>Ding, Bowen</creator><creator>Xie, Jingwen</creator><creator>Xie, Mingchen</creator><creator>Li, Peilong</creator><creator>Zheng, Feng</creator><creator>He, Liang</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20250102</creationdate><title>Enhancing Uncertainty Modeling with Semantic Graph for Hallucination Detection</title><author>Chen, Kedi ; Chen, Qin ; Zhou, Jie ; Tao, Xinqi ; Ding, Bowen ; Xie, Jingwen ; Xie, Mingchen ; Li, Peilong ; Zheng, Feng ; He, Liang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2501_020203</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2025</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Chen, Kedi</creatorcontrib><creatorcontrib>Chen, Qin</creatorcontrib><creatorcontrib>Zhou, Jie</creatorcontrib><creatorcontrib>Tao, Xinqi</creatorcontrib><creatorcontrib>Ding, Bowen</creatorcontrib><creatorcontrib>Xie, Jingwen</creatorcontrib><creatorcontrib>Xie, Mingchen</creatorcontrib><creatorcontrib>Li, Peilong</creatorcontrib><creatorcontrib>Zheng, Feng</creatorcontrib><creatorcontrib>He, Liang</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chen, Kedi</au><au>Chen, Qin</au><au>Zhou, Jie</au><au>Tao, Xinqi</au><au>Ding, Bowen</au><au>Xie, Jingwen</au><au>Xie, Mingchen</au><au>Li, Peilong</au><au>Zheng, Feng</au><au>He, Liang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Enhancing Uncertainty Modeling with Semantic Graph for Hallucination Detection</atitle><date>2025-01-02</date><risdate>2025</risdate><abstract>Large Language Models (LLMs) are prone to hallucination with non-factual or unfaithful statements, which undermines the applications in real-world scenarios. Recent researches focus on uncertainty-based hallucination detection, which utilizes the output probability of LLMs for uncertainty calculation and does not rely on external knowledge or frequent sampling from LLMs. Whereas, most approaches merely consider the uncertainty of each independent token, while the intricate semantic relations among tokens and sentences are not well studied, which limits the detection of hallucination that spans over multiple tokens and sentences in the passage. In this paper, we propose a method to enhance uncertainty modeling with semantic graph for hallucination detection. Specifically, we first construct a semantic graph that well captures the relations among entity tokens and sentences. Then, we incorporate the relations between two entities for uncertainty propagation to enhance sentence-level hallucination detection. Given that hallucination occurs due to the conflict between sentences, we further present a graph-based uncertainty calibration method that integrates the contradiction probability of the sentence with its neighbors in the semantic graph for uncertainty calculation. Extensive experiments on two datasets show the great advantages of our proposed approach. In particular, we obtain substantial improvements with 19.78% in passage-level hallucination detection.</abstract><doi>10.48550/arxiv.2501.02020</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2501.02020
ispartof
issn
language eng
recordid cdi_arxiv_primary_2501_02020
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computation and Language
title Enhancing Uncertainty Modeling with Semantic Graph for Hallucination Detection
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-30T03%3A42%3A52IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Enhancing%20Uncertainty%20Modeling%20with%20Semantic%20Graph%20for%20Hallucination%20Detection&rft.au=Chen,%20Kedi&rft.date=2025-01-02&rft_id=info:doi/10.48550/arxiv.2501.02020&rft_dat=%3Carxiv_GOX%3E2501_02020%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true