ANAH-v2: Scaling Analytical Hallucination Annotation of Large Language Models

Large language models (LLMs) exhibit hallucinations in long-form question-answering tasks across various domains and wide applications. Current hallucination detection and mitigation datasets are limited in domains and sizes, which struggle to scale due to prohibitive labor costs and insufficient re...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Gu, Yuzhe, Ji, Ziwei, Zhang, Wenwei, Lyu, Chengqi, Lin, Dahua, Chen, Kai
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Gu, Yuzhe
Ji, Ziwei
Zhang, Wenwei
Lyu, Chengqi
Lin, Dahua
Chen, Kai
description Large language models (LLMs) exhibit hallucinations in long-form question-answering tasks across various domains and wide applications. Current hallucination detection and mitigation datasets are limited in domains and sizes, which struggle to scale due to prohibitive labor costs and insufficient reliability of existing hallucination annotators. To facilitate the scalable oversight of LLM hallucinations, this paper introduces an iterative self-training framework that simultaneously and progressively scales up the hallucination annotation dataset and improves the accuracy of the hallucination annotator. Based on the Expectation Maximization (EM) algorithm, in each iteration, the framework first applies a hallucination annotation pipeline to annotate a scaled dataset and then trains a more accurate hallucination annotator on the dataset. This new hallucination annotator is adopted in the hallucination annotation pipeline used for the next iteration. Extensive experimental results demonstrate that the finally obtained hallucination annotator with only 7B parameters surpasses the performance of GPT-4 and obtains new state-of-the-art hallucination detection results on HaluEval and HalluQA by zero-shot inference. Such an annotator can not only evaluate the hallucination levels of various LLMs on the large-scale dataset but also help to mitigate the hallucination of LLMs generations, with the Natural Language Inference (NLI) metric increasing from 25% to 37% on HaluEval.
doi_str_mv 10.48550/arxiv.2407.04693
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2407_04693</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2407_04693</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2407_046933</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjEw1zMwMbM05mTwdfRz9NAtM7JSCE5OzMnMS1dwzEvMqSzJBPIUPBJzckqTM_MSSzLz84ASefklEGZ-moJPYlF6KpDMSy9NBDJ881NSc4p5GFjTEnOKU3mhNDeDvJtriLOHLtji-IKizNzEosp4kAPiwQ4wJqwCAJpOOoI</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>ANAH-v2: Scaling Analytical Hallucination Annotation of Large Language Models</title><source>arXiv.org</source><creator>Gu, Yuzhe ; Ji, Ziwei ; Zhang, Wenwei ; Lyu, Chengqi ; Lin, Dahua ; Chen, Kai</creator><creatorcontrib>Gu, Yuzhe ; Ji, Ziwei ; Zhang, Wenwei ; Lyu, Chengqi ; Lin, Dahua ; Chen, Kai</creatorcontrib><description>Large language models (LLMs) exhibit hallucinations in long-form question-answering tasks across various domains and wide applications. Current hallucination detection and mitigation datasets are limited in domains and sizes, which struggle to scale due to prohibitive labor costs and insufficient reliability of existing hallucination annotators. To facilitate the scalable oversight of LLM hallucinations, this paper introduces an iterative self-training framework that simultaneously and progressively scales up the hallucination annotation dataset and improves the accuracy of the hallucination annotator. Based on the Expectation Maximization (EM) algorithm, in each iteration, the framework first applies a hallucination annotation pipeline to annotate a scaled dataset and then trains a more accurate hallucination annotator on the dataset. This new hallucination annotator is adopted in the hallucination annotation pipeline used for the next iteration. Extensive experimental results demonstrate that the finally obtained hallucination annotator with only 7B parameters surpasses the performance of GPT-4 and obtains new state-of-the-art hallucination detection results on HaluEval and HalluQA by zero-shot inference. Such an annotator can not only evaluate the hallucination levels of various LLMs on the large-scale dataset but also help to mitigate the hallucination of LLMs generations, with the Natural Language Inference (NLI) metric increasing from 25% to 37% on HaluEval.</description><identifier>DOI: 10.48550/arxiv.2407.04693</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language</subject><creationdate>2024-07</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,777,882</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2407.04693$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2407.04693$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Gu, Yuzhe</creatorcontrib><creatorcontrib>Ji, Ziwei</creatorcontrib><creatorcontrib>Zhang, Wenwei</creatorcontrib><creatorcontrib>Lyu, Chengqi</creatorcontrib><creatorcontrib>Lin, Dahua</creatorcontrib><creatorcontrib>Chen, Kai</creatorcontrib><title>ANAH-v2: Scaling Analytical Hallucination Annotation of Large Language Models</title><description>Large language models (LLMs) exhibit hallucinations in long-form question-answering tasks across various domains and wide applications. Current hallucination detection and mitigation datasets are limited in domains and sizes, which struggle to scale due to prohibitive labor costs and insufficient reliability of existing hallucination annotators. To facilitate the scalable oversight of LLM hallucinations, this paper introduces an iterative self-training framework that simultaneously and progressively scales up the hallucination annotation dataset and improves the accuracy of the hallucination annotator. Based on the Expectation Maximization (EM) algorithm, in each iteration, the framework first applies a hallucination annotation pipeline to annotate a scaled dataset and then trains a more accurate hallucination annotator on the dataset. This new hallucination annotator is adopted in the hallucination annotation pipeline used for the next iteration. Extensive experimental results demonstrate that the finally obtained hallucination annotator with only 7B parameters surpasses the performance of GPT-4 and obtains new state-of-the-art hallucination detection results on HaluEval and HalluQA by zero-shot inference. Such an annotator can not only evaluate the hallucination levels of various LLMs on the large-scale dataset but also help to mitigate the hallucination of LLMs generations, with the Natural Language Inference (NLI) metric increasing from 25% to 37% on HaluEval.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjEw1zMwMbM05mTwdfRz9NAtM7JSCE5OzMnMS1dwzEvMqSzJBPIUPBJzckqTM_MSSzLz84ASefklEGZ-moJPYlF6KpDMSy9NBDJ881NSc4p5GFjTEnOKU3mhNDeDvJtriLOHLtji-IKizNzEosp4kAPiwQ4wJqwCAJpOOoI</recordid><startdate>20240705</startdate><enddate>20240705</enddate><creator>Gu, Yuzhe</creator><creator>Ji, Ziwei</creator><creator>Zhang, Wenwei</creator><creator>Lyu, Chengqi</creator><creator>Lin, Dahua</creator><creator>Chen, Kai</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240705</creationdate><title>ANAH-v2: Scaling Analytical Hallucination Annotation of Large Language Models</title><author>Gu, Yuzhe ; Ji, Ziwei ; Zhang, Wenwei ; Lyu, Chengqi ; Lin, Dahua ; Chen, Kai</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2407_046933</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Gu, Yuzhe</creatorcontrib><creatorcontrib>Ji, Ziwei</creatorcontrib><creatorcontrib>Zhang, Wenwei</creatorcontrib><creatorcontrib>Lyu, Chengqi</creatorcontrib><creatorcontrib>Lin, Dahua</creatorcontrib><creatorcontrib>Chen, Kai</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Gu, Yuzhe</au><au>Ji, Ziwei</au><au>Zhang, Wenwei</au><au>Lyu, Chengqi</au><au>Lin, Dahua</au><au>Chen, Kai</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>ANAH-v2: Scaling Analytical Hallucination Annotation of Large Language Models</atitle><date>2024-07-05</date><risdate>2024</risdate><abstract>Large language models (LLMs) exhibit hallucinations in long-form question-answering tasks across various domains and wide applications. Current hallucination detection and mitigation datasets are limited in domains and sizes, which struggle to scale due to prohibitive labor costs and insufficient reliability of existing hallucination annotators. To facilitate the scalable oversight of LLM hallucinations, this paper introduces an iterative self-training framework that simultaneously and progressively scales up the hallucination annotation dataset and improves the accuracy of the hallucination annotator. Based on the Expectation Maximization (EM) algorithm, in each iteration, the framework first applies a hallucination annotation pipeline to annotate a scaled dataset and then trains a more accurate hallucination annotator on the dataset. This new hallucination annotator is adopted in the hallucination annotation pipeline used for the next iteration. Extensive experimental results demonstrate that the finally obtained hallucination annotator with only 7B parameters surpasses the performance of GPT-4 and obtains new state-of-the-art hallucination detection results on HaluEval and HalluQA by zero-shot inference. Such an annotator can not only evaluate the hallucination levels of various LLMs on the large-scale dataset but also help to mitigate the hallucination of LLMs generations, with the Natural Language Inference (NLI) metric increasing from 25% to 37% on HaluEval.</abstract><doi>10.48550/arxiv.2407.04693</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2407.04693
ispartof
issn
language eng
recordid cdi_arxiv_primary_2407_04693
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computation and Language
title ANAH-v2: Scaling Analytical Hallucination Annotation of Large Language Models
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-18T00%3A24%3A05IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=ANAH-v2:%20Scaling%20Analytical%20Hallucination%20Annotation%20of%20Large%20Language%20Models&rft.au=Gu,%20Yuzhe&rft.date=2024-07-05&rft_id=info:doi/10.48550/arxiv.2407.04693&rft_dat=%3Carxiv_GOX%3E2407_04693%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true