Self-Evolution Fine-Tuning for Policy Optimization
The alignment of large language models (LLMs) is crucial not only for unlocking their potential in specific tasks but also for ensuring that responses meet human expectations and adhere to safety and ethical principles. Current alignment methodologies face considerable challenges. For instance, supe...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Chen, Ruijun Liang, Jiehao Gao, Shiping Wan, Fanqi Quan, Xiaojun |
description | The alignment of large language models (LLMs) is crucial not only for
unlocking their potential in specific tasks but also for ensuring that
responses meet human expectations and adhere to safety and ethical principles.
Current alignment methodologies face considerable challenges. For instance,
supervised fine-tuning (SFT) requires extensive, high-quality annotated
samples, while reinforcement learning from human feedback (RLHF) is complex and
often unstable. In this paper, we introduce self-evolution fine-tuning (SEFT)
for policy optimization, with the aim of eliminating the need for annotated
samples while retaining the stability and efficiency of SFT. SEFT first trains
an adaptive reviser to elevate low-quality responses while maintaining
high-quality ones. The reviser then gradually guides the policy's optimization
by fine-tuning it with enhanced responses. One of the prominent features of
this method is its ability to leverage unlimited amounts of unannotated data
for policy optimization through supervised fine-tuning. Our experiments on
AlpacaEval 2.0 and MT-Bench demonstrate the effectiveness of SEFT. We also
provide a comprehensive analysis of its advantages over existing alignment
techniques. |
doi_str_mv | 10.48550/arxiv.2406.10813 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2406_10813</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2406_10813</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-e22979a7e22371ddacf7c1d2657fcba8dd4ff131917a9beacfea5194c7fa7a083</originalsourceid><addsrcrecordid>eNotjskKwjAURbNxIeoHuLI_kJqXtH3NUsQJBAW7L88MEqit1AH16x1XZ3Evh8PYEESc5GkqxtTewy2WichiEDmoLpM7V3k-uzXV9RKaOpqH2vHiWof6EPmmjbZNFcwj2pwu4Rie9Pn0WcdTdXaDP3usmM-K6ZKvN4vVdLLmlKHiTkqNmvBNhWAtGY8GrMxS9GZPubWJ96BAA5Leu_fsKAWdGPSEJHLVY6Of9htdntpwpPZRfuLLb7x6AQGWP7A</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Self-Evolution Fine-Tuning for Policy Optimization</title><source>arXiv.org</source><creator>Chen, Ruijun ; Liang, Jiehao ; Gao, Shiping ; Wan, Fanqi ; Quan, Xiaojun</creator><creatorcontrib>Chen, Ruijun ; Liang, Jiehao ; Gao, Shiping ; Wan, Fanqi ; Quan, Xiaojun</creatorcontrib><description>The alignment of large language models (LLMs) is crucial not only for
unlocking their potential in specific tasks but also for ensuring that
responses meet human expectations and adhere to safety and ethical principles.
Current alignment methodologies face considerable challenges. For instance,
supervised fine-tuning (SFT) requires extensive, high-quality annotated
samples, while reinforcement learning from human feedback (RLHF) is complex and
often unstable. In this paper, we introduce self-evolution fine-tuning (SEFT)
for policy optimization, with the aim of eliminating the need for annotated
samples while retaining the stability and efficiency of SFT. SEFT first trains
an adaptive reviser to elevate low-quality responses while maintaining
high-quality ones. The reviser then gradually guides the policy's optimization
by fine-tuning it with enhanced responses. One of the prominent features of
this method is its ability to leverage unlimited amounts of unannotated data
for policy optimization through supervised fine-tuning. Our experiments on
AlpacaEval 2.0 and MT-Bench demonstrate the effectiveness of SEFT. We also
provide a comprehensive analysis of its advantages over existing alignment
techniques.</description><identifier>DOI: 10.48550/arxiv.2406.10813</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2024-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,777,882</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2406.10813$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2406.10813$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Chen, Ruijun</creatorcontrib><creatorcontrib>Liang, Jiehao</creatorcontrib><creatorcontrib>Gao, Shiping</creatorcontrib><creatorcontrib>Wan, Fanqi</creatorcontrib><creatorcontrib>Quan, Xiaojun</creatorcontrib><title>Self-Evolution Fine-Tuning for Policy Optimization</title><description>The alignment of large language models (LLMs) is crucial not only for
unlocking their potential in specific tasks but also for ensuring that
responses meet human expectations and adhere to safety and ethical principles.
Current alignment methodologies face considerable challenges. For instance,
supervised fine-tuning (SFT) requires extensive, high-quality annotated
samples, while reinforcement learning from human feedback (RLHF) is complex and
often unstable. In this paper, we introduce self-evolution fine-tuning (SEFT)
for policy optimization, with the aim of eliminating the need for annotated
samples while retaining the stability and efficiency of SFT. SEFT first trains
an adaptive reviser to elevate low-quality responses while maintaining
high-quality ones. The reviser then gradually guides the policy's optimization
by fine-tuning it with enhanced responses. One of the prominent features of
this method is its ability to leverage unlimited amounts of unannotated data
for policy optimization through supervised fine-tuning. Our experiments on
AlpacaEval 2.0 and MT-Bench demonstrate the effectiveness of SEFT. We also
provide a comprehensive analysis of its advantages over existing alignment
techniques.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotjskKwjAURbNxIeoHuLI_kJqXtH3NUsQJBAW7L88MEqit1AH16x1XZ3Evh8PYEESc5GkqxtTewy2WichiEDmoLpM7V3k-uzXV9RKaOpqH2vHiWof6EPmmjbZNFcwj2pwu4Rie9Pn0WcdTdXaDP3usmM-K6ZKvN4vVdLLmlKHiTkqNmvBNhWAtGY8GrMxS9GZPubWJ96BAA5Leu_fsKAWdGPSEJHLVY6Of9htdntpwpPZRfuLLb7x6AQGWP7A</recordid><startdate>20240616</startdate><enddate>20240616</enddate><creator>Chen, Ruijun</creator><creator>Liang, Jiehao</creator><creator>Gao, Shiping</creator><creator>Wan, Fanqi</creator><creator>Quan, Xiaojun</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240616</creationdate><title>Self-Evolution Fine-Tuning for Policy Optimization</title><author>Chen, Ruijun ; Liang, Jiehao ; Gao, Shiping ; Wan, Fanqi ; Quan, Xiaojun</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-e22979a7e22371ddacf7c1d2657fcba8dd4ff131917a9beacfea5194c7fa7a083</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Chen, Ruijun</creatorcontrib><creatorcontrib>Liang, Jiehao</creatorcontrib><creatorcontrib>Gao, Shiping</creatorcontrib><creatorcontrib>Wan, Fanqi</creatorcontrib><creatorcontrib>Quan, Xiaojun</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chen, Ruijun</au><au>Liang, Jiehao</au><au>Gao, Shiping</au><au>Wan, Fanqi</au><au>Quan, Xiaojun</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Self-Evolution Fine-Tuning for Policy Optimization</atitle><date>2024-06-16</date><risdate>2024</risdate><abstract>The alignment of large language models (LLMs) is crucial not only for
unlocking their potential in specific tasks but also for ensuring that
responses meet human expectations and adhere to safety and ethical principles.
Current alignment methodologies face considerable challenges. For instance,
supervised fine-tuning (SFT) requires extensive, high-quality annotated
samples, while reinforcement learning from human feedback (RLHF) is complex and
often unstable. In this paper, we introduce self-evolution fine-tuning (SEFT)
for policy optimization, with the aim of eliminating the need for annotated
samples while retaining the stability and efficiency of SFT. SEFT first trains
an adaptive reviser to elevate low-quality responses while maintaining
high-quality ones. The reviser then gradually guides the policy's optimization
by fine-tuning it with enhanced responses. One of the prominent features of
this method is its ability to leverage unlimited amounts of unannotated data
for policy optimization through supervised fine-tuning. Our experiments on
AlpacaEval 2.0 and MT-Bench demonstrate the effectiveness of SEFT. We also
provide a comprehensive analysis of its advantages over existing alignment
techniques.</abstract><doi>10.48550/arxiv.2406.10813</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2406.10813 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2406_10813 |
source | arXiv.org |
subjects | Computer Science - Computation and Language |
title | Self-Evolution Fine-Tuning for Policy Optimization |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-18T07%3A40%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Self-Evolution%20Fine-Tuning%20for%20Policy%20Optimization&rft.au=Chen,%20Ruijun&rft.date=2024-06-16&rft_id=info:doi/10.48550/arxiv.2406.10813&rft_dat=%3Carxiv_GOX%3E2406_10813%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |