Usr-mtl: an unsupervised sentence representation learning framework with multi-task learning
Developing the utilized intelligent systems is increasingly important to learn effective text representations, especially extract the sentence features. Numerous previous studies have been concentrated on the task of sentence representation learning based on deep learning approaches. However, the pr...
Gespeichert in:
Veröffentlicht in: | Applied intelligence (Dordrecht, Netherlands) Netherlands), 2021-06, Vol.51 (6), p.3506-3521 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 3521 |
---|---|
container_issue | 6 |
container_start_page | 3506 |
container_title | Applied intelligence (Dordrecht, Netherlands) |
container_volume | 51 |
creator | Xu, Wenshen Li, Shuangyin Lu, Yonghe |
description | Developing the utilized intelligent systems is increasingly important to learn effective text representations, especially extract the sentence features. Numerous previous studies have been concentrated on the task of sentence representation learning based on deep learning approaches. However, the present approaches are mostly proposed with the single task or replied on the labeled corpus when learning the embedding of the sentences. In this paper, we assess the factors in learning sentence representation and propose an efficient unsupervised learning framework with multi-task learning (USR-MTL), in which various text learning tasks are merged into the unitized framework. With the syntactic and semantic features of sentences, three different factors to some extent are reflected in the task of the sentence representation learning that is the wording, or the ordering of the neighbored sentences of a target sentence in other words. Hence, we integrate the word-order learning task, word prediction task, and the sentence-order learning task into the proposed framework to attain meaningful sentence embeddings. Here, the process of sentence embedding learning is reformulated as a multi-task learning framework of the sentence-level task and the two word-level tasks. Moreover, the proposed framework is motivated by an unsupervised learning algorithm utilizing the unlabeled corpus. Based on the experimental results, our approach achieves the state-of-the-art performances on the downstream natural language processing tasks compared to the popular unsupervised representation learning techniques. The experiments on representation visualization and task analysis demonstrate the effectiveness of the tasks in the proposed framework in creating reasonable sentence representations proving the capacity of the proposed unsupervised multi-task framework for the sentence representation learning. |
doi_str_mv | 10.1007/s10489-020-02042-2 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2529006838</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2529006838</sourcerecordid><originalsourceid>FETCH-LOGICAL-c319t-d71e44c1f1c441f297cdf022524e1c6202532194d062bd9f7f918dd0cde13f943</originalsourceid><addsrcrecordid>eNp9kE1LxDAQhoMouK7-AU8Bz9HJx7aNNxG_YMGLCx6EUJvJ2t1tWpPUxX9v14rePAzDwPO-Aw8hpxzOOUB-ETmoQjMQsBslmNgjEz7LJcuVzvfJBLRQLMv08yE5inEFAFICn5CXRQysSZtLWnra-9h3GD7qiJZG9Al9hTRgF3B3laluPd1gGXztl9SFssFtG9Z0W6c32vSbVLNUxvUvckwOXLmJePKzp2Rxe_N0fc_mj3cP11dzVkmuE7M5R6Uq7nilFHdC55V1IMRMKORVJkDMpOBaWcjEq9Uud5oX1kJlkUunlZySs7G3C-17jzGZVdsHP7w0Q4kGyApZDJQYqSq0MQZ0pgt1U4ZPw8HsLJrRohkMmm-LRgwhOYbiAPslhr_qf1JfnlF2Iw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2529006838</pqid></control><display><type>article</type><title>Usr-mtl: an unsupervised sentence representation learning framework with multi-task learning</title><source>SpringerLink Journals - AutoHoldings</source><creator>Xu, Wenshen ; Li, Shuangyin ; Lu, Yonghe</creator><creatorcontrib>Xu, Wenshen ; Li, Shuangyin ; Lu, Yonghe</creatorcontrib><description>Developing the utilized intelligent systems is increasingly important to learn effective text representations, especially extract the sentence features. Numerous previous studies have been concentrated on the task of sentence representation learning based on deep learning approaches. However, the present approaches are mostly proposed with the single task or replied on the labeled corpus when learning the embedding of the sentences. In this paper, we assess the factors in learning sentence representation and propose an efficient unsupervised learning framework with multi-task learning (USR-MTL), in which various text learning tasks are merged into the unitized framework. With the syntactic and semantic features of sentences, three different factors to some extent are reflected in the task of the sentence representation learning that is the wording, or the ordering of the neighbored sentences of a target sentence in other words. Hence, we integrate the word-order learning task, word prediction task, and the sentence-order learning task into the proposed framework to attain meaningful sentence embeddings. Here, the process of sentence embedding learning is reformulated as a multi-task learning framework of the sentence-level task and the two word-level tasks. Moreover, the proposed framework is motivated by an unsupervised learning algorithm utilizing the unlabeled corpus. Based on the experimental results, our approach achieves the state-of-the-art performances on the downstream natural language processing tasks compared to the popular unsupervised representation learning techniques. The experiments on representation visualization and task analysis demonstrate the effectiveness of the tasks in the proposed framework in creating reasonable sentence representations proving the capacity of the proposed unsupervised multi-task framework for the sentence representation learning.</description><identifier>ISSN: 0924-669X</identifier><identifier>EISSN: 1573-7497</identifier><identifier>DOI: 10.1007/s10489-020-02042-2</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Algorithms ; Artificial Intelligence ; Cognitive tasks ; Computer Science ; Downstream effects ; Embedding ; Feature extraction ; Machine learning ; Machines ; Manufacturing ; Mechanical Engineering ; Natural language processing ; Processes ; Representations ; Sentences ; System effectiveness ; Task analysis ; Words (language)</subject><ispartof>Applied intelligence (Dordrecht, Netherlands), 2021-06, Vol.51 (6), p.3506-3521</ispartof><rights>Springer Science+Business Media, LLC, part of Springer Nature 2020</rights><rights>Springer Science+Business Media, LLC, part of Springer Nature 2020.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c319t-d71e44c1f1c441f297cdf022524e1c6202532194d062bd9f7f918dd0cde13f943</citedby><cites>FETCH-LOGICAL-c319t-d71e44c1f1c441f297cdf022524e1c6202532194d062bd9f7f918dd0cde13f943</cites><orcidid>0000-0001-6404-3438</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s10489-020-02042-2$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s10489-020-02042-2$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,41488,42557,51319</link.rule.ids></links><search><creatorcontrib>Xu, Wenshen</creatorcontrib><creatorcontrib>Li, Shuangyin</creatorcontrib><creatorcontrib>Lu, Yonghe</creatorcontrib><title>Usr-mtl: an unsupervised sentence representation learning framework with multi-task learning</title><title>Applied intelligence (Dordrecht, Netherlands)</title><addtitle>Appl Intell</addtitle><description>Developing the utilized intelligent systems is increasingly important to learn effective text representations, especially extract the sentence features. Numerous previous studies have been concentrated on the task of sentence representation learning based on deep learning approaches. However, the present approaches are mostly proposed with the single task or replied on the labeled corpus when learning the embedding of the sentences. In this paper, we assess the factors in learning sentence representation and propose an efficient unsupervised learning framework with multi-task learning (USR-MTL), in which various text learning tasks are merged into the unitized framework. With the syntactic and semantic features of sentences, three different factors to some extent are reflected in the task of the sentence representation learning that is the wording, or the ordering of the neighbored sentences of a target sentence in other words. Hence, we integrate the word-order learning task, word prediction task, and the sentence-order learning task into the proposed framework to attain meaningful sentence embeddings. Here, the process of sentence embedding learning is reformulated as a multi-task learning framework of the sentence-level task and the two word-level tasks. Moreover, the proposed framework is motivated by an unsupervised learning algorithm utilizing the unlabeled corpus. Based on the experimental results, our approach achieves the state-of-the-art performances on the downstream natural language processing tasks compared to the popular unsupervised representation learning techniques. The experiments on representation visualization and task analysis demonstrate the effectiveness of the tasks in the proposed framework in creating reasonable sentence representations proving the capacity of the proposed unsupervised multi-task framework for the sentence representation learning.</description><subject>Algorithms</subject><subject>Artificial Intelligence</subject><subject>Cognitive tasks</subject><subject>Computer Science</subject><subject>Downstream effects</subject><subject>Embedding</subject><subject>Feature extraction</subject><subject>Machine learning</subject><subject>Machines</subject><subject>Manufacturing</subject><subject>Mechanical Engineering</subject><subject>Natural language processing</subject><subject>Processes</subject><subject>Representations</subject><subject>Sentences</subject><subject>System effectiveness</subject><subject>Task analysis</subject><subject>Words (language)</subject><issn>0924-669X</issn><issn>1573-7497</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><recordid>eNp9kE1LxDAQhoMouK7-AU8Bz9HJx7aNNxG_YMGLCx6EUJvJ2t1tWpPUxX9v14rePAzDwPO-Aw8hpxzOOUB-ETmoQjMQsBslmNgjEz7LJcuVzvfJBLRQLMv08yE5inEFAFICn5CXRQysSZtLWnra-9h3GD7qiJZG9Al9hTRgF3B3laluPd1gGXztl9SFssFtG9Z0W6c32vSbVLNUxvUvckwOXLmJePKzp2Rxe_N0fc_mj3cP11dzVkmuE7M5R6Uq7nilFHdC55V1IMRMKORVJkDMpOBaWcjEq9Uud5oX1kJlkUunlZySs7G3C-17jzGZVdsHP7w0Q4kGyApZDJQYqSq0MQZ0pgt1U4ZPw8HsLJrRohkMmm-LRgwhOYbiAPslhr_qf1JfnlF2Iw</recordid><startdate>20210601</startdate><enddate>20210601</enddate><creator>Xu, Wenshen</creator><creator>Li, Shuangyin</creator><creator>Lu, Yonghe</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L6V</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M7S</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PSYQQ</scope><scope>PTHSS</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0001-6404-3438</orcidid></search><sort><creationdate>20210601</creationdate><title>Usr-mtl: an unsupervised sentence representation learning framework with multi-task learning</title><author>Xu, Wenshen ; Li, Shuangyin ; Lu, Yonghe</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c319t-d71e44c1f1c441f297cdf022524e1c6202532194d062bd9f7f918dd0cde13f943</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Algorithms</topic><topic>Artificial Intelligence</topic><topic>Cognitive tasks</topic><topic>Computer Science</topic><topic>Downstream effects</topic><topic>Embedding</topic><topic>Feature extraction</topic><topic>Machine learning</topic><topic>Machines</topic><topic>Manufacturing</topic><topic>Mechanical Engineering</topic><topic>Natural language processing</topic><topic>Processes</topic><topic>Representations</topic><topic>Sentences</topic><topic>System effectiveness</topic><topic>Task analysis</topic><topic>Words (language)</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Xu, Wenshen</creatorcontrib><creatorcontrib>Li, Shuangyin</creatorcontrib><creatorcontrib>Lu, Yonghe</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>Access via ABI/INFORM (ProQuest)</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>ProQuest Engineering Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Engineering Database</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>ProQuest One Business</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest One Psychology</collection><collection>Engineering Collection</collection><collection>ProQuest Central Basic</collection><jtitle>Applied intelligence (Dordrecht, Netherlands)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Xu, Wenshen</au><au>Li, Shuangyin</au><au>Lu, Yonghe</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Usr-mtl: an unsupervised sentence representation learning framework with multi-task learning</atitle><jtitle>Applied intelligence (Dordrecht, Netherlands)</jtitle><stitle>Appl Intell</stitle><date>2021-06-01</date><risdate>2021</risdate><volume>51</volume><issue>6</issue><spage>3506</spage><epage>3521</epage><pages>3506-3521</pages><issn>0924-669X</issn><eissn>1573-7497</eissn><abstract>Developing the utilized intelligent systems is increasingly important to learn effective text representations, especially extract the sentence features. Numerous previous studies have been concentrated on the task of sentence representation learning based on deep learning approaches. However, the present approaches are mostly proposed with the single task or replied on the labeled corpus when learning the embedding of the sentences. In this paper, we assess the factors in learning sentence representation and propose an efficient unsupervised learning framework with multi-task learning (USR-MTL), in which various text learning tasks are merged into the unitized framework. With the syntactic and semantic features of sentences, three different factors to some extent are reflected in the task of the sentence representation learning that is the wording, or the ordering of the neighbored sentences of a target sentence in other words. Hence, we integrate the word-order learning task, word prediction task, and the sentence-order learning task into the proposed framework to attain meaningful sentence embeddings. Here, the process of sentence embedding learning is reformulated as a multi-task learning framework of the sentence-level task and the two word-level tasks. Moreover, the proposed framework is motivated by an unsupervised learning algorithm utilizing the unlabeled corpus. Based on the experimental results, our approach achieves the state-of-the-art performances on the downstream natural language processing tasks compared to the popular unsupervised representation learning techniques. The experiments on representation visualization and task analysis demonstrate the effectiveness of the tasks in the proposed framework in creating reasonable sentence representations proving the capacity of the proposed unsupervised multi-task framework for the sentence representation learning.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s10489-020-02042-2</doi><tpages>16</tpages><orcidid>https://orcid.org/0000-0001-6404-3438</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0924-669X |
ispartof | Applied intelligence (Dordrecht, Netherlands), 2021-06, Vol.51 (6), p.3506-3521 |
issn | 0924-669X 1573-7497 |
language | eng |
recordid | cdi_proquest_journals_2529006838 |
source | SpringerLink Journals - AutoHoldings |
subjects | Algorithms Artificial Intelligence Cognitive tasks Computer Science Downstream effects Embedding Feature extraction Machine learning Machines Manufacturing Mechanical Engineering Natural language processing Processes Representations Sentences System effectiveness Task analysis Words (language) |
title | Usr-mtl: an unsupervised sentence representation learning framework with multi-task learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-19T19%3A31%3A21IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Usr-mtl:%20an%20unsupervised%20sentence%20representation%20learning%20framework%20with%20multi-task%20learning&rft.jtitle=Applied%20intelligence%20(Dordrecht,%20Netherlands)&rft.au=Xu,%20Wenshen&rft.date=2021-06-01&rft.volume=51&rft.issue=6&rft.spage=3506&rft.epage=3521&rft.pages=3506-3521&rft.issn=0924-669X&rft.eissn=1573-7497&rft_id=info:doi/10.1007/s10489-020-02042-2&rft_dat=%3Cproquest_cross%3E2529006838%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2529006838&rft_id=info:pmid/&rfr_iscdi=true |