An unsupervised method for extractive multi-document summarization based on centroid approach and sentence embeddings
Extractive multi-document summarization (MDS) is the process of automatically summarizing a collection of documents by ranking sentences according to their importance and informativeness. Text representation is a fundamental process that affects the effectiveness of many text summarization methods....
Gespeichert in:
Veröffentlicht in: | Expert systems with applications 2021-04, Vol.167, p.114152, Article 114152 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | 114152 |
container_title | Expert systems with applications |
container_volume | 167 |
creator | Lamsiyah, Salima El Mahdaouy, Abdelkader Espinasse, Bernard El Alaoui Ouatik, Saïd |
description | Extractive multi-document summarization (MDS) is the process of automatically summarizing a collection of documents by ranking sentences according to their importance and informativeness. Text representation is a fundamental process that affects the effectiveness of many text summarization methods. Word embedding representations have shown to be effective for several Natural Language Processing (NLP) tasks including Automatic Text Summarization (ATS). However, most of these representations do not consider the order and the semantic relationships between words in a sentence. This does not fully allow grasping the sentence semantics and the syntactic relationships between sentences constituents. In this paper, to overcome this problem, we propose an unsupervised method for generic extractive multi-document summarization based on the sentence embedding representations and the centroid approach. The proposed method selects relevant sentences according to the final score obtained by combining three scores: sentence content relevance, sentence novelty, and sentence position scores. The sentence content relevance score is computed as the cosine similarity between the centroid embedding vector of the cluster of documents and the sentence embedding vectors. The sentence novelty metric is explicitly adopted to deal with redundancy. The sentence position metric assumes that the first sentences of a document are more relevant to the summary, and it assigns high scores to these sentences. Moreover, this paper provides a comparative analysis of nine sentence embedding models used to represent sentences as dense vectors in a low dimensional vector space in the context of extractive multi-document summarization. Experiments are performed on the standard DUC’2002–2004 benchmark datasets and the Multi-News dataset. The overall obtained results have shown that our method outperforms several state-of-the-art methods and achieves promising results compared to the best performing methods including supervised deep learning based methods.
•An unsupervised method for extractive multi-document summarization.•Pre-trained sentence embedding models are used for sentences representations.•Centroid approach is applied to compute the sentence content relevance score.•Sentence selection based on sentence relevance, novelty and position scores.•The use of sentence embedding methods leads to significant improvements. |
doi_str_mv | 10.1016/j.eswa.2020.114152 |
format | Article |
fullrecord | <record><control><sourceid>proquest_hal_p</sourceid><recordid>TN_cdi_hal_primary_oai_HAL_hal_03604013v1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417420308952</els_id><sourcerecordid>2503172547</sourcerecordid><originalsourceid>FETCH-LOGICAL-c406t-5688617185f3b6fd85c1a8c49ff0d5ed9fc3dd126ef78efe2a860e82af143ee93</originalsourceid><addsrcrecordid>eNp9UU1rGzEQFaWBumn-QE6CnnJYVx-7khZyMSFpAoZe2rOQpVEs4125ktZN8uurZUuOPc3w5r3hzTyErilZU0LFt8Ma8h-zZoRVgLa0Yx_QiirJGyF7_hGtSN_JpqWy_YQ-53wghEpC5ApNmxFPY55OkM4hg8MDlH102MeE4aUkY0s4Ax6mYwmNi3YaYCw4T8NgUngzJcQR78wsrI2tsxSDw-Z0StHYPTajw7miMFrAMOzAuTA-5y_owptjhqt_9RL9erj_effYbH98f7rbbBvbElGaTiglqKSq83wnvFOdpUbZtveeuA5c7y13jjIBXirwwIwSBBQznrYcoOeX6GbZuzdHfUqhmn7V0QT9uNnqGSNckJZQfqaV-3XhVuu_J8hFH-KUxmpPs45wKlnXyspiC8ummHMC_76WEj1HoQ96jkLPUegliiq6XURQbz0HSDrbML_EhQS2aBfD_-R_AVNglC0</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2503172547</pqid></control><display><type>article</type><title>An unsupervised method for extractive multi-document summarization based on centroid approach and sentence embeddings</title><source>Access via ScienceDirect (Elsevier)</source><creator>Lamsiyah, Salima ; El Mahdaouy, Abdelkader ; Espinasse, Bernard ; El Alaoui Ouatik, Saïd</creator><creatorcontrib>Lamsiyah, Salima ; El Mahdaouy, Abdelkader ; Espinasse, Bernard ; El Alaoui Ouatik, Saïd</creatorcontrib><description>Extractive multi-document summarization (MDS) is the process of automatically summarizing a collection of documents by ranking sentences according to their importance and informativeness. Text representation is a fundamental process that affects the effectiveness of many text summarization methods. Word embedding representations have shown to be effective for several Natural Language Processing (NLP) tasks including Automatic Text Summarization (ATS). However, most of these representations do not consider the order and the semantic relationships between words in a sentence. This does not fully allow grasping the sentence semantics and the syntactic relationships between sentences constituents. In this paper, to overcome this problem, we propose an unsupervised method for generic extractive multi-document summarization based on the sentence embedding representations and the centroid approach. The proposed method selects relevant sentences according to the final score obtained by combining three scores: sentence content relevance, sentence novelty, and sentence position scores. The sentence content relevance score is computed as the cosine similarity between the centroid embedding vector of the cluster of documents and the sentence embedding vectors. The sentence novelty metric is explicitly adopted to deal with redundancy. The sentence position metric assumes that the first sentences of a document are more relevant to the summary, and it assigns high scores to these sentences. Moreover, this paper provides a comparative analysis of nine sentence embedding models used to represent sentences as dense vectors in a low dimensional vector space in the context of extractive multi-document summarization. Experiments are performed on the standard DUC’2002–2004 benchmark datasets and the Multi-News dataset. The overall obtained results have shown that our method outperforms several state-of-the-art methods and achieves promising results compared to the best performing methods including supervised deep learning based methods.
•An unsupervised method for extractive multi-document summarization.•Pre-trained sentence embedding models are used for sentences representations.•Centroid approach is applied to compute the sentence content relevance score.•Sentence selection based on sentence relevance, novelty and position scores.•The use of sentence embedding methods leads to significant improvements.</description><identifier>ISSN: 0957-4174</identifier><identifier>EISSN: 1873-6793</identifier><identifier>DOI: 10.1016/j.eswa.2020.114152</identifier><language>eng</language><publisher>New York: Elsevier Ltd</publisher><subject>Artificial Intelligence ; Centroid approach ; Centroids ; Computation and Language ; Computer Science ; Datasets ; Embedding ; Extractive text summarization ; Methods ; Natural language processing ; Redundancy ; Representations ; Semantics ; Sentence embeddings ; Sentences ; Transfer learning ; Word embeddings</subject><ispartof>Expert systems with applications, 2021-04, Vol.167, p.114152, Article 114152</ispartof><rights>2020 Elsevier Ltd</rights><rights>Copyright Elsevier BV Apr 1, 2021</rights><rights>Distributed under a Creative Commons Attribution 4.0 International License</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c406t-5688617185f3b6fd85c1a8c49ff0d5ed9fc3dd126ef78efe2a860e82af143ee93</citedby><cites>FETCH-LOGICAL-c406t-5688617185f3b6fd85c1a8c49ff0d5ed9fc3dd126ef78efe2a860e82af143ee93</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.eswa.2020.114152$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>230,314,780,784,885,3550,27924,27925,45995</link.rule.ids><backlink>$$Uhttps://hal.science/hal-03604013$$DView record in HAL$$Hfree_for_read</backlink></links><search><creatorcontrib>Lamsiyah, Salima</creatorcontrib><creatorcontrib>El Mahdaouy, Abdelkader</creatorcontrib><creatorcontrib>Espinasse, Bernard</creatorcontrib><creatorcontrib>El Alaoui Ouatik, Saïd</creatorcontrib><title>An unsupervised method for extractive multi-document summarization based on centroid approach and sentence embeddings</title><title>Expert systems with applications</title><description>Extractive multi-document summarization (MDS) is the process of automatically summarizing a collection of documents by ranking sentences according to their importance and informativeness. Text representation is a fundamental process that affects the effectiveness of many text summarization methods. Word embedding representations have shown to be effective for several Natural Language Processing (NLP) tasks including Automatic Text Summarization (ATS). However, most of these representations do not consider the order and the semantic relationships between words in a sentence. This does not fully allow grasping the sentence semantics and the syntactic relationships between sentences constituents. In this paper, to overcome this problem, we propose an unsupervised method for generic extractive multi-document summarization based on the sentence embedding representations and the centroid approach. The proposed method selects relevant sentences according to the final score obtained by combining three scores: sentence content relevance, sentence novelty, and sentence position scores. The sentence content relevance score is computed as the cosine similarity between the centroid embedding vector of the cluster of documents and the sentence embedding vectors. The sentence novelty metric is explicitly adopted to deal with redundancy. The sentence position metric assumes that the first sentences of a document are more relevant to the summary, and it assigns high scores to these sentences. Moreover, this paper provides a comparative analysis of nine sentence embedding models used to represent sentences as dense vectors in a low dimensional vector space in the context of extractive multi-document summarization. Experiments are performed on the standard DUC’2002–2004 benchmark datasets and the Multi-News dataset. The overall obtained results have shown that our method outperforms several state-of-the-art methods and achieves promising results compared to the best performing methods including supervised deep learning based methods.
•An unsupervised method for extractive multi-document summarization.•Pre-trained sentence embedding models are used for sentences representations.•Centroid approach is applied to compute the sentence content relevance score.•Sentence selection based on sentence relevance, novelty and position scores.•The use of sentence embedding methods leads to significant improvements.</description><subject>Artificial Intelligence</subject><subject>Centroid approach</subject><subject>Centroids</subject><subject>Computation and Language</subject><subject>Computer Science</subject><subject>Datasets</subject><subject>Embedding</subject><subject>Extractive text summarization</subject><subject>Methods</subject><subject>Natural language processing</subject><subject>Redundancy</subject><subject>Representations</subject><subject>Semantics</subject><subject>Sentence embeddings</subject><subject>Sentences</subject><subject>Transfer learning</subject><subject>Word embeddings</subject><issn>0957-4174</issn><issn>1873-6793</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNp9UU1rGzEQFaWBumn-QE6CnnJYVx-7khZyMSFpAoZe2rOQpVEs4125ktZN8uurZUuOPc3w5r3hzTyErilZU0LFt8Ma8h-zZoRVgLa0Yx_QiirJGyF7_hGtSN_JpqWy_YQ-53wghEpC5ApNmxFPY55OkM4hg8MDlH102MeE4aUkY0s4Ax6mYwmNi3YaYCw4T8NgUngzJcQR78wsrI2tsxSDw-Z0StHYPTajw7miMFrAMOzAuTA-5y_owptjhqt_9RL9erj_effYbH98f7rbbBvbElGaTiglqKSq83wnvFOdpUbZtveeuA5c7y13jjIBXirwwIwSBBQznrYcoOeX6GbZuzdHfUqhmn7V0QT9uNnqGSNckJZQfqaV-3XhVuu_J8hFH-KUxmpPs45wKlnXyspiC8ummHMC_76WEj1HoQ96jkLPUegliiq6XURQbz0HSDrbML_EhQS2aBfD_-R_AVNglC0</recordid><startdate>20210401</startdate><enddate>20210401</enddate><creator>Lamsiyah, Salima</creator><creator>El Mahdaouy, Abdelkader</creator><creator>Espinasse, Bernard</creator><creator>El Alaoui Ouatik, Saïd</creator><general>Elsevier Ltd</general><general>Elsevier BV</general><general>Elsevier</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>1XC</scope></search><sort><creationdate>20210401</creationdate><title>An unsupervised method for extractive multi-document summarization based on centroid approach and sentence embeddings</title><author>Lamsiyah, Salima ; El Mahdaouy, Abdelkader ; Espinasse, Bernard ; El Alaoui Ouatik, Saïd</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c406t-5688617185f3b6fd85c1a8c49ff0d5ed9fc3dd126ef78efe2a860e82af143ee93</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Artificial Intelligence</topic><topic>Centroid approach</topic><topic>Centroids</topic><topic>Computation and Language</topic><topic>Computer Science</topic><topic>Datasets</topic><topic>Embedding</topic><topic>Extractive text summarization</topic><topic>Methods</topic><topic>Natural language processing</topic><topic>Redundancy</topic><topic>Representations</topic><topic>Semantics</topic><topic>Sentence embeddings</topic><topic>Sentences</topic><topic>Transfer learning</topic><topic>Word embeddings</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Lamsiyah, Salima</creatorcontrib><creatorcontrib>El Mahdaouy, Abdelkader</creatorcontrib><creatorcontrib>Espinasse, Bernard</creatorcontrib><creatorcontrib>El Alaoui Ouatik, Saïd</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Hyper Article en Ligne (HAL)</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Lamsiyah, Salima</au><au>El Mahdaouy, Abdelkader</au><au>Espinasse, Bernard</au><au>El Alaoui Ouatik, Saïd</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>An unsupervised method for extractive multi-document summarization based on centroid approach and sentence embeddings</atitle><jtitle>Expert systems with applications</jtitle><date>2021-04-01</date><risdate>2021</risdate><volume>167</volume><spage>114152</spage><pages>114152-</pages><artnum>114152</artnum><issn>0957-4174</issn><eissn>1873-6793</eissn><abstract>Extractive multi-document summarization (MDS) is the process of automatically summarizing a collection of documents by ranking sentences according to their importance and informativeness. Text representation is a fundamental process that affects the effectiveness of many text summarization methods. Word embedding representations have shown to be effective for several Natural Language Processing (NLP) tasks including Automatic Text Summarization (ATS). However, most of these representations do not consider the order and the semantic relationships between words in a sentence. This does not fully allow grasping the sentence semantics and the syntactic relationships between sentences constituents. In this paper, to overcome this problem, we propose an unsupervised method for generic extractive multi-document summarization based on the sentence embedding representations and the centroid approach. The proposed method selects relevant sentences according to the final score obtained by combining three scores: sentence content relevance, sentence novelty, and sentence position scores. The sentence content relevance score is computed as the cosine similarity between the centroid embedding vector of the cluster of documents and the sentence embedding vectors. The sentence novelty metric is explicitly adopted to deal with redundancy. The sentence position metric assumes that the first sentences of a document are more relevant to the summary, and it assigns high scores to these sentences. Moreover, this paper provides a comparative analysis of nine sentence embedding models used to represent sentences as dense vectors in a low dimensional vector space in the context of extractive multi-document summarization. Experiments are performed on the standard DUC’2002–2004 benchmark datasets and the Multi-News dataset. The overall obtained results have shown that our method outperforms several state-of-the-art methods and achieves promising results compared to the best performing methods including supervised deep learning based methods.
•An unsupervised method for extractive multi-document summarization.•Pre-trained sentence embedding models are used for sentences representations.•Centroid approach is applied to compute the sentence content relevance score.•Sentence selection based on sentence relevance, novelty and position scores.•The use of sentence embedding methods leads to significant improvements.</abstract><cop>New York</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2020.114152</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0957-4174 |
ispartof | Expert systems with applications, 2021-04, Vol.167, p.114152, Article 114152 |
issn | 0957-4174 1873-6793 |
language | eng |
recordid | cdi_hal_primary_oai_HAL_hal_03604013v1 |
source | Access via ScienceDirect (Elsevier) |
subjects | Artificial Intelligence Centroid approach Centroids Computation and Language Computer Science Datasets Embedding Extractive text summarization Methods Natural language processing Redundancy Representations Semantics Sentence embeddings Sentences Transfer learning Word embeddings |
title | An unsupervised method for extractive multi-document summarization based on centroid approach and sentence embeddings |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-20T01%3A27%3A31IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_hal_p&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=An%20unsupervised%20method%20for%20extractive%20multi-document%20summarization%20based%20on%20centroid%20approach%20and%20sentence%20embeddings&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Lamsiyah,%20Salima&rft.date=2021-04-01&rft.volume=167&rft.spage=114152&rft.pages=114152-&rft.artnum=114152&rft.issn=0957-4174&rft.eissn=1873-6793&rft_id=info:doi/10.1016/j.eswa.2020.114152&rft_dat=%3Cproquest_hal_p%3E2503172547%3C/proquest_hal_p%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2503172547&rft_id=info:pmid/&rft_els_id=S0957417420308952&rfr_iscdi=true |