Attention, please! A survey of Neural Attention Models in Deep Learning

In humans, Attention is a core property of all perceptual and cognitive operations. Given our limited ability to process competing sources, attention mechanisms select, modulate, and focus on the information most relevant to behavior. For decades, concepts and functions of attention have been studie...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Correia, Alana de Santana, Colombini, Esther Luna
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Correia, Alana de Santana
Colombini, Esther Luna
description In humans, Attention is a core property of all perceptual and cognitive operations. Given our limited ability to process competing sources, attention mechanisms select, modulate, and focus on the information most relevant to behavior. For decades, concepts and functions of attention have been studied in philosophy, psychology, neuroscience, and computing. For the last six years, this property has been widely explored in deep neural networks. Currently, the state-of-the-art in Deep Learning is represented by neural attention models in several application domains. This survey provides a comprehensive overview and analysis of developments in neural attention models. We systematically reviewed hundreds of architectures in the area, identifying and discussing those in which attention has shown a significant impact. We also developed and made public an automated methodology to facilitate the development of reviews in the area. By critically analyzing 650 works, we describe the primary uses of attention in convolutional, recurrent networks and generative models, identifying common subgroups of uses and applications. Furthermore, we describe the impact of attention in different application domains and their impact on neural networks' interpretability. Finally, we list possible trends and opportunities for further research, hoping that this review will provide a succinct overview of the main attentional models in the area and guide researchers in developing future approaches that will drive further improvements.
doi_str_mv 10.48550/arxiv.2103.16775
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2103_16775</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2103_16775</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-c9f9d7e8cb844c3c461c33cbf8315fd6cc5687a200c999a5cb930b25067274913</originalsourceid><addsrcrecordid>eNo9zz1PwzAUhWEvDKjwA5i43Umw4-8xKlCQAl26R45zjSwFJ3LSiv57REGdzvLqSA8hd4yWwkhJH13-jseyYpSXTGktr8m2XhZMSxzTA0wDuhnXUMN8yEc8wRjgAw_ZDXCp4H3scZghJnhCnKBBl1NMnzfkKrhhxtv_XZH9y_N-81o0u-3bpm4Kp7QsvA2212h8Z4Tw3AvFPOe-C4YzGXrlvVRGu4pSb6110neW066SVOlKC8v4itz_3Z4l7ZTjl8un9lfUnkX8B7KHRQo</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Attention, please! A survey of Neural Attention Models in Deep Learning</title><source>arXiv.org</source><creator>Correia, Alana de Santana ; Colombini, Esther Luna</creator><creatorcontrib>Correia, Alana de Santana ; Colombini, Esther Luna</creatorcontrib><description>In humans, Attention is a core property of all perceptual and cognitive operations. Given our limited ability to process competing sources, attention mechanisms select, modulate, and focus on the information most relevant to behavior. For decades, concepts and functions of attention have been studied in philosophy, psychology, neuroscience, and computing. For the last six years, this property has been widely explored in deep neural networks. Currently, the state-of-the-art in Deep Learning is represented by neural attention models in several application domains. This survey provides a comprehensive overview and analysis of developments in neural attention models. We systematically reviewed hundreds of architectures in the area, identifying and discussing those in which attention has shown a significant impact. We also developed and made public an automated methodology to facilitate the development of reviews in the area. By critically analyzing 650 works, we describe the primary uses of attention in convolutional, recurrent networks and generative models, identifying common subgroups of uses and applications. Furthermore, we describe the impact of attention in different application domains and their impact on neural networks' interpretability. Finally, we list possible trends and opportunities for further research, hoping that this review will provide a succinct overview of the main attentional models in the area and guide researchers in developing future approaches that will drive further improvements.</description><identifier>DOI: 10.48550/arxiv.2103.16775</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning ; Computer Science - Robotics</subject><creationdate>2021-03</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2103.16775$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2103.16775$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Correia, Alana de Santana</creatorcontrib><creatorcontrib>Colombini, Esther Luna</creatorcontrib><title>Attention, please! A survey of Neural Attention Models in Deep Learning</title><description>In humans, Attention is a core property of all perceptual and cognitive operations. Given our limited ability to process competing sources, attention mechanisms select, modulate, and focus on the information most relevant to behavior. For decades, concepts and functions of attention have been studied in philosophy, psychology, neuroscience, and computing. For the last six years, this property has been widely explored in deep neural networks. Currently, the state-of-the-art in Deep Learning is represented by neural attention models in several application domains. This survey provides a comprehensive overview and analysis of developments in neural attention models. We systematically reviewed hundreds of architectures in the area, identifying and discussing those in which attention has shown a significant impact. We also developed and made public an automated methodology to facilitate the development of reviews in the area. By critically analyzing 650 works, we describe the primary uses of attention in convolutional, recurrent networks and generative models, identifying common subgroups of uses and applications. Furthermore, we describe the impact of attention in different application domains and their impact on neural networks' interpretability. Finally, we list possible trends and opportunities for further research, hoping that this review will provide a succinct overview of the main attentional models in the area and guide researchers in developing future approaches that will drive further improvements.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><subject>Computer Science - Robotics</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNo9zz1PwzAUhWEvDKjwA5i43Umw4-8xKlCQAl26R45zjSwFJ3LSiv57REGdzvLqSA8hd4yWwkhJH13-jseyYpSXTGktr8m2XhZMSxzTA0wDuhnXUMN8yEc8wRjgAw_ZDXCp4H3scZghJnhCnKBBl1NMnzfkKrhhxtv_XZH9y_N-81o0u-3bpm4Kp7QsvA2212h8Z4Tw3AvFPOe-C4YzGXrlvVRGu4pSb6110neW066SVOlKC8v4itz_3Z4l7ZTjl8un9lfUnkX8B7KHRQo</recordid><startdate>20210330</startdate><enddate>20210330</enddate><creator>Correia, Alana de Santana</creator><creator>Colombini, Esther Luna</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20210330</creationdate><title>Attention, please! A survey of Neural Attention Models in Deep Learning</title><author>Correia, Alana de Santana ; Colombini, Esther Luna</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-c9f9d7e8cb844c3c461c33cbf8315fd6cc5687a200c999a5cb930b25067274913</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><topic>Computer Science - Robotics</topic><toplevel>online_resources</toplevel><creatorcontrib>Correia, Alana de Santana</creatorcontrib><creatorcontrib>Colombini, Esther Luna</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Correia, Alana de Santana</au><au>Colombini, Esther Luna</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Attention, please! A survey of Neural Attention Models in Deep Learning</atitle><date>2021-03-30</date><risdate>2021</risdate><abstract>In humans, Attention is a core property of all perceptual and cognitive operations. Given our limited ability to process competing sources, attention mechanisms select, modulate, and focus on the information most relevant to behavior. For decades, concepts and functions of attention have been studied in philosophy, psychology, neuroscience, and computing. For the last six years, this property has been widely explored in deep neural networks. Currently, the state-of-the-art in Deep Learning is represented by neural attention models in several application domains. This survey provides a comprehensive overview and analysis of developments in neural attention models. We systematically reviewed hundreds of architectures in the area, identifying and discussing those in which attention has shown a significant impact. We also developed and made public an automated methodology to facilitate the development of reviews in the area. By critically analyzing 650 works, we describe the primary uses of attention in convolutional, recurrent networks and generative models, identifying common subgroups of uses and applications. Furthermore, we describe the impact of attention in different application domains and their impact on neural networks' interpretability. Finally, we list possible trends and opportunities for further research, hoping that this review will provide a succinct overview of the main attentional models in the area and guide researchers in developing future approaches that will drive further improvements.</abstract><doi>10.48550/arxiv.2103.16775</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2103.16775
ispartof
issn
language eng
recordid cdi_arxiv_primary_2103_16775
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computer Vision and Pattern Recognition
Computer Science - Learning
Computer Science - Robotics
title Attention, please! A survey of Neural Attention Models in Deep Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-13T05%3A37%3A07IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Attention,%20please!%20A%20survey%20of%20Neural%20Attention%20Models%20in%20Deep%20Learning&rft.au=Correia,%20Alana%20de%20Santana&rft.date=2021-03-30&rft_id=info:doi/10.48550/arxiv.2103.16775&rft_dat=%3Carxiv_GOX%3E2103_16775%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true