Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM
•Bio-Inspired novel attention-based inception architecture is proposed that can adapt to different types of spatial contexts using convolution filters of different sizes. The characteristics of each dataset are unique, hence the attention mechanism helps focus on those features to improve classifica...
Gespeichert in:
Veröffentlicht in: | Computers & electrical engineering 2021-10, Vol.95, p.107395, Article 107395 |
---|---|
Hauptverfasser: | , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | 107395 |
container_title | Computers & electrical engineering |
container_volume | 95 |
creator | Abdul, Wadood Alsulaiman, Mansour Amin, Syed Umar Faisal, Mohammed Muhammad, Ghulam Albogamy, Fahad R. Bencherif, Mohamed A. Ghaleb, Hamid |
description | •Bio-Inspired novel attention-based inception architecture is proposed that can adapt to different types of spatial contexts using convolution filters of different sizes. The characteristics of each dataset are unique, hence the attention mechanism helps focus on those features to improve classification performance.•The shallow inception model is designed with a two-layer attention mechanism with fewer layers but with a large number of convolution filters that can address the overfitting problem caused by small dataset sizes.•LSTM-based recurrent neural network (RNN) module is proposed to extract temporal features after the inception module is applied.•The proposed model is lightweight with fewer parameters and has less processing time.•The proposed model achieves good performance for both dynamic and static signs and gestures.
Bio-inspired deep learning models have revolutionized sign language classification, achieving extraordinary accuracy and human-like video understanding. Recognition and classification of sign language videos in real-time are challenging because the duration and speed of each sign vary for different subjects, the background of videos is dynamic in most cases, and the classification result should be produced in real-time. This study proposes a model based on a convolution neural network (CNN) Inception model with an attention mechanism for extracting spatial features and Bi-LSTM (long short-term memory) for temporal feature extraction. The proposed model is tested on datasets with highly variable characteristics such as different clothing, variable lighting, and variable distance from the camera. Real-time classification achieves significant early detections while achieving performance comparable to the offline operation. The proposed model has fewer parameters, fewer deep learning layers, and requires significantly less processing time than state-of-the-art models.
The Inception model with an attention mechanism with two attention blocks [Display omitted] |
doi_str_mv | 10.1016/j.compeleceng.2021.107395 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2594202883</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0045790621003621</els_id><sourcerecordid>2594202883</sourcerecordid><originalsourceid>FETCH-LOGICAL-c415t-c989866d5ac3cf9e2cb2f8c4879dd5c1832a0868826f4d02b1758885a725357c3</originalsourceid><addsrcrecordid>eNqNkE9PwzAMxSMEEmPwHYI4dyRp06bHMfFn0hAHxjnKHLfK1KUjyZD49gTGgSMn69l-tt6PkGvOZpzx-nY7g3G3xwEBfT8TTPDcb8pWnpAJV01bsEbKUzJhrJJF07L6nFzEuGVZ11xNiF36hMPgevSJBjRDkdwO6TyYjQMaXe_pYHx_MD1SGEyMrnNgkhs9PUTne2pSytasi42JaKnzgPufufGW3rnV6_r5kpx1Zoh49Vun5O3hfr14KlYvj8vFfFVAxWUqoFWtqmsrDZTQtShgIzoFVY5hrQSuSmGYqpUSdVdZJja8kUopaRohS9lAOSU3x7v7ML4fMCa9HQ_B55dayLbKdJQq81Z73IIwxhiw0_vgdiZ8as70N1S91X-g6m-o-gg1exdHL-YYHw6DjuAwR7YuICRtR_ePK18D1oZd</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2594202883</pqid></control><display><type>article</type><title>Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM</title><source>Access via ScienceDirect (Elsevier)</source><creator>Abdul, Wadood ; Alsulaiman, Mansour ; Amin, Syed Umar ; Faisal, Mohammed ; Muhammad, Ghulam ; Albogamy, Fahad R. ; Bencherif, Mohamed A. ; Ghaleb, Hamid</creator><creatorcontrib>Abdul, Wadood ; Alsulaiman, Mansour ; Amin, Syed Umar ; Faisal, Mohammed ; Muhammad, Ghulam ; Albogamy, Fahad R. ; Bencherif, Mohamed A. ; Ghaleb, Hamid</creatorcontrib><description>•Bio-Inspired novel attention-based inception architecture is proposed that can adapt to different types of spatial contexts using convolution filters of different sizes. The characteristics of each dataset are unique, hence the attention mechanism helps focus on those features to improve classification performance.•The shallow inception model is designed with a two-layer attention mechanism with fewer layers but with a large number of convolution filters that can address the overfitting problem caused by small dataset sizes.•LSTM-based recurrent neural network (RNN) module is proposed to extract temporal features after the inception module is applied.•The proposed model is lightweight with fewer parameters and has less processing time.•The proposed model achieves good performance for both dynamic and static signs and gestures.
Bio-inspired deep learning models have revolutionized sign language classification, achieving extraordinary accuracy and human-like video understanding. Recognition and classification of sign language videos in real-time are challenging because the duration and speed of each sign vary for different subjects, the background of videos is dynamic in most cases, and the classification result should be produced in real-time. This study proposes a model based on a convolution neural network (CNN) Inception model with an attention mechanism for extracting spatial features and Bi-LSTM (long short-term memory) for temporal feature extraction. The proposed model is tested on datasets with highly variable characteristics such as different clothing, variable lighting, and variable distance from the camera. Real-time classification achieves significant early detections while achieving performance comparable to the offline operation. The proposed model has fewer parameters, fewer deep learning layers, and requires significantly less processing time than state-of-the-art models.
The Inception model with an attention mechanism with two attention blocks [Display omitted]</description><identifier>ISSN: 0045-7906</identifier><identifier>EISSN: 1879-0755</identifier><identifier>DOI: 10.1016/j.compeleceng.2021.107395</identifier><language>eng</language><publisher>Amsterdam: Elsevier Ltd</publisher><subject>Artificial neural networks ; BiLSTM ; Bio-inspired computing ; Biomimetics ; Classification ; Deep learning ; Feature extraction ; Inception ; Machine learning ; Real time ; Real-time classification ; Sign language ; Video data</subject><ispartof>Computers & electrical engineering, 2021-10, Vol.95, p.107395, Article 107395</ispartof><rights>2021</rights><rights>Copyright Elsevier BV Oct 2021</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c415t-c989866d5ac3cf9e2cb2f8c4879dd5c1832a0868826f4d02b1758885a725357c3</citedby><cites>FETCH-LOGICAL-c415t-c989866d5ac3cf9e2cb2f8c4879dd5c1832a0868826f4d02b1758885a725357c3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.compeleceng.2021.107395$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,780,784,3550,27924,27925,45995</link.rule.ids></links><search><creatorcontrib>Abdul, Wadood</creatorcontrib><creatorcontrib>Alsulaiman, Mansour</creatorcontrib><creatorcontrib>Amin, Syed Umar</creatorcontrib><creatorcontrib>Faisal, Mohammed</creatorcontrib><creatorcontrib>Muhammad, Ghulam</creatorcontrib><creatorcontrib>Albogamy, Fahad R.</creatorcontrib><creatorcontrib>Bencherif, Mohamed A.</creatorcontrib><creatorcontrib>Ghaleb, Hamid</creatorcontrib><title>Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM</title><title>Computers & electrical engineering</title><description>•Bio-Inspired novel attention-based inception architecture is proposed that can adapt to different types of spatial contexts using convolution filters of different sizes. The characteristics of each dataset are unique, hence the attention mechanism helps focus on those features to improve classification performance.•The shallow inception model is designed with a two-layer attention mechanism with fewer layers but with a large number of convolution filters that can address the overfitting problem caused by small dataset sizes.•LSTM-based recurrent neural network (RNN) module is proposed to extract temporal features after the inception module is applied.•The proposed model is lightweight with fewer parameters and has less processing time.•The proposed model achieves good performance for both dynamic and static signs and gestures.
Bio-inspired deep learning models have revolutionized sign language classification, achieving extraordinary accuracy and human-like video understanding. Recognition and classification of sign language videos in real-time are challenging because the duration and speed of each sign vary for different subjects, the background of videos is dynamic in most cases, and the classification result should be produced in real-time. This study proposes a model based on a convolution neural network (CNN) Inception model with an attention mechanism for extracting spatial features and Bi-LSTM (long short-term memory) for temporal feature extraction. The proposed model is tested on datasets with highly variable characteristics such as different clothing, variable lighting, and variable distance from the camera. Real-time classification achieves significant early detections while achieving performance comparable to the offline operation. The proposed model has fewer parameters, fewer deep learning layers, and requires significantly less processing time than state-of-the-art models.
The Inception model with an attention mechanism with two attention blocks [Display omitted]</description><subject>Artificial neural networks</subject><subject>BiLSTM</subject><subject>Bio-inspired computing</subject><subject>Biomimetics</subject><subject>Classification</subject><subject>Deep learning</subject><subject>Feature extraction</subject><subject>Inception</subject><subject>Machine learning</subject><subject>Real time</subject><subject>Real-time classification</subject><subject>Sign language</subject><subject>Video data</subject><issn>0045-7906</issn><issn>1879-0755</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNqNkE9PwzAMxSMEEmPwHYI4dyRp06bHMfFn0hAHxjnKHLfK1KUjyZD49gTGgSMn69l-tt6PkGvOZpzx-nY7g3G3xwEBfT8TTPDcb8pWnpAJV01bsEbKUzJhrJJF07L6nFzEuGVZ11xNiF36hMPgevSJBjRDkdwO6TyYjQMaXe_pYHx_MD1SGEyMrnNgkhs9PUTne2pSytasi42JaKnzgPufufGW3rnV6_r5kpx1Zoh49Vun5O3hfr14KlYvj8vFfFVAxWUqoFWtqmsrDZTQtShgIzoFVY5hrQSuSmGYqpUSdVdZJja8kUopaRohS9lAOSU3x7v7ML4fMCa9HQ_B55dayLbKdJQq81Z73IIwxhiw0_vgdiZ8as70N1S91X-g6m-o-gg1exdHL-YYHw6DjuAwR7YuICRtR_ePK18D1oZd</recordid><startdate>202110</startdate><enddate>202110</enddate><creator>Abdul, Wadood</creator><creator>Alsulaiman, Mansour</creator><creator>Amin, Syed Umar</creator><creator>Faisal, Mohammed</creator><creator>Muhammad, Ghulam</creator><creator>Albogamy, Fahad R.</creator><creator>Bencherif, Mohamed A.</creator><creator>Ghaleb, Hamid</creator><general>Elsevier Ltd</general><general>Elsevier BV</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>202110</creationdate><title>Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM</title><author>Abdul, Wadood ; Alsulaiman, Mansour ; Amin, Syed Umar ; Faisal, Mohammed ; Muhammad, Ghulam ; Albogamy, Fahad R. ; Bencherif, Mohamed A. ; Ghaleb, Hamid</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c415t-c989866d5ac3cf9e2cb2f8c4879dd5c1832a0868826f4d02b1758885a725357c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Artificial neural networks</topic><topic>BiLSTM</topic><topic>Bio-inspired computing</topic><topic>Biomimetics</topic><topic>Classification</topic><topic>Deep learning</topic><topic>Feature extraction</topic><topic>Inception</topic><topic>Machine learning</topic><topic>Real time</topic><topic>Real-time classification</topic><topic>Sign language</topic><topic>Video data</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Abdul, Wadood</creatorcontrib><creatorcontrib>Alsulaiman, Mansour</creatorcontrib><creatorcontrib>Amin, Syed Umar</creatorcontrib><creatorcontrib>Faisal, Mohammed</creatorcontrib><creatorcontrib>Muhammad, Ghulam</creatorcontrib><creatorcontrib>Albogamy, Fahad R.</creatorcontrib><creatorcontrib>Bencherif, Mohamed A.</creatorcontrib><creatorcontrib>Ghaleb, Hamid</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Computers & electrical engineering</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Abdul, Wadood</au><au>Alsulaiman, Mansour</au><au>Amin, Syed Umar</au><au>Faisal, Mohammed</au><au>Muhammad, Ghulam</au><au>Albogamy, Fahad R.</au><au>Bencherif, Mohamed A.</au><au>Ghaleb, Hamid</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM</atitle><jtitle>Computers & electrical engineering</jtitle><date>2021-10</date><risdate>2021</risdate><volume>95</volume><spage>107395</spage><pages>107395-</pages><artnum>107395</artnum><issn>0045-7906</issn><eissn>1879-0755</eissn><abstract>•Bio-Inspired novel attention-based inception architecture is proposed that can adapt to different types of spatial contexts using convolution filters of different sizes. The characteristics of each dataset are unique, hence the attention mechanism helps focus on those features to improve classification performance.•The shallow inception model is designed with a two-layer attention mechanism with fewer layers but with a large number of convolution filters that can address the overfitting problem caused by small dataset sizes.•LSTM-based recurrent neural network (RNN) module is proposed to extract temporal features after the inception module is applied.•The proposed model is lightweight with fewer parameters and has less processing time.•The proposed model achieves good performance for both dynamic and static signs and gestures.
Bio-inspired deep learning models have revolutionized sign language classification, achieving extraordinary accuracy and human-like video understanding. Recognition and classification of sign language videos in real-time are challenging because the duration and speed of each sign vary for different subjects, the background of videos is dynamic in most cases, and the classification result should be produced in real-time. This study proposes a model based on a convolution neural network (CNN) Inception model with an attention mechanism for extracting spatial features and Bi-LSTM (long short-term memory) for temporal feature extraction. The proposed model is tested on datasets with highly variable characteristics such as different clothing, variable lighting, and variable distance from the camera. Real-time classification achieves significant early detections while achieving performance comparable to the offline operation. The proposed model has fewer parameters, fewer deep learning layers, and requires significantly less processing time than state-of-the-art models.
The Inception model with an attention mechanism with two attention blocks [Display omitted]</abstract><cop>Amsterdam</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.compeleceng.2021.107395</doi></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0045-7906 |
ispartof | Computers & electrical engineering, 2021-10, Vol.95, p.107395, Article 107395 |
issn | 0045-7906 1879-0755 |
language | eng |
recordid | cdi_proquest_journals_2594202883 |
source | Access via ScienceDirect (Elsevier) |
subjects | Artificial neural networks BiLSTM Bio-inspired computing Biomimetics Classification Deep learning Feature extraction Inception Machine learning Real time Real-time classification Sign language Video data |
title | Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-21T08%3A25%3A08IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Intelligent%20real-time%20Arabic%20sign%20language%20classification%20using%20attention-based%20inception%20and%20BiLSTM&rft.jtitle=Computers%20&%20electrical%20engineering&rft.au=Abdul,%20Wadood&rft.date=2021-10&rft.volume=95&rft.spage=107395&rft.pages=107395-&rft.artnum=107395&rft.issn=0045-7906&rft.eissn=1879-0755&rft_id=info:doi/10.1016/j.compeleceng.2021.107395&rft_dat=%3Cproquest_cross%3E2594202883%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2594202883&rft_id=info:pmid/&rft_els_id=S0045790621003621&rfr_iscdi=true |