Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM

•Bio-Inspired novel attention-based inception architecture is proposed that can adapt to different types of spatial contexts using convolution filters of different sizes. The characteristics of each dataset are unique, hence the attention mechanism helps focus on those features to improve classifica...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Computers & electrical engineering 2021-10, Vol.95, p.107395, Article 107395
Hauptverfasser: Abdul, Wadood, Alsulaiman, Mansour, Amin, Syed Umar, Faisal, Mohammed, Muhammad, Ghulam, Albogamy, Fahad R., Bencherif, Mohamed A., Ghaleb, Hamid
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page 107395
container_title Computers & electrical engineering
container_volume 95
creator Abdul, Wadood
Alsulaiman, Mansour
Amin, Syed Umar
Faisal, Mohammed
Muhammad, Ghulam
Albogamy, Fahad R.
Bencherif, Mohamed A.
Ghaleb, Hamid
description •Bio-Inspired novel attention-based inception architecture is proposed that can adapt to different types of spatial contexts using convolution filters of different sizes. The characteristics of each dataset are unique, hence the attention mechanism helps focus on those features to improve classification performance.•The shallow inception model is designed with a two-layer attention mechanism with fewer layers but with a large number of convolution filters that can address the overfitting problem caused by small dataset sizes.•LSTM-based recurrent neural network (RNN) module is proposed to extract temporal features after the inception module is applied.•The proposed model is lightweight with fewer parameters and has less processing time.•The proposed model achieves good performance for both dynamic and static signs and gestures. Bio-inspired deep learning models have revolutionized sign language classification, achieving extraordinary accuracy and human-like video understanding. Recognition and classification of sign language videos in real-time are challenging because the duration and speed of each sign vary for different subjects, the background of videos is dynamic in most cases, and the classification result should be produced in real-time. This study proposes a model based on a convolution neural network (CNN) Inception model with an attention mechanism for extracting spatial features and Bi-LSTM (long short-term memory) for temporal feature extraction. The proposed model is tested on datasets with highly variable characteristics such as different clothing, variable lighting, and variable distance from the camera. Real-time classification achieves significant early detections while achieving performance comparable to the offline operation. The proposed model has fewer parameters, fewer deep learning layers, and requires significantly less processing time than state-of-the-art models. The Inception model with an attention mechanism with two attention blocks [Display omitted]
doi_str_mv 10.1016/j.compeleceng.2021.107395
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2594202883</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0045790621003621</els_id><sourcerecordid>2594202883</sourcerecordid><originalsourceid>FETCH-LOGICAL-c415t-c989866d5ac3cf9e2cb2f8c4879dd5c1832a0868826f4d02b1758885a725357c3</originalsourceid><addsrcrecordid>eNqNkE9PwzAMxSMEEmPwHYI4dyRp06bHMfFn0hAHxjnKHLfK1KUjyZD49gTGgSMn69l-tt6PkGvOZpzx-nY7g3G3xwEBfT8TTPDcb8pWnpAJV01bsEbKUzJhrJJF07L6nFzEuGVZ11xNiF36hMPgevSJBjRDkdwO6TyYjQMaXe_pYHx_MD1SGEyMrnNgkhs9PUTne2pSytasi42JaKnzgPufufGW3rnV6_r5kpx1Zoh49Vun5O3hfr14KlYvj8vFfFVAxWUqoFWtqmsrDZTQtShgIzoFVY5hrQSuSmGYqpUSdVdZJja8kUopaRohS9lAOSU3x7v7ML4fMCa9HQ_B55dayLbKdJQq81Z73IIwxhiw0_vgdiZ8as70N1S91X-g6m-o-gg1exdHL-YYHw6DjuAwR7YuICRtR_ePK18D1oZd</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2594202883</pqid></control><display><type>article</type><title>Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM</title><source>Access via ScienceDirect (Elsevier)</source><creator>Abdul, Wadood ; Alsulaiman, Mansour ; Amin, Syed Umar ; Faisal, Mohammed ; Muhammad, Ghulam ; Albogamy, Fahad R. ; Bencherif, Mohamed A. ; Ghaleb, Hamid</creator><creatorcontrib>Abdul, Wadood ; Alsulaiman, Mansour ; Amin, Syed Umar ; Faisal, Mohammed ; Muhammad, Ghulam ; Albogamy, Fahad R. ; Bencherif, Mohamed A. ; Ghaleb, Hamid</creatorcontrib><description>•Bio-Inspired novel attention-based inception architecture is proposed that can adapt to different types of spatial contexts using convolution filters of different sizes. The characteristics of each dataset are unique, hence the attention mechanism helps focus on those features to improve classification performance.•The shallow inception model is designed with a two-layer attention mechanism with fewer layers but with a large number of convolution filters that can address the overfitting problem caused by small dataset sizes.•LSTM-based recurrent neural network (RNN) module is proposed to extract temporal features after the inception module is applied.•The proposed model is lightweight with fewer parameters and has less processing time.•The proposed model achieves good performance for both dynamic and static signs and gestures. Bio-inspired deep learning models have revolutionized sign language classification, achieving extraordinary accuracy and human-like video understanding. Recognition and classification of sign language videos in real-time are challenging because the duration and speed of each sign vary for different subjects, the background of videos is dynamic in most cases, and the classification result should be produced in real-time. This study proposes a model based on a convolution neural network (CNN) Inception model with an attention mechanism for extracting spatial features and Bi-LSTM (long short-term memory) for temporal feature extraction. The proposed model is tested on datasets with highly variable characteristics such as different clothing, variable lighting, and variable distance from the camera. Real-time classification achieves significant early detections while achieving performance comparable to the offline operation. The proposed model has fewer parameters, fewer deep learning layers, and requires significantly less processing time than state-of-the-art models. The Inception model with an attention mechanism with two attention blocks [Display omitted]</description><identifier>ISSN: 0045-7906</identifier><identifier>EISSN: 1879-0755</identifier><identifier>DOI: 10.1016/j.compeleceng.2021.107395</identifier><language>eng</language><publisher>Amsterdam: Elsevier Ltd</publisher><subject>Artificial neural networks ; BiLSTM ; Bio-inspired computing ; Biomimetics ; Classification ; Deep learning ; Feature extraction ; Inception ; Machine learning ; Real time ; Real-time classification ; Sign language ; Video data</subject><ispartof>Computers &amp; electrical engineering, 2021-10, Vol.95, p.107395, Article 107395</ispartof><rights>2021</rights><rights>Copyright Elsevier BV Oct 2021</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c415t-c989866d5ac3cf9e2cb2f8c4879dd5c1832a0868826f4d02b1758885a725357c3</citedby><cites>FETCH-LOGICAL-c415t-c989866d5ac3cf9e2cb2f8c4879dd5c1832a0868826f4d02b1758885a725357c3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.compeleceng.2021.107395$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,780,784,3550,27924,27925,45995</link.rule.ids></links><search><creatorcontrib>Abdul, Wadood</creatorcontrib><creatorcontrib>Alsulaiman, Mansour</creatorcontrib><creatorcontrib>Amin, Syed Umar</creatorcontrib><creatorcontrib>Faisal, Mohammed</creatorcontrib><creatorcontrib>Muhammad, Ghulam</creatorcontrib><creatorcontrib>Albogamy, Fahad R.</creatorcontrib><creatorcontrib>Bencherif, Mohamed A.</creatorcontrib><creatorcontrib>Ghaleb, Hamid</creatorcontrib><title>Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM</title><title>Computers &amp; electrical engineering</title><description>•Bio-Inspired novel attention-based inception architecture is proposed that can adapt to different types of spatial contexts using convolution filters of different sizes. The characteristics of each dataset are unique, hence the attention mechanism helps focus on those features to improve classification performance.•The shallow inception model is designed with a two-layer attention mechanism with fewer layers but with a large number of convolution filters that can address the overfitting problem caused by small dataset sizes.•LSTM-based recurrent neural network (RNN) module is proposed to extract temporal features after the inception module is applied.•The proposed model is lightweight with fewer parameters and has less processing time.•The proposed model achieves good performance for both dynamic and static signs and gestures. Bio-inspired deep learning models have revolutionized sign language classification, achieving extraordinary accuracy and human-like video understanding. Recognition and classification of sign language videos in real-time are challenging because the duration and speed of each sign vary for different subjects, the background of videos is dynamic in most cases, and the classification result should be produced in real-time. This study proposes a model based on a convolution neural network (CNN) Inception model with an attention mechanism for extracting spatial features and Bi-LSTM (long short-term memory) for temporal feature extraction. The proposed model is tested on datasets with highly variable characteristics such as different clothing, variable lighting, and variable distance from the camera. Real-time classification achieves significant early detections while achieving performance comparable to the offline operation. The proposed model has fewer parameters, fewer deep learning layers, and requires significantly less processing time than state-of-the-art models. The Inception model with an attention mechanism with two attention blocks [Display omitted]</description><subject>Artificial neural networks</subject><subject>BiLSTM</subject><subject>Bio-inspired computing</subject><subject>Biomimetics</subject><subject>Classification</subject><subject>Deep learning</subject><subject>Feature extraction</subject><subject>Inception</subject><subject>Machine learning</subject><subject>Real time</subject><subject>Real-time classification</subject><subject>Sign language</subject><subject>Video data</subject><issn>0045-7906</issn><issn>1879-0755</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNqNkE9PwzAMxSMEEmPwHYI4dyRp06bHMfFn0hAHxjnKHLfK1KUjyZD49gTGgSMn69l-tt6PkGvOZpzx-nY7g3G3xwEBfT8TTPDcb8pWnpAJV01bsEbKUzJhrJJF07L6nFzEuGVZ11xNiF36hMPgevSJBjRDkdwO6TyYjQMaXe_pYHx_MD1SGEyMrnNgkhs9PUTne2pSytasi42JaKnzgPufufGW3rnV6_r5kpx1Zoh49Vun5O3hfr14KlYvj8vFfFVAxWUqoFWtqmsrDZTQtShgIzoFVY5hrQSuSmGYqpUSdVdZJja8kUopaRohS9lAOSU3x7v7ML4fMCa9HQ_B55dayLbKdJQq81Z73IIwxhiw0_vgdiZ8as70N1S91X-g6m-o-gg1exdHL-YYHw6DjuAwR7YuICRtR_ePK18D1oZd</recordid><startdate>202110</startdate><enddate>202110</enddate><creator>Abdul, Wadood</creator><creator>Alsulaiman, Mansour</creator><creator>Amin, Syed Umar</creator><creator>Faisal, Mohammed</creator><creator>Muhammad, Ghulam</creator><creator>Albogamy, Fahad R.</creator><creator>Bencherif, Mohamed A.</creator><creator>Ghaleb, Hamid</creator><general>Elsevier Ltd</general><general>Elsevier BV</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>202110</creationdate><title>Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM</title><author>Abdul, Wadood ; Alsulaiman, Mansour ; Amin, Syed Umar ; Faisal, Mohammed ; Muhammad, Ghulam ; Albogamy, Fahad R. ; Bencherif, Mohamed A. ; Ghaleb, Hamid</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c415t-c989866d5ac3cf9e2cb2f8c4879dd5c1832a0868826f4d02b1758885a725357c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Artificial neural networks</topic><topic>BiLSTM</topic><topic>Bio-inspired computing</topic><topic>Biomimetics</topic><topic>Classification</topic><topic>Deep learning</topic><topic>Feature extraction</topic><topic>Inception</topic><topic>Machine learning</topic><topic>Real time</topic><topic>Real-time classification</topic><topic>Sign language</topic><topic>Video data</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Abdul, Wadood</creatorcontrib><creatorcontrib>Alsulaiman, Mansour</creatorcontrib><creatorcontrib>Amin, Syed Umar</creatorcontrib><creatorcontrib>Faisal, Mohammed</creatorcontrib><creatorcontrib>Muhammad, Ghulam</creatorcontrib><creatorcontrib>Albogamy, Fahad R.</creatorcontrib><creatorcontrib>Bencherif, Mohamed A.</creatorcontrib><creatorcontrib>Ghaleb, Hamid</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Computers &amp; electrical engineering</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Abdul, Wadood</au><au>Alsulaiman, Mansour</au><au>Amin, Syed Umar</au><au>Faisal, Mohammed</au><au>Muhammad, Ghulam</au><au>Albogamy, Fahad R.</au><au>Bencherif, Mohamed A.</au><au>Ghaleb, Hamid</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM</atitle><jtitle>Computers &amp; electrical engineering</jtitle><date>2021-10</date><risdate>2021</risdate><volume>95</volume><spage>107395</spage><pages>107395-</pages><artnum>107395</artnum><issn>0045-7906</issn><eissn>1879-0755</eissn><abstract>•Bio-Inspired novel attention-based inception architecture is proposed that can adapt to different types of spatial contexts using convolution filters of different sizes. The characteristics of each dataset are unique, hence the attention mechanism helps focus on those features to improve classification performance.•The shallow inception model is designed with a two-layer attention mechanism with fewer layers but with a large number of convolution filters that can address the overfitting problem caused by small dataset sizes.•LSTM-based recurrent neural network (RNN) module is proposed to extract temporal features after the inception module is applied.•The proposed model is lightweight with fewer parameters and has less processing time.•The proposed model achieves good performance for both dynamic and static signs and gestures. Bio-inspired deep learning models have revolutionized sign language classification, achieving extraordinary accuracy and human-like video understanding. Recognition and classification of sign language videos in real-time are challenging because the duration and speed of each sign vary for different subjects, the background of videos is dynamic in most cases, and the classification result should be produced in real-time. This study proposes a model based on a convolution neural network (CNN) Inception model with an attention mechanism for extracting spatial features and Bi-LSTM (long short-term memory) for temporal feature extraction. The proposed model is tested on datasets with highly variable characteristics such as different clothing, variable lighting, and variable distance from the camera. Real-time classification achieves significant early detections while achieving performance comparable to the offline operation. The proposed model has fewer parameters, fewer deep learning layers, and requires significantly less processing time than state-of-the-art models. The Inception model with an attention mechanism with two attention blocks [Display omitted]</abstract><cop>Amsterdam</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.compeleceng.2021.107395</doi></addata></record>
fulltext fulltext
identifier ISSN: 0045-7906
ispartof Computers & electrical engineering, 2021-10, Vol.95, p.107395, Article 107395
issn 0045-7906
1879-0755
language eng
recordid cdi_proquest_journals_2594202883
source Access via ScienceDirect (Elsevier)
subjects Artificial neural networks
BiLSTM
Bio-inspired computing
Biomimetics
Classification
Deep learning
Feature extraction
Inception
Machine learning
Real time
Real-time classification
Sign language
Video data
title Intelligent real-time Arabic sign language classification using attention-based inception and BiLSTM
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-21T08%3A25%3A08IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Intelligent%20real-time%20Arabic%20sign%20language%20classification%20using%20attention-based%20inception%20and%20BiLSTM&rft.jtitle=Computers%20&%20electrical%20engineering&rft.au=Abdul,%20Wadood&rft.date=2021-10&rft.volume=95&rft.spage=107395&rft.pages=107395-&rft.artnum=107395&rft.issn=0045-7906&rft.eissn=1879-0755&rft_id=info:doi/10.1016/j.compeleceng.2021.107395&rft_dat=%3Cproquest_cross%3E2594202883%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2594202883&rft_id=info:pmid/&rft_els_id=S0045790621003621&rfr_iscdi=true