Extraction of multimodal features from depth and RGB images for recognising hand gestures

In this study, we propose a multimodal feature based framework for recognising hand gestures from RGB and depth images. In addition to the features from the RGB image, the depth image features are explored into constructing the discriminative feature labels of various gestures. Depth maps having pow...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of Future Robot Life 2021-12, Vol.2 (1-2), p.1-21
Hauptverfasser: Suni, S.S., Gopakumar, K.
Format: Artikel
Sprache:eng
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 21
container_issue 1-2
container_start_page 1
container_title Journal of Future Robot Life
container_volume 2
creator Suni, S.S.
Gopakumar, K.
description In this study, we propose a multimodal feature based framework for recognising hand gestures from RGB and depth images. In addition to the features from the RGB image, the depth image features are explored into constructing the discriminative feature labels of various gestures. Depth maps having powerful source of information, increases the performance level of various computer vision problems. A newly refined Gradient-Local Binary Pattern (G-LBP) is applied to extract the features from depth images and histogram of gradients (HOG) features are extracted from RGB images. The components from both RGB and depth channels, are concatenated to form a multimodal feature vector. In the final process, classification is performed using K-Nearest Neighbour and multi-class Support Vector Machines. The designed system is invariant to scale, rotation and illumination. The newly developed feature combination method is helpful to achieve superior recognition rates for future innovations.
doi_str_mv 10.3233/FRL-200010
format Article
fullrecord <record><control><sourceid>crossref</sourceid><recordid>TN_cdi_crossref_primary_10_3233_FRL_200010</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>10_3233_FRL_200010</sourcerecordid><originalsourceid>FETCH-LOGICAL-c710-ffea14e66acfd112af8c1c8db25702ba2f309ca641c47e063a022a0206db4aba3</originalsourceid><addsrcrecordid>eNo9kEtLw0AQgBdRsNRe_AV7FqKzu8kmOWppqxAQSi-ewmQfaSTJlt0U9N-7seJhmIFvXnyE3DN4FFyIp-2-SjgAMLgiC54VZVKWkl3_15m4JasQPmMLL0omsnxBPjZfk0c1dW6kztLh3E_d4DT21Bqczt4Ear0bqDan6Uhx1HS_e6HdgO1MnKfeKNeOXejGlh5nHsHv3B25sdgHs_rLS3LYbg7r16R6372tn6tE5QwSG8-w1EiJymrGONpCMVXohmc58Aa5FVAqlClTaW5ACgTOY4DUTYoNiiV5uKxV3oXgja1PPr7nv2sG9ayljlrqixbxA8-DVhs</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Extraction of multimodal features from depth and RGB images for recognising hand gestures</title><source>DOAJ Directory of Open Access Journals</source><source>EZB-FREE-00999 freely available EZB journals</source><creator>Suni, S.S. ; Gopakumar, K.</creator><creatorcontrib>Suni, S.S. ; Gopakumar, K.</creatorcontrib><description>In this study, we propose a multimodal feature based framework for recognising hand gestures from RGB and depth images. In addition to the features from the RGB image, the depth image features are explored into constructing the discriminative feature labels of various gestures. Depth maps having powerful source of information, increases the performance level of various computer vision problems. A newly refined Gradient-Local Binary Pattern (G-LBP) is applied to extract the features from depth images and histogram of gradients (HOG) features are extracted from RGB images. The components from both RGB and depth channels, are concatenated to form a multimodal feature vector. In the final process, classification is performed using K-Nearest Neighbour and multi-class Support Vector Machines. The designed system is invariant to scale, rotation and illumination. The newly developed feature combination method is helpful to achieve superior recognition rates for future innovations.</description><identifier>ISSN: 2589-9953</identifier><identifier>EISSN: 2589-9961</identifier><identifier>DOI: 10.3233/FRL-200010</identifier><language>eng</language><ispartof>Journal of Future Robot Life, 2021-12, Vol.2 (1-2), p.1-21</ispartof><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c710-ffea14e66acfd112af8c1c8db25702ba2f309ca641c47e063a022a0206db4aba3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,864,27922,27923</link.rule.ids></links><search><creatorcontrib>Suni, S.S.</creatorcontrib><creatorcontrib>Gopakumar, K.</creatorcontrib><title>Extraction of multimodal features from depth and RGB images for recognising hand gestures</title><title>Journal of Future Robot Life</title><description>In this study, we propose a multimodal feature based framework for recognising hand gestures from RGB and depth images. In addition to the features from the RGB image, the depth image features are explored into constructing the discriminative feature labels of various gestures. Depth maps having powerful source of information, increases the performance level of various computer vision problems. A newly refined Gradient-Local Binary Pattern (G-LBP) is applied to extract the features from depth images and histogram of gradients (HOG) features are extracted from RGB images. The components from both RGB and depth channels, are concatenated to form a multimodal feature vector. In the final process, classification is performed using K-Nearest Neighbour and multi-class Support Vector Machines. The designed system is invariant to scale, rotation and illumination. The newly developed feature combination method is helpful to achieve superior recognition rates for future innovations.</description><issn>2589-9953</issn><issn>2589-9961</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNo9kEtLw0AQgBdRsNRe_AV7FqKzu8kmOWppqxAQSi-ewmQfaSTJlt0U9N-7seJhmIFvXnyE3DN4FFyIp-2-SjgAMLgiC54VZVKWkl3_15m4JasQPmMLL0omsnxBPjZfk0c1dW6kztLh3E_d4DT21Bqczt4Ear0bqDan6Uhx1HS_e6HdgO1MnKfeKNeOXejGlh5nHsHv3B25sdgHs_rLS3LYbg7r16R6372tn6tE5QwSG8-w1EiJymrGONpCMVXohmc58Aa5FVAqlClTaW5ACgTOY4DUTYoNiiV5uKxV3oXgja1PPr7nv2sG9ayljlrqixbxA8-DVhs</recordid><startdate>20211207</startdate><enddate>20211207</enddate><creator>Suni, S.S.</creator><creator>Gopakumar, K.</creator><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>20211207</creationdate><title>Extraction of multimodal features from depth and RGB images for recognising hand gestures</title><author>Suni, S.S. ; Gopakumar, K.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c710-ffea14e66acfd112af8c1c8db25702ba2f309ca641c47e063a022a0206db4aba3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Suni, S.S.</creatorcontrib><creatorcontrib>Gopakumar, K.</creatorcontrib><collection>CrossRef</collection><jtitle>Journal of Future Robot Life</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Suni, S.S.</au><au>Gopakumar, K.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Extraction of multimodal features from depth and RGB images for recognising hand gestures</atitle><jtitle>Journal of Future Robot Life</jtitle><date>2021-12-07</date><risdate>2021</risdate><volume>2</volume><issue>1-2</issue><spage>1</spage><epage>21</epage><pages>1-21</pages><issn>2589-9953</issn><eissn>2589-9961</eissn><abstract>In this study, we propose a multimodal feature based framework for recognising hand gestures from RGB and depth images. In addition to the features from the RGB image, the depth image features are explored into constructing the discriminative feature labels of various gestures. Depth maps having powerful source of information, increases the performance level of various computer vision problems. A newly refined Gradient-Local Binary Pattern (G-LBP) is applied to extract the features from depth images and histogram of gradients (HOG) features are extracted from RGB images. The components from both RGB and depth channels, are concatenated to form a multimodal feature vector. In the final process, classification is performed using K-Nearest Neighbour and multi-class Support Vector Machines. The designed system is invariant to scale, rotation and illumination. The newly developed feature combination method is helpful to achieve superior recognition rates for future innovations.</abstract><doi>10.3233/FRL-200010</doi><tpages>21</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2589-9953
ispartof Journal of Future Robot Life, 2021-12, Vol.2 (1-2), p.1-21
issn 2589-9953
2589-9961
language eng
recordid cdi_crossref_primary_10_3233_FRL_200010
source DOAJ Directory of Open Access Journals; EZB-FREE-00999 freely available EZB journals
title Extraction of multimodal features from depth and RGB images for recognising hand gestures
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-13T15%3A33%3A04IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-crossref&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Extraction%20of%20multimodal%20features%20from%20depth%20and%20RGB%20images%20for%20recognising%20hand%20gestures&rft.jtitle=Journal%20of%20Future%20Robot%20Life&rft.au=Suni,%20S.S.&rft.date=2021-12-07&rft.volume=2&rft.issue=1-2&rft.spage=1&rft.epage=21&rft.pages=1-21&rft.issn=2589-9953&rft.eissn=2589-9961&rft_id=info:doi/10.3233/FRL-200010&rft_dat=%3Ccrossref%3E10_3233_FRL_200010%3C/crossref%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true