Action recognition using kinematics posture feature on 3D skeleton joint locations

•Motion information from skeleton can be efficiently encoded by considering the body joints as wearable kinematics sensor.•Proposed Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) from positions and bone-angles.•Proposed Kinematics Posture Feature (KPF) from LJPF and A...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Pattern recognition letters 2021-05, Vol.145, p.216-224
Hauptverfasser: Ahad, Md Atiqur Rahman, Ahmed, Masud, Das Antar, Anindya, Makihara, Yasushi, Yagi, Yasushi
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 224
container_issue
container_start_page 216
container_title Pattern recognition letters
container_volume 145
creator Ahad, Md Atiqur Rahman
Ahmed, Masud
Das Antar, Anindya
Makihara, Yasushi
Yagi, Yasushi
description •Motion information from skeleton can be efficiently encoded by considering the body joints as wearable kinematics sensor.•Proposed Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) from positions and bone-angles.•Proposed Kinematics Posture Feature (KPF) from LJPF and AJPF to encode motion and posture variation across frames.•Proposed Position-based Statistical Feature (PSF) from segmented KPF, that consists of temporal statistical features.•Five benchmark datasets are explored on PSF features with statistical and deep learning-based models. Action recognition is a very widely explored research area in computer vision and related fields. We propose Kinematics Posture Feature (KPF) extraction from 3D joint positions based on skeleton data for improving the performance of action recognition. In this approach, we consider the skeleton 3D joints as kinematics sensors. We propose Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) based on 3D linear joint positions and angles between bone segments. We then combine these two kinematics features for each video frame for each action to create the KPF feature sets. These feature sets encode the variation of motion in the temporal domain as if each body joint represents kinematics position and orientation sensors. In the next stage, we process the extracted KPF feature descriptor by using a low pass filter, and segment them by using sliding windows with optimized length. This concept resembles the approach of processing kinematics sensor data. From the segmented windows, we compute the Position-based Statistical Feature (PSF). These features consist of temporal domain statistical features (e.g., mean, standard deviation, variance, etc.). These statistical features encode the variation of postures (i.e., joint positions and angles) across the video frames. For performing classification, we explore Support Vector Machine (Linear), RNN, CNNRNN, and ConvRNN model. The proposed PSF feature sets demonstrate prominent performance in both statistical machine learning- and deep learning-based models. For evaluation, we explore five benchmark datasets namely UTKinect-Action3D, Kinect Activity Recognition Dataset (KARD), MSR 3D Action Pairs, Florence 3D, and Office Activity Dataset (OAD). To prevent overfitting, we consider the leave-one-subject-out framework as the experimental setup and perform 10-fold cross-validation. Our approach outperforms several existing methods
doi_str_mv 10.1016/j.patrec.2021.02.013
format Article
fullrecord <record><control><sourceid>proquest_webof</sourceid><recordid>TN_cdi_webofscience_primary_000636965200029CitationCount</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0167865521000751</els_id><sourcerecordid>2528501124</sourcerecordid><originalsourceid>FETCH-LOGICAL-c446t-b6b1a59d35c7e651029ffa655372a82fc83f869ab69ca0e9a4d6f9d19362b2273</originalsourceid><addsrcrecordid>eNqNkE1LxDAURYMoOI7-AxcFl9KajzZtNsIwfsKAILoOaZoM6cw0NUkV_72ZdnAprt5d3JO8dwC4RDBDENGbNutFcEpmGGKUQZxBRI7ADFUlTkuS58dgFmtlWtGiOAVn3rcQQkpYNQOvCxmM7ZJI23Vnxjx4062TjenUTgQjfdJbHwanEq3EOGOH3CV-o7YqxNxa04Vka6XY4_4cnGix9eriMOfg_eH-bfmUrl4en5eLVSrznIa0pjUSBWtIIUtFCwQx01rEDUmJRYW1rIiuKBM1ZVJAxUTeUM0axAjFNcYlmYOr6d3e2Y9B-cBbO7gufslxgasCIoTz2MqnlnTWe6c0753ZCffNEeR7e7zlkz2-t8ch5tFexKoJ-1K11V4a1Un1i476KKMFjgmzpQnj6Us7dCGi1_9HY_t2aquo6tMoxw9EY-JOgTfW_L3pD5khm2A</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2528501124</pqid></control><display><type>article</type><title>Action recognition using kinematics posture feature on 3D skeleton joint locations</title><source>Web of Science - Science Citation Index Expanded - 2021&lt;img src="https://exlibris-pub.s3.amazonaws.com/fromwos-v2.jpg" /&gt;</source><source>Access via ScienceDirect (Elsevier)</source><creator>Ahad, Md Atiqur Rahman ; Ahmed, Masud ; Das Antar, Anindya ; Makihara, Yasushi ; Yagi, Yasushi</creator><creatorcontrib>Ahad, Md Atiqur Rahman ; Ahmed, Masud ; Das Antar, Anindya ; Makihara, Yasushi ; Yagi, Yasushi</creatorcontrib><description>•Motion information from skeleton can be efficiently encoded by considering the body joints as wearable kinematics sensor.•Proposed Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) from positions and bone-angles.•Proposed Kinematics Posture Feature (KPF) from LJPF and AJPF to encode motion and posture variation across frames.•Proposed Position-based Statistical Feature (PSF) from segmented KPF, that consists of temporal statistical features.•Five benchmark datasets are explored on PSF features with statistical and deep learning-based models. Action recognition is a very widely explored research area in computer vision and related fields. We propose Kinematics Posture Feature (KPF) extraction from 3D joint positions based on skeleton data for improving the performance of action recognition. In this approach, we consider the skeleton 3D joints as kinematics sensors. We propose Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) based on 3D linear joint positions and angles between bone segments. We then combine these two kinematics features for each video frame for each action to create the KPF feature sets. These feature sets encode the variation of motion in the temporal domain as if each body joint represents kinematics position and orientation sensors. In the next stage, we process the extracted KPF feature descriptor by using a low pass filter, and segment them by using sliding windows with optimized length. This concept resembles the approach of processing kinematics sensor data. From the segmented windows, we compute the Position-based Statistical Feature (PSF). These features consist of temporal domain statistical features (e.g., mean, standard deviation, variance, etc.). These statistical features encode the variation of postures (i.e., joint positions and angles) across the video frames. For performing classification, we explore Support Vector Machine (Linear), RNN, CNNRNN, and ConvRNN model. The proposed PSF feature sets demonstrate prominent performance in both statistical machine learning- and deep learning-based models. For evaluation, we explore five benchmark datasets namely UTKinect-Action3D, Kinect Activity Recognition Dataset (KARD), MSR 3D Action Pairs, Florence 3D, and Office Activity Dataset (OAD). To prevent overfitting, we consider the leave-one-subject-out framework as the experimental setup and perform 10-fold cross-validation. Our approach outperforms several existing methods in these benchmark datasets and achieves very promising classification performance.</description><identifier>ISSN: 0167-8655</identifier><identifier>EISSN: 1872-7344</identifier><identifier>DOI: 10.1016/j.patrec.2021.02.013</identifier><language>eng</language><publisher>AMSTERDAM: Elsevier B.V</publisher><subject>Action recognition ; Activity recognition ; Angular joint position feature (AJPF) ; Angular position ; Benchmark datasets ; Benchmarks ; Classification ; Computer Science ; Computer Science, Artificial Intelligence ; Computer vision ; Convrnn ; Datasets ; Deep learning ; Deep neural network ; Domains ; Ensemble architecture ; Feature extraction ; Joint angle ; Joint position ; Joints (anatomy) ; Kinematics ; Kinematics posture feature (KPF) ; Learning algorithms ; Linear joint position feature (LJPF) ; Low pass filters ; Machine learning ; Position sensing ; Position-based statistical feature (PSF) ; Posture ; Science &amp; Technology ; Segments ; Sensors ; Skeleton ; Skeleton data ; Statistical analysis ; Statistics ; Support vector machines ; Technology ; Video data</subject><ispartof>Pattern recognition letters, 2021-05, Vol.145, p.216-224</ispartof><rights>2021 Elsevier B.V.</rights><rights>Copyright Elsevier Science Ltd. May 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>true</woscitedreferencessubscribed><woscitedreferencescount>43</woscitedreferencescount><woscitedreferencesoriginalsourcerecordid>wos000636965200029</woscitedreferencesoriginalsourcerecordid><citedby>FETCH-LOGICAL-c446t-b6b1a59d35c7e651029ffa655372a82fc83f869ab69ca0e9a4d6f9d19362b2273</citedby><cites>FETCH-LOGICAL-c446t-b6b1a59d35c7e651029ffa655372a82fc83f869ab69ca0e9a4d6f9d19362b2273</cites><orcidid>0000-0001-9912-8757 ; 0000-0001-8355-7004</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.patrec.2021.02.013$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>315,781,785,3551,27929,27930,39263,46000</link.rule.ids></links><search><creatorcontrib>Ahad, Md Atiqur Rahman</creatorcontrib><creatorcontrib>Ahmed, Masud</creatorcontrib><creatorcontrib>Das Antar, Anindya</creatorcontrib><creatorcontrib>Makihara, Yasushi</creatorcontrib><creatorcontrib>Yagi, Yasushi</creatorcontrib><title>Action recognition using kinematics posture feature on 3D skeleton joint locations</title><title>Pattern recognition letters</title><addtitle>PATTERN RECOGN LETT</addtitle><description>•Motion information from skeleton can be efficiently encoded by considering the body joints as wearable kinematics sensor.•Proposed Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) from positions and bone-angles.•Proposed Kinematics Posture Feature (KPF) from LJPF and AJPF to encode motion and posture variation across frames.•Proposed Position-based Statistical Feature (PSF) from segmented KPF, that consists of temporal statistical features.•Five benchmark datasets are explored on PSF features with statistical and deep learning-based models. Action recognition is a very widely explored research area in computer vision and related fields. We propose Kinematics Posture Feature (KPF) extraction from 3D joint positions based on skeleton data for improving the performance of action recognition. In this approach, we consider the skeleton 3D joints as kinematics sensors. We propose Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) based on 3D linear joint positions and angles between bone segments. We then combine these two kinematics features for each video frame for each action to create the KPF feature sets. These feature sets encode the variation of motion in the temporal domain as if each body joint represents kinematics position and orientation sensors. In the next stage, we process the extracted KPF feature descriptor by using a low pass filter, and segment them by using sliding windows with optimized length. This concept resembles the approach of processing kinematics sensor data. From the segmented windows, we compute the Position-based Statistical Feature (PSF). These features consist of temporal domain statistical features (e.g., mean, standard deviation, variance, etc.). These statistical features encode the variation of postures (i.e., joint positions and angles) across the video frames. For performing classification, we explore Support Vector Machine (Linear), RNN, CNNRNN, and ConvRNN model. The proposed PSF feature sets demonstrate prominent performance in both statistical machine learning- and deep learning-based models. For evaluation, we explore five benchmark datasets namely UTKinect-Action3D, Kinect Activity Recognition Dataset (KARD), MSR 3D Action Pairs, Florence 3D, and Office Activity Dataset (OAD). To prevent overfitting, we consider the leave-one-subject-out framework as the experimental setup and perform 10-fold cross-validation. Our approach outperforms several existing methods in these benchmark datasets and achieves very promising classification performance.</description><subject>Action recognition</subject><subject>Activity recognition</subject><subject>Angular joint position feature (AJPF)</subject><subject>Angular position</subject><subject>Benchmark datasets</subject><subject>Benchmarks</subject><subject>Classification</subject><subject>Computer Science</subject><subject>Computer Science, Artificial Intelligence</subject><subject>Computer vision</subject><subject>Convrnn</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Deep neural network</subject><subject>Domains</subject><subject>Ensemble architecture</subject><subject>Feature extraction</subject><subject>Joint angle</subject><subject>Joint position</subject><subject>Joints (anatomy)</subject><subject>Kinematics</subject><subject>Kinematics posture feature (KPF)</subject><subject>Learning algorithms</subject><subject>Linear joint position feature (LJPF)</subject><subject>Low pass filters</subject><subject>Machine learning</subject><subject>Position sensing</subject><subject>Position-based statistical feature (PSF)</subject><subject>Posture</subject><subject>Science &amp; Technology</subject><subject>Segments</subject><subject>Sensors</subject><subject>Skeleton</subject><subject>Skeleton data</subject><subject>Statistical analysis</subject><subject>Statistics</subject><subject>Support vector machines</subject><subject>Technology</subject><subject>Video data</subject><issn>0167-8655</issn><issn>1872-7344</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>HGBXW</sourceid><recordid>eNqNkE1LxDAURYMoOI7-AxcFl9KajzZtNsIwfsKAILoOaZoM6cw0NUkV_72ZdnAprt5d3JO8dwC4RDBDENGbNutFcEpmGGKUQZxBRI7ADFUlTkuS58dgFmtlWtGiOAVn3rcQQkpYNQOvCxmM7ZJI23Vnxjx4062TjenUTgQjfdJbHwanEq3EOGOH3CV-o7YqxNxa04Vka6XY4_4cnGix9eriMOfg_eH-bfmUrl4en5eLVSrznIa0pjUSBWtIIUtFCwQx01rEDUmJRYW1rIiuKBM1ZVJAxUTeUM0axAjFNcYlmYOr6d3e2Y9B-cBbO7gufslxgasCIoTz2MqnlnTWe6c0753ZCffNEeR7e7zlkz2-t8ch5tFexKoJ-1K11V4a1Un1i476KKMFjgmzpQnj6Us7dCGi1_9HY_t2aquo6tMoxw9EY-JOgTfW_L3pD5khm2A</recordid><startdate>202105</startdate><enddate>202105</enddate><creator>Ahad, Md Atiqur Rahman</creator><creator>Ahmed, Masud</creator><creator>Das Antar, Anindya</creator><creator>Makihara, Yasushi</creator><creator>Yagi, Yasushi</creator><general>Elsevier B.V</general><general>Elsevier</general><general>Elsevier Science Ltd</general><scope>BLEPL</scope><scope>DTL</scope><scope>HGBXW</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7TK</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0001-9912-8757</orcidid><orcidid>https://orcid.org/0000-0001-8355-7004</orcidid></search><sort><creationdate>202105</creationdate><title>Action recognition using kinematics posture feature on 3D skeleton joint locations</title><author>Ahad, Md Atiqur Rahman ; Ahmed, Masud ; Das Antar, Anindya ; Makihara, Yasushi ; Yagi, Yasushi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c446t-b6b1a59d35c7e651029ffa655372a82fc83f869ab69ca0e9a4d6f9d19362b2273</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Action recognition</topic><topic>Activity recognition</topic><topic>Angular joint position feature (AJPF)</topic><topic>Angular position</topic><topic>Benchmark datasets</topic><topic>Benchmarks</topic><topic>Classification</topic><topic>Computer Science</topic><topic>Computer Science, Artificial Intelligence</topic><topic>Computer vision</topic><topic>Convrnn</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Deep neural network</topic><topic>Domains</topic><topic>Ensemble architecture</topic><topic>Feature extraction</topic><topic>Joint angle</topic><topic>Joint position</topic><topic>Joints (anatomy)</topic><topic>Kinematics</topic><topic>Kinematics posture feature (KPF)</topic><topic>Learning algorithms</topic><topic>Linear joint position feature (LJPF)</topic><topic>Low pass filters</topic><topic>Machine learning</topic><topic>Position sensing</topic><topic>Position-based statistical feature (PSF)</topic><topic>Posture</topic><topic>Science &amp; Technology</topic><topic>Segments</topic><topic>Sensors</topic><topic>Skeleton</topic><topic>Skeleton data</topic><topic>Statistical analysis</topic><topic>Statistics</topic><topic>Support vector machines</topic><topic>Technology</topic><topic>Video data</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Ahad, Md Atiqur Rahman</creatorcontrib><creatorcontrib>Ahmed, Masud</creatorcontrib><creatorcontrib>Das Antar, Anindya</creatorcontrib><creatorcontrib>Makihara, Yasushi</creatorcontrib><creatorcontrib>Yagi, Yasushi</creatorcontrib><collection>Web of Science Core Collection</collection><collection>Science Citation Index Expanded</collection><collection>Web of Science - Science Citation Index Expanded - 2021</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Pattern recognition letters</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ahad, Md Atiqur Rahman</au><au>Ahmed, Masud</au><au>Das Antar, Anindya</au><au>Makihara, Yasushi</au><au>Yagi, Yasushi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Action recognition using kinematics posture feature on 3D skeleton joint locations</atitle><jtitle>Pattern recognition letters</jtitle><stitle>PATTERN RECOGN LETT</stitle><date>2021-05</date><risdate>2021</risdate><volume>145</volume><spage>216</spage><epage>224</epage><pages>216-224</pages><issn>0167-8655</issn><eissn>1872-7344</eissn><abstract>•Motion information from skeleton can be efficiently encoded by considering the body joints as wearable kinematics sensor.•Proposed Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) from positions and bone-angles.•Proposed Kinematics Posture Feature (KPF) from LJPF and AJPF to encode motion and posture variation across frames.•Proposed Position-based Statistical Feature (PSF) from segmented KPF, that consists of temporal statistical features.•Five benchmark datasets are explored on PSF features with statistical and deep learning-based models. Action recognition is a very widely explored research area in computer vision and related fields. We propose Kinematics Posture Feature (KPF) extraction from 3D joint positions based on skeleton data for improving the performance of action recognition. In this approach, we consider the skeleton 3D joints as kinematics sensors. We propose Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) based on 3D linear joint positions and angles between bone segments. We then combine these two kinematics features for each video frame for each action to create the KPF feature sets. These feature sets encode the variation of motion in the temporal domain as if each body joint represents kinematics position and orientation sensors. In the next stage, we process the extracted KPF feature descriptor by using a low pass filter, and segment them by using sliding windows with optimized length. This concept resembles the approach of processing kinematics sensor data. From the segmented windows, we compute the Position-based Statistical Feature (PSF). These features consist of temporal domain statistical features (e.g., mean, standard deviation, variance, etc.). These statistical features encode the variation of postures (i.e., joint positions and angles) across the video frames. For performing classification, we explore Support Vector Machine (Linear), RNN, CNNRNN, and ConvRNN model. The proposed PSF feature sets demonstrate prominent performance in both statistical machine learning- and deep learning-based models. For evaluation, we explore five benchmark datasets namely UTKinect-Action3D, Kinect Activity Recognition Dataset (KARD), MSR 3D Action Pairs, Florence 3D, and Office Activity Dataset (OAD). To prevent overfitting, we consider the leave-one-subject-out framework as the experimental setup and perform 10-fold cross-validation. Our approach outperforms several existing methods in these benchmark datasets and achieves very promising classification performance.</abstract><cop>AMSTERDAM</cop><pub>Elsevier B.V</pub><doi>10.1016/j.patrec.2021.02.013</doi><tpages>9</tpages><orcidid>https://orcid.org/0000-0001-9912-8757</orcidid><orcidid>https://orcid.org/0000-0001-8355-7004</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0167-8655
ispartof Pattern recognition letters, 2021-05, Vol.145, p.216-224
issn 0167-8655
1872-7344
language eng
recordid cdi_webofscience_primary_000636965200029CitationCount
source Web of Science - Science Citation Index Expanded - 2021<img src="https://exlibris-pub.s3.amazonaws.com/fromwos-v2.jpg" />; Access via ScienceDirect (Elsevier)
subjects Action recognition
Activity recognition
Angular joint position feature (AJPF)
Angular position
Benchmark datasets
Benchmarks
Classification
Computer Science
Computer Science, Artificial Intelligence
Computer vision
Convrnn
Datasets
Deep learning
Deep neural network
Domains
Ensemble architecture
Feature extraction
Joint angle
Joint position
Joints (anatomy)
Kinematics
Kinematics posture feature (KPF)
Learning algorithms
Linear joint position feature (LJPF)
Low pass filters
Machine learning
Position sensing
Position-based statistical feature (PSF)
Posture
Science & Technology
Segments
Sensors
Skeleton
Skeleton data
Statistical analysis
Statistics
Support vector machines
Technology
Video data
title Action recognition using kinematics posture feature on 3D skeleton joint locations
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-12T13%3A47%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_webof&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Action%20recognition%20using%20kinematics%20posture%20feature%20on%203D%20skeleton%20joint%20locations&rft.jtitle=Pattern%20recognition%20letters&rft.au=Ahad,%20Md%20Atiqur%20Rahman&rft.date=2021-05&rft.volume=145&rft.spage=216&rft.epage=224&rft.pages=216-224&rft.issn=0167-8655&rft.eissn=1872-7344&rft_id=info:doi/10.1016/j.patrec.2021.02.013&rft_dat=%3Cproquest_webof%3E2528501124%3C/proquest_webof%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2528501124&rft_id=info:pmid/&rft_els_id=S0167865521000751&rfr_iscdi=true