Action recognition using kinematics posture feature on 3D skeleton joint locations
•Motion information from skeleton can be efficiently encoded by considering the body joints as wearable kinematics sensor.•Proposed Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) from positions and bone-angles.•Proposed Kinematics Posture Feature (KPF) from LJPF and A...
Gespeichert in:
Veröffentlicht in: | Pattern recognition letters 2021-05, Vol.145, p.216-224 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 224 |
---|---|
container_issue | |
container_start_page | 216 |
container_title | Pattern recognition letters |
container_volume | 145 |
creator | Ahad, Md Atiqur Rahman Ahmed, Masud Das Antar, Anindya Makihara, Yasushi Yagi, Yasushi |
description | •Motion information from skeleton can be efficiently encoded by considering the body joints as wearable kinematics sensor.•Proposed Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) from positions and bone-angles.•Proposed Kinematics Posture Feature (KPF) from LJPF and AJPF to encode motion and posture variation across frames.•Proposed Position-based Statistical Feature (PSF) from segmented KPF, that consists of temporal statistical features.•Five benchmark datasets are explored on PSF features with statistical and deep learning-based models.
Action recognition is a very widely explored research area in computer vision and related fields. We propose Kinematics Posture Feature (KPF) extraction from 3D joint positions based on skeleton data for improving the performance of action recognition. In this approach, we consider the skeleton 3D joints as kinematics sensors. We propose Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) based on 3D linear joint positions and angles between bone segments. We then combine these two kinematics features for each video frame for each action to create the KPF feature sets. These feature sets encode the variation of motion in the temporal domain as if each body joint represents kinematics position and orientation sensors. In the next stage, we process the extracted KPF feature descriptor by using a low pass filter, and segment them by using sliding windows with optimized length. This concept resembles the approach of processing kinematics sensor data. From the segmented windows, we compute the Position-based Statistical Feature (PSF). These features consist of temporal domain statistical features (e.g., mean, standard deviation, variance, etc.). These statistical features encode the variation of postures (i.e., joint positions and angles) across the video frames. For performing classification, we explore Support Vector Machine (Linear), RNN, CNNRNN, and ConvRNN model. The proposed PSF feature sets demonstrate prominent performance in both statistical machine learning- and deep learning-based models. For evaluation, we explore five benchmark datasets namely UTKinect-Action3D, Kinect Activity Recognition Dataset (KARD), MSR 3D Action Pairs, Florence 3D, and Office Activity Dataset (OAD). To prevent overfitting, we consider the leave-one-subject-out framework as the experimental setup and perform 10-fold cross-validation. Our approach outperforms several existing methods |
doi_str_mv | 10.1016/j.patrec.2021.02.013 |
format | Article |
fullrecord | <record><control><sourceid>proquest_webof</sourceid><recordid>TN_cdi_webofscience_primary_000636965200029CitationCount</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0167865521000751</els_id><sourcerecordid>2528501124</sourcerecordid><originalsourceid>FETCH-LOGICAL-c446t-b6b1a59d35c7e651029ffa655372a82fc83f869ab69ca0e9a4d6f9d19362b2273</originalsourceid><addsrcrecordid>eNqNkE1LxDAURYMoOI7-AxcFl9KajzZtNsIwfsKAILoOaZoM6cw0NUkV_72ZdnAprt5d3JO8dwC4RDBDENGbNutFcEpmGGKUQZxBRI7ADFUlTkuS58dgFmtlWtGiOAVn3rcQQkpYNQOvCxmM7ZJI23Vnxjx4062TjenUTgQjfdJbHwanEq3EOGOH3CV-o7YqxNxa04Vka6XY4_4cnGix9eriMOfg_eH-bfmUrl4en5eLVSrznIa0pjUSBWtIIUtFCwQx01rEDUmJRYW1rIiuKBM1ZVJAxUTeUM0axAjFNcYlmYOr6d3e2Y9B-cBbO7gufslxgasCIoTz2MqnlnTWe6c0753ZCffNEeR7e7zlkz2-t8ch5tFexKoJ-1K11V4a1Un1i476KKMFjgmzpQnj6Us7dCGi1_9HY_t2aquo6tMoxw9EY-JOgTfW_L3pD5khm2A</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2528501124</pqid></control><display><type>article</type><title>Action recognition using kinematics posture feature on 3D skeleton joint locations</title><source>Web of Science - Science Citation Index Expanded - 2021<img src="https://exlibris-pub.s3.amazonaws.com/fromwos-v2.jpg" /></source><source>Access via ScienceDirect (Elsevier)</source><creator>Ahad, Md Atiqur Rahman ; Ahmed, Masud ; Das Antar, Anindya ; Makihara, Yasushi ; Yagi, Yasushi</creator><creatorcontrib>Ahad, Md Atiqur Rahman ; Ahmed, Masud ; Das Antar, Anindya ; Makihara, Yasushi ; Yagi, Yasushi</creatorcontrib><description>•Motion information from skeleton can be efficiently encoded by considering the body joints as wearable kinematics sensor.•Proposed Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) from positions and bone-angles.•Proposed Kinematics Posture Feature (KPF) from LJPF and AJPF to encode motion and posture variation across frames.•Proposed Position-based Statistical Feature (PSF) from segmented KPF, that consists of temporal statistical features.•Five benchmark datasets are explored on PSF features with statistical and deep learning-based models.
Action recognition is a very widely explored research area in computer vision and related fields. We propose Kinematics Posture Feature (KPF) extraction from 3D joint positions based on skeleton data for improving the performance of action recognition. In this approach, we consider the skeleton 3D joints as kinematics sensors. We propose Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) based on 3D linear joint positions and angles between bone segments. We then combine these two kinematics features for each video frame for each action to create the KPF feature sets. These feature sets encode the variation of motion in the temporal domain as if each body joint represents kinematics position and orientation sensors. In the next stage, we process the extracted KPF feature descriptor by using a low pass filter, and segment them by using sliding windows with optimized length. This concept resembles the approach of processing kinematics sensor data. From the segmented windows, we compute the Position-based Statistical Feature (PSF). These features consist of temporal domain statistical features (e.g., mean, standard deviation, variance, etc.). These statistical features encode the variation of postures (i.e., joint positions and angles) across the video frames. For performing classification, we explore Support Vector Machine (Linear), RNN, CNNRNN, and ConvRNN model. The proposed PSF feature sets demonstrate prominent performance in both statistical machine learning- and deep learning-based models. For evaluation, we explore five benchmark datasets namely UTKinect-Action3D, Kinect Activity Recognition Dataset (KARD), MSR 3D Action Pairs, Florence 3D, and Office Activity Dataset (OAD). To prevent overfitting, we consider the leave-one-subject-out framework as the experimental setup and perform 10-fold cross-validation. Our approach outperforms several existing methods in these benchmark datasets and achieves very promising classification performance.</description><identifier>ISSN: 0167-8655</identifier><identifier>EISSN: 1872-7344</identifier><identifier>DOI: 10.1016/j.patrec.2021.02.013</identifier><language>eng</language><publisher>AMSTERDAM: Elsevier B.V</publisher><subject>Action recognition ; Activity recognition ; Angular joint position feature (AJPF) ; Angular position ; Benchmark datasets ; Benchmarks ; Classification ; Computer Science ; Computer Science, Artificial Intelligence ; Computer vision ; Convrnn ; Datasets ; Deep learning ; Deep neural network ; Domains ; Ensemble architecture ; Feature extraction ; Joint angle ; Joint position ; Joints (anatomy) ; Kinematics ; Kinematics posture feature (KPF) ; Learning algorithms ; Linear joint position feature (LJPF) ; Low pass filters ; Machine learning ; Position sensing ; Position-based statistical feature (PSF) ; Posture ; Science & Technology ; Segments ; Sensors ; Skeleton ; Skeleton data ; Statistical analysis ; Statistics ; Support vector machines ; Technology ; Video data</subject><ispartof>Pattern recognition letters, 2021-05, Vol.145, p.216-224</ispartof><rights>2021 Elsevier B.V.</rights><rights>Copyright Elsevier Science Ltd. May 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>true</woscitedreferencessubscribed><woscitedreferencescount>43</woscitedreferencescount><woscitedreferencesoriginalsourcerecordid>wos000636965200029</woscitedreferencesoriginalsourcerecordid><citedby>FETCH-LOGICAL-c446t-b6b1a59d35c7e651029ffa655372a82fc83f869ab69ca0e9a4d6f9d19362b2273</citedby><cites>FETCH-LOGICAL-c446t-b6b1a59d35c7e651029ffa655372a82fc83f869ab69ca0e9a4d6f9d19362b2273</cites><orcidid>0000-0001-9912-8757 ; 0000-0001-8355-7004</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.patrec.2021.02.013$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>315,781,785,3551,27929,27930,39263,46000</link.rule.ids></links><search><creatorcontrib>Ahad, Md Atiqur Rahman</creatorcontrib><creatorcontrib>Ahmed, Masud</creatorcontrib><creatorcontrib>Das Antar, Anindya</creatorcontrib><creatorcontrib>Makihara, Yasushi</creatorcontrib><creatorcontrib>Yagi, Yasushi</creatorcontrib><title>Action recognition using kinematics posture feature on 3D skeleton joint locations</title><title>Pattern recognition letters</title><addtitle>PATTERN RECOGN LETT</addtitle><description>•Motion information from skeleton can be efficiently encoded by considering the body joints as wearable kinematics sensor.•Proposed Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) from positions and bone-angles.•Proposed Kinematics Posture Feature (KPF) from LJPF and AJPF to encode motion and posture variation across frames.•Proposed Position-based Statistical Feature (PSF) from segmented KPF, that consists of temporal statistical features.•Five benchmark datasets are explored on PSF features with statistical and deep learning-based models.
Action recognition is a very widely explored research area in computer vision and related fields. We propose Kinematics Posture Feature (KPF) extraction from 3D joint positions based on skeleton data for improving the performance of action recognition. In this approach, we consider the skeleton 3D joints as kinematics sensors. We propose Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) based on 3D linear joint positions and angles between bone segments. We then combine these two kinematics features for each video frame for each action to create the KPF feature sets. These feature sets encode the variation of motion in the temporal domain as if each body joint represents kinematics position and orientation sensors. In the next stage, we process the extracted KPF feature descriptor by using a low pass filter, and segment them by using sliding windows with optimized length. This concept resembles the approach of processing kinematics sensor data. From the segmented windows, we compute the Position-based Statistical Feature (PSF). These features consist of temporal domain statistical features (e.g., mean, standard deviation, variance, etc.). These statistical features encode the variation of postures (i.e., joint positions and angles) across the video frames. For performing classification, we explore Support Vector Machine (Linear), RNN, CNNRNN, and ConvRNN model. The proposed PSF feature sets demonstrate prominent performance in both statistical machine learning- and deep learning-based models. For evaluation, we explore five benchmark datasets namely UTKinect-Action3D, Kinect Activity Recognition Dataset (KARD), MSR 3D Action Pairs, Florence 3D, and Office Activity Dataset (OAD). To prevent overfitting, we consider the leave-one-subject-out framework as the experimental setup and perform 10-fold cross-validation. Our approach outperforms several existing methods in these benchmark datasets and achieves very promising classification performance.</description><subject>Action recognition</subject><subject>Activity recognition</subject><subject>Angular joint position feature (AJPF)</subject><subject>Angular position</subject><subject>Benchmark datasets</subject><subject>Benchmarks</subject><subject>Classification</subject><subject>Computer Science</subject><subject>Computer Science, Artificial Intelligence</subject><subject>Computer vision</subject><subject>Convrnn</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Deep neural network</subject><subject>Domains</subject><subject>Ensemble architecture</subject><subject>Feature extraction</subject><subject>Joint angle</subject><subject>Joint position</subject><subject>Joints (anatomy)</subject><subject>Kinematics</subject><subject>Kinematics posture feature (KPF)</subject><subject>Learning algorithms</subject><subject>Linear joint position feature (LJPF)</subject><subject>Low pass filters</subject><subject>Machine learning</subject><subject>Position sensing</subject><subject>Position-based statistical feature (PSF)</subject><subject>Posture</subject><subject>Science & Technology</subject><subject>Segments</subject><subject>Sensors</subject><subject>Skeleton</subject><subject>Skeleton data</subject><subject>Statistical analysis</subject><subject>Statistics</subject><subject>Support vector machines</subject><subject>Technology</subject><subject>Video data</subject><issn>0167-8655</issn><issn>1872-7344</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>HGBXW</sourceid><recordid>eNqNkE1LxDAURYMoOI7-AxcFl9KajzZtNsIwfsKAILoOaZoM6cw0NUkV_72ZdnAprt5d3JO8dwC4RDBDENGbNutFcEpmGGKUQZxBRI7ADFUlTkuS58dgFmtlWtGiOAVn3rcQQkpYNQOvCxmM7ZJI23Vnxjx4062TjenUTgQjfdJbHwanEq3EOGOH3CV-o7YqxNxa04Vka6XY4_4cnGix9eriMOfg_eH-bfmUrl4en5eLVSrznIa0pjUSBWtIIUtFCwQx01rEDUmJRYW1rIiuKBM1ZVJAxUTeUM0axAjFNcYlmYOr6d3e2Y9B-cBbO7gufslxgasCIoTz2MqnlnTWe6c0753ZCffNEeR7e7zlkz2-t8ch5tFexKoJ-1K11V4a1Un1i476KKMFjgmzpQnj6Us7dCGi1_9HY_t2aquo6tMoxw9EY-JOgTfW_L3pD5khm2A</recordid><startdate>202105</startdate><enddate>202105</enddate><creator>Ahad, Md Atiqur Rahman</creator><creator>Ahmed, Masud</creator><creator>Das Antar, Anindya</creator><creator>Makihara, Yasushi</creator><creator>Yagi, Yasushi</creator><general>Elsevier B.V</general><general>Elsevier</general><general>Elsevier Science Ltd</general><scope>BLEPL</scope><scope>DTL</scope><scope>HGBXW</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7TK</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0001-9912-8757</orcidid><orcidid>https://orcid.org/0000-0001-8355-7004</orcidid></search><sort><creationdate>202105</creationdate><title>Action recognition using kinematics posture feature on 3D skeleton joint locations</title><author>Ahad, Md Atiqur Rahman ; Ahmed, Masud ; Das Antar, Anindya ; Makihara, Yasushi ; Yagi, Yasushi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c446t-b6b1a59d35c7e651029ffa655372a82fc83f869ab69ca0e9a4d6f9d19362b2273</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Action recognition</topic><topic>Activity recognition</topic><topic>Angular joint position feature (AJPF)</topic><topic>Angular position</topic><topic>Benchmark datasets</topic><topic>Benchmarks</topic><topic>Classification</topic><topic>Computer Science</topic><topic>Computer Science, Artificial Intelligence</topic><topic>Computer vision</topic><topic>Convrnn</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Deep neural network</topic><topic>Domains</topic><topic>Ensemble architecture</topic><topic>Feature extraction</topic><topic>Joint angle</topic><topic>Joint position</topic><topic>Joints (anatomy)</topic><topic>Kinematics</topic><topic>Kinematics posture feature (KPF)</topic><topic>Learning algorithms</topic><topic>Linear joint position feature (LJPF)</topic><topic>Low pass filters</topic><topic>Machine learning</topic><topic>Position sensing</topic><topic>Position-based statistical feature (PSF)</topic><topic>Posture</topic><topic>Science & Technology</topic><topic>Segments</topic><topic>Sensors</topic><topic>Skeleton</topic><topic>Skeleton data</topic><topic>Statistical analysis</topic><topic>Statistics</topic><topic>Support vector machines</topic><topic>Technology</topic><topic>Video data</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Ahad, Md Atiqur Rahman</creatorcontrib><creatorcontrib>Ahmed, Masud</creatorcontrib><creatorcontrib>Das Antar, Anindya</creatorcontrib><creatorcontrib>Makihara, Yasushi</creatorcontrib><creatorcontrib>Yagi, Yasushi</creatorcontrib><collection>Web of Science Core Collection</collection><collection>Science Citation Index Expanded</collection><collection>Web of Science - Science Citation Index Expanded - 2021</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Pattern recognition letters</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ahad, Md Atiqur Rahman</au><au>Ahmed, Masud</au><au>Das Antar, Anindya</au><au>Makihara, Yasushi</au><au>Yagi, Yasushi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Action recognition using kinematics posture feature on 3D skeleton joint locations</atitle><jtitle>Pattern recognition letters</jtitle><stitle>PATTERN RECOGN LETT</stitle><date>2021-05</date><risdate>2021</risdate><volume>145</volume><spage>216</spage><epage>224</epage><pages>216-224</pages><issn>0167-8655</issn><eissn>1872-7344</eissn><abstract>•Motion information from skeleton can be efficiently encoded by considering the body joints as wearable kinematics sensor.•Proposed Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) from positions and bone-angles.•Proposed Kinematics Posture Feature (KPF) from LJPF and AJPF to encode motion and posture variation across frames.•Proposed Position-based Statistical Feature (PSF) from segmented KPF, that consists of temporal statistical features.•Five benchmark datasets are explored on PSF features with statistical and deep learning-based models.
Action recognition is a very widely explored research area in computer vision and related fields. We propose Kinematics Posture Feature (KPF) extraction from 3D joint positions based on skeleton data for improving the performance of action recognition. In this approach, we consider the skeleton 3D joints as kinematics sensors. We propose Linear Joint Position Feature (LJPF) and Angular Joint Position Feature (AJPF) based on 3D linear joint positions and angles between bone segments. We then combine these two kinematics features for each video frame for each action to create the KPF feature sets. These feature sets encode the variation of motion in the temporal domain as if each body joint represents kinematics position and orientation sensors. In the next stage, we process the extracted KPF feature descriptor by using a low pass filter, and segment them by using sliding windows with optimized length. This concept resembles the approach of processing kinematics sensor data. From the segmented windows, we compute the Position-based Statistical Feature (PSF). These features consist of temporal domain statistical features (e.g., mean, standard deviation, variance, etc.). These statistical features encode the variation of postures (i.e., joint positions and angles) across the video frames. For performing classification, we explore Support Vector Machine (Linear), RNN, CNNRNN, and ConvRNN model. The proposed PSF feature sets demonstrate prominent performance in both statistical machine learning- and deep learning-based models. For evaluation, we explore five benchmark datasets namely UTKinect-Action3D, Kinect Activity Recognition Dataset (KARD), MSR 3D Action Pairs, Florence 3D, and Office Activity Dataset (OAD). To prevent overfitting, we consider the leave-one-subject-out framework as the experimental setup and perform 10-fold cross-validation. Our approach outperforms several existing methods in these benchmark datasets and achieves very promising classification performance.</abstract><cop>AMSTERDAM</cop><pub>Elsevier B.V</pub><doi>10.1016/j.patrec.2021.02.013</doi><tpages>9</tpages><orcidid>https://orcid.org/0000-0001-9912-8757</orcidid><orcidid>https://orcid.org/0000-0001-8355-7004</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0167-8655 |
ispartof | Pattern recognition letters, 2021-05, Vol.145, p.216-224 |
issn | 0167-8655 1872-7344 |
language | eng |
recordid | cdi_webofscience_primary_000636965200029CitationCount |
source | Web of Science - Science Citation Index Expanded - 2021<img src="https://exlibris-pub.s3.amazonaws.com/fromwos-v2.jpg" />; Access via ScienceDirect (Elsevier) |
subjects | Action recognition Activity recognition Angular joint position feature (AJPF) Angular position Benchmark datasets Benchmarks Classification Computer Science Computer Science, Artificial Intelligence Computer vision Convrnn Datasets Deep learning Deep neural network Domains Ensemble architecture Feature extraction Joint angle Joint position Joints (anatomy) Kinematics Kinematics posture feature (KPF) Learning algorithms Linear joint position feature (LJPF) Low pass filters Machine learning Position sensing Position-based statistical feature (PSF) Posture Science & Technology Segments Sensors Skeleton Skeleton data Statistical analysis Statistics Support vector machines Technology Video data |
title | Action recognition using kinematics posture feature on 3D skeleton joint locations |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-12T13%3A47%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_webof&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Action%20recognition%20using%20kinematics%20posture%20feature%20on%203D%20skeleton%20joint%20locations&rft.jtitle=Pattern%20recognition%20letters&rft.au=Ahad,%20Md%20Atiqur%20Rahman&rft.date=2021-05&rft.volume=145&rft.spage=216&rft.epage=224&rft.pages=216-224&rft.issn=0167-8655&rft.eissn=1872-7344&rft_id=info:doi/10.1016/j.patrec.2021.02.013&rft_dat=%3Cproquest_webof%3E2528501124%3C/proquest_webof%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2528501124&rft_id=info:pmid/&rft_els_id=S0167865521000751&rfr_iscdi=true |