Robust video content analysis schemes for human action recognition

Introduction: Action recognition is a challenging time series classification task that has received much attention in the recent past due to its importance in critical applications, such as surveillance, visual behavior study, topic discovery, security, and content retrieval. Objectives: The main ob...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Science progress (1916) 2021-04, Vol.104 (2), p.368504211005480-368504211005480
Hauptverfasser: Aly, Cherry A., Abas, Fazly S., Ann, Goh H.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 368504211005480
container_issue 2
container_start_page 368504211005480
container_title Science progress (1916)
container_volume 104
creator Aly, Cherry A.
Abas, Fazly S.
Ann, Goh H.
description Introduction: Action recognition is a challenging time series classification task that has received much attention in the recent past due to its importance in critical applications, such as surveillance, visual behavior study, topic discovery, security, and content retrieval. Objectives: The main objective of the research is to develop a robust and high-performance human action recognition techniques. A combination of local and holistic feature extraction methods used through analyzing the most effective features to extract to reach the objective, followed by using simple and high-performance machine learning algorithms. Methods: This paper presents three robust action recognition techniques based on a series of image analysis methods to detect activities in different scenes. The general scheme architecture consists of shot boundary detection, shot frame rate re-sampling, and compact feature vector extraction. This process is achieved by emphasizing variations and extracting strong patterns in feature vectors before classification. Results: The proposed schemes are tested on datasets with cluttered backgrounds, low- or high-resolution videos, different viewpoints, and different camera motion conditions, namely, the Hollywood-2, KTH, UCF11 (YouTube actions), and Weizmann datasets. The proposed schemes resulted in highly accurate video analysis results compared to those of other works based on four widely used datasets. The First, Second, and Third Schemes provides recognition accuracies of 57.8%, 73.6%, and 52.0% on Hollywood2, 94.5%, 97.0%, and 59.3% on KTH, 94.5%, 95.6%, and 94.2% on UCF11, and 98.9%, 97.8% and 100% on Weizmann. Conclusion: Each of the proposed schemes provides high recognition accuracy compared to other state-of-art methods. Especially, the Second Scheme as it gives excellent comparable results to other benchmarked approaches.
doi_str_mv 10.1177/00368504211005480
format Article
fullrecord <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_10455027</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sage_id>10.1177_00368504211005480</sage_id><sourcerecordid>2551526760</sourcerecordid><originalsourceid>FETCH-LOGICAL-c3820-4f855ddc7fdeca2761866637f17fb5735afa99436f3348585343add5608e86543</originalsourceid><addsrcrecordid>eNp1kUtLxDAQx4Mouj4-gBcpePHSddI8exIVX7AgiJ5DNk12K22zJu3CfntbVtcXnmZgfvOfxx-hYwxjjIU4ByBcMqAZxgCMSthCowyoSAXmZBuNhno6AHtoP8ZXAMwwl7toj5AcEyLkCF09-WkX22RZFtYnxjetbdpEN7paxTIm0cxtbWPifEjmXa2bRJu29E0SrPGzphzyQ7TjdBXt0Uc8QC-3N8_X9-nk8e7h-nKSGiIzSKmTjBWFEa6wRmeCY8k5J8Jh4aZMEKadznNKuCOESiYZoUQXBeMgreSMkgN0sdZddNPaFqZfNOhKLUJZ67BSXpfqZ6Up52rmlwoDZQwy0SucfSgE_9bZ2Kq6jMZWlW6s76LKGM5FLiXwHj39hb76LvRvGSiGWcYFh57Ca8oEH2OwbrMNBjVYpP5Y1PecfD9j0_HpSQ-M10DUM_s19n_Fd7AumA8</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2551526760</pqid></control><display><type>article</type><title>Robust video content analysis schemes for human action recognition</title><source>MEDLINE</source><source>DOAJ Directory of Open Access Journals</source><source>Sage Journals GOLD Open Access 2024</source><source>EZB-FREE-00999 freely available EZB journals</source><source>PubMed Central</source><source>Alma/SFX Local Collection</source><creator>Aly, Cherry A. ; Abas, Fazly S. ; Ann, Goh H.</creator><creatorcontrib>Aly, Cherry A. ; Abas, Fazly S. ; Ann, Goh H.</creatorcontrib><description>Introduction: Action recognition is a challenging time series classification task that has received much attention in the recent past due to its importance in critical applications, such as surveillance, visual behavior study, topic discovery, security, and content retrieval. Objectives: The main objective of the research is to develop a robust and high-performance human action recognition techniques. A combination of local and holistic feature extraction methods used through analyzing the most effective features to extract to reach the objective, followed by using simple and high-performance machine learning algorithms. Methods: This paper presents three robust action recognition techniques based on a series of image analysis methods to detect activities in different scenes. The general scheme architecture consists of shot boundary detection, shot frame rate re-sampling, and compact feature vector extraction. This process is achieved by emphasizing variations and extracting strong patterns in feature vectors before classification. Results: The proposed schemes are tested on datasets with cluttered backgrounds, low- or high-resolution videos, different viewpoints, and different camera motion conditions, namely, the Hollywood-2, KTH, UCF11 (YouTube actions), and Weizmann datasets. The proposed schemes resulted in highly accurate video analysis results compared to those of other works based on four widely used datasets. The First, Second, and Third Schemes provides recognition accuracies of 57.8%, 73.6%, and 52.0% on Hollywood2, 94.5%, 97.0%, and 59.3% on KTH, 94.5%, 95.6%, and 94.2% on UCF11, and 98.9%, 97.8% and 100% on Weizmann. Conclusion: Each of the proposed schemes provides high recognition accuracy compared to other state-of-art methods. Especially, the Second Scheme as it gives excellent comparable results to other benchmarked approaches.</description><identifier>ISSN: 0036-8504</identifier><identifier>EISSN: 2047-7163</identifier><identifier>DOI: 10.1177/00368504211005480</identifier><identifier>PMID: 33913378</identifier><language>eng</language><publisher>London, England: SAGE Publications</publisher><subject>Algorithms ; Classification ; Content analysis ; Datasets ; Feature extraction ; Human Activities ; Human activity recognition ; Human motion ; Human performance ; Humans ; Image analysis ; Image processing ; Image Processing, Computer-Assisted - methods ; Learning algorithms ; Machine learning ; Object recognition ; Pattern Recognition, Automated - methods ; Robustness</subject><ispartof>Science progress (1916), 2021-04, Vol.104 (2), p.368504211005480-368504211005480</ispartof><rights>The Author(s) 2021</rights><rights>2021. This article is distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 License ( https://creativecommons.org/licenses/by-nc/4.0/ ) which permits non-commercial use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access pages ( https://us.sagepub.com/en-us/nam/open-access-at-sage ). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>The Author(s) 2021 2021 SAGE Publications</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c3820-4f855ddc7fdeca2761866637f17fb5735afa99436f3348585343add5608e86543</citedby><cites>FETCH-LOGICAL-c3820-4f855ddc7fdeca2761866637f17fb5735afa99436f3348585343add5608e86543</cites><orcidid>0000-0002-8744-8793</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC10455027/pdf/$$EPDF$$P50$$Gpubmedcentral$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC10455027/$$EHTML$$P50$$Gpubmedcentral$$Hfree_for_read</linktohtml><link.rule.ids>230,315,729,782,786,866,887,21973,27860,27931,27932,44952,45340,53798,53800</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/33913378$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Aly, Cherry A.</creatorcontrib><creatorcontrib>Abas, Fazly S.</creatorcontrib><creatorcontrib>Ann, Goh H.</creatorcontrib><title>Robust video content analysis schemes for human action recognition</title><title>Science progress (1916)</title><addtitle>Sci Prog</addtitle><description>Introduction: Action recognition is a challenging time series classification task that has received much attention in the recent past due to its importance in critical applications, such as surveillance, visual behavior study, topic discovery, security, and content retrieval. Objectives: The main objective of the research is to develop a robust and high-performance human action recognition techniques. A combination of local and holistic feature extraction methods used through analyzing the most effective features to extract to reach the objective, followed by using simple and high-performance machine learning algorithms. Methods: This paper presents three robust action recognition techniques based on a series of image analysis methods to detect activities in different scenes. The general scheme architecture consists of shot boundary detection, shot frame rate re-sampling, and compact feature vector extraction. This process is achieved by emphasizing variations and extracting strong patterns in feature vectors before classification. Results: The proposed schemes are tested on datasets with cluttered backgrounds, low- or high-resolution videos, different viewpoints, and different camera motion conditions, namely, the Hollywood-2, KTH, UCF11 (YouTube actions), and Weizmann datasets. The proposed schemes resulted in highly accurate video analysis results compared to those of other works based on four widely used datasets. The First, Second, and Third Schemes provides recognition accuracies of 57.8%, 73.6%, and 52.0% on Hollywood2, 94.5%, 97.0%, and 59.3% on KTH, 94.5%, 95.6%, and 94.2% on UCF11, and 98.9%, 97.8% and 100% on Weizmann. Conclusion: Each of the proposed schemes provides high recognition accuracy compared to other state-of-art methods. Especially, the Second Scheme as it gives excellent comparable results to other benchmarked approaches.</description><subject>Algorithms</subject><subject>Classification</subject><subject>Content analysis</subject><subject>Datasets</subject><subject>Feature extraction</subject><subject>Human Activities</subject><subject>Human activity recognition</subject><subject>Human motion</subject><subject>Human performance</subject><subject>Humans</subject><subject>Image analysis</subject><subject>Image processing</subject><subject>Image Processing, Computer-Assisted - methods</subject><subject>Learning algorithms</subject><subject>Machine learning</subject><subject>Object recognition</subject><subject>Pattern Recognition, Automated - methods</subject><subject>Robustness</subject><issn>0036-8504</issn><issn>2047-7163</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>AFRWT</sourceid><sourceid>EIF</sourceid><recordid>eNp1kUtLxDAQx4Mouj4-gBcpePHSddI8exIVX7AgiJ5DNk12K22zJu3CfntbVtcXnmZgfvOfxx-hYwxjjIU4ByBcMqAZxgCMSthCowyoSAXmZBuNhno6AHtoP8ZXAMwwl7toj5AcEyLkCF09-WkX22RZFtYnxjetbdpEN7paxTIm0cxtbWPifEjmXa2bRJu29E0SrPGzphzyQ7TjdBXt0Uc8QC-3N8_X9-nk8e7h-nKSGiIzSKmTjBWFEa6wRmeCY8k5J8Jh4aZMEKadznNKuCOESiYZoUQXBeMgreSMkgN0sdZddNPaFqZfNOhKLUJZ67BSXpfqZ6Up52rmlwoDZQwy0SucfSgE_9bZ2Kq6jMZWlW6s76LKGM5FLiXwHj39hb76LvRvGSiGWcYFh57Ca8oEH2OwbrMNBjVYpP5Y1PecfD9j0_HpSQ-M10DUM_s19n_Fd7AumA8</recordid><startdate>202104</startdate><enddate>202104</enddate><creator>Aly, Cherry A.</creator><creator>Abas, Fazly S.</creator><creator>Ann, Goh H.</creator><general>SAGE Publications</general><general>Sage Publications Ltd</general><scope>AFRWT</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>JQ2</scope><scope>K9.</scope><scope>7X8</scope><scope>5PM</scope><orcidid>https://orcid.org/0000-0002-8744-8793</orcidid></search><sort><creationdate>202104</creationdate><title>Robust video content analysis schemes for human action recognition</title><author>Aly, Cherry A. ; Abas, Fazly S. ; Ann, Goh H.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c3820-4f855ddc7fdeca2761866637f17fb5735afa99436f3348585343add5608e86543</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Algorithms</topic><topic>Classification</topic><topic>Content analysis</topic><topic>Datasets</topic><topic>Feature extraction</topic><topic>Human Activities</topic><topic>Human activity recognition</topic><topic>Human motion</topic><topic>Human performance</topic><topic>Humans</topic><topic>Image analysis</topic><topic>Image processing</topic><topic>Image Processing, Computer-Assisted - methods</topic><topic>Learning algorithms</topic><topic>Machine learning</topic><topic>Object recognition</topic><topic>Pattern Recognition, Automated - methods</topic><topic>Robustness</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Aly, Cherry A.</creatorcontrib><creatorcontrib>Abas, Fazly S.</creatorcontrib><creatorcontrib>Ann, Goh H.</creatorcontrib><collection>Sage Journals GOLD Open Access 2024</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology &amp; Engineering</collection><collection>Engineering Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Health &amp; Medical Complete (Alumni)</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Science progress (1916)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Aly, Cherry A.</au><au>Abas, Fazly S.</au><au>Ann, Goh H.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Robust video content analysis schemes for human action recognition</atitle><jtitle>Science progress (1916)</jtitle><addtitle>Sci Prog</addtitle><date>2021-04</date><risdate>2021</risdate><volume>104</volume><issue>2</issue><spage>368504211005480</spage><epage>368504211005480</epage><pages>368504211005480-368504211005480</pages><issn>0036-8504</issn><eissn>2047-7163</eissn><abstract>Introduction: Action recognition is a challenging time series classification task that has received much attention in the recent past due to its importance in critical applications, such as surveillance, visual behavior study, topic discovery, security, and content retrieval. Objectives: The main objective of the research is to develop a robust and high-performance human action recognition techniques. A combination of local and holistic feature extraction methods used through analyzing the most effective features to extract to reach the objective, followed by using simple and high-performance machine learning algorithms. Methods: This paper presents three robust action recognition techniques based on a series of image analysis methods to detect activities in different scenes. The general scheme architecture consists of shot boundary detection, shot frame rate re-sampling, and compact feature vector extraction. This process is achieved by emphasizing variations and extracting strong patterns in feature vectors before classification. Results: The proposed schemes are tested on datasets with cluttered backgrounds, low- or high-resolution videos, different viewpoints, and different camera motion conditions, namely, the Hollywood-2, KTH, UCF11 (YouTube actions), and Weizmann datasets. The proposed schemes resulted in highly accurate video analysis results compared to those of other works based on four widely used datasets. The First, Second, and Third Schemes provides recognition accuracies of 57.8%, 73.6%, and 52.0% on Hollywood2, 94.5%, 97.0%, and 59.3% on KTH, 94.5%, 95.6%, and 94.2% on UCF11, and 98.9%, 97.8% and 100% on Weizmann. Conclusion: Each of the proposed schemes provides high recognition accuracy compared to other state-of-art methods. Especially, the Second Scheme as it gives excellent comparable results to other benchmarked approaches.</abstract><cop>London, England</cop><pub>SAGE Publications</pub><pmid>33913378</pmid><doi>10.1177/00368504211005480</doi><orcidid>https://orcid.org/0000-0002-8744-8793</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0036-8504
ispartof Science progress (1916), 2021-04, Vol.104 (2), p.368504211005480-368504211005480
issn 0036-8504
2047-7163
language eng
recordid cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_10455027
source MEDLINE; DOAJ Directory of Open Access Journals; Sage Journals GOLD Open Access 2024; EZB-FREE-00999 freely available EZB journals; PubMed Central; Alma/SFX Local Collection
subjects Algorithms
Classification
Content analysis
Datasets
Feature extraction
Human Activities
Human activity recognition
Human motion
Human performance
Humans
Image analysis
Image processing
Image Processing, Computer-Assisted - methods
Learning algorithms
Machine learning
Object recognition
Pattern Recognition, Automated - methods
Robustness
title Robust video content analysis schemes for human action recognition
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-06T09%3A36%3A49IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Robust%20video%20content%20analysis%20schemes%20for%20human%20action%20recognition&rft.jtitle=Science%20progress%20(1916)&rft.au=Aly,%20Cherry%20A.&rft.date=2021-04&rft.volume=104&rft.issue=2&rft.spage=368504211005480&rft.epage=368504211005480&rft.pages=368504211005480-368504211005480&rft.issn=0036-8504&rft.eissn=2047-7163&rft_id=info:doi/10.1177/00368504211005480&rft_dat=%3Cproquest_pubme%3E2551526760%3C/proquest_pubme%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2551526760&rft_id=info:pmid/33913378&rft_sage_id=10.1177_00368504211005480&rfr_iscdi=true