Improving Human Activity Recognition Integrating LSTM with Different Data Sources: Features, Object Detection and Skeleton Tracking

Over the past few years, technologies in the field of computer vision have greatly advanced. The use of deep neural networks, together with the development of computing capabilities, has made it possible to solve problems of great interest to society. In this work, we focus on one such problem that...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2022, Vol.10, p.1-1
Hauptverfasser: Domingo, Jaime Duque, Gomez-Garcia-Bermejo, Jaime, Zalama, Eduardo
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1
container_issue
container_start_page 1
container_title IEEE access
container_volume 10
creator Domingo, Jaime Duque
Gomez-Garcia-Bermejo, Jaime
Zalama, Eduardo
description Over the past few years, technologies in the field of computer vision have greatly advanced. The use of deep neural networks, together with the development of computing capabilities, has made it possible to solve problems of great interest to society. In this work, we focus on one such problem that has seen a great development, the recognition of actions in live videos. Although the problem has been oriented in different ways in the literature, we have focused on indoor residential environments, such as a house or a nursing home. Our system can be used to understand what actions a person or group of people are carrying out. Two of the approaches used to solve the problem have been 3D convolution networks and recurrent networks. In our case, we have created a model that accurately combines several recurrent networks with processed data from different techniques: image feature extraction, object detection and people's skeletons. The need to integrate these three techniques arises from the search to improve the detection of certain actions by taking advantage of the best recognition offered by each of the methods. In a complete experimentation, where several techniques have been evaluated against different datasets, the classification of the actions has been improved with respect to the existing models.
doi_str_mv 10.1109/ACCESS.2022.3186465
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2685163673</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9807301</ieee_id><doaj_id>oai_doaj_org_article_72d2abf8418c4ce6a88dcef4dde9a2e2</doaj_id><sourcerecordid>2685163673</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-eaf5d0c054e450304640ebf55245bc119acebe1dc6d24c2210f42a9a2d75fbf73</originalsourceid><addsrcrecordid>eNpNkUtvEzEUhUcIJKq2v6AbS2xJ8Hs87KK0pZGCKjFhbd2xr4PTZKZ4nKKu-eN1mKrCm2tfn_P5carqitE5Y7T5slgub9p2zinnc8GMllq9q844081MKKHf_zf_WF2O446WYUpL1WfV39XhMQ1Psd-Su-MBerJwOT7F_Ex-oBu2fcxx6Mmqz7hNkE-ydbv5Tv7E_ItcxxAwYZ_JNWQg7XBMDsev5BYhHxOOn8l9t0NXtjGXcgJB70n7gHvMZbFJ4B4K8qL6EGA_4uVrPa9-3t5slnez9f231XKxnjlJTZ4hBOWpo0qiVFRQqSXFLijFpeocYw047JB5pz2XjnNGg-TQAPe1Cl2oxXm1mrh-gJ19TPEA6dkOEO2_xpC2FlKObo-25p5DF4xkxkmHGozxDoP0HgsQeWF9mljl934fccx2V57fl-tbro1iWuhaFJWYVC4N45gwvJ3KqD2FZ6fw7Ck8-xpecV1NroiIb47G0FpQJl4Ag-eXqw</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2685163673</pqid></control><display><type>article</type><title>Improving Human Activity Recognition Integrating LSTM with Different Data Sources: Features, Object Detection and Skeleton Tracking</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>EZB-FREE-00999 freely available EZB journals</source><creator>Domingo, Jaime Duque ; Gomez-Garcia-Bermejo, Jaime ; Zalama, Eduardo</creator><creatorcontrib>Domingo, Jaime Duque ; Gomez-Garcia-Bermejo, Jaime ; Zalama, Eduardo</creatorcontrib><description>Over the past few years, technologies in the field of computer vision have greatly advanced. The use of deep neural networks, together with the development of computing capabilities, has made it possible to solve problems of great interest to society. In this work, we focus on one such problem that has seen a great development, the recognition of actions in live videos. Although the problem has been oriented in different ways in the literature, we have focused on indoor residential environments, such as a house or a nursing home. Our system can be used to understand what actions a person or group of people are carrying out. Two of the approaches used to solve the problem have been 3D convolution networks and recurrent networks. In our case, we have created a model that accurately combines several recurrent networks with processed data from different techniques: image feature extraction, object detection and people's skeletons. The need to integrate these three techniques arises from the search to improve the detection of certain actions by taking advantage of the best recognition offered by each of the methods. In a complete experimentation, where several techniques have been evaluated against different datasets, the classification of the actions has been improved with respect to the existing models.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2022.3186465</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Artificial neural networks ; Computer architecture ; Computer vision ; Experimentation ; Feature extraction ; Feature recognition ; HAR ; Human Activity Recognition ; Indoor environments ; LRCN ; LSTM ; Object recognition ; OpenPose ; Recurrent Neural Network ; Semantics ; Sensors ; Skeleton ; Three-dimensional displays ; Videos ; YOLO</subject><ispartof>IEEE access, 2022, Vol.10, p.1-1</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-eaf5d0c054e450304640ebf55245bc119acebe1dc6d24c2210f42a9a2d75fbf73</citedby><cites>FETCH-LOGICAL-c408t-eaf5d0c054e450304640ebf55245bc119acebe1dc6d24c2210f42a9a2d75fbf73</cites><orcidid>0000-0001-6649-5550 ; 0000-0003-4763-5356</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9807301$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,864,2102,4024,27633,27923,27924,27925,54933</link.rule.ids></links><search><creatorcontrib>Domingo, Jaime Duque</creatorcontrib><creatorcontrib>Gomez-Garcia-Bermejo, Jaime</creatorcontrib><creatorcontrib>Zalama, Eduardo</creatorcontrib><title>Improving Human Activity Recognition Integrating LSTM with Different Data Sources: Features, Object Detection and Skeleton Tracking</title><title>IEEE access</title><addtitle>Access</addtitle><description>Over the past few years, technologies in the field of computer vision have greatly advanced. The use of deep neural networks, together with the development of computing capabilities, has made it possible to solve problems of great interest to society. In this work, we focus on one such problem that has seen a great development, the recognition of actions in live videos. Although the problem has been oriented in different ways in the literature, we have focused on indoor residential environments, such as a house or a nursing home. Our system can be used to understand what actions a person or group of people are carrying out. Two of the approaches used to solve the problem have been 3D convolution networks and recurrent networks. In our case, we have created a model that accurately combines several recurrent networks with processed data from different techniques: image feature extraction, object detection and people's skeletons. The need to integrate these three techniques arises from the search to improve the detection of certain actions by taking advantage of the best recognition offered by each of the methods. In a complete experimentation, where several techniques have been evaluated against different datasets, the classification of the actions has been improved with respect to the existing models.</description><subject>Artificial neural networks</subject><subject>Computer architecture</subject><subject>Computer vision</subject><subject>Experimentation</subject><subject>Feature extraction</subject><subject>Feature recognition</subject><subject>HAR</subject><subject>Human Activity Recognition</subject><subject>Indoor environments</subject><subject>LRCN</subject><subject>LSTM</subject><subject>Object recognition</subject><subject>OpenPose</subject><subject>Recurrent Neural Network</subject><subject>Semantics</subject><subject>Sensors</subject><subject>Skeleton</subject><subject>Three-dimensional displays</subject><subject>Videos</subject><subject>YOLO</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNkUtvEzEUhUcIJKq2v6AbS2xJ8Hs87KK0pZGCKjFhbd2xr4PTZKZ4nKKu-eN1mKrCm2tfn_P5carqitE5Y7T5slgub9p2zinnc8GMllq9q844081MKKHf_zf_WF2O446WYUpL1WfV39XhMQ1Psd-Su-MBerJwOT7F_Ex-oBu2fcxx6Mmqz7hNkE-ydbv5Tv7E_ItcxxAwYZ_JNWQg7XBMDsev5BYhHxOOn8l9t0NXtjGXcgJB70n7gHvMZbFJ4B4K8qL6EGA_4uVrPa9-3t5slnez9f231XKxnjlJTZ4hBOWpo0qiVFRQqSXFLijFpeocYw047JB5pz2XjnNGg-TQAPe1Cl2oxXm1mrh-gJ19TPEA6dkOEO2_xpC2FlKObo-25p5DF4xkxkmHGozxDoP0HgsQeWF9mljl934fccx2V57fl-tbro1iWuhaFJWYVC4N45gwvJ3KqD2FZ6fw7Ck8-xpecV1NroiIb47G0FpQJl4Ag-eXqw</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Domingo, Jaime Duque</creator><creator>Gomez-Garcia-Bermejo, Jaime</creator><creator>Zalama, Eduardo</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0001-6649-5550</orcidid><orcidid>https://orcid.org/0000-0003-4763-5356</orcidid></search><sort><creationdate>2022</creationdate><title>Improving Human Activity Recognition Integrating LSTM with Different Data Sources: Features, Object Detection and Skeleton Tracking</title><author>Domingo, Jaime Duque ; Gomez-Garcia-Bermejo, Jaime ; Zalama, Eduardo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-eaf5d0c054e450304640ebf55245bc119acebe1dc6d24c2210f42a9a2d75fbf73</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Artificial neural networks</topic><topic>Computer architecture</topic><topic>Computer vision</topic><topic>Experimentation</topic><topic>Feature extraction</topic><topic>Feature recognition</topic><topic>HAR</topic><topic>Human Activity Recognition</topic><topic>Indoor environments</topic><topic>LRCN</topic><topic>LSTM</topic><topic>Object recognition</topic><topic>OpenPose</topic><topic>Recurrent Neural Network</topic><topic>Semantics</topic><topic>Sensors</topic><topic>Skeleton</topic><topic>Three-dimensional displays</topic><topic>Videos</topic><topic>YOLO</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Domingo, Jaime Duque</creatorcontrib><creatorcontrib>Gomez-Garcia-Bermejo, Jaime</creatorcontrib><creatorcontrib>Zalama, Eduardo</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Domingo, Jaime Duque</au><au>Gomez-Garcia-Bermejo, Jaime</au><au>Zalama, Eduardo</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Improving Human Activity Recognition Integrating LSTM with Different Data Sources: Features, Object Detection and Skeleton Tracking</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2022</date><risdate>2022</risdate><volume>10</volume><spage>1</spage><epage>1</epage><pages>1-1</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>Over the past few years, technologies in the field of computer vision have greatly advanced. The use of deep neural networks, together with the development of computing capabilities, has made it possible to solve problems of great interest to society. In this work, we focus on one such problem that has seen a great development, the recognition of actions in live videos. Although the problem has been oriented in different ways in the literature, we have focused on indoor residential environments, such as a house or a nursing home. Our system can be used to understand what actions a person or group of people are carrying out. Two of the approaches used to solve the problem have been 3D convolution networks and recurrent networks. In our case, we have created a model that accurately combines several recurrent networks with processed data from different techniques: image feature extraction, object detection and people's skeletons. The need to integrate these three techniques arises from the search to improve the detection of certain actions by taking advantage of the best recognition offered by each of the methods. In a complete experimentation, where several techniques have been evaluated against different datasets, the classification of the actions has been improved with respect to the existing models.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2022.3186465</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0001-6649-5550</orcidid><orcidid>https://orcid.org/0000-0003-4763-5356</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2169-3536
ispartof IEEE access, 2022, Vol.10, p.1-1
issn 2169-3536
2169-3536
language eng
recordid cdi_proquest_journals_2685163673
source IEEE Open Access Journals; DOAJ Directory of Open Access Journals; EZB-FREE-00999 freely available EZB journals
subjects Artificial neural networks
Computer architecture
Computer vision
Experimentation
Feature extraction
Feature recognition
HAR
Human Activity Recognition
Indoor environments
LRCN
LSTM
Object recognition
OpenPose
Recurrent Neural Network
Semantics
Sensors
Skeleton
Three-dimensional displays
Videos
YOLO
title Improving Human Activity Recognition Integrating LSTM with Different Data Sources: Features, Object Detection and Skeleton Tracking
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-22T23%3A29%3A50IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Improving%20Human%20Activity%20Recognition%20Integrating%20LSTM%20with%20Different%20Data%20Sources:%20Features,%20Object%20Detection%20and%20Skeleton%20Tracking&rft.jtitle=IEEE%20access&rft.au=Domingo,%20Jaime%20Duque&rft.date=2022&rft.volume=10&rft.spage=1&rft.epage=1&rft.pages=1-1&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2022.3186465&rft_dat=%3Cproquest_cross%3E2685163673%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2685163673&rft_id=info:pmid/&rft_ieee_id=9807301&rft_doaj_id=oai_doaj_org_article_72d2abf8418c4ce6a88dcef4dde9a2e2&rfr_iscdi=true