Edge Computing-Based Video Action Recognition Method and Its Application in Online Physical Education Teaching
Due to the impact of COVID-19, online physical education (PE) teaching has garnered increasing attention. Given the characteristics of online PE teaching, introducing artificial intelligence technology to automatically detect or recognize students' actions or behaviors has gradually emerged as...
Gespeichert in:
Veröffentlicht in: | IEEE access 2024, Vol.12, p.148666-148676 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 148676 |
---|---|
container_issue | |
container_start_page | 148666 |
container_title | IEEE access |
container_volume | 12 |
creator | Han, Jinzhu Zhao, Jinjin Yue, Yan Che, Xinrui |
description | Due to the impact of COVID-19, online physical education (PE) teaching has garnered increasing attention. Given the characteristics of online PE teaching, introducing artificial intelligence technology to automatically detect or recognize students' actions or behaviors has gradually emerged as a trend. However, traditional cloud computing-based intelligent online PE teaching systems often face various challenging issues, such as computational complexity and latency. Edge computing can address these problems. However, edge devices typically have limited computing power, while existing deep action recognition models often contain a large number of parameters and require significant computational resources, making them difficult to deploy on edge devices. To address the above issues, this paper proposes a lightweight video recognition method, named the lightweight video ViT (LWV-ViT) network. More specifically, based on the standard ViT model, the video-based ViT (VBViT) network is first introduced by developing a cross-temporal token interaction module to effectively process temporal information in videos. Furthermore, the LWV-ViT network is proposed by implementing a spatial-temporal pruning scheme to reduce the number of parameters. Finally, the proposed LWV-ViT network is deployed in an edge computing-based online PE teaching system, where it is installed on each edge device. This setup enables fast data processing, reduces transmission latency, and protects sensitive data. Experimental results show that the proposed LWV-ViT network achieves the best recognition rates for both behavior detection (96.5%, 95.73%) and action recognition (97.9%, 88.3%, 79.9%) tasks, and has the fewest trainable parameters (2.7 M), which means it performs well in edge computing-based online PE teaching systems. |
doi_str_mv | 10.1109/ACCESS.2024.3475372 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_3118091812</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10706898</ieee_id><doaj_id>oai_doaj_org_article_249c14ab695147df98e2e1cbd1fc22d6</doaj_id><sourcerecordid>3118091812</sourcerecordid><originalsourceid>FETCH-LOGICAL-c289t-9fb295fd6ab871985313237521b4bb0be62874d7a763b6268d333c1f9fd890f33</originalsourceid><addsrcrecordid>eNpNUU1P4zAQjVasBGL5BXCwtOd0_ZH441iiwlYCgYDlavmzdRXsEKcH_v2apkLMZUZv5r0ZzauqSwQXCEHxZ9l1q-fnBYa4WZCGtYThH9UZRlTUpCX05Ft9Wl3kvIMleIFadlbFld040KW3YT-FuKmvVXYWvAbrEliaKaQInpxJmxgO9b2btskCFS1YTxksh6EPRh1aIYKH2IfowOP2Ixe0Byu7PzZfnDLbov-r-ulVn93FMZ9X_25WL93f-u7hdt0t72qDuZhq4TUWrbdUac6Q4C1BBBPWYqQbraF2FHPWWKYYJZpiyi0hxCAvvOUCekLOq_Wsa5PayWEMb2r8kEkFeQDSuJFqnILpncSNMKhRuvwDNcx6wR12yGiLvMHY0qL1e9YaxvS-d3mSu7QfYzlfEoQ4FIgjXKbIPGXGlPPo_NdWBOWnT3L2SX76JI8-FdbVzArOuW8MBikXnPwHMSWN9w</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3118091812</pqid></control><display><type>article</type><title>Edge Computing-Based Video Action Recognition Method and Its Application in Online Physical Education Teaching</title><source>DOAJ Directory of Open Access Journals</source><source>IEEE Xplore Open Access Journals</source><source>EZB Electronic Journals Library</source><creator>Han, Jinzhu ; Zhao, Jinjin ; Yue, Yan ; Che, Xinrui</creator><creatorcontrib>Han, Jinzhu ; Zhao, Jinjin ; Yue, Yan ; Che, Xinrui</creatorcontrib><description>Due to the impact of COVID-19, online physical education (PE) teaching has garnered increasing attention. Given the characteristics of online PE teaching, introducing artificial intelligence technology to automatically detect or recognize students' actions or behaviors has gradually emerged as a trend. However, traditional cloud computing-based intelligent online PE teaching systems often face various challenging issues, such as computational complexity and latency. Edge computing can address these problems. However, edge devices typically have limited computing power, while existing deep action recognition models often contain a large number of parameters and require significant computational resources, making them difficult to deploy on edge devices. To address the above issues, this paper proposes a lightweight video recognition method, named the lightweight video ViT (LWV-ViT) network. More specifically, based on the standard ViT model, the video-based ViT (VBViT) network is first introduced by developing a cross-temporal token interaction module to effectively process temporal information in videos. Furthermore, the LWV-ViT network is proposed by implementing a spatial-temporal pruning scheme to reduce the number of parameters. Finally, the proposed LWV-ViT network is deployed in an edge computing-based online PE teaching system, where it is installed on each edge device. This setup enables fast data processing, reduces transmission latency, and protects sensitive data. Experimental results show that the proposed LWV-ViT network achieves the best recognition rates for both behavior detection (96.5%, 95.73%) and action recognition (97.9%, 88.3%, 79.9%) tasks, and has the fewest trainable parameters (2.7 M), which means it performs well in edge computing-based online PE teaching systems.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2024.3475372</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Activity recognition ; Artificial intelligence ; Cloud computing ; Computational complexity ; Convolutional neural networks ; cross-temporal token interaction ; Data processing ; Edge computing ; Education ; Electronic learning ; Feature extraction ; Image recognition ; Lightweight ; lightweight network ; online PE teaching ; Parameter sensitivity ; Physical education ; Servers ; spatial-temporal pruning ; Spatiotemporal phenomena ; Three-dimensional displays ; Transformers ; Video action recognition ; Weight reduction</subject><ispartof>IEEE access, 2024, Vol.12, p.148666-148676</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c289t-9fb295fd6ab871985313237521b4bb0be62874d7a763b6268d333c1f9fd890f33</cites><orcidid>0009-0007-1106-5925</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10706898$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,860,2096,4010,27610,27900,27901,27902,54908</link.rule.ids></links><search><creatorcontrib>Han, Jinzhu</creatorcontrib><creatorcontrib>Zhao, Jinjin</creatorcontrib><creatorcontrib>Yue, Yan</creatorcontrib><creatorcontrib>Che, Xinrui</creatorcontrib><title>Edge Computing-Based Video Action Recognition Method and Its Application in Online Physical Education Teaching</title><title>IEEE access</title><addtitle>Access</addtitle><description>Due to the impact of COVID-19, online physical education (PE) teaching has garnered increasing attention. Given the characteristics of online PE teaching, introducing artificial intelligence technology to automatically detect or recognize students' actions or behaviors has gradually emerged as a trend. However, traditional cloud computing-based intelligent online PE teaching systems often face various challenging issues, such as computational complexity and latency. Edge computing can address these problems. However, edge devices typically have limited computing power, while existing deep action recognition models often contain a large number of parameters and require significant computational resources, making them difficult to deploy on edge devices. To address the above issues, this paper proposes a lightweight video recognition method, named the lightweight video ViT (LWV-ViT) network. More specifically, based on the standard ViT model, the video-based ViT (VBViT) network is first introduced by developing a cross-temporal token interaction module to effectively process temporal information in videos. Furthermore, the LWV-ViT network is proposed by implementing a spatial-temporal pruning scheme to reduce the number of parameters. Finally, the proposed LWV-ViT network is deployed in an edge computing-based online PE teaching system, where it is installed on each edge device. This setup enables fast data processing, reduces transmission latency, and protects sensitive data. Experimental results show that the proposed LWV-ViT network achieves the best recognition rates for both behavior detection (96.5%, 95.73%) and action recognition (97.9%, 88.3%, 79.9%) tasks, and has the fewest trainable parameters (2.7 M), which means it performs well in edge computing-based online PE teaching systems.</description><subject>Activity recognition</subject><subject>Artificial intelligence</subject><subject>Cloud computing</subject><subject>Computational complexity</subject><subject>Convolutional neural networks</subject><subject>cross-temporal token interaction</subject><subject>Data processing</subject><subject>Edge computing</subject><subject>Education</subject><subject>Electronic learning</subject><subject>Feature extraction</subject><subject>Image recognition</subject><subject>Lightweight</subject><subject>lightweight network</subject><subject>online PE teaching</subject><subject>Parameter sensitivity</subject><subject>Physical education</subject><subject>Servers</subject><subject>spatial-temporal pruning</subject><subject>Spatiotemporal phenomena</subject><subject>Three-dimensional displays</subject><subject>Transformers</subject><subject>Video action recognition</subject><subject>Weight reduction</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNUU1P4zAQjVasBGL5BXCwtOd0_ZH441iiwlYCgYDlavmzdRXsEKcH_v2apkLMZUZv5r0ZzauqSwQXCEHxZ9l1q-fnBYa4WZCGtYThH9UZRlTUpCX05Ft9Wl3kvIMleIFadlbFld040KW3YT-FuKmvVXYWvAbrEliaKaQInpxJmxgO9b2btskCFS1YTxksh6EPRh1aIYKH2IfowOP2Ixe0Byu7PzZfnDLbov-r-ulVn93FMZ9X_25WL93f-u7hdt0t72qDuZhq4TUWrbdUac6Q4C1BBBPWYqQbraF2FHPWWKYYJZpiyi0hxCAvvOUCekLOq_Wsa5PayWEMb2r8kEkFeQDSuJFqnILpncSNMKhRuvwDNcx6wR12yGiLvMHY0qL1e9YaxvS-d3mSu7QfYzlfEoQ4FIgjXKbIPGXGlPPo_NdWBOWnT3L2SX76JI8-FdbVzArOuW8MBikXnPwHMSWN9w</recordid><startdate>2024</startdate><enddate>2024</enddate><creator>Han, Jinzhu</creator><creator>Zhao, Jinjin</creator><creator>Yue, Yan</creator><creator>Che, Xinrui</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0009-0007-1106-5925</orcidid></search><sort><creationdate>2024</creationdate><title>Edge Computing-Based Video Action Recognition Method and Its Application in Online Physical Education Teaching</title><author>Han, Jinzhu ; Zhao, Jinjin ; Yue, Yan ; Che, Xinrui</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c289t-9fb295fd6ab871985313237521b4bb0be62874d7a763b6268d333c1f9fd890f33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Activity recognition</topic><topic>Artificial intelligence</topic><topic>Cloud computing</topic><topic>Computational complexity</topic><topic>Convolutional neural networks</topic><topic>cross-temporal token interaction</topic><topic>Data processing</topic><topic>Edge computing</topic><topic>Education</topic><topic>Electronic learning</topic><topic>Feature extraction</topic><topic>Image recognition</topic><topic>Lightweight</topic><topic>lightweight network</topic><topic>online PE teaching</topic><topic>Parameter sensitivity</topic><topic>Physical education</topic><topic>Servers</topic><topic>spatial-temporal pruning</topic><topic>Spatiotemporal phenomena</topic><topic>Three-dimensional displays</topic><topic>Transformers</topic><topic>Video action recognition</topic><topic>Weight reduction</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Han, Jinzhu</creatorcontrib><creatorcontrib>Zhao, Jinjin</creatorcontrib><creatorcontrib>Yue, Yan</creatorcontrib><creatorcontrib>Che, Xinrui</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005–Present</collection><collection>IEEE Xplore Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998–Present</collection><collection>IEL</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Han, Jinzhu</au><au>Zhao, Jinjin</au><au>Yue, Yan</au><au>Che, Xinrui</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Edge Computing-Based Video Action Recognition Method and Its Application in Online Physical Education Teaching</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2024</date><risdate>2024</risdate><volume>12</volume><spage>148666</spage><epage>148676</epage><pages>148666-148676</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>Due to the impact of COVID-19, online physical education (PE) teaching has garnered increasing attention. Given the characteristics of online PE teaching, introducing artificial intelligence technology to automatically detect or recognize students' actions or behaviors has gradually emerged as a trend. However, traditional cloud computing-based intelligent online PE teaching systems often face various challenging issues, such as computational complexity and latency. Edge computing can address these problems. However, edge devices typically have limited computing power, while existing deep action recognition models often contain a large number of parameters and require significant computational resources, making them difficult to deploy on edge devices. To address the above issues, this paper proposes a lightweight video recognition method, named the lightweight video ViT (LWV-ViT) network. More specifically, based on the standard ViT model, the video-based ViT (VBViT) network is first introduced by developing a cross-temporal token interaction module to effectively process temporal information in videos. Furthermore, the LWV-ViT network is proposed by implementing a spatial-temporal pruning scheme to reduce the number of parameters. Finally, the proposed LWV-ViT network is deployed in an edge computing-based online PE teaching system, where it is installed on each edge device. This setup enables fast data processing, reduces transmission latency, and protects sensitive data. Experimental results show that the proposed LWV-ViT network achieves the best recognition rates for both behavior detection (96.5%, 95.73%) and action recognition (97.9%, 88.3%, 79.9%) tasks, and has the fewest trainable parameters (2.7 M), which means it performs well in edge computing-based online PE teaching systems.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2024.3475372</doi><tpages>11</tpages><orcidid>https://orcid.org/0009-0007-1106-5925</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2169-3536 |
ispartof | IEEE access, 2024, Vol.12, p.148666-148676 |
issn | 2169-3536 2169-3536 |
language | eng |
recordid | cdi_proquest_journals_3118091812 |
source | DOAJ Directory of Open Access Journals; IEEE Xplore Open Access Journals; EZB Electronic Journals Library |
subjects | Activity recognition Artificial intelligence Cloud computing Computational complexity Convolutional neural networks cross-temporal token interaction Data processing Edge computing Education Electronic learning Feature extraction Image recognition Lightweight lightweight network online PE teaching Parameter sensitivity Physical education Servers spatial-temporal pruning Spatiotemporal phenomena Three-dimensional displays Transformers Video action recognition Weight reduction |
title | Edge Computing-Based Video Action Recognition Method and Its Application in Online Physical Education Teaching |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-01T08%3A23%3A53IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Edge%20Computing-Based%20Video%20Action%20Recognition%20Method%20and%20Its%20Application%20in%20Online%20Physical%20Education%20Teaching&rft.jtitle=IEEE%20access&rft.au=Han,%20Jinzhu&rft.date=2024&rft.volume=12&rft.spage=148666&rft.epage=148676&rft.pages=148666-148676&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2024.3475372&rft_dat=%3Cproquest_cross%3E3118091812%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3118091812&rft_id=info:pmid/&rft_ieee_id=10706898&rft_doaj_id=oai_doaj_org_article_249c14ab695147df98e2e1cbd1fc22d6&rfr_iscdi=true |