Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data

Deep learning methods have achieved a lot of success in various applications involving converting wearable sensor data to actionable health insights. A common application areas is activity recognition, where deep-learning methods still suffer from limitations such as sensitivity to signal quality, s...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2024-07
Hauptverfasser: Eun Som Jeon, Choi, Hongjun, Shukla, Ankita, Wang, Yuan, Lee, Hyunglae, Buman, Matthew P, Turaga, Pavan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Eun Som Jeon
Choi, Hongjun
Shukla, Ankita
Wang, Yuan
Lee, Hyunglae
Buman, Matthew P
Turaga, Pavan
description Deep learning methods have achieved a lot of success in various applications involving converting wearable sensor data to actionable health insights. A common application areas is activity recognition, where deep-learning methods still suffer from limitations such as sensitivity to signal quality, sensor characteristic variations, and variability between subjects. To mitigate these issues, robust features obtained by topological data analysis (TDA) have been suggested as a potential solution. However, there are two significant obstacles to using topological features in deep learning: (1) large computational load to extract topological features using TDA, and (2) different signal representations obtained from deep learning and TDA which makes fusion difficult. In this paper, to enable integration of the strengths of topological methods in deep-learning for time-series data, we propose to use two teacher networks, one trained on the raw time-series data, and another trained on persistence images generated by TDA methods. The distilled student model utilizes only the raw time-series data at test-time. This approach addresses both issues. The use of KD with multiple teachers utilizes complementary information, and results in a compact model with strong supervisory features and an integrated richer representation. To assimilate desirable information from different modalities, we design new constraints, including orthogonality imposed on feature correlation maps for improving feature expressiveness and allowing the student to easily learn from the teacher. Also, we apply an annealing strategy in KD for fast saturation and better accommodation from different features, while the knowledge gap between the teachers and student is reduced. Finally, a robust student model is distilled, which uses only the time-series data as an input, while implicitly preserving topological features.
doi_str_mv 10.48550/arxiv.2407.05315
format Article
fullrecord <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2407_05315</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3077525667</sourcerecordid><originalsourceid>FETCH-LOGICAL-a527-b40d7137ee36a87ec261079811e43a70c3d7b32bebfd1b2dd0f065f59e1f464a3</originalsourceid><addsrcrecordid>eNotj01LAzEYhIMgWGp_gCcDnrfmO9ujtFrFooIFj8u7m3dLStzUZOvHv3dtPQ3DDMM8hFxwNlWl1uwa0rf_nArF7JRpyfUJGQkpeVEqIc7IJOctY0wYK7SWI_K0jrsY4sY3EOgLpuxzj12DdLn3Dh197OJXQLdBuhgSHwL0Pna0jYm-ISSoA9JX7PLgF9DDOTltIWSc_OuYrO9u1_P7YvW8fJjfrArQwha1Ys5yaRGlgdJiIwxndlZyjkqCZY10tpaixrp1vBbOsZYZ3eoZ8lYZBXJMLo-zB9hql_w7pJ_qD7o6QA-Nq2Njl-LHHnNfbeM-dcOnSjJrtdDGWPkLuwNaGQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3077525667</pqid></control><display><type>article</type><title>Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Eun Som Jeon ; Choi, Hongjun ; Shukla, Ankita ; Wang, Yuan ; Lee, Hyunglae ; Buman, Matthew P ; Turaga, Pavan</creator><creatorcontrib>Eun Som Jeon ; Choi, Hongjun ; Shukla, Ankita ; Wang, Yuan ; Lee, Hyunglae ; Buman, Matthew P ; Turaga, Pavan</creatorcontrib><description>Deep learning methods have achieved a lot of success in various applications involving converting wearable sensor data to actionable health insights. A common application areas is activity recognition, where deep-learning methods still suffer from limitations such as sensitivity to signal quality, sensor characteristic variations, and variability between subjects. To mitigate these issues, robust features obtained by topological data analysis (TDA) have been suggested as a potential solution. However, there are two significant obstacles to using topological features in deep learning: (1) large computational load to extract topological features using TDA, and (2) different signal representations obtained from deep learning and TDA which makes fusion difficult. In this paper, to enable integration of the strengths of topological methods in deep-learning for time-series data, we propose to use two teacher networks, one trained on the raw time-series data, and another trained on persistence images generated by TDA methods. The distilled student model utilizes only the raw time-series data at test-time. This approach addresses both issues. The use of KD with multiple teachers utilizes complementary information, and results in a compact model with strong supervisory features and an integrated richer representation. To assimilate desirable information from different modalities, we design new constraints, including orthogonality imposed on feature correlation maps for improving feature expressiveness and allowing the student to easily learn from the teacher. Also, we apply an annealing strategy in KD for fast saturation and better accommodation from different features, while the knowledge gap between the teachers and student is reduced. Finally, a robust student model is distilled, which uses only the time-series data as an input, while implicitly preserving topological features.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2407.05315</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Activity recognition ; Computer Science - Learning ; Data analysis ; Deep learning ; Distillation ; Machine learning ; Mathematics - Algebraic Topology ; Orthogonality ; Representations ; Robustness (mathematics) ; Sensors ; Signal quality ; Teachers ; Testing time ; Time series ; Topology ; Wearable technology</subject><ispartof>arXiv.org, 2024-07</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,780,881,27902</link.rule.ids><backlink>$$Uhttps://doi.org/10.1016/j.engappai.2023.107719$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.48550/arXiv.2407.05315$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Eun Som Jeon</creatorcontrib><creatorcontrib>Choi, Hongjun</creatorcontrib><creatorcontrib>Shukla, Ankita</creatorcontrib><creatorcontrib>Wang, Yuan</creatorcontrib><creatorcontrib>Lee, Hyunglae</creatorcontrib><creatorcontrib>Buman, Matthew P</creatorcontrib><creatorcontrib>Turaga, Pavan</creatorcontrib><title>Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data</title><title>arXiv.org</title><description>Deep learning methods have achieved a lot of success in various applications involving converting wearable sensor data to actionable health insights. A common application areas is activity recognition, where deep-learning methods still suffer from limitations such as sensitivity to signal quality, sensor characteristic variations, and variability between subjects. To mitigate these issues, robust features obtained by topological data analysis (TDA) have been suggested as a potential solution. However, there are two significant obstacles to using topological features in deep learning: (1) large computational load to extract topological features using TDA, and (2) different signal representations obtained from deep learning and TDA which makes fusion difficult. In this paper, to enable integration of the strengths of topological methods in deep-learning for time-series data, we propose to use two teacher networks, one trained on the raw time-series data, and another trained on persistence images generated by TDA methods. The distilled student model utilizes only the raw time-series data at test-time. This approach addresses both issues. The use of KD with multiple teachers utilizes complementary information, and results in a compact model with strong supervisory features and an integrated richer representation. To assimilate desirable information from different modalities, we design new constraints, including orthogonality imposed on feature correlation maps for improving feature expressiveness and allowing the student to easily learn from the teacher. Also, we apply an annealing strategy in KD for fast saturation and better accommodation from different features, while the knowledge gap between the teachers and student is reduced. Finally, a robust student model is distilled, which uses only the time-series data as an input, while implicitly preserving topological features.</description><subject>Activity recognition</subject><subject>Computer Science - Learning</subject><subject>Data analysis</subject><subject>Deep learning</subject><subject>Distillation</subject><subject>Machine learning</subject><subject>Mathematics - Algebraic Topology</subject><subject>Orthogonality</subject><subject>Representations</subject><subject>Robustness (mathematics)</subject><subject>Sensors</subject><subject>Signal quality</subject><subject>Teachers</subject><subject>Testing time</subject><subject>Time series</subject><subject>Topology</subject><subject>Wearable technology</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><sourceid>GOX</sourceid><recordid>eNotj01LAzEYhIMgWGp_gCcDnrfmO9ujtFrFooIFj8u7m3dLStzUZOvHv3dtPQ3DDMM8hFxwNlWl1uwa0rf_nArF7JRpyfUJGQkpeVEqIc7IJOctY0wYK7SWI_K0jrsY4sY3EOgLpuxzj12DdLn3Dh197OJXQLdBuhgSHwL0Pna0jYm-ISSoA9JX7PLgF9DDOTltIWSc_OuYrO9u1_P7YvW8fJjfrArQwha1Ys5yaRGlgdJiIwxndlZyjkqCZY10tpaixrp1vBbOsZYZ3eoZ8lYZBXJMLo-zB9hql_w7pJ_qD7o6QA-Nq2Njl-LHHnNfbeM-dcOnSjJrtdDGWPkLuwNaGQ</recordid><startdate>20240707</startdate><enddate>20240707</enddate><creator>Eun Som Jeon</creator><creator>Choi, Hongjun</creator><creator>Shukla, Ankita</creator><creator>Wang, Yuan</creator><creator>Lee, Hyunglae</creator><creator>Buman, Matthew P</creator><creator>Turaga, Pavan</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>AKZ</scope><scope>GOX</scope></search><sort><creationdate>20240707</creationdate><title>Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data</title><author>Eun Som Jeon ; Choi, Hongjun ; Shukla, Ankita ; Wang, Yuan ; Lee, Hyunglae ; Buman, Matthew P ; Turaga, Pavan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a527-b40d7137ee36a87ec261079811e43a70c3d7b32bebfd1b2dd0f065f59e1f464a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Activity recognition</topic><topic>Computer Science - Learning</topic><topic>Data analysis</topic><topic>Deep learning</topic><topic>Distillation</topic><topic>Machine learning</topic><topic>Mathematics - Algebraic Topology</topic><topic>Orthogonality</topic><topic>Representations</topic><topic>Robustness (mathematics)</topic><topic>Sensors</topic><topic>Signal quality</topic><topic>Teachers</topic><topic>Testing time</topic><topic>Time series</topic><topic>Topology</topic><topic>Wearable technology</topic><toplevel>online_resources</toplevel><creatorcontrib>Eun Som Jeon</creatorcontrib><creatorcontrib>Choi, Hongjun</creatorcontrib><creatorcontrib>Shukla, Ankita</creatorcontrib><creatorcontrib>Wang, Yuan</creatorcontrib><creatorcontrib>Lee, Hyunglae</creatorcontrib><creatorcontrib>Buman, Matthew P</creatorcontrib><creatorcontrib>Turaga, Pavan</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv Mathematics</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Eun Som Jeon</au><au>Choi, Hongjun</au><au>Shukla, Ankita</au><au>Wang, Yuan</au><au>Lee, Hyunglae</au><au>Buman, Matthew P</au><au>Turaga, Pavan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data</atitle><jtitle>arXiv.org</jtitle><date>2024-07-07</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Deep learning methods have achieved a lot of success in various applications involving converting wearable sensor data to actionable health insights. A common application areas is activity recognition, where deep-learning methods still suffer from limitations such as sensitivity to signal quality, sensor characteristic variations, and variability between subjects. To mitigate these issues, robust features obtained by topological data analysis (TDA) have been suggested as a potential solution. However, there are two significant obstacles to using topological features in deep learning: (1) large computational load to extract topological features using TDA, and (2) different signal representations obtained from deep learning and TDA which makes fusion difficult. In this paper, to enable integration of the strengths of topological methods in deep-learning for time-series data, we propose to use two teacher networks, one trained on the raw time-series data, and another trained on persistence images generated by TDA methods. The distilled student model utilizes only the raw time-series data at test-time. This approach addresses both issues. The use of KD with multiple teachers utilizes complementary information, and results in a compact model with strong supervisory features and an integrated richer representation. To assimilate desirable information from different modalities, we design new constraints, including orthogonality imposed on feature correlation maps for improving feature expressiveness and allowing the student to easily learn from the teacher. Also, we apply an annealing strategy in KD for fast saturation and better accommodation from different features, while the knowledge gap between the teachers and student is reduced. Finally, a robust student model is distilled, which uses only the time-series data as an input, while implicitly preserving topological features.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2407.05315</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-07
issn 2331-8422
language eng
recordid cdi_arxiv_primary_2407_05315
source arXiv.org; Free E- Journals
subjects Activity recognition
Computer Science - Learning
Data analysis
Deep learning
Distillation
Machine learning
Mathematics - Algebraic Topology
Orthogonality
Representations
Robustness (mathematics)
Sensors
Signal quality
Teachers
Testing time
Time series
Topology
Wearable technology
title Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-31T05%3A04%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Topological%20Persistence%20Guided%20Knowledge%20Distillation%20for%20Wearable%20Sensor%20Data&rft.jtitle=arXiv.org&rft.au=Eun%20Som%20Jeon&rft.date=2024-07-07&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2407.05315&rft_dat=%3Cproquest_arxiv%3E3077525667%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3077525667&rft_id=info:pmid/&rfr_iscdi=true