Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data
Deep learning methods have achieved a lot of success in various applications involving converting wearable sensor data to actionable health insights. A common application areas is activity recognition, where deep-learning methods still suffer from limitations such as sensitivity to signal quality, s...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2024-07 |
---|---|
Hauptverfasser: | , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Eun Som Jeon Choi, Hongjun Shukla, Ankita Wang, Yuan Lee, Hyunglae Buman, Matthew P Turaga, Pavan |
description | Deep learning methods have achieved a lot of success in various applications involving converting wearable sensor data to actionable health insights. A common application areas is activity recognition, where deep-learning methods still suffer from limitations such as sensitivity to signal quality, sensor characteristic variations, and variability between subjects. To mitigate these issues, robust features obtained by topological data analysis (TDA) have been suggested as a potential solution. However, there are two significant obstacles to using topological features in deep learning: (1) large computational load to extract topological features using TDA, and (2) different signal representations obtained from deep learning and TDA which makes fusion difficult. In this paper, to enable integration of the strengths of topological methods in deep-learning for time-series data, we propose to use two teacher networks, one trained on the raw time-series data, and another trained on persistence images generated by TDA methods. The distilled student model utilizes only the raw time-series data at test-time. This approach addresses both issues. The use of KD with multiple teachers utilizes complementary information, and results in a compact model with strong supervisory features and an integrated richer representation. To assimilate desirable information from different modalities, we design new constraints, including orthogonality imposed on feature correlation maps for improving feature expressiveness and allowing the student to easily learn from the teacher. Also, we apply an annealing strategy in KD for fast saturation and better accommodation from different features, while the knowledge gap between the teachers and student is reduced. Finally, a robust student model is distilled, which uses only the time-series data as an input, while implicitly preserving topological features. |
doi_str_mv | 10.48550/arxiv.2407.05315 |
format | Article |
fullrecord | <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2407_05315</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3077525667</sourcerecordid><originalsourceid>FETCH-LOGICAL-a527-b40d7137ee36a87ec261079811e43a70c3d7b32bebfd1b2dd0f065f59e1f464a3</originalsourceid><addsrcrecordid>eNotj01LAzEYhIMgWGp_gCcDnrfmO9ujtFrFooIFj8u7m3dLStzUZOvHv3dtPQ3DDMM8hFxwNlWl1uwa0rf_nArF7JRpyfUJGQkpeVEqIc7IJOctY0wYK7SWI_K0jrsY4sY3EOgLpuxzj12DdLn3Dh197OJXQLdBuhgSHwL0Pna0jYm-ISSoA9JX7PLgF9DDOTltIWSc_OuYrO9u1_P7YvW8fJjfrArQwha1Ys5yaRGlgdJiIwxndlZyjkqCZY10tpaixrp1vBbOsZYZ3eoZ8lYZBXJMLo-zB9hql_w7pJ_qD7o6QA-Nq2Njl-LHHnNfbeM-dcOnSjJrtdDGWPkLuwNaGQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3077525667</pqid></control><display><type>article</type><title>Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Eun Som Jeon ; Choi, Hongjun ; Shukla, Ankita ; Wang, Yuan ; Lee, Hyunglae ; Buman, Matthew P ; Turaga, Pavan</creator><creatorcontrib>Eun Som Jeon ; Choi, Hongjun ; Shukla, Ankita ; Wang, Yuan ; Lee, Hyunglae ; Buman, Matthew P ; Turaga, Pavan</creatorcontrib><description>Deep learning methods have achieved a lot of success in various applications involving converting wearable sensor data to actionable health insights. A common application areas is activity recognition, where deep-learning methods still suffer from limitations such as sensitivity to signal quality, sensor characteristic variations, and variability between subjects. To mitigate these issues, robust features obtained by topological data analysis (TDA) have been suggested as a potential solution. However, there are two significant obstacles to using topological features in deep learning: (1) large computational load to extract topological features using TDA, and (2) different signal representations obtained from deep learning and TDA which makes fusion difficult. In this paper, to enable integration of the strengths of topological methods in deep-learning for time-series data, we propose to use two teacher networks, one trained on the raw time-series data, and another trained on persistence images generated by TDA methods. The distilled student model utilizes only the raw time-series data at test-time. This approach addresses both issues. The use of KD with multiple teachers utilizes complementary information, and results in a compact model with strong supervisory features and an integrated richer representation. To assimilate desirable information from different modalities, we design new constraints, including orthogonality imposed on feature correlation maps for improving feature expressiveness and allowing the student to easily learn from the teacher. Also, we apply an annealing strategy in KD for fast saturation and better accommodation from different features, while the knowledge gap between the teachers and student is reduced. Finally, a robust student model is distilled, which uses only the time-series data as an input, while implicitly preserving topological features.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2407.05315</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Activity recognition ; Computer Science - Learning ; Data analysis ; Deep learning ; Distillation ; Machine learning ; Mathematics - Algebraic Topology ; Orthogonality ; Representations ; Robustness (mathematics) ; Sensors ; Signal quality ; Teachers ; Testing time ; Time series ; Topology ; Wearable technology</subject><ispartof>arXiv.org, 2024-07</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,780,881,27902</link.rule.ids><backlink>$$Uhttps://doi.org/10.1016/j.engappai.2023.107719$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.48550/arXiv.2407.05315$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Eun Som Jeon</creatorcontrib><creatorcontrib>Choi, Hongjun</creatorcontrib><creatorcontrib>Shukla, Ankita</creatorcontrib><creatorcontrib>Wang, Yuan</creatorcontrib><creatorcontrib>Lee, Hyunglae</creatorcontrib><creatorcontrib>Buman, Matthew P</creatorcontrib><creatorcontrib>Turaga, Pavan</creatorcontrib><title>Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data</title><title>arXiv.org</title><description>Deep learning methods have achieved a lot of success in various applications involving converting wearable sensor data to actionable health insights. A common application areas is activity recognition, where deep-learning methods still suffer from limitations such as sensitivity to signal quality, sensor characteristic variations, and variability between subjects. To mitigate these issues, robust features obtained by topological data analysis (TDA) have been suggested as a potential solution. However, there are two significant obstacles to using topological features in deep learning: (1) large computational load to extract topological features using TDA, and (2) different signal representations obtained from deep learning and TDA which makes fusion difficult. In this paper, to enable integration of the strengths of topological methods in deep-learning for time-series data, we propose to use two teacher networks, one trained on the raw time-series data, and another trained on persistence images generated by TDA methods. The distilled student model utilizes only the raw time-series data at test-time. This approach addresses both issues. The use of KD with multiple teachers utilizes complementary information, and results in a compact model with strong supervisory features and an integrated richer representation. To assimilate desirable information from different modalities, we design new constraints, including orthogonality imposed on feature correlation maps for improving feature expressiveness and allowing the student to easily learn from the teacher. Also, we apply an annealing strategy in KD for fast saturation and better accommodation from different features, while the knowledge gap between the teachers and student is reduced. Finally, a robust student model is distilled, which uses only the time-series data as an input, while implicitly preserving topological features.</description><subject>Activity recognition</subject><subject>Computer Science - Learning</subject><subject>Data analysis</subject><subject>Deep learning</subject><subject>Distillation</subject><subject>Machine learning</subject><subject>Mathematics - Algebraic Topology</subject><subject>Orthogonality</subject><subject>Representations</subject><subject>Robustness (mathematics)</subject><subject>Sensors</subject><subject>Signal quality</subject><subject>Teachers</subject><subject>Testing time</subject><subject>Time series</subject><subject>Topology</subject><subject>Wearable technology</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><sourceid>GOX</sourceid><recordid>eNotj01LAzEYhIMgWGp_gCcDnrfmO9ujtFrFooIFj8u7m3dLStzUZOvHv3dtPQ3DDMM8hFxwNlWl1uwa0rf_nArF7JRpyfUJGQkpeVEqIc7IJOctY0wYK7SWI_K0jrsY4sY3EOgLpuxzj12DdLn3Dh197OJXQLdBuhgSHwL0Pna0jYm-ISSoA9JX7PLgF9DDOTltIWSc_OuYrO9u1_P7YvW8fJjfrArQwha1Ys5yaRGlgdJiIwxndlZyjkqCZY10tpaixrp1vBbOsZYZ3eoZ8lYZBXJMLo-zB9hql_w7pJ_qD7o6QA-Nq2Njl-LHHnNfbeM-dcOnSjJrtdDGWPkLuwNaGQ</recordid><startdate>20240707</startdate><enddate>20240707</enddate><creator>Eun Som Jeon</creator><creator>Choi, Hongjun</creator><creator>Shukla, Ankita</creator><creator>Wang, Yuan</creator><creator>Lee, Hyunglae</creator><creator>Buman, Matthew P</creator><creator>Turaga, Pavan</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>AKZ</scope><scope>GOX</scope></search><sort><creationdate>20240707</creationdate><title>Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data</title><author>Eun Som Jeon ; Choi, Hongjun ; Shukla, Ankita ; Wang, Yuan ; Lee, Hyunglae ; Buman, Matthew P ; Turaga, Pavan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a527-b40d7137ee36a87ec261079811e43a70c3d7b32bebfd1b2dd0f065f59e1f464a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Activity recognition</topic><topic>Computer Science - Learning</topic><topic>Data analysis</topic><topic>Deep learning</topic><topic>Distillation</topic><topic>Machine learning</topic><topic>Mathematics - Algebraic Topology</topic><topic>Orthogonality</topic><topic>Representations</topic><topic>Robustness (mathematics)</topic><topic>Sensors</topic><topic>Signal quality</topic><topic>Teachers</topic><topic>Testing time</topic><topic>Time series</topic><topic>Topology</topic><topic>Wearable technology</topic><toplevel>online_resources</toplevel><creatorcontrib>Eun Som Jeon</creatorcontrib><creatorcontrib>Choi, Hongjun</creatorcontrib><creatorcontrib>Shukla, Ankita</creatorcontrib><creatorcontrib>Wang, Yuan</creatorcontrib><creatorcontrib>Lee, Hyunglae</creatorcontrib><creatorcontrib>Buman, Matthew P</creatorcontrib><creatorcontrib>Turaga, Pavan</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv Mathematics</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Eun Som Jeon</au><au>Choi, Hongjun</au><au>Shukla, Ankita</au><au>Wang, Yuan</au><au>Lee, Hyunglae</au><au>Buman, Matthew P</au><au>Turaga, Pavan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data</atitle><jtitle>arXiv.org</jtitle><date>2024-07-07</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Deep learning methods have achieved a lot of success in various applications involving converting wearable sensor data to actionable health insights. A common application areas is activity recognition, where deep-learning methods still suffer from limitations such as sensitivity to signal quality, sensor characteristic variations, and variability between subjects. To mitigate these issues, robust features obtained by topological data analysis (TDA) have been suggested as a potential solution. However, there are two significant obstacles to using topological features in deep learning: (1) large computational load to extract topological features using TDA, and (2) different signal representations obtained from deep learning and TDA which makes fusion difficult. In this paper, to enable integration of the strengths of topological methods in deep-learning for time-series data, we propose to use two teacher networks, one trained on the raw time-series data, and another trained on persistence images generated by TDA methods. The distilled student model utilizes only the raw time-series data at test-time. This approach addresses both issues. The use of KD with multiple teachers utilizes complementary information, and results in a compact model with strong supervisory features and an integrated richer representation. To assimilate desirable information from different modalities, we design new constraints, including orthogonality imposed on feature correlation maps for improving feature expressiveness and allowing the student to easily learn from the teacher. Also, we apply an annealing strategy in KD for fast saturation and better accommodation from different features, while the knowledge gap between the teachers and student is reduced. Finally, a robust student model is distilled, which uses only the time-series data as an input, while implicitly preserving topological features.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2407.05315</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2024-07 |
issn | 2331-8422 |
language | eng |
recordid | cdi_arxiv_primary_2407_05315 |
source | arXiv.org; Free E- Journals |
subjects | Activity recognition Computer Science - Learning Data analysis Deep learning Distillation Machine learning Mathematics - Algebraic Topology Orthogonality Representations Robustness (mathematics) Sensors Signal quality Teachers Testing time Time series Topology Wearable technology |
title | Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-31T05%3A04%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Topological%20Persistence%20Guided%20Knowledge%20Distillation%20for%20Wearable%20Sensor%20Data&rft.jtitle=arXiv.org&rft.au=Eun%20Som%20Jeon&rft.date=2024-07-07&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2407.05315&rft_dat=%3Cproquest_arxiv%3E3077525667%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3077525667&rft_id=info:pmid/&rfr_iscdi=true |