Ultrasensitive Textile Strain Sensors Redefine Wearable Silent Speech Interfaces with High Machine Learning Efficiency

Our research presents a wearable Silent Speech Interface (SSI) technology that excels in device comfort, time-energy efficiency, and speech decoding accuracy for real-world use. We developed a biocompatible, durable textile choker with an embedded graphene-based strain sensor, capable of accurately...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2023-12
Hauptverfasser: Tang, Chenyu, Xu, Muzi, Wentian Yi, Zhang, Zibo, Occhipinti, Edoardo, Dong, Chaoqun, Ravenscroft, Dafydd, Sung-Min, Jung, Lee, Sanghyo, Gao, Shuo, Kim, Jong Min, Occhipinti, Luigi G
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Tang, Chenyu
Xu, Muzi
Wentian Yi
Zhang, Zibo
Occhipinti, Edoardo
Dong, Chaoqun
Ravenscroft, Dafydd
Sung-Min, Jung
Lee, Sanghyo
Gao, Shuo
Kim, Jong Min
Occhipinti, Luigi G
description Our research presents a wearable Silent Speech Interface (SSI) technology that excels in device comfort, time-energy efficiency, and speech decoding accuracy for real-world use. We developed a biocompatible, durable textile choker with an embedded graphene-based strain sensor, capable of accurately detecting subtle throat movements. This sensor, surpassing other strain sensors in sensitivity by 420%, simplifies signal processing compared to traditional voice recognition methods. Our system uses a computationally efficient neural network, specifically a one-dimensional convolutional neural network with residual structures, to decode speech signals. This network is energy and time-efficient, reducing computational load by 90% while achieving 95.25% accuracy for a 20-word lexicon and swiftly adapting to new users and words with minimal samples. This innovation demonstrates a practical, sensitive, and precise wearable SSI suitable for daily communication applications.
doi_str_mv 10.48550/arxiv.2311.15683
format Article
fullrecord <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2311_15683</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2894592102</sourcerecordid><originalsourceid>FETCH-LOGICAL-a953-7397adbe394db055df77428cf24d2aca49ef8d5600f5475329f2f6b378ae31f13</originalsourceid><addsrcrecordid>eNotkE1rAjEQhkOhULH-gJ4a6HltPnc3xyK2CpZCtfS4ZLMTN2KjTaLVf--qPQ2887zD8CD0QMlQlFKSZx0Obj9knNIhlXnJb1CPcU6zUjB2hwYxrgghLC-YlLyH9l_rFHQEH11ye8ALOCS3BjzvUufxvFtsQsSf0IB1HvA36KDrM9BRPuH5FsC0eOoTBKsNRPznUosnbtnid23ac2fWdbzzSzy21hkH3hzv0a3V6wiD_9lHi9fxYjTJZh9v09HLLNNK8qzgqtBNDVyJpiZSNrYoBCuNZaJh2mihwJaNzAmxUhSSM2WZzWtelBo4tZT30eP17EVKtQ3uR4djdZZTXeR0xNOV2IbN7w5iqlabXfDdTxUrlZCKUcL4CZEqaNg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2894592102</pqid></control><display><type>article</type><title>Ultrasensitive Textile Strain Sensors Redefine Wearable Silent Speech Interfaces with High Machine Learning Efficiency</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Tang, Chenyu ; Xu, Muzi ; Wentian Yi ; Zhang, Zibo ; Occhipinti, Edoardo ; Dong, Chaoqun ; Ravenscroft, Dafydd ; Sung-Min, Jung ; Lee, Sanghyo ; Gao, Shuo ; Kim, Jong Min ; Occhipinti, Luigi G</creator><creatorcontrib>Tang, Chenyu ; Xu, Muzi ; Wentian Yi ; Zhang, Zibo ; Occhipinti, Edoardo ; Dong, Chaoqun ; Ravenscroft, Dafydd ; Sung-Min, Jung ; Lee, Sanghyo ; Gao, Shuo ; Kim, Jong Min ; Occhipinti, Luigi G</creatorcontrib><description>Our research presents a wearable Silent Speech Interface (SSI) technology that excels in device comfort, time-energy efficiency, and speech decoding accuracy for real-world use. We developed a biocompatible, durable textile choker with an embedded graphene-based strain sensor, capable of accurately detecting subtle throat movements. This sensor, surpassing other strain sensors in sensitivity by 420%, simplifies signal processing compared to traditional voice recognition methods. Our system uses a computationally efficient neural network, specifically a one-dimensional convolutional neural network with residual structures, to decode speech signals. This network is energy and time-efficient, reducing computational load by 90% while achieving 95.25% accuracy for a 20-word lexicon and swiftly adapting to new users and words with minimal samples. This innovation demonstrates a practical, sensitive, and precise wearable SSI suitable for daily communication applications.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2311.15683</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Artificial neural networks ; Biocompatibility ; Computational efficiency ; Computer Science - Sound ; Decoding ; Graphene ; Machine learning ; Sensors ; Signal processing ; Speech ; Voice communication ; Voice recognition ; Wearable technology</subject><ispartof>arXiv.org, 2023-12</ispartof><rights>2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,780,881,27902</link.rule.ids><backlink>$$Uhttps://doi.org/10.48550/arXiv.2311.15683$$DView paper in arXiv$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.1038/s41528-024-00315-1$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink></links><search><creatorcontrib>Tang, Chenyu</creatorcontrib><creatorcontrib>Xu, Muzi</creatorcontrib><creatorcontrib>Wentian Yi</creatorcontrib><creatorcontrib>Zhang, Zibo</creatorcontrib><creatorcontrib>Occhipinti, Edoardo</creatorcontrib><creatorcontrib>Dong, Chaoqun</creatorcontrib><creatorcontrib>Ravenscroft, Dafydd</creatorcontrib><creatorcontrib>Sung-Min, Jung</creatorcontrib><creatorcontrib>Lee, Sanghyo</creatorcontrib><creatorcontrib>Gao, Shuo</creatorcontrib><creatorcontrib>Kim, Jong Min</creatorcontrib><creatorcontrib>Occhipinti, Luigi G</creatorcontrib><title>Ultrasensitive Textile Strain Sensors Redefine Wearable Silent Speech Interfaces with High Machine Learning Efficiency</title><title>arXiv.org</title><description>Our research presents a wearable Silent Speech Interface (SSI) technology that excels in device comfort, time-energy efficiency, and speech decoding accuracy for real-world use. We developed a biocompatible, durable textile choker with an embedded graphene-based strain sensor, capable of accurately detecting subtle throat movements. This sensor, surpassing other strain sensors in sensitivity by 420%, simplifies signal processing compared to traditional voice recognition methods. Our system uses a computationally efficient neural network, specifically a one-dimensional convolutional neural network with residual structures, to decode speech signals. This network is energy and time-efficient, reducing computational load by 90% while achieving 95.25% accuracy for a 20-word lexicon and swiftly adapting to new users and words with minimal samples. This innovation demonstrates a practical, sensitive, and precise wearable SSI suitable for daily communication applications.</description><subject>Artificial neural networks</subject><subject>Biocompatibility</subject><subject>Computational efficiency</subject><subject>Computer Science - Sound</subject><subject>Decoding</subject><subject>Graphene</subject><subject>Machine learning</subject><subject>Sensors</subject><subject>Signal processing</subject><subject>Speech</subject><subject>Voice communication</subject><subject>Voice recognition</subject><subject>Wearable technology</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><sourceid>GOX</sourceid><recordid>eNotkE1rAjEQhkOhULH-gJ4a6HltPnc3xyK2CpZCtfS4ZLMTN2KjTaLVf--qPQ2887zD8CD0QMlQlFKSZx0Obj9knNIhlXnJb1CPcU6zUjB2hwYxrgghLC-YlLyH9l_rFHQEH11ye8ALOCS3BjzvUufxvFtsQsSf0IB1HvA36KDrM9BRPuH5FsC0eOoTBKsNRPznUosnbtnid23ac2fWdbzzSzy21hkH3hzv0a3V6wiD_9lHi9fxYjTJZh9v09HLLNNK8qzgqtBNDVyJpiZSNrYoBCuNZaJh2mihwJaNzAmxUhSSM2WZzWtelBo4tZT30eP17EVKtQ3uR4djdZZTXeR0xNOV2IbN7w5iqlabXfDdTxUrlZCKUcL4CZEqaNg</recordid><startdate>20231207</startdate><enddate>20231207</enddate><creator>Tang, Chenyu</creator><creator>Xu, Muzi</creator><creator>Wentian Yi</creator><creator>Zhang, Zibo</creator><creator>Occhipinti, Edoardo</creator><creator>Dong, Chaoqun</creator><creator>Ravenscroft, Dafydd</creator><creator>Sung-Min, Jung</creator><creator>Lee, Sanghyo</creator><creator>Gao, Shuo</creator><creator>Kim, Jong Min</creator><creator>Occhipinti, Luigi G</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231207</creationdate><title>Ultrasensitive Textile Strain Sensors Redefine Wearable Silent Speech Interfaces with High Machine Learning Efficiency</title><author>Tang, Chenyu ; Xu, Muzi ; Wentian Yi ; Zhang, Zibo ; Occhipinti, Edoardo ; Dong, Chaoqun ; Ravenscroft, Dafydd ; Sung-Min, Jung ; Lee, Sanghyo ; Gao, Shuo ; Kim, Jong Min ; Occhipinti, Luigi G</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a953-7397adbe394db055df77428cf24d2aca49ef8d5600f5475329f2f6b378ae31f13</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Artificial neural networks</topic><topic>Biocompatibility</topic><topic>Computational efficiency</topic><topic>Computer Science - Sound</topic><topic>Decoding</topic><topic>Graphene</topic><topic>Machine learning</topic><topic>Sensors</topic><topic>Signal processing</topic><topic>Speech</topic><topic>Voice communication</topic><topic>Voice recognition</topic><topic>Wearable technology</topic><toplevel>online_resources</toplevel><creatorcontrib>Tang, Chenyu</creatorcontrib><creatorcontrib>Xu, Muzi</creatorcontrib><creatorcontrib>Wentian Yi</creatorcontrib><creatorcontrib>Zhang, Zibo</creatorcontrib><creatorcontrib>Occhipinti, Edoardo</creatorcontrib><creatorcontrib>Dong, Chaoqun</creatorcontrib><creatorcontrib>Ravenscroft, Dafydd</creatorcontrib><creatorcontrib>Sung-Min, Jung</creatorcontrib><creatorcontrib>Lee, Sanghyo</creatorcontrib><creatorcontrib>Gao, Shuo</creatorcontrib><creatorcontrib>Kim, Jong Min</creatorcontrib><creatorcontrib>Occhipinti, Luigi G</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Tang, Chenyu</au><au>Xu, Muzi</au><au>Wentian Yi</au><au>Zhang, Zibo</au><au>Occhipinti, Edoardo</au><au>Dong, Chaoqun</au><au>Ravenscroft, Dafydd</au><au>Sung-Min, Jung</au><au>Lee, Sanghyo</au><au>Gao, Shuo</au><au>Kim, Jong Min</au><au>Occhipinti, Luigi G</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Ultrasensitive Textile Strain Sensors Redefine Wearable Silent Speech Interfaces with High Machine Learning Efficiency</atitle><jtitle>arXiv.org</jtitle><date>2023-12-07</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>Our research presents a wearable Silent Speech Interface (SSI) technology that excels in device comfort, time-energy efficiency, and speech decoding accuracy for real-world use. We developed a biocompatible, durable textile choker with an embedded graphene-based strain sensor, capable of accurately detecting subtle throat movements. This sensor, surpassing other strain sensors in sensitivity by 420%, simplifies signal processing compared to traditional voice recognition methods. Our system uses a computationally efficient neural network, specifically a one-dimensional convolutional neural network with residual structures, to decode speech signals. This network is energy and time-efficient, reducing computational load by 90% while achieving 95.25% accuracy for a 20-word lexicon and swiftly adapting to new users and words with minimal samples. This innovation demonstrates a practical, sensitive, and precise wearable SSI suitable for daily communication applications.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2311.15683</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2023-12
issn 2331-8422
language eng
recordid cdi_arxiv_primary_2311_15683
source arXiv.org; Free E- Journals
subjects Artificial neural networks
Biocompatibility
Computational efficiency
Computer Science - Sound
Decoding
Graphene
Machine learning
Sensors
Signal processing
Speech
Voice communication
Voice recognition
Wearable technology
title Ultrasensitive Textile Strain Sensors Redefine Wearable Silent Speech Interfaces with High Machine Learning Efficiency
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-14T03%3A11%3A42IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Ultrasensitive%20Textile%20Strain%20Sensors%20Redefine%20Wearable%20Silent%20Speech%20Interfaces%20with%20High%20Machine%20Learning%20Efficiency&rft.jtitle=arXiv.org&rft.au=Tang,%20Chenyu&rft.date=2023-12-07&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2311.15683&rft_dat=%3Cproquest_arxiv%3E2894592102%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2894592102&rft_id=info:pmid/&rfr_iscdi=true