Evaluation of text-to-gesture generation model using convolutional neural network
Conversational gestures have a crucial role in realizing natural interactions with virtual agents and robots. Data-driven approaches, such as deep learning and machine learning, are promising in constructing the gesture generation model, which automatically provides the gesture motion for speech or...
Gespeichert in:
Veröffentlicht in: | Neural networks 2022-07, Vol.151, p.365-375 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 375 |
---|---|
container_issue | |
container_start_page | 365 |
container_title | Neural networks |
container_volume | 151 |
creator | Asakawa, Eiichi Kaneko, Naoshi Hasegawa, Dai Shirakawa, Shinichi |
description | Conversational gestures have a crucial role in realizing natural interactions with virtual agents and robots. Data-driven approaches, such as deep learning and machine learning, are promising in constructing the gesture generation model, which automatically provides the gesture motion for speech or spoken texts. This study experimentally analyzes a deep learning-based gesture generation model from spoken text using a convolutional neural network. The proposed model takes a sequence of spoken words as the input and outputs a sequence of 2D joint coordinates representing the conversational gesture motion. We prepare a dataset consisting of gesture motions and spoken texts by adding text information to an existing dataset and train the models using specific speaker’s data. The quality of the generated gestures is compared with those from an existing speech-to-gesture generation model through a user perceptual study. The subjective evaluation shows that the model performance is comparable or superior to those by the existing speech-to-gesture generation model. In addition, we investigate the importance of data cleansing and loss function selection in the text-to-gesture generation model. We further examine the model transferability between speakers. The experimental results demonstrate successful model transferability of the proposed model. Finally, we show that the text-to-gesture generation model can produce good quality gestures even when using a transformer architecture.
•The quality of text-to-gesture generation models is evaluated by human perceptual studies.•The quality of text-to-gesture generation models is comparable to speech-to-gesture models.•Data cleansing and loss function selection are important in text-to-gesture generation models.•The possibility of model transfer between speakers is demonstrated. |
doi_str_mv | 10.1016/j.neunet.2022.03.041 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2656200768</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0893608022001198</els_id><sourcerecordid>2656200768</sourcerecordid><originalsourceid>FETCH-LOGICAL-c474t-38d44ccb169ce5f940a5d9eb52979bee51aa92e86a3fc875f199e5ec5849dae33</originalsourceid><addsrcrecordid>eNp9kFtLxDAQhYMoul7-gUgffWmd3NrkRRDxBoII-hyy6XTp2m00SVf993at-ujTgZkzc2Y-Qo4pFBRoebYsehx6TAUDxgrgBQi6RWZUVTpnlWLbZAZK87wEBXtkP8YlAJRK8F2yx6WoWMVhRh6v1rYbbGp9n_kmS_iR8uTzBcY0BMwW2GOYuitfY5cNse0XmfP92nfDpm67bLwjfEt69-HlkOw0tot49KMH5Pn66unyNr9_uLm7vLjPnahEyrmqhXBuTkvtUDZagJW1xrlkutJzREmt1QxVaXnjVCUbqjVKdFIJXVvk_ICcTntfg38bxnvNqo0Ou8726IdoWClLBlCVarSKyeqCjzFgY15Du7Lh01AwG5hmaSaYZgPTADcjzHHs5CdhmK-w_hv6pTcazicDjn-uWwwmuhZ7h3Ub0CVT-_b_hC8TpomP</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2656200768</pqid></control><display><type>article</type><title>Evaluation of text-to-gesture generation model using convolutional neural network</title><source>Elsevier ScienceDirect Journals Complete</source><creator>Asakawa, Eiichi ; Kaneko, Naoshi ; Hasegawa, Dai ; Shirakawa, Shinichi</creator><creatorcontrib>Asakawa, Eiichi ; Kaneko, Naoshi ; Hasegawa, Dai ; Shirakawa, Shinichi</creatorcontrib><description>Conversational gestures have a crucial role in realizing natural interactions with virtual agents and robots. Data-driven approaches, such as deep learning and machine learning, are promising in constructing the gesture generation model, which automatically provides the gesture motion for speech or spoken texts. This study experimentally analyzes a deep learning-based gesture generation model from spoken text using a convolutional neural network. The proposed model takes a sequence of spoken words as the input and outputs a sequence of 2D joint coordinates representing the conversational gesture motion. We prepare a dataset consisting of gesture motions and spoken texts by adding text information to an existing dataset and train the models using specific speaker’s data. The quality of the generated gestures is compared with those from an existing speech-to-gesture generation model through a user perceptual study. The subjective evaluation shows that the model performance is comparable or superior to those by the existing speech-to-gesture generation model. In addition, we investigate the importance of data cleansing and loss function selection in the text-to-gesture generation model. We further examine the model transferability between speakers. The experimental results demonstrate successful model transferability of the proposed model. Finally, we show that the text-to-gesture generation model can produce good quality gestures even when using a transformer architecture.
•The quality of text-to-gesture generation models is evaluated by human perceptual studies.•The quality of text-to-gesture generation models is comparable to speech-to-gesture models.•Data cleansing and loss function selection are important in text-to-gesture generation models.•The possibility of model transfer between speakers is demonstrated.</description><identifier>ISSN: 0893-6080</identifier><identifier>EISSN: 1879-2782</identifier><identifier>DOI: 10.1016/j.neunet.2022.03.041</identifier><identifier>PMID: 35472730</identifier><language>eng</language><publisher>United States: Elsevier Ltd</publisher><subject>Convolutional neural network ; Deep learning ; Gesture generation ; Spoken text ; Transformer architecture</subject><ispartof>Neural networks, 2022-07, Vol.151, p.365-375</ispartof><rights>2022 The Author(s)</rights><rights>Copyright © 2022 The Author(s). Published by Elsevier Ltd.. All rights reserved.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c474t-38d44ccb169ce5f940a5d9eb52979bee51aa92e86a3fc875f199e5ec5849dae33</citedby><cites>FETCH-LOGICAL-c474t-38d44ccb169ce5f940a5d9eb52979bee51aa92e86a3fc875f199e5ec5849dae33</cites><orcidid>0000-0002-5638-2509 ; 0000-0002-4659-6108</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S0893608022001198$$EHTML$$P50$$Gelsevier$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,3537,27901,27902,65534</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/35472730$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Asakawa, Eiichi</creatorcontrib><creatorcontrib>Kaneko, Naoshi</creatorcontrib><creatorcontrib>Hasegawa, Dai</creatorcontrib><creatorcontrib>Shirakawa, Shinichi</creatorcontrib><title>Evaluation of text-to-gesture generation model using convolutional neural network</title><title>Neural networks</title><addtitle>Neural Netw</addtitle><description>Conversational gestures have a crucial role in realizing natural interactions with virtual agents and robots. Data-driven approaches, such as deep learning and machine learning, are promising in constructing the gesture generation model, which automatically provides the gesture motion for speech or spoken texts. This study experimentally analyzes a deep learning-based gesture generation model from spoken text using a convolutional neural network. The proposed model takes a sequence of spoken words as the input and outputs a sequence of 2D joint coordinates representing the conversational gesture motion. We prepare a dataset consisting of gesture motions and spoken texts by adding text information to an existing dataset and train the models using specific speaker’s data. The quality of the generated gestures is compared with those from an existing speech-to-gesture generation model through a user perceptual study. The subjective evaluation shows that the model performance is comparable or superior to those by the existing speech-to-gesture generation model. In addition, we investigate the importance of data cleansing and loss function selection in the text-to-gesture generation model. We further examine the model transferability between speakers. The experimental results demonstrate successful model transferability of the proposed model. Finally, we show that the text-to-gesture generation model can produce good quality gestures even when using a transformer architecture.
•The quality of text-to-gesture generation models is evaluated by human perceptual studies.•The quality of text-to-gesture generation models is comparable to speech-to-gesture models.•Data cleansing and loss function selection are important in text-to-gesture generation models.•The possibility of model transfer between speakers is demonstrated.</description><subject>Convolutional neural network</subject><subject>Deep learning</subject><subject>Gesture generation</subject><subject>Spoken text</subject><subject>Transformer architecture</subject><issn>0893-6080</issn><issn>1879-2782</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNp9kFtLxDAQhYMoul7-gUgffWmd3NrkRRDxBoII-hyy6XTp2m00SVf993at-ujTgZkzc2Y-Qo4pFBRoebYsehx6TAUDxgrgBQi6RWZUVTpnlWLbZAZK87wEBXtkP8YlAJRK8F2yx6WoWMVhRh6v1rYbbGp9n_kmS_iR8uTzBcY0BMwW2GOYuitfY5cNse0XmfP92nfDpm67bLwjfEt69-HlkOw0tot49KMH5Pn66unyNr9_uLm7vLjPnahEyrmqhXBuTkvtUDZagJW1xrlkutJzREmt1QxVaXnjVCUbqjVKdFIJXVvk_ICcTntfg38bxnvNqo0Ou8726IdoWClLBlCVarSKyeqCjzFgY15Du7Lh01AwG5hmaSaYZgPTADcjzHHs5CdhmK-w_hv6pTcazicDjn-uWwwmuhZ7h3Ub0CVT-_b_hC8TpomP</recordid><startdate>202207</startdate><enddate>202207</enddate><creator>Asakawa, Eiichi</creator><creator>Kaneko, Naoshi</creator><creator>Hasegawa, Dai</creator><creator>Shirakawa, Shinichi</creator><general>Elsevier Ltd</general><scope>6I.</scope><scope>AAFTH</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-5638-2509</orcidid><orcidid>https://orcid.org/0000-0002-4659-6108</orcidid></search><sort><creationdate>202207</creationdate><title>Evaluation of text-to-gesture generation model using convolutional neural network</title><author>Asakawa, Eiichi ; Kaneko, Naoshi ; Hasegawa, Dai ; Shirakawa, Shinichi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c474t-38d44ccb169ce5f940a5d9eb52979bee51aa92e86a3fc875f199e5ec5849dae33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Convolutional neural network</topic><topic>Deep learning</topic><topic>Gesture generation</topic><topic>Spoken text</topic><topic>Transformer architecture</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Asakawa, Eiichi</creatorcontrib><creatorcontrib>Kaneko, Naoshi</creatorcontrib><creatorcontrib>Hasegawa, Dai</creatorcontrib><creatorcontrib>Shirakawa, Shinichi</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Neural networks</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Asakawa, Eiichi</au><au>Kaneko, Naoshi</au><au>Hasegawa, Dai</au><au>Shirakawa, Shinichi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Evaluation of text-to-gesture generation model using convolutional neural network</atitle><jtitle>Neural networks</jtitle><addtitle>Neural Netw</addtitle><date>2022-07</date><risdate>2022</risdate><volume>151</volume><spage>365</spage><epage>375</epage><pages>365-375</pages><issn>0893-6080</issn><eissn>1879-2782</eissn><abstract>Conversational gestures have a crucial role in realizing natural interactions with virtual agents and robots. Data-driven approaches, such as deep learning and machine learning, are promising in constructing the gesture generation model, which automatically provides the gesture motion for speech or spoken texts. This study experimentally analyzes a deep learning-based gesture generation model from spoken text using a convolutional neural network. The proposed model takes a sequence of spoken words as the input and outputs a sequence of 2D joint coordinates representing the conversational gesture motion. We prepare a dataset consisting of gesture motions and spoken texts by adding text information to an existing dataset and train the models using specific speaker’s data. The quality of the generated gestures is compared with those from an existing speech-to-gesture generation model through a user perceptual study. The subjective evaluation shows that the model performance is comparable or superior to those by the existing speech-to-gesture generation model. In addition, we investigate the importance of data cleansing and loss function selection in the text-to-gesture generation model. We further examine the model transferability between speakers. The experimental results demonstrate successful model transferability of the proposed model. Finally, we show that the text-to-gesture generation model can produce good quality gestures even when using a transformer architecture.
•The quality of text-to-gesture generation models is evaluated by human perceptual studies.•The quality of text-to-gesture generation models is comparable to speech-to-gesture models.•Data cleansing and loss function selection are important in text-to-gesture generation models.•The possibility of model transfer between speakers is demonstrated.</abstract><cop>United States</cop><pub>Elsevier Ltd</pub><pmid>35472730</pmid><doi>10.1016/j.neunet.2022.03.041</doi><tpages>11</tpages><orcidid>https://orcid.org/0000-0002-5638-2509</orcidid><orcidid>https://orcid.org/0000-0002-4659-6108</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0893-6080 |
ispartof | Neural networks, 2022-07, Vol.151, p.365-375 |
issn | 0893-6080 1879-2782 |
language | eng |
recordid | cdi_proquest_miscellaneous_2656200768 |
source | Elsevier ScienceDirect Journals Complete |
subjects | Convolutional neural network Deep learning Gesture generation Spoken text Transformer architecture |
title | Evaluation of text-to-gesture generation model using convolutional neural network |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-21T15%3A01%3A15IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Evaluation%20of%20text-to-gesture%20generation%20model%20using%20convolutional%20neural%20network&rft.jtitle=Neural%20networks&rft.au=Asakawa,%20Eiichi&rft.date=2022-07&rft.volume=151&rft.spage=365&rft.epage=375&rft.pages=365-375&rft.issn=0893-6080&rft.eissn=1879-2782&rft_id=info:doi/10.1016/j.neunet.2022.03.041&rft_dat=%3Cproquest_cross%3E2656200768%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2656200768&rft_id=info:pmid/35472730&rft_els_id=S0893608022001198&rfr_iscdi=true |