Creating speech-synchronized animation
We present a facial model designed primarily to support animated speech. Our facial model takes facial geometry as input and transforms it into a parametric deformable model. The facial model uses a muscle-based parameterization, allowing for easier integration between speech synchrony and facial ex...
Gespeichert in:
Veröffentlicht in: | IEEE transactions on visualization and computer graphics 2005-05, Vol.11 (3), p.341-352 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 352 |
---|---|
container_issue | 3 |
container_start_page | 341 |
container_title | IEEE transactions on visualization and computer graphics |
container_volume | 11 |
creator | King, S.A. Parent, R.E. |
description | We present a facial model designed primarily to support animated speech. Our facial model takes facial geometry as input and transforms it into a parametric deformable model. The facial model uses a muscle-based parameterization, allowing for easier integration between speech synchrony and facial expressions. Our facial model has a highly deformable lip model that is grafted onto the input facial geometry to provide the necessary geometric complexity needed for creating lip shapes and high-quality renderings. Our facial model also includes a highly deformable tongue model that can represent the shapes the tongue undergoes during speech. We add teeth, gums, and upper palate geometry to complete the inner mouth. To decrease the processing time, we hierarchically deform the facial surface. We also present a method to animate the facial model over time to create animated speech using a model of coarticulation that blends visemes together using dominance functions. We treat visemes as a dynamic shaping of the vocal tract by describing visemes as curves instead of keyframes. We show the utility of the techniques described in this paper by implementing them in a text-to-audiovisual-speech system that creates animation of speech from unrestricted text. The facial and coarticulation models must first be interactively initialized. The system then automatically creates accurate real-time animated speech from the input text. It is capable of cheaply producing tremendous amounts of animated speech with very low resource requirements. |
doi_str_mv | 10.1109/TVCG.2005.43 |
format | Article |
fullrecord | <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_miscellaneous_67787195</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>1407866</ieee_id><sourcerecordid>28007851</sourcerecordid><originalsourceid>FETCH-LOGICAL-c399t-e8303a0e9e3ad06b7dea542d47c7a03c9596e2f774409b156b04070f086793a43</originalsourceid><addsrcrecordid>eNqF0U1Lw0AQBuBFFFurN2-CFA_1Yurs9-5Rglah4KV6XTbJxKa0Sc02h_rr3dKC4EFPOzAPw868hFxSGFMK9n72nk7GDECOBT8ifWoFTUCCOo41aJ0wxVSPnIWwAKBCGHtKelQaZQznfTJKW_Sbqv4YhjViPk_Cts7nbVNXX1gMfV2tYrepz8lJ6ZcBLw7vgLw9Pc7S52T6OnlJH6ZJzq3dJGg4cA9okfsCVKYL9FKwQuhce-C5lVYhK7UWAmxGpcpAgIYSjNKWe8EH5HY_d902nx2GjVtVIcfl0tfYdMEZqxiPO6goR39KpbXR1Mp_ITMA2kga4c0vuGi6to7ruvg9QRlnLKK7PcrbJoQWS7du443araPgdnm4XR5ul4cTPPLrw8wuW2Hxgw8BRHC1BxUi_rTjVYxS_Bt4u4sw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>867412322</pqid></control><display><type>article</type><title>Creating speech-synchronized animation</title><source>IEEE Electronic Library (IEL)</source><creator>King, S.A. ; Parent, R.E.</creator><creatorcontrib>King, S.A. ; Parent, R.E.</creatorcontrib><description>We present a facial model designed primarily to support animated speech. Our facial model takes facial geometry as input and transforms it into a parametric deformable model. The facial model uses a muscle-based parameterization, allowing for easier integration between speech synchrony and facial expressions. Our facial model has a highly deformable lip model that is grafted onto the input facial geometry to provide the necessary geometric complexity needed for creating lip shapes and high-quality renderings. Our facial model also includes a highly deformable tongue model that can represent the shapes the tongue undergoes during speech. We add teeth, gums, and upper palate geometry to complete the inner mouth. To decrease the processing time, we hierarchically deform the facial surface. We also present a method to animate the facial model over time to create animated speech using a model of coarticulation that blends visemes together using dominance functions. We treat visemes as a dynamic shaping of the vocal tract by describing visemes as curves instead of keyframes. We show the utility of the techniques described in this paper by implementing them in a text-to-audiovisual-speech system that creates animation of speech from unrestricted text. The facial and coarticulation models must first be interactively initialized. The system then automatically creates accurate real-time animated speech from the input text. It is capable of cheaply producing tremendous amounts of animated speech with very low resource requirements.</description><identifier>ISSN: 1077-2626</identifier><identifier>EISSN: 1941-0506</identifier><identifier>DOI: 10.1109/TVCG.2005.43</identifier><identifier>PMID: 15868833</identifier><identifier>CODEN: ITVGEA</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Algorithms ; Animated ; animation ; Artificial Intelligence ; coarticulation ; Computer Graphics ; Computer Simulation ; Deformable models ; Deformation ; Face - anatomy & histology ; Face - physiology ; Facial ; Facial animation ; facial modeling ; Facial Muscles - anatomy & histology ; Facial Muscles - physiology ; Formability ; Geometry ; Humans ; Image Enhancement - methods ; Image Interpretation, Computer-Assisted - methods ; Imaging, Three-Dimensional - methods ; Index Terms- Facial animation ; Information Storage and Retrieval - methods ; Lip - anatomy & histology ; Lip - physiology ; lip synchronization ; Mathematical models ; Models, Anatomic ; Models, Biological ; Mouth ; Numerical Analysis, Computer-Assisted ; Parametrization ; Shape ; Signal Processing, Computer-Assisted ; Solid modeling ; Speech ; Speech - physiology ; speech synchronization ; Surface treatment ; Teeth ; Texts ; Tongue ; User-Computer Interface ; Video Recording - methods ; visual speech synthesis</subject><ispartof>IEEE transactions on visualization and computer graphics, 2005-05, Vol.11 (3), p.341-352</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2005</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c399t-e8303a0e9e3ad06b7dea542d47c7a03c9596e2f774409b156b04070f086793a43</citedby><cites>FETCH-LOGICAL-c399t-e8303a0e9e3ad06b7dea542d47c7a03c9596e2f774409b156b04070f086793a43</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/1407866$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,792,27901,27902,54733</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/1407866$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/15868833$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>King, S.A.</creatorcontrib><creatorcontrib>Parent, R.E.</creatorcontrib><title>Creating speech-synchronized animation</title><title>IEEE transactions on visualization and computer graphics</title><addtitle>TVCG</addtitle><addtitle>IEEE Trans Vis Comput Graph</addtitle><description>We present a facial model designed primarily to support animated speech. Our facial model takes facial geometry as input and transforms it into a parametric deformable model. The facial model uses a muscle-based parameterization, allowing for easier integration between speech synchrony and facial expressions. Our facial model has a highly deformable lip model that is grafted onto the input facial geometry to provide the necessary geometric complexity needed for creating lip shapes and high-quality renderings. Our facial model also includes a highly deformable tongue model that can represent the shapes the tongue undergoes during speech. We add teeth, gums, and upper palate geometry to complete the inner mouth. To decrease the processing time, we hierarchically deform the facial surface. We also present a method to animate the facial model over time to create animated speech using a model of coarticulation that blends visemes together using dominance functions. We treat visemes as a dynamic shaping of the vocal tract by describing visemes as curves instead of keyframes. We show the utility of the techniques described in this paper by implementing them in a text-to-audiovisual-speech system that creates animation of speech from unrestricted text. The facial and coarticulation models must first be interactively initialized. The system then automatically creates accurate real-time animated speech from the input text. It is capable of cheaply producing tremendous amounts of animated speech with very low resource requirements.</description><subject>Algorithms</subject><subject>Animated</subject><subject>animation</subject><subject>Artificial Intelligence</subject><subject>coarticulation</subject><subject>Computer Graphics</subject><subject>Computer Simulation</subject><subject>Deformable models</subject><subject>Deformation</subject><subject>Face - anatomy & histology</subject><subject>Face - physiology</subject><subject>Facial</subject><subject>Facial animation</subject><subject>facial modeling</subject><subject>Facial Muscles - anatomy & histology</subject><subject>Facial Muscles - physiology</subject><subject>Formability</subject><subject>Geometry</subject><subject>Humans</subject><subject>Image Enhancement - methods</subject><subject>Image Interpretation, Computer-Assisted - methods</subject><subject>Imaging, Three-Dimensional - methods</subject><subject>Index Terms- Facial animation</subject><subject>Information Storage and Retrieval - methods</subject><subject>Lip - anatomy & histology</subject><subject>Lip - physiology</subject><subject>lip synchronization</subject><subject>Mathematical models</subject><subject>Models, Anatomic</subject><subject>Models, Biological</subject><subject>Mouth</subject><subject>Numerical Analysis, Computer-Assisted</subject><subject>Parametrization</subject><subject>Shape</subject><subject>Signal Processing, Computer-Assisted</subject><subject>Solid modeling</subject><subject>Speech</subject><subject>Speech - physiology</subject><subject>speech synchronization</subject><subject>Surface treatment</subject><subject>Teeth</subject><subject>Texts</subject><subject>Tongue</subject><subject>User-Computer Interface</subject><subject>Video Recording - methods</subject><subject>visual speech synthesis</subject><issn>1077-2626</issn><issn>1941-0506</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2005</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><sourceid>EIF</sourceid><recordid>eNqF0U1Lw0AQBuBFFFurN2-CFA_1Yurs9-5Rglah4KV6XTbJxKa0Sc02h_rr3dKC4EFPOzAPw868hFxSGFMK9n72nk7GDECOBT8ifWoFTUCCOo41aJ0wxVSPnIWwAKBCGHtKelQaZQznfTJKW_Sbqv4YhjViPk_Cts7nbVNXX1gMfV2tYrepz8lJ6ZcBLw7vgLw9Pc7S52T6OnlJH6ZJzq3dJGg4cA9okfsCVKYL9FKwQuhce-C5lVYhK7UWAmxGpcpAgIYSjNKWe8EH5HY_d902nx2GjVtVIcfl0tfYdMEZqxiPO6goR39KpbXR1Mp_ITMA2kga4c0vuGi6to7ruvg9QRlnLKK7PcrbJoQWS7du443araPgdnm4XR5ul4cTPPLrw8wuW2Hxgw8BRHC1BxUi_rTjVYxS_Bt4u4sw</recordid><startdate>20050501</startdate><enddate>20050501</enddate><creator>King, S.A.</creator><creator>Parent, R.E.</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>7X8</scope><scope>F28</scope><scope>FR3</scope></search><sort><creationdate>20050501</creationdate><title>Creating speech-synchronized animation</title><author>King, S.A. ; Parent, R.E.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c399t-e8303a0e9e3ad06b7dea542d47c7a03c9596e2f774409b156b04070f086793a43</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2005</creationdate><topic>Algorithms</topic><topic>Animated</topic><topic>animation</topic><topic>Artificial Intelligence</topic><topic>coarticulation</topic><topic>Computer Graphics</topic><topic>Computer Simulation</topic><topic>Deformable models</topic><topic>Deformation</topic><topic>Face - anatomy & histology</topic><topic>Face - physiology</topic><topic>Facial</topic><topic>Facial animation</topic><topic>facial modeling</topic><topic>Facial Muscles - anatomy & histology</topic><topic>Facial Muscles - physiology</topic><topic>Formability</topic><topic>Geometry</topic><topic>Humans</topic><topic>Image Enhancement - methods</topic><topic>Image Interpretation, Computer-Assisted - methods</topic><topic>Imaging, Three-Dimensional - methods</topic><topic>Index Terms- Facial animation</topic><topic>Information Storage and Retrieval - methods</topic><topic>Lip - anatomy & histology</topic><topic>Lip - physiology</topic><topic>lip synchronization</topic><topic>Mathematical models</topic><topic>Models, Anatomic</topic><topic>Models, Biological</topic><topic>Mouth</topic><topic>Numerical Analysis, Computer-Assisted</topic><topic>Parametrization</topic><topic>Shape</topic><topic>Signal Processing, Computer-Assisted</topic><topic>Solid modeling</topic><topic>Speech</topic><topic>Speech - physiology</topic><topic>speech synchronization</topic><topic>Surface treatment</topic><topic>Teeth</topic><topic>Texts</topic><topic>Tongue</topic><topic>User-Computer Interface</topic><topic>Video Recording - methods</topic><topic>visual speech synthesis</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>King, S.A.</creatorcontrib><creatorcontrib>Parent, R.E.</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>MEDLINE - Academic</collection><collection>ANTE: Abstracts in New Technology & Engineering</collection><collection>Engineering Research Database</collection><jtitle>IEEE transactions on visualization and computer graphics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>King, S.A.</au><au>Parent, R.E.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Creating speech-synchronized animation</atitle><jtitle>IEEE transactions on visualization and computer graphics</jtitle><stitle>TVCG</stitle><addtitle>IEEE Trans Vis Comput Graph</addtitle><date>2005-05-01</date><risdate>2005</risdate><volume>11</volume><issue>3</issue><spage>341</spage><epage>352</epage><pages>341-352</pages><issn>1077-2626</issn><eissn>1941-0506</eissn><coden>ITVGEA</coden><abstract>We present a facial model designed primarily to support animated speech. Our facial model takes facial geometry as input and transforms it into a parametric deformable model. The facial model uses a muscle-based parameterization, allowing for easier integration between speech synchrony and facial expressions. Our facial model has a highly deformable lip model that is grafted onto the input facial geometry to provide the necessary geometric complexity needed for creating lip shapes and high-quality renderings. Our facial model also includes a highly deformable tongue model that can represent the shapes the tongue undergoes during speech. We add teeth, gums, and upper palate geometry to complete the inner mouth. To decrease the processing time, we hierarchically deform the facial surface. We also present a method to animate the facial model over time to create animated speech using a model of coarticulation that blends visemes together using dominance functions. We treat visemes as a dynamic shaping of the vocal tract by describing visemes as curves instead of keyframes. We show the utility of the techniques described in this paper by implementing them in a text-to-audiovisual-speech system that creates animation of speech from unrestricted text. The facial and coarticulation models must first be interactively initialized. The system then automatically creates accurate real-time animated speech from the input text. It is capable of cheaply producing tremendous amounts of animated speech with very low resource requirements.</abstract><cop>United States</cop><pub>IEEE</pub><pmid>15868833</pmid><doi>10.1109/TVCG.2005.43</doi><tpages>12</tpages></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | ISSN: 1077-2626 |
ispartof | IEEE transactions on visualization and computer graphics, 2005-05, Vol.11 (3), p.341-352 |
issn | 1077-2626 1941-0506 |
language | eng |
recordid | cdi_proquest_miscellaneous_67787195 |
source | IEEE Electronic Library (IEL) |
subjects | Algorithms Animated animation Artificial Intelligence coarticulation Computer Graphics Computer Simulation Deformable models Deformation Face - anatomy & histology Face - physiology Facial Facial animation facial modeling Facial Muscles - anatomy & histology Facial Muscles - physiology Formability Geometry Humans Image Enhancement - methods Image Interpretation, Computer-Assisted - methods Imaging, Three-Dimensional - methods Index Terms- Facial animation Information Storage and Retrieval - methods Lip - anatomy & histology Lip - physiology lip synchronization Mathematical models Models, Anatomic Models, Biological Mouth Numerical Analysis, Computer-Assisted Parametrization Shape Signal Processing, Computer-Assisted Solid modeling Speech Speech - physiology speech synchronization Surface treatment Teeth Texts Tongue User-Computer Interface Video Recording - methods visual speech synthesis |
title | Creating speech-synchronized animation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T20%3A43%3A16IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Creating%20speech-synchronized%20animation&rft.jtitle=IEEE%20transactions%20on%20visualization%20and%20computer%20graphics&rft.au=King,%20S.A.&rft.date=2005-05-01&rft.volume=11&rft.issue=3&rft.spage=341&rft.epage=352&rft.pages=341-352&rft.issn=1077-2626&rft.eissn=1941-0506&rft.coden=ITVGEA&rft_id=info:doi/10.1109/TVCG.2005.43&rft_dat=%3Cproquest_RIE%3E28007851%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=867412322&rft_id=info:pmid/15868833&rft_ieee_id=1407866&rfr_iscdi=true |