Creating speech-synchronized animation

We present a facial model designed primarily to support animated speech. Our facial model takes facial geometry as input and transforms it into a parametric deformable model. The facial model uses a muscle-based parameterization, allowing for easier integration between speech synchrony and facial ex...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on visualization and computer graphics 2005-05, Vol.11 (3), p.341-352
Hauptverfasser: King, S.A., Parent, R.E.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 352
container_issue 3
container_start_page 341
container_title IEEE transactions on visualization and computer graphics
container_volume 11
creator King, S.A.
Parent, R.E.
description We present a facial model designed primarily to support animated speech. Our facial model takes facial geometry as input and transforms it into a parametric deformable model. The facial model uses a muscle-based parameterization, allowing for easier integration between speech synchrony and facial expressions. Our facial model has a highly deformable lip model that is grafted onto the input facial geometry to provide the necessary geometric complexity needed for creating lip shapes and high-quality renderings. Our facial model also includes a highly deformable tongue model that can represent the shapes the tongue undergoes during speech. We add teeth, gums, and upper palate geometry to complete the inner mouth. To decrease the processing time, we hierarchically deform the facial surface. We also present a method to animate the facial model over time to create animated speech using a model of coarticulation that blends visemes together using dominance functions. We treat visemes as a dynamic shaping of the vocal tract by describing visemes as curves instead of keyframes. We show the utility of the techniques described in this paper by implementing them in a text-to-audiovisual-speech system that creates animation of speech from unrestricted text. The facial and coarticulation models must first be interactively initialized. The system then automatically creates accurate real-time animated speech from the input text. It is capable of cheaply producing tremendous amounts of animated speech with very low resource requirements.
doi_str_mv 10.1109/TVCG.2005.43
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_miscellaneous_67787195</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>1407866</ieee_id><sourcerecordid>28007851</sourcerecordid><originalsourceid>FETCH-LOGICAL-c399t-e8303a0e9e3ad06b7dea542d47c7a03c9596e2f774409b156b04070f086793a43</originalsourceid><addsrcrecordid>eNqF0U1Lw0AQBuBFFFurN2-CFA_1Yurs9-5Rglah4KV6XTbJxKa0Sc02h_rr3dKC4EFPOzAPw868hFxSGFMK9n72nk7GDECOBT8ifWoFTUCCOo41aJ0wxVSPnIWwAKBCGHtKelQaZQznfTJKW_Sbqv4YhjViPk_Cts7nbVNXX1gMfV2tYrepz8lJ6ZcBLw7vgLw9Pc7S52T6OnlJH6ZJzq3dJGg4cA9okfsCVKYL9FKwQuhce-C5lVYhK7UWAmxGpcpAgIYSjNKWe8EH5HY_d902nx2GjVtVIcfl0tfYdMEZqxiPO6goR39KpbXR1Mp_ITMA2kga4c0vuGi6to7ruvg9QRlnLKK7PcrbJoQWS7du443araPgdnm4XR5ul4cTPPLrw8wuW2Hxgw8BRHC1BxUi_rTjVYxS_Bt4u4sw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>867412322</pqid></control><display><type>article</type><title>Creating speech-synchronized animation</title><source>IEEE Electronic Library (IEL)</source><creator>King, S.A. ; Parent, R.E.</creator><creatorcontrib>King, S.A. ; Parent, R.E.</creatorcontrib><description>We present a facial model designed primarily to support animated speech. Our facial model takes facial geometry as input and transforms it into a parametric deformable model. The facial model uses a muscle-based parameterization, allowing for easier integration between speech synchrony and facial expressions. Our facial model has a highly deformable lip model that is grafted onto the input facial geometry to provide the necessary geometric complexity needed for creating lip shapes and high-quality renderings. Our facial model also includes a highly deformable tongue model that can represent the shapes the tongue undergoes during speech. We add teeth, gums, and upper palate geometry to complete the inner mouth. To decrease the processing time, we hierarchically deform the facial surface. We also present a method to animate the facial model over time to create animated speech using a model of coarticulation that blends visemes together using dominance functions. We treat visemes as a dynamic shaping of the vocal tract by describing visemes as curves instead of keyframes. We show the utility of the techniques described in this paper by implementing them in a text-to-audiovisual-speech system that creates animation of speech from unrestricted text. The facial and coarticulation models must first be interactively initialized. The system then automatically creates accurate real-time animated speech from the input text. It is capable of cheaply producing tremendous amounts of animated speech with very low resource requirements.</description><identifier>ISSN: 1077-2626</identifier><identifier>EISSN: 1941-0506</identifier><identifier>DOI: 10.1109/TVCG.2005.43</identifier><identifier>PMID: 15868833</identifier><identifier>CODEN: ITVGEA</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Algorithms ; Animated ; animation ; Artificial Intelligence ; coarticulation ; Computer Graphics ; Computer Simulation ; Deformable models ; Deformation ; Face - anatomy &amp; histology ; Face - physiology ; Facial ; Facial animation ; facial modeling ; Facial Muscles - anatomy &amp; histology ; Facial Muscles - physiology ; Formability ; Geometry ; Humans ; Image Enhancement - methods ; Image Interpretation, Computer-Assisted - methods ; Imaging, Three-Dimensional - methods ; Index Terms- Facial animation ; Information Storage and Retrieval - methods ; Lip - anatomy &amp; histology ; Lip - physiology ; lip synchronization ; Mathematical models ; Models, Anatomic ; Models, Biological ; Mouth ; Numerical Analysis, Computer-Assisted ; Parametrization ; Shape ; Signal Processing, Computer-Assisted ; Solid modeling ; Speech ; Speech - physiology ; speech synchronization ; Surface treatment ; Teeth ; Texts ; Tongue ; User-Computer Interface ; Video Recording - methods ; visual speech synthesis</subject><ispartof>IEEE transactions on visualization and computer graphics, 2005-05, Vol.11 (3), p.341-352</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2005</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c399t-e8303a0e9e3ad06b7dea542d47c7a03c9596e2f774409b156b04070f086793a43</citedby><cites>FETCH-LOGICAL-c399t-e8303a0e9e3ad06b7dea542d47c7a03c9596e2f774409b156b04070f086793a43</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/1407866$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,792,27901,27902,54733</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/1407866$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/15868833$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>King, S.A.</creatorcontrib><creatorcontrib>Parent, R.E.</creatorcontrib><title>Creating speech-synchronized animation</title><title>IEEE transactions on visualization and computer graphics</title><addtitle>TVCG</addtitle><addtitle>IEEE Trans Vis Comput Graph</addtitle><description>We present a facial model designed primarily to support animated speech. Our facial model takes facial geometry as input and transforms it into a parametric deformable model. The facial model uses a muscle-based parameterization, allowing for easier integration between speech synchrony and facial expressions. Our facial model has a highly deformable lip model that is grafted onto the input facial geometry to provide the necessary geometric complexity needed for creating lip shapes and high-quality renderings. Our facial model also includes a highly deformable tongue model that can represent the shapes the tongue undergoes during speech. We add teeth, gums, and upper palate geometry to complete the inner mouth. To decrease the processing time, we hierarchically deform the facial surface. We also present a method to animate the facial model over time to create animated speech using a model of coarticulation that blends visemes together using dominance functions. We treat visemes as a dynamic shaping of the vocal tract by describing visemes as curves instead of keyframes. We show the utility of the techniques described in this paper by implementing them in a text-to-audiovisual-speech system that creates animation of speech from unrestricted text. The facial and coarticulation models must first be interactively initialized. The system then automatically creates accurate real-time animated speech from the input text. It is capable of cheaply producing tremendous amounts of animated speech with very low resource requirements.</description><subject>Algorithms</subject><subject>Animated</subject><subject>animation</subject><subject>Artificial Intelligence</subject><subject>coarticulation</subject><subject>Computer Graphics</subject><subject>Computer Simulation</subject><subject>Deformable models</subject><subject>Deformation</subject><subject>Face - anatomy &amp; histology</subject><subject>Face - physiology</subject><subject>Facial</subject><subject>Facial animation</subject><subject>facial modeling</subject><subject>Facial Muscles - anatomy &amp; histology</subject><subject>Facial Muscles - physiology</subject><subject>Formability</subject><subject>Geometry</subject><subject>Humans</subject><subject>Image Enhancement - methods</subject><subject>Image Interpretation, Computer-Assisted - methods</subject><subject>Imaging, Three-Dimensional - methods</subject><subject>Index Terms- Facial animation</subject><subject>Information Storage and Retrieval - methods</subject><subject>Lip - anatomy &amp; histology</subject><subject>Lip - physiology</subject><subject>lip synchronization</subject><subject>Mathematical models</subject><subject>Models, Anatomic</subject><subject>Models, Biological</subject><subject>Mouth</subject><subject>Numerical Analysis, Computer-Assisted</subject><subject>Parametrization</subject><subject>Shape</subject><subject>Signal Processing, Computer-Assisted</subject><subject>Solid modeling</subject><subject>Speech</subject><subject>Speech - physiology</subject><subject>speech synchronization</subject><subject>Surface treatment</subject><subject>Teeth</subject><subject>Texts</subject><subject>Tongue</subject><subject>User-Computer Interface</subject><subject>Video Recording - methods</subject><subject>visual speech synthesis</subject><issn>1077-2626</issn><issn>1941-0506</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2005</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><sourceid>EIF</sourceid><recordid>eNqF0U1Lw0AQBuBFFFurN2-CFA_1Yurs9-5Rglah4KV6XTbJxKa0Sc02h_rr3dKC4EFPOzAPw868hFxSGFMK9n72nk7GDECOBT8ifWoFTUCCOo41aJ0wxVSPnIWwAKBCGHtKelQaZQznfTJKW_Sbqv4YhjViPk_Cts7nbVNXX1gMfV2tYrepz8lJ6ZcBLw7vgLw9Pc7S52T6OnlJH6ZJzq3dJGg4cA9okfsCVKYL9FKwQuhce-C5lVYhK7UWAmxGpcpAgIYSjNKWe8EH5HY_d902nx2GjVtVIcfl0tfYdMEZqxiPO6goR39KpbXR1Mp_ITMA2kga4c0vuGi6to7ruvg9QRlnLKK7PcrbJoQWS7du443araPgdnm4XR5ul4cTPPLrw8wuW2Hxgw8BRHC1BxUi_rTjVYxS_Bt4u4sw</recordid><startdate>20050501</startdate><enddate>20050501</enddate><creator>King, S.A.</creator><creator>Parent, R.E.</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>7X8</scope><scope>F28</scope><scope>FR3</scope></search><sort><creationdate>20050501</creationdate><title>Creating speech-synchronized animation</title><author>King, S.A. ; Parent, R.E.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c399t-e8303a0e9e3ad06b7dea542d47c7a03c9596e2f774409b156b04070f086793a43</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2005</creationdate><topic>Algorithms</topic><topic>Animated</topic><topic>animation</topic><topic>Artificial Intelligence</topic><topic>coarticulation</topic><topic>Computer Graphics</topic><topic>Computer Simulation</topic><topic>Deformable models</topic><topic>Deformation</topic><topic>Face - anatomy &amp; histology</topic><topic>Face - physiology</topic><topic>Facial</topic><topic>Facial animation</topic><topic>facial modeling</topic><topic>Facial Muscles - anatomy &amp; histology</topic><topic>Facial Muscles - physiology</topic><topic>Formability</topic><topic>Geometry</topic><topic>Humans</topic><topic>Image Enhancement - methods</topic><topic>Image Interpretation, Computer-Assisted - methods</topic><topic>Imaging, Three-Dimensional - methods</topic><topic>Index Terms- Facial animation</topic><topic>Information Storage and Retrieval - methods</topic><topic>Lip - anatomy &amp; histology</topic><topic>Lip - physiology</topic><topic>lip synchronization</topic><topic>Mathematical models</topic><topic>Models, Anatomic</topic><topic>Models, Biological</topic><topic>Mouth</topic><topic>Numerical Analysis, Computer-Assisted</topic><topic>Parametrization</topic><topic>Shape</topic><topic>Signal Processing, Computer-Assisted</topic><topic>Solid modeling</topic><topic>Speech</topic><topic>Speech - physiology</topic><topic>speech synchronization</topic><topic>Surface treatment</topic><topic>Teeth</topic><topic>Texts</topic><topic>Tongue</topic><topic>User-Computer Interface</topic><topic>Video Recording - methods</topic><topic>visual speech synthesis</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>King, S.A.</creatorcontrib><creatorcontrib>Parent, R.E.</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>MEDLINE - Academic</collection><collection>ANTE: Abstracts in New Technology &amp; Engineering</collection><collection>Engineering Research Database</collection><jtitle>IEEE transactions on visualization and computer graphics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>King, S.A.</au><au>Parent, R.E.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Creating speech-synchronized animation</atitle><jtitle>IEEE transactions on visualization and computer graphics</jtitle><stitle>TVCG</stitle><addtitle>IEEE Trans Vis Comput Graph</addtitle><date>2005-05-01</date><risdate>2005</risdate><volume>11</volume><issue>3</issue><spage>341</spage><epage>352</epage><pages>341-352</pages><issn>1077-2626</issn><eissn>1941-0506</eissn><coden>ITVGEA</coden><abstract>We present a facial model designed primarily to support animated speech. Our facial model takes facial geometry as input and transforms it into a parametric deformable model. The facial model uses a muscle-based parameterization, allowing for easier integration between speech synchrony and facial expressions. Our facial model has a highly deformable lip model that is grafted onto the input facial geometry to provide the necessary geometric complexity needed for creating lip shapes and high-quality renderings. Our facial model also includes a highly deformable tongue model that can represent the shapes the tongue undergoes during speech. We add teeth, gums, and upper palate geometry to complete the inner mouth. To decrease the processing time, we hierarchically deform the facial surface. We also present a method to animate the facial model over time to create animated speech using a model of coarticulation that blends visemes together using dominance functions. We treat visemes as a dynamic shaping of the vocal tract by describing visemes as curves instead of keyframes. We show the utility of the techniques described in this paper by implementing them in a text-to-audiovisual-speech system that creates animation of speech from unrestricted text. The facial and coarticulation models must first be interactively initialized. The system then automatically creates accurate real-time animated speech from the input text. It is capable of cheaply producing tremendous amounts of animated speech with very low resource requirements.</abstract><cop>United States</cop><pub>IEEE</pub><pmid>15868833</pmid><doi>10.1109/TVCG.2005.43</doi><tpages>12</tpages></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 1077-2626
ispartof IEEE transactions on visualization and computer graphics, 2005-05, Vol.11 (3), p.341-352
issn 1077-2626
1941-0506
language eng
recordid cdi_proquest_miscellaneous_67787195
source IEEE Electronic Library (IEL)
subjects Algorithms
Animated
animation
Artificial Intelligence
coarticulation
Computer Graphics
Computer Simulation
Deformable models
Deformation
Face - anatomy & histology
Face - physiology
Facial
Facial animation
facial modeling
Facial Muscles - anatomy & histology
Facial Muscles - physiology
Formability
Geometry
Humans
Image Enhancement - methods
Image Interpretation, Computer-Assisted - methods
Imaging, Three-Dimensional - methods
Index Terms- Facial animation
Information Storage and Retrieval - methods
Lip - anatomy & histology
Lip - physiology
lip synchronization
Mathematical models
Models, Anatomic
Models, Biological
Mouth
Numerical Analysis, Computer-Assisted
Parametrization
Shape
Signal Processing, Computer-Assisted
Solid modeling
Speech
Speech - physiology
speech synchronization
Surface treatment
Teeth
Texts
Tongue
User-Computer Interface
Video Recording - methods
visual speech synthesis
title Creating speech-synchronized animation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T20%3A43%3A16IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Creating%20speech-synchronized%20animation&rft.jtitle=IEEE%20transactions%20on%20visualization%20and%20computer%20graphics&rft.au=King,%20S.A.&rft.date=2005-05-01&rft.volume=11&rft.issue=3&rft.spage=341&rft.epage=352&rft.pages=341-352&rft.issn=1077-2626&rft.eissn=1941-0506&rft.coden=ITVGEA&rft_id=info:doi/10.1109/TVCG.2005.43&rft_dat=%3Cproquest_RIE%3E28007851%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=867412322&rft_id=info:pmid/15868833&rft_ieee_id=1407866&rfr_iscdi=true