Skip to Main Content
Frequently Asked Questions
Submit an ETD
Global Search Box
Need Help?
Keyword Search
Participating Institutions
Advanced Search
School Logo
Files
File List
osu1148973645.pdf (21.08 MB)
Supplemental Files
File List
coartCompare_QuickTm.mov
(15.41 MB)
coartWords_QuickTm.mov
(8.7 MB)
complexExpressions_DivX.avi
(562.88 KB)
constraintsCoart_DivX.avi
(6.42 MB)
headCompare_QuickTm.mov
(13.21 MB)
headMotionExtract_DivX.avi
(2.87 MB)
mrksToSkin_DivX.avi
(3.02 MB)
muscles_DivX.avi
(9.34 MB)
sampleMocap_DivX.avi
(8.7 MB)
stylesHappy_DivX.avi
(442.72 KB)
synthResultsMocap_QuickTm.mov
(32.1 MB)
synthResultsNoMocap_DivX.avi
(12.63 MB)
ETD Abstract Container
Abstract Header
A facial animation model for expressive audio-visual speech
Author Info
Somasundaram, Arunachalam
Permalink:
http://rave.ohiolink.edu/etdc/view?acc_num=osu1148973645
Abstract Details
Year and Degree
2006, Doctor of Philosophy, Ohio State University, Computer and Information Science.
Abstract
Expressive facial speech animation is a challenging topic of great interest to the computer graphics community. Adding emotions to audio-visual speech animation is very important for realistic facial animation. The complexity of neutral visual speech synthesis is mainly attributed to co-articulation. Co-articulation is the phenomenon due to which the facial pose of the current segment of speech is affected by the neigboring segments of speech. The inclusion of emotions and fluency effects in speech adds to that complexity because of the corresponding shape and timing modifications brought about in speech. Speech is often accompanied by supportive visual prosodic elements such as motion of the head, eyes, and eyebrow, which improve the intelligibility of speech, and they need to be synthesized. In this dissertation, we present a technique to modify input neutral audio and synthesize visual speech incorporating effects of emotion and fluency. Visemes, which are visual counterpart of phonemes, are used to animate speech. We motion capture 3-D facial motion and extract facial muscle positions of expressive visemes. Our expressive visemes capture the pose of the entire face. The expressive visemes are blended using a novel constraint-based co-articulation technique that can easily accommodate the effects of emotion. We also present a visual prosody model for emotional speech, based on motion capture data, that exhibits non-verbal behaviors such as eyebrow motion and overall head motion.
Committee
Richard Parent (Advisor)
Pages
155 p.
Subject Headings
Computer Science
Keywords
expressive facial speech animation
;
expressive audio-visual speech
;
facial animation
;
speech animation
;
facial expressions
;
face
;
speech
;
emotions
Recommended Citations
Refworks
EndNote
RIS
Mendeley
Citations
Somasundaram, A. (2006).
A facial animation model for expressive audio-visual speech
[Doctoral dissertation, Ohio State University]. OhioLINK Electronic Theses and Dissertations Center. http://rave.ohiolink.edu/etdc/view?acc_num=osu1148973645
APA Style (7th edition)
Somasundaram, Arunachalam.
A facial animation model for expressive audio-visual speech.
2006. Ohio State University, Doctoral dissertation.
OhioLINK Electronic Theses and Dissertations Center
, http://rave.ohiolink.edu/etdc/view?acc_num=osu1148973645.
MLA Style (8th edition)
Somasundaram, Arunachalam. "A facial animation model for expressive audio-visual speech." Doctoral dissertation, Ohio State University, 2006. http://rave.ohiolink.edu/etdc/view?acc_num=osu1148973645
Chicago Manual of Style (17th edition)
Abstract Footer
Document number:
osu1148973645
Download Count:
809
Copyright Info
© 2006, all rights reserved.
This open access ETD is published by The Ohio State University and OhioLINK.