Documents
Poster
PHOTOREALISTIC ADAPTATION AND INTERPOLATION OF FACIAL EXPRESSIONS USING HMMS AND AAMS FOR AUDIO-VISUAL SPEECH SYNTHESIS
- Citation Author(s):
- Submitted by:
- Panagiotis Filntisis
- Last updated:
- 16 September 2017 - 12:51pm
- Document Type:
- Poster
- Document Year:
- 2017
- Event:
- Presenters:
- PANAGIOTIS PARASKEVAS FILNTISIS
- Paper Code:
- TA-PA.5
- Categories:
- Log in to post comments
In this paper, motivated by the continuously increasing presence of intelligent agents in everyday life, we address the problem of expres-sive photorealistic audio-visual speech synthesis, with a strong focus on the visual modality. Emotion constitutes one of the main driving factors of social life and it is expressed mainly through facial expres-sions. Synthesis of a talking head capable of expressive audio-visual speech is challenging due to the data overhead that arises when con-sidering the vast number of emotions we would like the talking head to express. In order to tackle this challenge, we propose the usage of two methods, namely Hidden Markov Model (HMM) adaptation and interpolation, with HMMs modeling visual parameters via an Active Appearance Model (AAM) of the face. We show that through HMM adaptation we can successfully adapt a “neutral” talking head to a target emotion with a small amount of adaptation data, as well as that through HMM interpolation we can robustly achieve different levels of intensity for an emotion.