Sorry, you need to enable JavaScript to visit this website.

PHOTOREALISTIC ADAPTATION AND INTERPOLATION OF FACIAL EXPRESSIONS USING HMMS AND AAMS FOR AUDIO-VISUAL SPEECH SYNTHESIS

Citation Author(s):
Athanasios Katsamanis, Petros Maragos
Submitted by:
Panagiotis Filntisis
Last updated:
16 September 2017 - 12:51pm
Document Type:
Poster
Document Year:
2017
Event:
Presenters:
PANAGIOTIS PARASKEVAS FILNTISIS
Paper Code:
TA-PA.5
 

In this paper, motivated by the continuously increasing presence of intelligent agents in everyday life, we address the problem of expres-sive photorealistic audio-visual speech synthesis, with a strong focus on the visual modality. Emotion constitutes one of the main driving factors of social life and it is expressed mainly through facial expres-sions. Synthesis of a talking head capable of expressive audio-visual speech is challenging due to the data overhead that arises when con-sidering the vast number of emotions we would like the talking head to express. In order to tackle this challenge, we propose the usage of two methods, namely Hidden Markov Model (HMM) adaptation and interpolation, with HMMs modeling visual parameters via an Active Appearance Model (AAM) of the face. We show that through HMM adaptation we can successfully adapt a “neutral” talking head to a target emotion with a small amount of adaptation data, as well as that through HMM interpolation we can robustly achieve different levels of intensity for an emotion.

up
0 users have voted: