Documents
Poster
Retrieving Speech Samples with Similar Emotional Content Using a Triplet Loss Function
- Citation Author(s):
- Submitted by:
- John Harvill
- Last updated:
- 9 May 2019 - 12:32pm
- Document Type:
- Poster
- Document Year:
- 2019
- Event:
- Presenters:
- John Harvill
- Paper Code:
- HLT-P4.5
- Categories:
- Keywords:
- Log in to post comments
The ability to identify speech with similar emotional content is valuable to many applications, including speech retrieval, surveillance, and emotional speech synthesis. While current formulations in speech emotion recognition based on classification or regression are not appropriate for this task, solutions based on preference learning offer appealing approaches for this task. This paper aims to find speech samples that are emotionally similar to an anchor speech sample provided as a query. This novel formulation opens interesting research questions. How well can a machine complete this task? How does the accuracy of automatic algorithms compare to the performance of a human performing this task? This study addresses these questions by training a deep learning model using a triplet loss function, mapping the acoustic features into an embedding that is discriminative for this task. The network receives an anchor speech sample and two competing speech samples, and the task is to determine which of the candidate speech sample conveys the closest emotional content to the emotion conveyed by the anchor. By comparing the results from our model with human perceptual evaluations, this study demonstrates that the proposed approach has performance very close to human performance in retrieving samples with similar emotional content.