Documents
Poster
Crowdsourcing Emotional Speech
- Citation Author(s):
- Submitted by:
- Jennifer Smith
- Last updated:
- 13 April 2018 - 10:55pm
- Document Type:
- Poster
- Document Year:
- 2018
- Event:
- Presenters:
- Jennifer Smith
- Paper Code:
- 3622
- Categories:
- Log in to post comments
We describe the methodology for the collection and annotation of a large corpus of emotional speech data through crowdsourcing. The corpus offers 187 hours of data from 2,965 subjects. Data includes non-emotional recordings from each subject as well as recordings for five emotions: angry, happy-low-arousal, happy-high-arousal, neutral,
and sad. The data consist of spontaneous speech elicited from subjects via a web-based tool. Subjects used their own personal recording equipment, resulting in a data set that contains variation in room acoustics, microphone, etc. This offers the advantage of matching the type of variation one would expect to see when exposing
speech technology in the wild in a web-based environment. The annotation scheme covers the quality of emotion expressed through the tone of voice and what was said, along with common audioquality issues. We discuss lessons learned in the process of the creation of this corpus.