<< Chapter < Page Chapter >> Page >

Emotion database

We use the Emotional Prosody Speech and Transcripts Database from the Linguistic Data Consortium. The data set consists of labeled emotional utterances of semantically neutral content, intoned by professional actors.

Specifically, the utterances were four-syllable dates and numbers, in the English language, from 15 emotion categories. Actors were provided with descriptions of each emotional context and were to reproduce the utterance in that manner. For example,

9 Actors participated, with a yield of ~3 hours of high quality speech samples. This translated to ~3000 speech samples overall, with ~200 samples per emotion. This was the largest English data set we could find with a high emotion resolution.

We split the database into two sets – a training set and test set, 70% - 30% respectively. These were used to train the neural network then test its performance.

Source

M. Liberman, K. Davis, M. Grossman, N. Martey, and J. Bell. “Emotional Prosody Speech and Transcripts.” Linguistic Data Consortium, 2002.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Robust classification of highly-specific emotion in human speech. OpenStax CNX. Dec 14, 2012 Download for free at http://cnx.org/content/col11465/1.1
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Robust classification of highly-specific emotion in human speech' conversation and receive update notifications?

Ask