Abstract
Acoustic features, in the examination of emotions occurring in pronouncing English and Chinese Mandarin words, are investigated in this study, then different emotion recognition experiments are presented. To this end, the sound recordings for 91 speakers were analyzed. In the test experiment, a linguistic data set was used to examine which acoustic features are most important for the emotional representation in signal acquisition, segmentation, construction, and encoding. In doing so, words, syllables, phonemes (which contain vowels and consonants), stress and frequency tones were taken into consideration. The types of emotions considered in the experiment included neutral, happy, and sad. Time duration differences, F0 frequency, and dB intensity levels variables were used in conjunction with unsupervised and supervised machine learning approaches for emotion recognition.
Original language | English |
---|---|
Pages (from-to) | 417-432 |
Number of pages | 16 |
Journal | International Journal of Speech Technology |
Volume | 26 |
Issue number | 2 |
Early online date | 18 Mar 2023 |
DOIs | |
Publication status | Published - Jul 2023 |
Scopus Subject Areas
- Artificial Intelligence
- Software
- Language and Linguistics
- Human-Computer Interaction
- Computer Vision and Pattern Recognition
- Linguistics and Language