Abstract
Emotion recognition has been well researched in mono-modality in the past decade. However, people express their emotion or feelings naturally via more than one modalities like voice, facial expressions, text, and behaviors. In this paper, we propose a new method to model deep interactive learning and dual modalities (e.g., speech and text) to conduct multimodal emotion recognition. An unsupervised triplet-loss objective function is constructed to learn representation of emotional information from speech audio. We extract text emotional feature representation by transfer learning of text-To-Text embedding from T5 pre-Trained model. Human-machine interaction like user feedback plays a vital role in improve multimodal emotion recognition in dialogue system. Deep interactive learning model is constructed by explicit and implicit feedback. Human-machine interactive learning enhanced transformer model can achieve higher levels of accuracy and precision than their non-interactive counterparts.
| Original language | English |
|---|---|
| Title of host publication | ACAI 2022 - Conference Proceedings |
| Subtitle of host publication | 2022 5th International Conference on Algorithms, Computing and Artificial Intelligence |
| Place of Publication | New York |
| Publisher | Association for Computing Machinery (ACM) |
| Number of pages | 7 |
| ISBN (Electronic) | 9781450398343 |
| ISBN (Print) | 9781450398336 |
| DOIs | |
| Publication status | Published - 23 Dec 2022 |
| Event | 5th International Conference on Algorithms, Computing and Artificial Intelligence, ACAI 2022 - Sanya, China Duration: 23 Dec 2022 → 25 Dec 2022 |
Publication series
| Name | ACM International Conference Proceeding Series |
|---|
Conference
| Conference | 5th International Conference on Algorithms, Computing and Artificial Intelligence, ACAI 2022 |
|---|---|
| Country/Territory | China |
| City | Sanya |
| Period | 23/12/22 → 25/12/22 |
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 9 Industry, Innovation, and Infrastructure
User-Defined Keywords
- human-machine interaction
- interactive learning
- Multimodal emotion recognition
- transformer model
Fingerprint
Dive into the research topics of 'Enhanced Human-Machine Interactive Learning for Multimodal Emotion Recognition in Dialogue System'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver