Enhanced Human-Machine Interactive Learning for Multimodal Emotion Recognition in Dialogue System

Clement H.C. Leung, James J. Deng*, Yuanxi Li

*Corresponding author for this work

Research output: Chapter in book/report/conference proceedingConference proceedingpeer-review

2 Citations (Scopus)

Abstract

Emotion recognition has been well researched in mono-modality in the past decade. However, people express their emotion or feelings naturally via more than one modalities like voice, facial expressions, text, and behaviors. In this paper, we propose a new method to model deep interactive learning and dual modalities (e.g., speech and text) to conduct multimodal emotion recognition. An unsupervised triplet-loss objective function is constructed to learn representation of emotional information from speech audio. We extract text emotional feature representation by transfer learning of text-To-Text embedding from T5 pre-Trained model. Human-machine interaction like user feedback plays a vital role in improve multimodal emotion recognition in dialogue system. Deep interactive learning model is constructed by explicit and implicit feedback. Human-machine interactive learning enhanced transformer model can achieve higher levels of accuracy and precision than their non-interactive counterparts.

Original languageEnglish
Title of host publicationACAI 2022 - Conference Proceedings
Subtitle of host publication2022 5th International Conference on Algorithms, Computing and Artificial Intelligence
Place of PublicationNew York
PublisherAssociation for Computing Machinery (ACM)
Number of pages7
ISBN (Electronic)9781450398343
ISBN (Print)9781450398336
DOIs
Publication statusPublished - 23 Dec 2022
Event5th International Conference on Algorithms, Computing and Artificial Intelligence, ACAI 2022 - Sanya, China
Duration: 23 Dec 202225 Dec 2022

Publication series

NameACM International Conference Proceeding Series

Conference

Conference5th International Conference on Algorithms, Computing and Artificial Intelligence, ACAI 2022
Country/TerritoryChina
CitySanya
Period23/12/2225/12/22

User-Defined Keywords

  • human-machine interaction
  • interactive learning
  • Multimodal emotion recognition
  • transformer model

Fingerprint

Dive into the research topics of 'Enhanced Human-Machine Interactive Learning for Multimodal Emotion Recognition in Dialogue System'. Together they form a unique fingerprint.

Cite this