Skip to main navigation Skip to search Skip to main content

MERCI: A Multimodal Dataset for Personalised and Emotionally-Aware Dialogues

  • Mohammed Althubyani
  • , Zhijin Meng
  • , Shengyuan Xie
  • , Francisco Cruz
  • , Imran Razzak
  • , Mukesh Prasad
  • , Eduardo B. Sandoval
  • , Baki Kocaballi
  • University of Technology Sydney
  • UNSW Sydney
  • Mohamed Bin Zayed University of Artificial Intelligence

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

The integration of conversational agents into daily life has become increasingly common. However, sustaining deeply engaging and natural interactions remains challenging due to a lack of multimodal datasets capturing personal and emotional nuances. In this paper, we introduce MERCI (Multimodal dataset for Emotionally-aware peRsonalised Conversational In-teractions), a dataset derived from user-robot dialogues involving thirty participants who completed user profile questionnaires covering ten personal topics (e.g., hobbies, music). A conver-sational system called PERCY then engaged with each partici-pant in open-domain conversations, leveraging GPT-4, real-time facial-expression and sentiment analysis to generate contextu-ally appropriate, empathetic responses. MERCI contains 1860 utterances, equating to about 12.5 hours of aligned audio, three-view video, transcripts with timestamps, emotion labels, and sentiment scores. This dataset serves as a reproducible test-bed for tasks such as emotion-aware response generation, multimodal affect recognition, and personalised policy learning. Baseline performance results have been established using advanced models such as BERT, T5, BART, and GPT-3.5/4/4o-mini across gener-ation, regression, and classification. Evaluations through human and automated methods have demonstrated strong naturalness, relevance, and consistency in responses while indicating areas for enhanced personalisation and empathic depth. We expect that MERCI will enhance the development of emotionally intelligent, user-centric conversational AI applications, potentially ranging from social robotics to mental health support.

Original languageEnglish
Title of host publicationCBMI 2025 - 2025 International Conference on Content-Based Multimedia Indexing, Conference Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798331555009
DOIs
StatePublished - 2025
Event2025 International Conference on Content-Based Multimedia Indexing, CBMI 2025 - Dublin, Ireland
Duration: 22 Oct 202524 Oct 2025

Publication series

NameCBMI 2025 - 2025 International Conference on Content-Based Multimedia Indexing, Conference Proceedings

Conference

Conference2025 International Conference on Content-Based Multimedia Indexing, CBMI 2025
Country/TerritoryIreland
CityDublin
Period22/10/2524/10/25

Keywords

  • dialogue systems
  • empathetic conversational agents
  • human-robot interaction
  • multimodal dataset
  • personalisation

Fingerprint

Dive into the research topics of 'MERCI: A Multimodal Dataset for Personalised and Emotionally-Aware Dialogues'. Together they form a unique fingerprint.

Cite this