TY - GEN
T1 - PERCY
T2 - 38th Australasian Joint Conference on Artificial Intelligence, AI 2025
AU - Meng, Zhijin
AU - Althubyani, Mohammed
AU - Xie, Shengyuan
AU - Razzak, Imran
AU - Sandoval, Eduardo B.
AU - Bamdad, Mahdi
AU - Cruz, Francisco
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2026.
PY - 2026
Y1 - 2026
N2 - Traditional rule-based conversational robots, constrained by fixed scripts and static response mappings, fundamentally lack adaptability for sustained personalized human interaction. Although large language models (LLMs) such as GPT-4 enable open-domain dialogue capabilities, most existing social robot approaches remain deficient in emotional awareness and longitudinal personalization continuity. To address this critical gap, we present PERCY (Personal Emotional Robotic Conversational sYstem) – an innovative framework that dynamically integrates: (1) real-time affective signals through facial expression recognition, (2) semantic content of user utterances, and (3) contextual profile data, synthesizing these multimodal inputs into emotion-aware prompt engineering for GPT-4. This integration drives both contextually appropriate verbal responses and synchronized non-verbal robot behaviors. PERCY utilizes GPT-4 to dynamically model the robot’s internal affective state, with non-verbal feedback primarily expressed through facial expressions. The system architecture leverages ROS-based multimodal processing: visual emotion recognition via fine-tuned MobileNetV2, textual sentiment analysis using NLTK’s VADER, decision-level sensor fusion, and GPT-4 prompt conditioning to orchestrate ARI robot behaviors. Empirical evaluation with 30 human participants demonstrated statistically significant improvements in dialogue coherence, contextual relevance, and response diversity compared to baseline systems. PERCY highlights the potential of integrating advanced multimodal perception and personalization to build a scalable foundation for next-generation emotionally intelligent human-robot interaction systems, rooted in contextually conditioned, multimodal affective computing.
AB - Traditional rule-based conversational robots, constrained by fixed scripts and static response mappings, fundamentally lack adaptability for sustained personalized human interaction. Although large language models (LLMs) such as GPT-4 enable open-domain dialogue capabilities, most existing social robot approaches remain deficient in emotional awareness and longitudinal personalization continuity. To address this critical gap, we present PERCY (Personal Emotional Robotic Conversational sYstem) – an innovative framework that dynamically integrates: (1) real-time affective signals through facial expression recognition, (2) semantic content of user utterances, and (3) contextual profile data, synthesizing these multimodal inputs into emotion-aware prompt engineering for GPT-4. This integration drives both contextually appropriate verbal responses and synchronized non-verbal robot behaviors. PERCY utilizes GPT-4 to dynamically model the robot’s internal affective state, with non-verbal feedback primarily expressed through facial expressions. The system architecture leverages ROS-based multimodal processing: visual emotion recognition via fine-tuned MobileNetV2, textual sentiment analysis using NLTK’s VADER, decision-level sensor fusion, and GPT-4 prompt conditioning to orchestrate ARI robot behaviors. Empirical evaluation with 30 human participants demonstrated statistically significant improvements in dialogue coherence, contextual relevance, and response diversity compared to baseline systems. PERCY highlights the potential of integrating advanced multimodal perception and personalization to build a scalable foundation for next-generation emotionally intelligent human-robot interaction systems, rooted in contextually conditioned, multimodal affective computing.
KW - Cognitive modelling and computer-human interaction
KW - Human-Robot Interaction
KW - Social Robotics
UR - https://www.scopus.com/pages/publications/105023828238
U2 - 10.1007/978-981-95-4972-6_36
DO - 10.1007/978-981-95-4972-6_36
M3 - Conference contribution
AN - SCOPUS:105023828238
SN - 9789819549719
T3 - Lecture Notes in Computer Science
SP - 466
EP - 478
BT - AI 2025
A2 - Liu, Miaomiao
A2 - Yu, Xin
A2 - Xu, Chang
A2 - Song, Yiliao
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 1 December 2025 through 5 December 2025
ER -