Skip to main navigation Skip to search Skip to main content

AI apology: interactive multi-objective reinforcement learning for human-aligned AI

  • Hadassah Harland
  • , Richard Dazeley
  • , Bahareh Nakisa
  • , Francisco Cruz
  • , Peter Vamplew
  • Deakin University
  • UNSW Sydney
  • Federation University Australia

Research output: Contribution to journalArticlepeer-review

7 Scopus citations

Abstract

For an Artificially Intelligent (AI) system to maintain alignment between human desires and its behaviour, it is important that the AI account for human preferences. This paper proposes and empirically evaluates the first approach to aligning agent behaviour to human preference via an apologetic framework. In practice, an apology may consist of an acknowledgement, an explanation and an intention for the improvement of future behaviour. We propose that such an apology, provided in response to recognition of undesirable behaviour, is one way in which an AI agent may both be transparent and trustworthy to a human user. Furthermore, that behavioural adaptation as part of apology is a viable approach to correct against undesirable behaviours. The Act-Assess-Apologise framework potentially could address both the practical and social needs of a human user, to recognise and make reparations against prior undesirable behaviour and adjust for the future. Applied to a dual-auxiliary impact minimisation problem, the apologetic agent had a near perfect determination and apology provision accuracy in several non-trivial configurations. The agent subsequently demonstrated behaviour alignment with success that included up to complete avoidance of the impacts described by these objectives in some scenarios.

Original languageEnglish
Pages (from-to)16917-16930
Number of pages14
JournalNeural Computing and Applications
Volume35
Issue number23
DOIs
StatePublished - Aug 2023

Keywords

  • AI apology
  • AI safety
  • Human alignment
  • Impact minimisation
  • Multi-objective reinforcement learning

Fingerprint

Dive into the research topics of 'AI apology: interactive multi-objective reinforcement learning for human-aligned AI'. Together they form a unique fingerprint.

Cite this