Ethical Alignment Decision Making for Connected Autonomous Vehicle in Traffic Dilemmas via Reinforcement Learning From Human Feedback

Xin Gao, Tian Luan, Xueyuan Li*, Qi Liu, Zhaoyang Ma, Xiaoqiang Meng, Zirui Li*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Since the introduction of the trolley problem, the ethical decision-making conundrum has evolved from autonomous vehicles (AVs) to connected AVs (CAVs), continuing as a prominent challenge. When confronted with ethical dilemmas, CAVs must align their responses not merely with value-neutral human preferences but also with broader moral and ethical frameworks. Consequently, to ensure that CAVs do not engage in actions that contravene established human moral principles, it is imperative that the ethical considerations are meticulously integrated into their decision-making systems. In this article, we introduce an innovative multiscale multimodal ethical network (M2ENet), which aims to align the AV decision-making system with human ethical feedback in ethical dilemma scenarios. First, we extract morphological and dynamic features from sensory information and signal data, respectively, using multiscale multimodal representation. Additionally, ethical policy-based network is devised to enable AVs to comprehend ethical information, which includes the introduction of an ethical alignment factor to ethically align the feature matrix from human feedback. Furthermore, the accuracy of ethical interaction information is improved through coupled ethical module informed by human feedback. Finally, the efficacy of the system is demonstrated through three representative ethical dilemmas in traffic scenarios, employing both simulation experiments and hardware-in-the-loop testing. The simulation experiments reveal that our proposed model can generate decision-making strategies more aligned with human preferences in ethical traffic scenarios. In addition, in our hardware-in-the-loop tests, it is observed that the average percentage of ethical bias weights decreases by 45.06% after 150 episodes of training.

Original languageEnglish
Pages (from-to)38585-38600
Number of pages16
JournalIEEE Internet of Things Journal
Volume11
Issue number23
DOIs
Publication statusPublished - 2024

Keywords

  • Coupled ethical module
  • ethical alignment decision making
  • multiscale multimodal ethical network
  • reinforcement learning from human feedback

Fingerprint

Dive into the research topics of 'Ethical Alignment Decision Making for Connected Autonomous Vehicle in Traffic Dilemmas via Reinforcement Learning From Human Feedback'. Together they form a unique fingerprint.

Cite this