Personalized decision-making for agents in face-to-face interaction in virtual reality

Xiaonuo Dongye, Dongdong Weng*, Haiyan Jiang*, Zeyu Tian, Yihua Bao, Pukun Chen

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Intelligent agents for face-to-face interaction in virtual reality are expected to make decisions and provide appropriate feedback based on the user’s multimodal interaction inputs. Designing the agent’s decision-making process poses a significant challenge owing to the limited availability of multimodal interaction decision-making datasets and the complexities associated with providing personalized interaction feedback to diverse users. To overcome these challenges, we propose a novel design framework that involves generating and labeling symbolic interaction data, pre-training a small-scale real-time decision-making network, collecting personalized interaction data within interactions, and fine-tuning the network using personalized data. We develop a prototype system to demonstrate our design framework, which utilizes interaction distances, head orientations, and hand postures as inputs in virtual reality. The agent is capable of delivering personalized feedback from different users. We evaluate the proposed design framework by demonstrating the utilization of large language models for data labeling, emphasizing reliability and robustness. Furthermore, we evaluate the incorporation of personalized data fine-tuning for decision-making networks within our design framework, highlighting its importance in improving the user interaction experience. The design principles of this framework can be further explored and applied to various domains involving virtual agents.

Original languageEnglish
Article number28
JournalMultimedia Systems
Volume31
Issue number1
DOIs
Publication statusPublished - Feb 2025

Keywords

  • Face-to-face interaction
  • Human–agent interaction
  • Large language model
  • Virtual reality

Cite this