TY - GEN
T1 - Context-Aware Head-and-Eye Motion Generation with Diffusion Model
AU - Shen, Yuxin
AU - Xu, Manjie
AU - Liang, Wei
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - In humanity's ongoing quest to craft natural and realistic avatars within virtual environments, the generation of authentic eye gaze behaviors stands paramount. Eye gaze not only serves as a primary non-verbal communication cue, but it also reflects cognitive processes, intent, and attentiveness, making it a crucial element in ensuring immersive interactions. However, automatically generating these intricate gaze behaviors presents significant challenges. Traditional methods can be both time-consuming and lack the precision to align gaze behaviors with the intricate nuances of the environment in which the avatar resides. To overcome these challenges, we introduce a novel two-stage approach to generate context-aware head-and-eye motions across diverse scenes. By harnessing the capabilities of advanced diffusion models, our approach adeptly produces contextually appropriate eye gaze points, further leading to the generation of natural head-and-eye movements. Utilizing Head-Mounted Display (HMD) eye-tracking technology, we also present a comprehensive dataset, which captures human eye gaze behaviors in tandem with associated scene features. We show that our approach consistently delivers intuitive and lifelike head-and-eye motions and demonstrates superior performance in terms of motion fluidity, alignment with contextual cues, and overall user satisfaction.
AB - In humanity's ongoing quest to craft natural and realistic avatars within virtual environments, the generation of authentic eye gaze behaviors stands paramount. Eye gaze not only serves as a primary non-verbal communication cue, but it also reflects cognitive processes, intent, and attentiveness, making it a crucial element in ensuring immersive interactions. However, automatically generating these intricate gaze behaviors presents significant challenges. Traditional methods can be both time-consuming and lack the precision to align gaze behaviors with the intricate nuances of the environment in which the avatar resides. To overcome these challenges, we introduce a novel two-stage approach to generate context-aware head-and-eye motions across diverse scenes. By harnessing the capabilities of advanced diffusion models, our approach adeptly produces contextually appropriate eye gaze points, further leading to the generation of natural head-and-eye movements. Utilizing Head-Mounted Display (HMD) eye-tracking technology, we also present a comprehensive dataset, which captures human eye gaze behaviors in tandem with associated scene features. We show that our approach consistently delivers intuitive and lifelike head-and-eye motions and demonstrates superior performance in terms of motion fluidity, alignment with contextual cues, and overall user satisfaction.
KW - Human computer interaction (HCI)
KW - Human-centered computing
KW - Interaction paradigms
KW - Virtual reality
UR - http://www.scopus.com/inward/record.url?scp=85191455492&partnerID=8YFLogxK
U2 - 10.1109/VR58804.2024.00039
DO - 10.1109/VR58804.2024.00039
M3 - Conference contribution
AN - SCOPUS:85191455492
T3 - Proceedings - 2024 IEEE Conference on Virtual Reality and 3D User Interfaces, VR 2024
SP - 157
EP - 167
BT - Proceedings - 2024 IEEE Conference on Virtual Reality and 3D User Interfaces, VR 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 31st IEEE Conference on Virtual Reality and 3D User Interfaces, VR 2024
Y2 - 16 March 2024 through 21 March 2024
ER -