Modeling the gaze point distribution to assist eye-based target selection in head-mounted displays

Ting Lei, Jing Chen*, Jixiang Chen, Bo Liu

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

Due to its natural and fast characteristics, eye tracking, as a novel input modality, has been widely used in head-mounted displays for interaction. However, because of the inadvertent jitter of eyes and limitations of eye tracking devices, the eye-based selection often performs poorly in accuracy and stability compared with other input modalities, especially for small targets. To address this issue, we built a likelihood model by modeling the gaze point distribution and then combined it with Bayesian rules to infer the intended target from the perspective of probability as an alternative to the traditional selection criteria based on boundary judgment. Our investigation shows that using our model improves the selection performance significantly over the conventional ray-casting selection method and using the existing optimal likelihood model, especially in the selection of small targets.

Original languageEnglish
Pages (from-to)25069-25081
Number of pages13
JournalNeural Computing and Applications
Volume35
Issue number36
DOIs
Publication statusPublished - Dec 2023

Keywords

  • Augmented reality
  • Eye tracking
  • Gaze interaction
  • Virtual reality

Fingerprint

Dive into the research topics of 'Modeling the gaze point distribution to assist eye-based target selection in head-mounted displays'. Together they form a unique fingerprint.

Cite this