Advanced Search
Li Guoan, Liu Junchen, Wang Miao. Staged Gaze Prediction in Virtual Scene Interaction Tasks[J]. Journal of Computer-Aided Design & Computer Graphics, 2025, 37(2): 207-215. DOI: 10.3724/SP.J.1089.2023-00333
Citation: Li Guoan, Liu Junchen, Wang Miao. Staged Gaze Prediction in Virtual Scene Interaction Tasks[J]. Journal of Computer-Aided Design & Computer Graphics, 2025, 37(2): 207-215. DOI: 10.3724/SP.J.1089.2023-00333

Staged Gaze Prediction in Virtual Scene Interaction Tasks

  • Gaze prediction method refers to an algorithm model that predicts the user’s current gaze direction through various types of user information. However, current methods for predicting gaze in virtual scenes typically rely on generalized models and still have considerable room for improvement in specific interactive tasks. This paper focuses on improving gaze prediction for the interactive task flow of finding-locking onto-approaching target objects in virtual scenes. We first construct the first dataset for this task, consisting of gaze recordings, object, helmet and controller parameters, as well as recorded videos, during five interacting tasks performed by 21 users in three interactive scenes. The users’ interaction process is divided into three stages: finding target objects, locking onto target objects, and approaching target objects. We then conduct phase-wise correlation analysis, selecting the parameter set with the highest correlation with gaze to input into the network for training. The proposed method is validated on the self-constructed dataset, achieving a gaze prediction error of 2.60°, which represents a 21.45% improvement over the current SOTA method’s error of 3.31°, significantly enhancing gaze prediction accuracy for this task scenario.
  • loading

Catalog

    Turn off MathJax
    Article Contents

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return