The changes that occur in pupil sizes during a virtual scene are indicative of user engagement and cognitive load for different tasks in an environment such as the Virtual scene. It may also serve as a preferred physiological measure for an applied setting. In scenarios where scene brightness cannot be controlled, the pupil response can be correlated to cognitive-emotional response as leverage in pupil indexing in users. The possibility of predicting the pupillary response of an individual’s pupil by factoring out measurement is a major challenge, but this can be modelled through a virtual scene. In this paper, we evaluate the optimal response in pupil response to the virtual scenes through a non-immerse method of augmented reality environment. The result shows error in pupil response is less visible when compared to the visual output in mapped fixation of the XYZ-coordinate and an exponential coordinate of the participant’s
user-friendly calibration to the virtual scene. This procedure can lead to a pilot study involving training in a 3D environment.