Scientific-Grade Eye Tracking Software for VR Human Behavior Research
March 22, 2021
Peter Schlueer & Sado Rabaudi
The WorldViz VR Eye Tracking Analytics Lab lets you set up VR eye tracking experiments - and, it’s a perfect fit with the Vizard VR software and the StarVR One headset with built-in eye tracking.
Tracking eye movement in virtual reality at research quality has recently opened wide possibilities for academic research. At WorldViz, our goal is to make eye tracking data available to human behavior researchers in a standard manner and with as much low-level control exposed as possible across manufacturers of leading hardware.
An ideal match for WorldViz VR software, the StarVR One VR headset with its built-in eye tracking gives you a seamless workflow for creating human behavioral experiments and analyzing eye tracking data.
“With a 210-degree horizontal and 130-degree vertical field-of-view, the StarVR One headset opens new avenues for studying human behavior under real-world conditions, while the integrated Tobii eye-tracking technology along with our Eye-Tracking Analytics Lab software delivers a powerhouse of scientific-grade data analysis capabilities.” - Andy Beall, WorldViz Chief Scientist and Co-Founder
Let’s dive right in how we put eye tracking data at your fingertips for research purposes. Our VR Eye Tracking Analytics Lab is a simple yet powerful tool for setting up eye tracking experiments in VR that include examples for common eye tracking tasks, multi user environments, heat maps, access to a support ticketing system for professional users, and more. The VR Eye Tracking Analytics Lab runs on our Vizard VR software for researchers, which allows native integration with thousands of Python libraries.
Eye Tracking Software Features
What can you do with our VR Eye Tracking Analytics Lab? We’re most excited about these features:
Recording and playback eye tracking behavior for "after action review" including 3D path review
Extensive data analytics
3D Visualization of data, such as heat maps
User performance triggered feedback loops with eye or physiological sensor data
Precise timing experimental control and device synchronization
360 videos and 3D files from a wide array of sources as customizable stimuli
Recording of gaze direction data, pupil size, fixation timings, and other low level parameters