Hello, esteemed tech developer. I am using the Apple Vision Pro to create an AR assist system about the da Vinci Surgical Robot in a medical surgical suite, and would like to capture eye movement data with tester uniformity. Although the Apple Vision Pro has a superb infrared sensor to monitor eye movement status, Apple does not seem to have open access officially. (I'm aware of many existing discussions about this, but I was still wondering if there might be an option, particularly for research labs.)Here's my FB number.FB16603687
Hi @DarkBule
Thanks for filing the feedback request. While you can't obtain gaze information, you can make entities visually respond to it. Here are 2 ways:
- HoverEffectComponent: A component that applies a visual effect to a hierarchy of entities when a person looks at or selects an entity.
- Hover State (Shader Graph Node): Provides access to gaze information in a shader. This gives you more precise control than
HoverEffectComponent
.
I realize this may not enable your use case, but want to make you (and future readers of this post) aware of your current options.