Member-only story
Today I’m going to show you how to build a minimal Xcode project that does hand and head tracking on VisionOS 1.0 and 1.1. Since there are no virtual hands in the Vision Pro simulator, you will need a real device to try this out.
We’re going to start from Andy Kong’s Hand and Head Tracking Demo on Github.

First, clone the project locally or download as ZIP.

Open it and double-click the HeadHandsDemo.xcodeproj
file. This should launch Xcode, and you should select the “Trust and Open” option when prompted about security concerns.

You’ll need to set up the Vision Pro to connect to your computer. In the Vision Pro, navigate to General -> Settings -> Remote Devices, then follow the steps there. Afterwards, select your Vision Pro from the build targets dropdown. It should appear where I’m highlighting, but I do not have it connected at the moment.
The Vision Pro takes code uploads over WiFi, but your computer and the headset must be on the same network. If it’s a corporate network (like school or a workplace) then your computer may not detect it — in this case I recommend using a phone hotspot or taking it to a different network.

If you get build errors that mention realityos or something like that, you probably have not downloaded the necessary visionOS SDK. You can do that in the Xcode settings under the “Platforms” tab. If it doesn’t appear for you there, then either your MacOS or Xcode needs to be updated in order to get it.

Finally, you may get issues about signing and licenses — to solve this, click on the project main folder…