Sample Code

Creating an Immersive AR Experience with Audio

Use sound effects and environmental sound layers to create an engaging AR experience.



This sample app uses SceneKit’s node-based audio API to associate environmental sounds with a virtual object that’s placed in the real world. Because audio is 3D positional in SceneKit by default, volume is automatically mixed based on the user’s distance from a node.

Getting Started

  • This sample code supports Relocalization and therefore, it requires ARKit 1.5 (iOS 11.3) or greater

  • ARKit is not available in the iOS Simulator

  • Building the sample requires Xcode 9.3 or later

Run an AR Session and Place Virtual Content

Before you can use audio, you need to set up a session and place the object from which to play sound. For simplicity, this sample runs a world tracking configuration and places a virtual object on the first horizontal plane that it detects. For more detail about this kind of session setup, see Tracking and Visualizing Planes. The object placement approach in this sample is similar to the one demonstrated in Placing Objects and Handling 3D Interaction.

Add 3D Audio to the Scene

To play audio from a given position in 3D space, create an SCNAudioSource from an audio file. This sample loads the file from the bundle in viewDidLoad:

// Instantiate the audio source
audioSource = SCNAudioSource(fileNamed: "fireplace.mp3")!

Then, the audio source is configured and prepared:

// As an environmental sound layer, audio should play indefinitely
audioSource.loops = true
// Decode the audio from disk ahead of time to prevent a delay in playback

When you’re ready to play the sound, create an SCNAudioPlayer, passing it the audio source:

// Create a player from the source and add it to `objectNode`
objectNode.addAudioPlayer(SCNAudioPlayer(source: audioSource))