Apple has been granted a patent for a “File Format for Spatial Audio.”

About the patent

Apple says Spatial Audio gives artists the opportunity to create immersive audio experiences for their fans with true multidimensional sound and clarity. Dolby Atmos is an immersive audio experience that enables artists to mix music so the sound comes from all around and from above. The newly granted patent relates generally to computerized data processing systems and methods for audio processing, and in particular to spatial audio processing in a simulated reality environment.

In the patent, Apple says that producing three-dimensional (3D) sound effects in augmented reality (AR), virtual reality (VR), and mixed reality (MR) applications, all of which in this disclosure are encompassed by the term simulated reality (SR), is challenging because existing audio formats were originally designed for producing 3D sound in a physical environment with fixed speaker locations and stationary listeners, such as in a movie theater. Examples of spatial audio formats designed to produce 3D sound include MPEG-H (Moving Picture Experts Group) 3D Audio standards, HOA (Higher-order Ambisonics) spatial audio techniques, and DOLBY ATMOS surround sound technology.

One alternative for producing 3D sound effects in SR environments is to manipulate individual discrete sounds contained in audio objects that can be virtually located anywhere in the 3D environment. Composing audio for SR applications using existing spatial audio formats and objects is difficult since there is no uniform way to access a variety of sound sources and incorporate them into a dynamic SR environment. Apple’s new patent addresses such issues.

Summary of the patent

Here’s Apple’s abstract of the patent: “An audio asset library containing audio assets formatted in accordance with a file format for spatial audio includes asset metadata that enables simulated reality (SR) application developers to compose sounds for use in SR applications. The audio assets are formatted to include audio data encoding a sound capable of being composed into a SR application along with asset metadata describing not only how the sound was encoded, but also how a listener in SR environment experiences the sound. 

“A SR developer platform is configured so that developers can compose sound for SR objects using audio assets stored in the audio library, including editing the asset metadata to include transformation parameters that support dynamic transformation of the asset metadata in the SR environment to alter how the SR listener experiences the composed sound. Other embodiments are also described and claimed.”




Article provided with permission from AppleWorld.Today