This is actually quite amazing and one of the fields I'm most interested in.
Spatial audio / 3d audio is much more immersive and less gimmicky than 3D video, but I feel like people don't really care too much because they were so underwhelmed with 3D video. It's one of those things with high fidelity audio you can't tell a difference between what is real and what is fake.
Edgar Choueiri, a Princeton professor, was the first I know of that tried to commercialize his algorithm for spatial audio. I believe he got a deal with Jawbone to put in Jambox as a feature called "live audio mode". He did try to sell some hardware for audiophiles to build out home systems, though I'm pretty sure they were insanely expensive (10-100k range). https://www.princeton.edu/3D3A/index.html
There's also a great podcast that tells stories using only spatial audio. They're really immersive and sometimes I can't tell if the audio is happening in the podcast or outside my room. It's called The Truth: http://www.thetruthpodcast.com/
Makes me happy to see tools like this are open to developers now. I expect to see a lot more content featuring this technology now.
If you want to hear a great example of spatial audio, check out The Quiet American [0] and be sure to use earbuds. He records his environment using two in-ear microphones to get all of the effects discussed in the article: sound shaped by the external parts of the ear, timing differences between the left and right ear, etc.
I remember listening to one recording where he is at the edge of a lake or ocean, and listening while standing with my eyes closed I almost expected to feel the waves around my ankles. It's incredibly immersive, and will work wonders for VR.
Back in the day 'spatial audio' was excellent for what it was in VRML, as played through the SGI viewer. I was sad VRML never happened as I quite enjoyed worlds where you could play different songs in different virtual rooms able to mix albeit not to the beat between different tracks by moving around. A fantastic music experience could be had.
This new take on 3D audio just makes me want to go back to 1998 and what was achievable then in 3D on the web if you were willing to do a 20Mb download over a modem to get the plugin required to 'see' VRML. The actual hurdle was use case, instead of a 3D web we ended up with a 2D flat table web. If you got good at VRML and used all of the features and built significant sized worlds then you soon realised that you were on your own, few others in the world had got that far, their models were more like Duplo Lego rather than Lego Technical.
The alleged 'bar' back then was the download for the plugin, that allegedly prevented VRML from catching on. The authorship tools were awesome - Maya like - and yet there were few takers. CosmoWorlds was awesome, has to be said. The lack of content creators was the problem, 2D Flash was the way people went, nobody was thinking in 3D.
So when can we start doing cool 3D just on the web in a web browser that you don't need specialist hardware for? With 3D spatial audio working too?
I've never really been impressed by HRTFs compared to true Binaural recordings (e.g. virtual barbershop) There's something happening in the binaural recordings for positioning that just isn't captured in HRTFs it seems.
Nice to see audio finally getting the attention it deserves! And seems like VR is the driving factor. I wonder how this compares to NVIDIA VRWorks Audio[1]?
When you think about it, 3D audio has many similarities to 3D video: rays of sound(/light) traveling through open space and attenuating in the process, bouncing off surfaces whose material filters different frequencies. From this point of view, in hindsight, the accelerated sound cards from ’00s seem a bit of a misstep – isn’t this something that’s better computed on a GPU?
Pretty much every third party sound library for games implements 3D audio support (commercial ones probably for the last 10 years or more) but in order to use more than just the spatial sound source location/user facing vector requires inputting the material types/locations/sizes/audio properties for everything in the game world (to calculate audio propagation loss/occlusion) and using a certain percentage of cycles per frame to this - not everyone chooses to dedicate this much work to audio or has the extra CPU cycles to do this after graphics/physics. Also they have to be able to support simulating the head transfer function via speaker (many possible configurations) or headphone so that's doubled or tripled the amount of QA for audio sfx instead of relying on a VR headset for a standard setup for VR, in addition to possibly adding some sort of customization for each individuals head transfer function. If your games don't really use it, it's because the developer didn't have the time/money/CPU budget for it because it probably wasn't a key feature.
Not sure what is different about Google's offering in VR space versus tradtional game audio engines here.
yes, HRTF, just typing hastily. If one enables HRTF support for instance, as this quick googleing finds support in an eight year old query on FMOD,
http://www.fmod.org/questions/question/forum-29423/
one needs to support/not support/test all the cases, with the HRTF enabled/disabled, user deciding to use headphones/speakers with whatever configuration they have. Maybe one will decide it sounds better with HRTF on with speakers or decide that's not a good idea and not allow it or not even support HRTF. It's up to whomever made the game to determine how to support whatever the user decides to use. So in the general case of games not just VR, one might have a console version and a PC version and may need to support headphones/2 speakers/multiple speaker set ups and enabling/disabling features on demand per configuration. So back to the original query this requires signficantly more QA/testing than supporting just HRTF with VR headphones.
More than impossible, it's the wrong thing. Speakers in a room don't move with your head when you turn it. By the time the sound gets to your ears, it's already had an HRTF applied - the natural one. The thing to go for there is wavefield synthesis, not any flavor of HRTF.
> I don't think OpenAL simulates some of the more complex phenomena having to do with how sound interacts with your body and ears?
It does. This blog post [0] seems to be a nice reference. What you want is called "head-related transfer function". (There are tons of papers on HRTF)
One can even turn it on for any game that use openal, even old ones. You can go download The Dark Mod now and play it with HRTF [1].
Note: am not an expert
What openal doesn't do, while this google thing does, is reverb (reflection of waves, echo). One way to do reverb is.. well simply raytrace it. Just send out a bunch of rays from the source, probably do some filtering of the results, and apply to the audio. Another way is to take a real room that you are "porting" to digital, ping somewhere in it (play the impulse response of an ideal 20kHz low pass filter, aka the sync function), then measure somewhere else. Then repeat in a.. grid or something. What you get is a bunch of filters (num_points if there is a single static "listener", or num_points^2 if there are more and or moving; as it is source->destination dependent). This can also be done by shooting rays in an arbitrary digital space, but i wanted people to imagine a room and waves going around in it. AFAIK this is how the valve bought software works.
Another thing is diffraction[2]... that idk if either gugl or volvo do. Calculate it with voxel grids then add to refraction ? Any renderers do proper Rayleigh scattering ?
Spatial audio / 3d audio is much more immersive and less gimmicky than 3D video, but I feel like people don't really care too much because they were so underwhelmed with 3D video. It's one of those things with high fidelity audio you can't tell a difference between what is real and what is fake.
Edgar Choueiri, a Princeton professor, was the first I know of that tried to commercialize his algorithm for spatial audio. I believe he got a deal with Jawbone to put in Jambox as a feature called "live audio mode". He did try to sell some hardware for audiophiles to build out home systems, though I'm pretty sure they were insanely expensive (10-100k range). https://www.princeton.edu/3D3A/index.html
There's also a great podcast that tells stories using only spatial audio. They're really immersive and sometimes I can't tell if the audio is happening in the podcast or outside my room. It's called The Truth: http://www.thetruthpodcast.com/
Makes me happy to see tools like this are open to developers now. I expect to see a lot more content featuring this technology now.