How we locate sound in nature is the basis for how well stereo imaging works AND is the core concept for surround sound systems.
The way the brain processes the aural signals from our ears is complicated, but we know some things extremely well based on research which goes back to the 1920s and has been refined and perfected since then. There are still researchers attempting to master sound localization in order to create headphone solutions for virtual reality and improve the lives of those with hearing issues and brain function issues.
Basically, we instinctively and without any conscious awareness use two methods to pinpoint the location of a sound on the horizontal plane.
Intra-aural delay
First is the difference in arrival of the sound to our ears. Our brains, rely on very high order skills to process the difference in time between the excitement of one ear from the other. A short sound, like a bump, click, knock, snap, tap, etc. to the right of us will reach the right ear a few milliseconds before it reaches the left ear. Our brains are VERY good at recognizing that the exact difference in that arrival time and fairly accurately predicting about how far to the right the sound came from. That's for one click. If there are additional clicks from the same location, we will ever so slightly move our heads and recognize how the time difference changes, and we will keep estimating location on subsequent clicks until we can hone in on the exact location. If we really want to pinpoint the sound, we will move our hears until the delay is no longer present then the sound should be directly in front of us and we should be able to see the source of the sound with our eyes by looking directly forward and changing our focus point and scanning up and down and a tiny amount side to side to make up for limitations in our aural location skills and any acoustical artifacts which might be reducing the accuracy of our hearing skills.
This use of delay time is limited to frequencies where the distance between our ears is more than one full wavelength of the sound source. So, if our ears are about 7 inches apart, then our ability to differentiate delay between our ears is limited to a high frequency of about 2,000Hz. It also happens that our heads block sound from wrapping around our heads at about the same frequency, so frequencies above about 2,000Hz are attenuated or completely blocked from reaching the opposite ear. As the frequency gets higher, our head is more effective at blocking the sound. Also, as the sound source moves farther to the side, the more of the head actually blocks the sound from reach the opposite ear. So, like all filters in the real world, our brain goes from relying entirely on intra-ear delay to locate a sound in the midrange to not being able to use time delay at all to help locate a sound. It may start getting more difficult to locate a sound around 2,000Hz and it gets more and more difficult as the frequency increases until at a certain point we cannot use delay at all. It turns out that for most humans we cannot use time delay at all above about 5,000Hz, or so. Also, it is believed that skill can be trained over long periods by using headphones with processed delay information at higher frequencies. This is helpful for fighter pilots who wear headphones and have a pressing reason to have more information about location.
We also know that this ability to rely on intra-aural delay information to locate sound sources is limited in the bass. However, it is almost impossible to pinpoint bass sounds in the bass more because of the way bass frequency acoustic wave propagate in a space AND the nature of bass sound sources. In nature, it is rare that a bass sound will be short with the majority of the energy at the very front of the waveform. When you drop a rock off a cliff, at 100 feet away the bass levels are louder when the echoes and reverberant sound arrives than in the initial impact. So our brains are trained to rely less heavily on delay in arrival times when it comes to bass frequencies. In an enclosed space, bass sounds tend to reverberate around for over a second, so it is even harder for our brains to locate a sound source based on low frequency information alone. Most studies have found that it is extremely rare for any human, no matter how well trained, to located a bass frequency sound source at any frequency below about 100Hz, or so. Most people cannot locate a bass source below about 150Hz.
Thus, intra-aural arrival time information is used by our brains in the range from about 150Hz to about 2,000Hz - with the common variations between people's physiology and recognizing it isn't a brick wall stop at those end frequencies.
That's one way we locate sounds.
The way the brain processes the aural signals from our ears is complicated, but we know some things extremely well based on research which goes back to the 1920s and has been refined and perfected since then. There are still researchers attempting to master sound localization in order to create headphone solutions for virtual reality and improve the lives of those with hearing issues and brain function issues.
Basically, we instinctively and without any conscious awareness use two methods to pinpoint the location of a sound on the horizontal plane.
Intra-aural delay
First is the difference in arrival of the sound to our ears. Our brains, rely on very high order skills to process the difference in time between the excitement of one ear from the other. A short sound, like a bump, click, knock, snap, tap, etc. to the right of us will reach the right ear a few milliseconds before it reaches the left ear. Our brains are VERY good at recognizing that the exact difference in that arrival time and fairly accurately predicting about how far to the right the sound came from. That's for one click. If there are additional clicks from the same location, we will ever so slightly move our heads and recognize how the time difference changes, and we will keep estimating location on subsequent clicks until we can hone in on the exact location. If we really want to pinpoint the sound, we will move our hears until the delay is no longer present then the sound should be directly in front of us and we should be able to see the source of the sound with our eyes by looking directly forward and changing our focus point and scanning up and down and a tiny amount side to side to make up for limitations in our aural location skills and any acoustical artifacts which might be reducing the accuracy of our hearing skills.
This use of delay time is limited to frequencies where the distance between our ears is more than one full wavelength of the sound source. So, if our ears are about 7 inches apart, then our ability to differentiate delay between our ears is limited to a high frequency of about 2,000Hz. It also happens that our heads block sound from wrapping around our heads at about the same frequency, so frequencies above about 2,000Hz are attenuated or completely blocked from reaching the opposite ear. As the frequency gets higher, our head is more effective at blocking the sound. Also, as the sound source moves farther to the side, the more of the head actually blocks the sound from reach the opposite ear. So, like all filters in the real world, our brain goes from relying entirely on intra-ear delay to locate a sound in the midrange to not being able to use time delay at all to help locate a sound. It may start getting more difficult to locate a sound around 2,000Hz and it gets more and more difficult as the frequency increases until at a certain point we cannot use delay at all. It turns out that for most humans we cannot use time delay at all above about 5,000Hz, or so. Also, it is believed that skill can be trained over long periods by using headphones with processed delay information at higher frequencies. This is helpful for fighter pilots who wear headphones and have a pressing reason to have more information about location.
We also know that this ability to rely on intra-aural delay information to locate sound sources is limited in the bass. However, it is almost impossible to pinpoint bass sounds in the bass more because of the way bass frequency acoustic wave propagate in a space AND the nature of bass sound sources. In nature, it is rare that a bass sound will be short with the majority of the energy at the very front of the waveform. When you drop a rock off a cliff, at 100 feet away the bass levels are louder when the echoes and reverberant sound arrives than in the initial impact. So our brains are trained to rely less heavily on delay in arrival times when it comes to bass frequencies. In an enclosed space, bass sounds tend to reverberate around for over a second, so it is even harder for our brains to locate a sound source based on low frequency information alone. Most studies have found that it is extremely rare for any human, no matter how well trained, to located a bass frequency sound source at any frequency below about 100Hz, or so. Most people cannot locate a bass source below about 150Hz.
Thus, intra-aural arrival time information is used by our brains in the range from about 150Hz to about 2,000Hz - with the common variations between people's physiology and recognizing it isn't a brick wall stop at those end frequencies.
That's one way we locate sounds.