You are misinformed.
Amplitude, spectral, and timing are all integrated into a positional / distance probability mapping. Humans can estimate the vector of a sound by about 2 degrees horizontal and 4 degrees vertical. Distance is also pretty accurate, especially in a room where direct and reflected sounds will arrive at different times, creating interference patterns.
The brain processes audio in a way not too dissimilar from the way that medical imaging scanners can use a small number of sensors to develop a detailed 3d image.
In a perfectly dark room, you can feel large objects by the void they make in the acoustic space of ambient noise and reflected sounds from your own body.
Interestingly, the shape of the ear is such that different phase shifts occur for front and rear positions of reflected and conducted sounds, further improving localization.
We often underestimate the information richness of the sonic sensome, as most live in a culture that deeply favors the visual environment, but some subcultures and also indigenous cultures have learned to more fully explore those sensory spaces.
People of the extreme northern latitudes may spend a much larger percentage of their waking hours in darkness or overwhelming white environments and learn to rely more on sound to sense their surroundings.
I learned to move around in dark rooms when I was young, I definitely can "feel large objects by the void they make" and people often turn on the lights because they thing I need them to "see" when I really don't.
> Humans can estimate the vector of a sound by about 2 degrees horizontal and 4 degrees vertical.
Yet I'm usually not even noticing whether a video has stereo or mono sound. So I highly doubt that ultra precise OLED loudspeakers would make a noticeable difference.