John
If I have this right the above from Craig is the argument about resolution or how many slices are being made at any given time.
From what I understand is it has no impact.
That's where the controversy lies. Of course reconstruction will reconstruct a waveform; no question about that, otherwise the outputs of digital audio systems would be stair-stepped instead of continuous. This isn't about reconstructing a waveform, but about reconstructing
a characteristic of the binaural listening experience which is, after all, how we hear sound.
The question is whether reconstruction is sufficiently precise to reconstruct the timing
difference between
two signals that are, say, 8 microseconds apart. I don't see how that's possible if the capture medium can't resolve differential timings under 21 microseconds.
Let me explain what I think is going on.
A 48kHz sample clock samples an incoming voltage, which is at "x" volts. So far, so good. 5 microseconds later, "y" volts is present at the input. 8 microseconds after that, "z" volts is present at the input. 5 microseconds later, "w" voltage is present at the input and that voltage lasts for 10 microseconds.
When the next sample occurs 21 microseconds after the first one, it will read the "w" voltage, but it will ignore the "y" and "z" values because they occurred between samples. I don't see any way the "y' and "z" values could factor into the encoding process because the system never sees them.
So then the question becomes does reconstruction reproduce those "ignored" variations successfully, and if not, does it matter? The argument that says it doesn't matter maintains that smoothing will accurately fill in the values between the "x" and "w" voltages, and will therefore reconstruct the frequency that was present at those times.
However, my understanding of Moorer's argument is that if there were
spatial cues in between "x" and "w," they will be lost. Whether that matters or not depends on whether you accept Moorer's contention that people can discriminate between extremely short time delays when signals hit both ears. As I doubt anyone in this thread has verified or disproven these experiments, I don't think it's possible to accept or dismiss them out of hand. However, if (I emphasize "if," although I don't know what his motivation would be for making things up) he is correct, then given the sampling scenario presented above, I simply don't see any way that a 21-microsecond window can encode incoming delay-based events whose duration is significantly less than that.