It can easily be proven that an ideal analogue delta sigma modulator shapes the quantisation noise, as explained for example in this AD note under EQ 3: https://www.analog.com/media/en/training-seminars/tutorials/MT-022.pdf
Is there a similar mathematical high level proof that shows that only the average of the bitstream approximates the input signal rather than instantaneous single values?