Jon, due to a setup error on my part, I once made a 16-bit live classical recording without effective dither, and the results weren't what I expected from general theory. The noise floor of the venue included the building's ventilation system and Manhattan's omnipresent traffic rumble, so I thought that the recording would be "self dithering." The lowest levels I saw on the meters were ~54 dB below full scale, and since the theoretical requirement is only that the LSB be fully randomized, I thought that there would be no problem.
But when I got home I found that the recording had quite audible granular noise (quantization distortion) on the fadeouts of sustained piano tones and even on some quietly sung vocal passages. It turned out that those -54 dB noise floor levels were due in large part to uncorrected DC offset in the A/D converter. Once that was subtracted out, the lowest levels were more like -72 dBFS, in part because the preamp gain was set lower than would have been optimal.
The preamp + A/D was a Grace Lunatec V3, and I'd forgotten to engage the 16-bit dithering function. From this experience I learned always to double-check that setting before recording. Also, once I realized what had caused the falsely reassuring level readings, I sent the V3 back to Grace, who had worked out a method to get rid of DC offset since my preamp had been made. In general I consider small amounts of DC offset to be harmless, but not if they throw off your estimate of the noise floor. (I'd set my overall record levels based on what I thought was the noise floor when I was setting up, and since that reading was artificially raised by the DC offset, my levels were lower than they would otherwise have been, thus exacerbating the problems caused by the lack of effective dither.)
Obviously these were practical mishaps and mistakes that don't disprove the theory, but they do show how mistakes and mishaps can get the upper hand in the real world.
--best regards