I need a refresher on the basics of delta sigma converters and none of the material has made it clear, to me anyway, about the relationship with resolution, frequency of sampling, and data rate.
Suppose I have a 16 bit delta sigma ADC, a sampling clock at 100kHz and an oversampling ratio of 64. I want to calculate samples per second.
A 16 bit converter requires 2^N clocks; 65536 clocks in this case. If my sampling frequency is 1ookHz and the OSR is set to 64 then I am sampling at 6.4MHz.
Samples per second would be 6.4MHz/2^16 or 97.65 samples per second.
Yes? No?
Suppose I have a 16 bit delta sigma ADC, a sampling clock at 100kHz and an oversampling ratio of 64. I want to calculate samples per second.
A 16 bit converter requires 2^N clocks; 65536 clocks in this case. If my sampling frequency is 1ookHz and the OSR is set to 64 then I am sampling at 6.4MHz.
Samples per second would be 6.4MHz/2^16 or 97.65 samples per second.
Yes? No?