Optimize requantizer to work in f32, then round

This commit is contained in:
Roderick van Domburg 2021-04-09 19:31:26 +02:00
parent 928a673653
commit d0ea9631d2

View file

@ -21,7 +21,16 @@ macro_rules! convert_samples_to {
$samples $samples
.iter() .iter()
.map(|sample| { .map(|sample| {
(*sample as f64 * (std::$type::MAX as f64 + 0.5) - 0.5) as $type >> $drop_bits // Losslessly represent [-1.0, 1.0] to [$type::MIN, $type::MAX]
// while maintaining DC linearity. There is nothing to be gained
// by doing this in f64, as the significand of a f32 is 24 bits,
// just like the maximum bit depth we are converting to.
let int_value = *sample * (std::$type::MAX as f32 + 0.5) - 0.5;
// Casting floats to ints truncates by default, which results
// in larger quantization error than rounding arithmetically.
// Flooring is faster, but again with larger error.
int_value.round() as $type >> $drop_bits
}) })
.collect() .collect()
}; };