Tesla does INT8 inference. Way more efficient than FP16, but took us a lot of effort to overcome quantization errors.— Elon Musk (@elonmusk) February 28, 2023
Tesla does INT8 inference. Way more efficient than FP16, but took us a lot of effort to overcome quantization errors.