Tesla does INT8 inference. Way more efficient than FP16。。。
文章来源: 蓝调2023-02-28 22:45:02