Contribute Media
A thank you to everyone who makes this possible: Read More

Quantization

Description

It’s important to make efficient use of both server-side and on-device compute resources when developing ML applications. To support more efficient deployment on servers and edge devices, PyTorch 1.3 now supports 8-bit model quantization using the familiar eager mode Python API.

Improve this page