Trained Quantization Thresholds for Accurate and Efficient Fixed-Point Inference of Deep Neural Networks

March 19, 2019 ยท Entered Twilight ยท ๐Ÿ› Conference on Machine Learning and Systems

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 6.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitattributes, .gitignore, LICENSE, README.md, graffitist, img, scripts, setup.py

Authors Sambhav R. Jain, Albert Gural, Michael Wu, Chris H. Dick arXiv ID 1903.08066 Category cs.CV: Computer Vision Cross-listed cs.AI, cs.LG Citations 164 Venue Conference on Machine Learning and Systems Repository https://github.com/Xilinx/graffitist โญ 170 Last Checked 2 months ago
Abstract
We propose a method of training quantization thresholds (TQT) for uniform symmetric quantizers using standard backpropagation and gradient descent. Contrary to prior work, we show that a careful analysis of the straight-through estimator for threshold gradients allows for a natural range-precision trade-off leading to better optima. Our quantizers are constrained to use power-of-2 scale-factors and per-tensor scaling of weights and activations to make it amenable for hardware implementations. We present analytical support for the general robustness of our methods and empirically validate them on various CNNs for ImageNet classification. We are able to achieve near-floating-point accuracy on traditionally difficult networks such as MobileNets with less than 5 epochs of quantized (8-bit) retraining. Finally, we present Graffitist, a framework that enables automatic quantization of TensorFlow graphs for TQT (available at https://github.com/Xilinx/graffitist ).
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision