Skip to content
#

quantization-aware-training

Here are 28 public repositories matching this topic...

micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape

  • Updated Oct 6, 2021
  • Python

Intel® Neural Compressor (formerly known as Intel® Low Precision Optimization Tool), targeting to provide unified APIs for network compression technologies, such as low precision quantization, sparsity, pruning, knowledge distillation, across different deep learning frameworks to pursue optimal inference performance.

  • Updated Jun 4, 2022
  • Python

Our work implements novel L2-Norm gradient (L2Grad) and variance of the weight distrbution (VarianceNorm) regularizers for quantization-aware training such that the distribution of weights are more compatible with post-training quantization especially for low bit-widths. We provide a theoretical basis that directly relates L2-Grad with post quantization test accuracy through a first order Taylor Series expansion followed by the reduction to an adversary with an L2 budget, in which we apply the Cauchy-Schwarz inequality to provide the desired bounds. We empirically show that L2Grad and VarianceNorm can both match the performance of L1Grad and outperform it on certain bit-widths. We also show that a regularization scheme that combines L2Grad and VarianceNorm in a novel "regularization scheduling" methodology can give even better results in terms of post-quantization accuracy, tested on uniform and piecewise linear quantization.

  • Updated May 15, 2021
  • Jupyter Notebook

Improve this page

Add a description, image, and links to the quantization-aware-training topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the quantization-aware-training topic, visit your repo's landing page and select "manage topics."

Learn more