-
- Downloads
Convert Distiller PTQ models to "native" PyTorch PTQ (#458)
Convert Distiller PTQ models to "native" PyTorch PTQ (#458) * New API: distiller.quantization.convert_distiller_ptq_model_to_pytorch() * Can also be called from PostTrainLinearQuantizer instance: quantizer.convert_to_pytorch() * Can also trigger from command line in image classification sample * Can save/load converted modules via apputils.load/save_checkpoint * Added Jupyter notebook tutorial * Converted modules have only the absolutely necessary quant-dequant operations. For a fully quantized model, this means just quantization of model input and de-quantization of model output. If a user keeps specific internal layers in FP32, quant-dequant operations are added as needed * Can configure either 'fbgemm' or 'qnnpack' backend. For 'fbgemm' we take care of preventing overflows (aka "reduce_range" in the PyTorch API)
Showing
- distiller/apputils/checkpoint.py 5 additions, 0 deletionsdistiller/apputils/checkpoint.py
- distiller/apputils/image_classifier.py 23 additions, 2 deletionsdistiller/apputils/image_classifier.py
- distiller/models/__init__.py 2 additions, 1 deletiondistiller/models/__init__.py
- distiller/quantization/__init__.py 2 additions, 0 deletionsdistiller/quantization/__init__.py
- distiller/quantization/pytorch_quant_conversion.py 436 additions, 0 deletionsdistiller/quantization/pytorch_quant_conversion.py
- distiller/quantization/quantizer.py 6 additions, 2 deletionsdistiller/quantization/quantizer.py
- distiller/quantization/range_linear.py 170 additions, 0 deletionsdistiller/quantization/range_linear.py
- examples/quantization/post_train_quant/command_line.md 13 additions, 1 deletionexamples/quantization/post_train_quant/command_line.md
- jupyter/post_train_quant_convert_pytorch.ipynb 481 additions, 0 deletionsjupyter/post_train_quant_convert_pytorch.ipynb
- tests/test_ptq_pytorch_convert.py 126 additions, 0 deletionstests/test_ptq_pytorch_convert.py
Loading
Please register or sign in to comment