- Jun 10, 2019
-
-
Neta Zmora authored
Some links have changed with the latest version of mkdocs. This closes issues #280 and #65 (reopened).
-
- Apr 14, 2019
-
-
Guy Jacob authored
-
- Feb 11, 2019
-
-
Guy Jacob authored
Summary of changes: (1) Post-train quantization based on pre-collected statistics (2) Quantized concat, element-wise addition / multiplication and embeddings (3) Move post-train quantization command line args out of sample code (4) Configure post-train quantization from YAML for more fine-grained control (See PR #136 for more detailed changes descriptions)
-
- Dec 04, 2018
-
-
Guy Jacob authored
* Asymmetric post-training quantization (only symmetric supported so until now) * Quantization aware training for range-based (min-max) symmetric and asymmetric quantization * Per-channel quantization support in both training and post-training * Added tests and examples * Updated documentation
-
- Nov 25, 2018
-
-
Neta Zmora authored
-
- Nov 24, 2018
-
-
Neta Zmora authored
Thanks to Dan Alistarh for bringing this issue to my attention. The activations of Linear layers have shape (batch_size, output_size) and those of Convolution layers have shape (batch_size, num_channels, width, height) and this distinction in shape was not correctly handled. This commit also fixes sparsity computation for very large activations, as seen in VGG16, which leads to memory exhaustion. One solution is to use smaller batch sizes, but this commit uses a different solution, which counts zeros “manually”, and using less space. Also in this commit: - Added a “caveats” section to the documentation. - Added more tests.
-
- Nov 21, 2018
-
-
Neta Zmora authored
-
- Jun 21, 2018
-
-
Guy Jacob authored
-
- May 22, 2018
-
-
Neta Zmora authored
Two places in the documentation gave the wrong path to the example Alexnet sensitivity pruning schedule.
-
- Apr 24, 2018
-
-
Neta Zmora authored
-
Neta Zmora authored
-
Neta Zmora authored
-