diff --git a/README.md b/README.md index 3dcede359edc0277b0c0e8535622410d6a036d30..466a8b80791962ac93bf5e91c5cd9d2f84ab348f 100755 --- a/README.md +++ b/README.md @@ -251,7 +251,7 @@ For more details, there are some other resources you can refer to: + [Tutorial: Using Distiller to prune a PyTorch language model](https://nervanasystems.github.io/distiller/tutorial-lang_model.html) + [Tutorial: Pruning Filters & Channels](https://nervanasystems.github.io/distiller/tutorial-struct_pruning.html) + [Tutorial: Post-Training Quantization of a Language Model](https://nervanasystems.github.io/distiller/tutorial-lang_model_quant.html) -+ [Tutorial: Post-Training Quantization of GNMT (translation model)](https://nervanasystems.github.io/distiller/tutorial-lang_model_quant.html) ++ [Tutorial: Post-Training Quantization of GNMT (translation model)](https://nervanasystems.github.io/distiller/tutorial-gnmt_quant.html) + [Post-training quantization command line examples](https://github.com/NervanaSystems/distiller/blob/master/examples/quantization/post_train_quant/command_line.md) ### Example invocations of the sample application