diff --git a/README.md b/README.md
index fbac8686e23bd75d6c1af13066facf08a98d7b30..4f1b835f77c4cbe598cef4068016e310ffb7549a 100755
--- a/README.md
+++ b/README.md
@@ -183,8 +183,8 @@ If you do not use CUDA 10.1 in your environment, please refer to [PyTorch websit
 
 Distiller comes with sample applications and tutorials covering a range of model types:
 
-| Model Type | Sparsity | Post-train quant | Quant-aware training | Auto Compression (AMC) |
-|------------|:--------:|:----------------:|:--------------------:|:----------------------:|
+| Model Type | Sparsity | Post-training quantization | Quantization-aware training | Auto Compression (AMC) |
+|------------|:--------:|:--------------------------:|:---------------------------:|:----------------------:|
 | [Image classification](https://github.com/NervanaSystems/distiller/tree/master/examples/classifier_compression) | :white_check_mark: | :white_check_mark: | :white_check_mark: | :white_check_mark: |
 | [Word-level language model](https://github.com/NervanaSystems/distiller/tree/master/examples/word_language_model)| :white_check_mark: | :white_check_mark: | | |
 | [Translation (GNMT)](https://github.com/NervanaSystems/distiller/tree/master/examples/GNMT) | | :white_check_mark: | | |