Skip to content
Snippets Groups Projects
Commit 3910b5bd authored by Neta Zmora's avatar Neta Zmora
Browse files

Documentation: small fix for RNN pruning image

parent e9e78530
No related branches found
No related tags found
No related merge requests found
...@@ -107,7 +107,7 @@ The authors of [Exploring Sparsity in Recurrent Neural Networks](https://arxiv.o ...@@ -107,7 +107,7 @@ The authors of [Exploring Sparsity in Recurrent Neural Networks](https://arxiv.o
Distiller's distiller.pruning.BaiduRNNPruner class implements this pruning algorithm. Distiller's distiller.pruning.BaiduRNNPruner class implements this pruning algorithm.
<center>![Gradual Pruning](imgs/baidu_rnn_pruning.png)</center> <center>![Baidu RNN Pruning](imgs/baidu_rnn_pruning.png)</center>
# Structure pruners # Structure pruners
Element-wise pruning can create very sparse models which can be compressed to consume less memory footprint and bandwidth, but without specialized hardware that can compute using the sparse representation of the tensors, we don't gain any speedup of the computation. Structure pruners, remove entire "structures", such as kernels, filters, and even entire feature-maps. Element-wise pruning can create very sparse models which can be compressed to consume less memory footprint and bandwidth, but without specialized hardware that can compute using the sparse representation of the tensors, we don't gain any speedup of the computation. Structure pruners, remove entire "structures", such as kernels, filters, and even entire feature-maps.
......
...@@ -274,7 +274,7 @@ abundant and gradually reduce the number of weights being pruned each time as th ...@@ -274,7 +274,7 @@ abundant and gradually reduce the number of weights being pruned each time as th
<h2 id="rnn-pruner">RNN pruner</h2> <h2 id="rnn-pruner">RNN pruner</h2>
<p>The authors of <a href="https://arxiv.org/abs/1704.05119">Exploring Sparsity in Recurrent Neural Networks</a>, Sharan Narang, Erich Elsen, Gregory Diamos, and Shubho Sengupta, "propose a technique to reduce the parameters of a network by pruning weights during the initial training of the network." They use a gradual pruning schedule which is reminiscent of the schedule used in AGP, for element-wise pruning of RNNs, which they also employ during training. They show pruning of RNN, GRU, LSTM and embedding layers.</p> <p>The authors of <a href="https://arxiv.org/abs/1704.05119">Exploring Sparsity in Recurrent Neural Networks</a>, Sharan Narang, Erich Elsen, Gregory Diamos, and Shubho Sengupta, "propose a technique to reduce the parameters of a network by pruning weights during the initial training of the network." They use a gradual pruning schedule which is reminiscent of the schedule used in AGP, for element-wise pruning of RNNs, which they also employ during training. They show pruning of RNN, GRU, LSTM and embedding layers.</p>
<p>Distiller's distiller.pruning.BaiduRNNPruner class implements this pruning algorithm.</p> <p>Distiller's distiller.pruning.BaiduRNNPruner class implements this pruning algorithm.</p>
<p><center><img alt="Gradual Pruning" src="../imgs/baidu_rnn_pruning.png" /></center></p> <p><center><img alt="Baidu RNN Pruning" src="../imgs/baidu_rnn_pruning.png" /></center></p>
<h1 id="structure-pruners">Structure pruners</h1> <h1 id="structure-pruners">Structure pruners</h1>
<p>Element-wise pruning can create very sparse models which can be compressed to consume less memory footprint and bandwidth, but without specialized hardware that can compute using the sparse representation of the tensors, we don't gain any speedup of the computation. Structure pruners, remove entire "structures", such as kernels, filters, and even entire feature-maps.</p> <p>Element-wise pruning can create very sparse models which can be compressed to consume less memory footprint and bandwidth, but without specialized hardware that can compute using the sparse representation of the tensors, we don't gain any speedup of the computation. Structure pruners, remove entire "structures", such as kernels, filters, and even entire feature-maps.</p>
<h2 id="ranked-structure-pruner">Ranked structure pruner</h2> <h2 id="ranked-structure-pruner">Ranked structure pruner</h2>
......
...@@ -246,5 +246,5 @@ And of course, if we used a sparse or compressed representation, then we are red ...@@ -246,5 +246,5 @@ And of course, if we used a sparse or compressed representation, then we are red
<!-- <!--
MkDocs version : 0.17.2 MkDocs version : 0.17.2
Build Date UTC : 2018-06-14 10:51:56 Build Date UTC : 2018-06-14 11:48:24
--> -->
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment