Skip to content
Snippets Groups Projects
  1. Feb 26, 2019
  2. Feb 13, 2019
  3. Nov 25, 2018
  4. Oct 29, 2018
  5. Oct 23, 2018
  6. Oct 22, 2018
    • Neta Zmora's avatar
      Activation statistics collection (#61) · 54a5867e
      Neta Zmora authored
      Activation statistics can be leveraged to make pruning and quantization decisions, and so
      We added support to collect these data.
      - Two types of activation statistics are supported: summary statistics, and detailed records 
      per activation.
      Currently we support the following summaries: 
      - Average activation sparsity, per layer
      - Average L1-norm for each activation channel, per layer
      - Average sparsity for each activation channel, per layer
      
      For the detailed records we collect some statistics per activation and store it in a record.  
      Using this collection method generates more detailed data, but consumes more time, so
      Beware.
      
      * You can collect activation data for the different training phases: training/validation/test.
      * You can access the data directly from each module that you chose to collect stats for.  
      * You can also create an Excel workbook with the stats.
      
      To demonstrate use of activation collection we added a sample schedule which prunes 
      weight filters by the activation APoZ according to:
      "Network Trimming: A Data-Driven Neuron Pruning Approach towards 
      Efficient Deep Architectures",
      Hengyuan Hu, Rui Peng, Yu-Wing Tai, Chi-Keung Tang, ICLR 2016
      https://arxiv.org/abs/1607.03250
      
      We also refactored the AGP code (AutomatedGradualPruner) to support structure pruning,
      and specifically we separated the AGP schedule from the filter pruning criterion.  We added
      examples of ranking filter importance based on activation APoZ (ActivationAPoZRankedFilterPruner),
      random (RandomRankedFilterPruner), filter gradients (GradientRankedFilterPruner), 
      and filter L1-norm (L1RankedStructureParameterPruner)
      Unverified
      54a5867e
    • Neta Zmora's avatar
  7. Oct 03, 2018
  8. Aug 29, 2018
  9. Aug 27, 2018
  10. Jul 11, 2018
  11. May 16, 2018
    • Neta Zmora's avatar
      PyTorch 0.4 improvement to SummaryGraph · 32c01c28
      Neta Zmora authored
      PyTorch 0.4 now fully supports the ONNX export features that are needed
      in order to create a SummaryGraph, which is sort of a "shadow graph" for
      PyTorch models.
      
      The big advantage of SummaryGraph is that it gives us information about
      the connectivity of nodes.  With connectivity information we can compute
      per-node MAC (compute) and BW, and better yet, we can remove channels,
      filters, and layers (more on this in future commits).
      
      In this commit we (1) replace the long and overly-verbose ONNX node names,
      with PyTorch names; and (2) move MAC and BW attributes from the Jupyter
      notebook to the SummaryGraph.
      32c01c28
    • Neta Zmora's avatar
      pytorch 0.4: adjustments to API changes · 957e6777
      Neta Zmora authored
      Various small changes due to the chamnges in the semantics and syntax of the
      PyTorch 0.4 API.
      
      Note that currently distiller.model_performance_summary() returns wrong results
      on graphs containing torch.nn.DataParallel layers.
      957e6777
    • Neta Zmora's avatar
      b9bf4282
  12. May 10, 2018
  13. Apr 25, 2018
  14. Apr 24, 2018
Loading