Skip to content
This repository has been archived by the owner on Sep 18, 2024. It is now read-only.

Release Note #4284

Merged
merged 6 commits into from
Nov 4, 2021
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
update release note links
  • Loading branch information
liuzhe committed Nov 3, 2021
commit 69834eb06676f7e588ec6eebd54f6f268aea042b
5 changes: 3 additions & 2 deletions docs/en_US/Release.rst
Original file line number Diff line number Diff line change
Expand Up @@ -21,16 +21,17 @@ Model Compression
* Optimized the memory usage of the pruners.

* MobileNetV2 end-to-end example `(notebook) <https://github.com/microsoft/nni/blob/v2.5/examples/model_compress/pruning/mobilenetv2_end2end/Compressing%20MobileNetV2%20with%20NNI%20Pruners.ipynb>`__
* Improved QAT quantizer
* Improved QAT quantizer `(doc) <https://nni.readthedocs.io/en/v2.5/Compression/Quantizer.html#qat-quantizer>`__

* support dtype and scheme customization
* support dp multi-gpu training
* support load_calibration_config

* Model speed-up now supports directly loading the mask
* Model speed-up now supports directly loading the mask `(doc) <https://nni.readthedocs.io/en/v2.5/Compression/ModelSpeedup.html#nni.compression.pytorch.ModelSpeedup>`__
* Support speed-up depth-wise convolution
* Support bn-folding for LSQ quantizer
* Support QAT and LSQ resume from PTQ
* Added doc for observer quantizer `(doc) <https://nni.readthedocs.io/en/v2.5/Compression/Quantizer.html#observer-quantizer>`__

Neural Architecture Search
^^^^^^^^^^^^^^^^^^^^^^^^^^
Expand Down