Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/A-suozhang/awesome-quantization-and-fixed-point-training
Neural Network Quantization & Low-Bit Fixed Point Training For Hardware-Friendly Algorithm Design
https://github.com/A-suozhang/awesome-quantization-and-fixed-point-training
List: awesome-quantization-and-fixed-point-training
fixed-point network-compression quantization
Last synced: 3 months ago
JSON representation
Neural Network Quantization & Low-Bit Fixed Point Training For Hardware-Friendly Algorithm Design
- Host: GitHub
- URL: https://github.com/A-suozhang/awesome-quantization-and-fixed-point-training
- Owner: A-suozhang
- Created: 2019-12-13T02:51:55.000Z (almost 5 years ago)
- Default Branch: master
- Last Pushed: 2020-12-18T01:26:30.000Z (almost 4 years ago)
- Last Synced: 2024-05-20T23:17:48.130Z (6 months ago)
- Topics: fixed-point, network-compression, quantization
- Size: 81.1 KB
- Stars: 151
- Watchers: 11
- Forks: 24
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
- ultimate-awesome - awesome-quantization-and-fixed-point-training - Neural Network Quantization & Low-Bit Fixed Point Training For Hardware-Friendly Algorithm Design. (Other Lists / PowerShell Lists)
README
# 初衷
* (表明该整理的目的,为后续工作展开做一些指引)
* 本文不完全包括了一些NN Fixed-Point Training的可能实现方向,为后续实验指路
* 本文的最终目的是面向FPGA等嵌入式设备的NN 部署/训练,的一个软件仿真工具。为了Scalability,会尝试实现一些较为General的定点训练技巧现方式
* 以及诸如MobileNet, ShuffleNet 等轻量化网络设计
* Quantization的两种主要方式
* 基于CodeBook的(Deep Compression) :实际参数还是高精度的,无法利用定点计算进行加速,仅能减少存储
* 基于定点数(Fixed Point表示),(IBM的FP8也可以归入此类) : 可利用定点计算加速,故本文主要采取该方式
* 目前预计的几种场景
* Post-Training Quantization : 在完成FP训练之后的压缩,产生定点的W/A进行部署
* Example:Distillation,EntropyConstraintQ, IncrementalQ
* Quantize-Aware Training : 在训练过程中考虑定点的影响,在训练中采取Fixed,产生定点的W/A进行部署
* Example: StraightThroughActivation的方法(训练时用Fixed Inference,但是梯度是对应的全精度副本做)
* Fixed-Point Training: 训练过程中进行**纯定点(W/G/A)**,模拟在纯定点设备上进行训练
* Example:WAGE
* ~~有点激进,不知道是否能实现~~
* Binary-Network# Methods
> 从自己的出发点对看到的一些定点相关工作的方法与思想的纯主观归纳(可能存在偏差甚至错误)
> 该种划分方式没有什么道理,只是我强行区分的罢了)
## A. Post-Training Quantization
> 该类方法最大的特点就是利用已经训练好的模型进行压缩,与Quantized-Aware Training相对比有几分 2-Stage的意味,优势在于可以利用已经基本确定的参数分布去分析,采取量化策略,对大模型比较有效,但是小模型会崩
* [1510-Deep Compression](https://arxiv.org/abs/1510.00149)
* 利用了参数分布的知识,采用K-Means
* [1702-Incremental Quantization](https://arxiv.org/abs/1702.03044)
* 分组-量化-Finetune的流程,每次retrain只有当前组被量化,迭代直到所有参数都被量化(介于A与B之间)
* [1802-Model compression via distillation and quantization](https://arxiv.org/abs/1802.05668)
* ICLR 2018
* 利用一个很大的Tecaher来引导低比特网络
* [1810-Post training 4-bit quantization of convolutional networks for rapid-deployment(NIPS 2019)](https://arxiv.org/abs/1810.05723)
* Intel, No Need To Finetune On Full Dataset
* 3 Methods,一些理论推导得到每channel比特数的分配以及ClippingValue可以依据已有的参数分布算出来
* [1907-And the Bit Goes Down: Revisiting the Quantization of Neural Networks](https://arxiv.org/abs/1907.05686)* Other Methods (~~More Of a Digest Not Important~~)
* [1511-Fixed Point Quantization of Deep Convolutional Network](https://arxiv.org/abs/1511.06393)
* 高通 基于SQNR,前Deep Compression的上古时期,没有什么大的亮点
* [Entropy Constraint Scalar Quantization](https://www.mdpi.com/1099-4300/18/12/449)
* 对每一个参数的Gradient做泰勒展开并且舍弃高阶项,化简得到哪些参数对最终Loss重要,以此作为剪枝或者量化的依据(选取聚类中心)
* ~~也可以作为剪枝的依据~~
* 和剪枝的这篇,有一定相关性[1611-Designing Energy-Efficient Convolutional Neural Networks using Energy-Aware Pruning](https://arxiv.org/abs/1611.05128)
* 贪心的剪去对最后Loss影响不大的 (也就是TCP(Transfer Channel Prunning)中的剪枝方式)
* [1711-NISP: Pruning Networks using Neuron Importance Score Propagation](https://arxiv.org/abs/1711.05908)也会涉及
* [1805-Retraining-Based Iterative Weight Quantization for Deep Neural Networks](https://arxiv.org/abs/1805.11233)
* [1906-Data-Free Quantization through weiht equailization & Bias Correction](https://arxiv.org/abs/1906.04721)## B. Quantize-Aware-Training
> 相比于第一类,该类方法的主要优势在于1-Stage,简化了训练过程
* 早期的一些Binary/XNORNet均属于此类,大部分基于StraightThroughActivation的思想,即认为定点过程的导数为1
* [1603-XNORNet](https://arxiv.org/abs/1603.05279)
* 对WA二值化,加上一个L1Norm的mean作为每层的ScalingFactor(其改进DoReFa加上了G的)
* [1605-TWN](https://arxiv.org/abs/1605.04711)
* TWN - 最小化全精度weight与Ternary Weight之间的L2 Norm
* [1709-WRPN-Intel-ICLR2018](https://arxiv.org/abs/1709.01134)
* 低比特WA(全精度G),但是让网络更wide(增多了FeatureMap数量)
* [1712-Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference](https://arxiv.org/abs/1712.05877)
* 包含了一个浮点的Scale Factor,最大化利用了动态范围,利用了(0.45-0.5)的这一段空间
* 非对称量化(有一个零点,以及一个浮点的Scale - 逐层)
* Merge Conv-BN
* [1805-PACT](https://arxiv.org/abs/1805.06085)
* 训练中Quantize Activation,训练一个activation clipping parameter(修改Relu的clip范围s)(也就是在训练中找FixScale)
* [1802-Mixed Precision Training Of ConvNets Using Integer Operations-ICLR2018](https://arxiv.org/pdf/1802.00930.pdf)
* 16bit Training
* [1805-Accurate & Efficient 2-bit QNN](https://www.semanticscholar.org/paper/ACCURATE-AND-EFFICIENT-2-BIT-QUANTIZED-NEURAL-Choi-Venkataramani/c3cb27f9ef7176658f37b607e75cc2c37f5e0ea8)
* PACT + SAWB (Statistics-Aware Weight bining)
* 文中有Analytic分析PACT和relu的表示能力一致
* [1808-Learning to Quantize Deep Networks by Optimizing Quantization Intervals with Task Loss](https://arxiv.org/abs/1808.05779)
* 学习量化的区间和非线性量化的间隔
* [1905-hawq: hessian aware quantization of neuralnetworks with mixed-precision](https://arxiv.org/pdf/1905.03696.pdf)
* [1911-V2](https://arxiv.org/pdf/1911.03852.pdf)
* [1901-Improving Neural Network Quantization with Retraining Outlier Channel Splitting](https://arxiv.org/pdf/1901.09504.pdf)
* [Code at](https://github.com/cornell-zhang/dnn-quant-ocs) - Implemented with Distiller Library* Other Works
* [1903-Training Quantized Network with Auxiliary Gradient Module](https://arxiv.org/abs/1903.11236)
* 额外的fullPrecision梯度模块(解决residue的skip connection不好定的问题,目的前向完全fix point),有几分用一个FP去杠杆起低比特网络的意味
* [1901-Accumulation bit-width Scaling](https://arxiv.org/abs/1901.06588)
* IBM ICLR 2019, 找Accumulator可以压到多少# C. (Full) Fixed-Point Training
> 纯定点的训练,大部分都是对一些经典的部分做一些简化。以及对梯度量化也会归入此类(目的是训练的加速,而不仅仅是为了部署)* [1606-DoReFa](https://arxiv.org/abs/1606.06160)
* 算是最先提出低比特训练
* 每个Layer有一个ScalingFactor
* [1802-WAGE - Training & Inference with Integers in DNN](https://arxiv.org/abs/1802.04680)
* 全定点,简化了很多部分,但是精度损失比较明显
* [1705-TernGrad](https://arxiv.org/abs/1705.07878)
* 将梯度定点为3值(用类似TWN的方式),目的是减少分布式训练中的通信传播
* [1805-Scalable Methods for 8-bit Training of Neural Networks](https://arxiv.org/abs/1805.11046)
* Intel(AIPG) NIPS2018 (WAG8)(RangeBN)
* [1812-Training Deep Neural Networks with 8-bit Floating Point Numbers](https://arxiv.org/abs/1812.08011)
* FP8 - Training
* [1905-mixed precision training with 8-bit floating point](https://arxiv.org/abs/1905.12334)
* wage all fp8
* 对比了rne(round2nearesteven)&sstochastic rounding
* 全部8bit训练Flow - [Towards Unified INT8 Training for Convolutional Neural Network](https://arxiv.org/pdf/1912.12607.pdf)
* SenseTime CVPR 2020# D. Binary
* 早期的一些Binary/XNORNet类似于Quantize-aware Training,但是大多依赖于Pretrained Model
* [1511-BinaryConnect](https://arxiv.org/abs/1511.00363)
* 最开始的二值网络
* [1602-Binarized Neural Networks](https://arxiv.org/abs/1602.02830)
* 同期最早的BNN
* [1603-XNORNet](https://arxiv.org/abs/1603.05279)
* 对WA二值化,加上一个L1Norm的mean作为每层的ScalingFactor(其改进DoReFa加上了G的)
* [1605-Ternary Weighted Network](https://arxiv.org/abs/1605.04711)
* TWN - 最小化全精度weight与Ternary Weight之间的L2 Norm
* [1709-WRPN-Intel-ICLR2018](https://arxiv.org/abs/1709.01134)
* 低比特WA(全精度G),但是让网络更wide(增多了FeatureMap数量)
* [1606-DoReFa](https://arxiv.org/abs/1709.01134)
* quantize gradient,提供了任意比特量化的实验
* 各种Binary follow-up
- [1709-Flexible Network Binarization with Layer-wise Priority](https://arxiv.org/abs/1709.04344)
- [1711-ReBNet: Residual Binarized Neural Network](https://arxiv.org/abs/1711.01243)
- Binary+Resnet
- [1711-Towards Accurate Binary Convolutional Neural Network](https://arxiv.org/abs/1711.11294)
- [1804-Training a Binary Weight Object Detector by Knowledge Transfer for Autonomous Driving](https://arxiv.org/abs/1804.06332)
- [1902-Self-Binarizing Networks](https://arxiv.org/abs/1902.00730)
- [1906-Latent Weights Do Not Exist: Rethinking Binarized Neural Network Optimization](https://arxiv.org/abs/1906.02107)
- [2001-Least squares binary quantization of neural networks](https://arxiv.org/abs/2001.02786v1)
- [2002-Widening and Squeezing: Towards Accurate and Efficient QNNs](https://arxiv.org/abs/2002.00555)
- 提精度
- [1808-Bi-Real Net: Enhancing the Performance of 1-bit CNNs With Improved Representational Capability and Advanced Training Algorithm](http://arxiv.org/abs/1808.00278)
* 从头开始训练+额外的shortcut+STE的修正
* 提到了训练BNN的两个问题
- [1812-Training Competitive Binary Neural Networks from Scratch](http://arxiv.org/abs/1812.01965)
- [1909-XNORNet++](http://arxiv.org/abs/1909.13863)
* fuse weight & activation scaling factor into onel)
- [2001-MeliusNet: Can Binary Neural Networks Achieve MobileNet-level Accuracy?](https://arxiv.org/abs/2001.05936)
* 本质上是提出了一种新的Block
* 指出了原本的BNN的优化方法有
- [1909-IRNet-Forward and Backward Information Retention for Accurate Binary Neural Networks](https://arxiv.org/pdf/1909.10788.pdf)* Other Methods
* [1812-Per-Tensor-Quantization of BackProp](https://arxiv.org/abs/1812.11732)
* ICLR2019, Precision Assignment (好多数学假设分析),给出了一种确定每层位宽的方法
* [1812-Hybrid 8-bit Training](https://arxiv.org/abs/1812.08011)
* FP8的后续,对不同的组件提出不同的exponential bit与mantissa bit的划分方式# Ideas
## Post-Training Quantization
* Important Weighing
* 通过评估每个参数对最终Loss的重要程度,作为剪枝或是量化的依据,比如[这篇文章](https://www.mdpi.com/1099-4300/18/12/449)
* Knowledge Distillation
* 通过一个高精度模型给低比特模型的Finetune提供SoftLabel
* Incremental & Iterative Quantization
* Gradually做Quantization,前者是每次选取一部分Weight,后者是逐渐降低比特数目
* Analytical Correction
* 从解析的角度去做一些矫正或者是限制,[代表文章](https://arxiv.org/abs/1810.05723)
* Data-Free Quantization
* 近几年兴起的一个领域,和👆的方法有一些相关性## Quantize-Aware Training
> 定点数化之后相比于浮点最大的问题就是**动态范围**,也是主要的优化目的
* 限制动态范围
* 从训练角度就不要让参数分布的太散,减少造成的影响,比如[PACT](https://arxiv.org/abs/1805.06085)
* 训练中加大WeightDecay(L1 Norm)之后训练出来的模型定点效果会更好
* 细粒度的确定量化区间
* 同一区间中会取同一个Range,通过减少被划入同一动态范围的参数量来减少Clamp的发生
* 从Layer-Wise到Channel-Wise再到Block-Wise
* 细粒度的分配bit数
* Mixed-Precision的一系列方法
* Hybrid FP8 - 对A/W采取不同的指数bit分配
* 扩大动态范围
* FP8 - 2次幂的浮点数
* 充分利用动态范围
* 引入浮点ScalingFactor,[这篇](https://arxiv.org/abs/1712.05877)
* 非线性量化
* 学习量化中的各种值(ClippingValue/QuantizeInterval)
* 用于替代解析方法求出来的一些解析解(解析方法的一些假设不是很精确)
* [PACT](https://arxiv.org/abs/1805.06085)学习了Relu的ClippingValue
* [KAIST的这篇文章](https://arxiv.org/abs/1808.05779)学习了量化的
* [TTQ](https://arxiv.org/pdf/1612.01064.pdf)学习了ScalingFactor### Fixed-Point Training
> 一般只有有加速训练的需求的时候才需要量化梯度,但是由于梯度的特殊性(动态范围大,且大值小值都很重要)采用一般的量化方式不足,一般需要额外的设计;但是👆部分的很多方法的目的是为了更好的量化,印次对该部分也有指导价值
* 量化梯度
* [TernGrad](https://papers.nips.cc/paper/6749-terngrad-ternary-gradients-to-reduce-communication-in-distributed-deep-learning)主要是利用了随机Rounding
* FP8/Fp16 扩大动态范围
* 更细粒度的划分Range
* 简化训练
* BN
* RangeBN / L1BN
* WAGE简化了各种(献祭了精度)
* 全部8bit训练Flow - [Towards Unified INT8 Training for Convolutional Neural Network](https://arxiv.org/pdf/1912.12607.pdf)### Binary
* Binarize本质上是一种极限的Quantize,所面临的问题其实也和Quantize类似,除了几个区别
* Overflow的Tradeoff不存在,由于只有一个量化区间
* 同理Rounding的操作也不存在
* 但是存在的Error都是前向的QuantizationError与等精度的模型进行对比 - Forward Error
* 另外是由于梯度的问题导致收敛不到一个有效的结果 - Gradient Error
* 解决以上两种问题的方法主要有
* 前向误差 - 本质上是通过微调(整体的)在有限的表达能力下尽可能的去近似Counter的一个全精度模型
* Scaling Factor(更细粒度的)
* Adjustable/Learnable的NonLinear Function - PACT
* Gradual/Iterative/Recursive(交替的优化binary函数以及参数)-有点类似于Quantize-aware training的思想
* 通过引入一定的随机性(个人臆测比较玄学,不知道效果如何) - 类似Stochastic Roudning思想的(抖色深的) - Half-wave Gaussian
* 增加模型的表达能力(增加模型大小,堆叠)
* 线性组合(multiple binary basis) - ABCNet
* KD
* Wide
* 反向误差 - 本质上是修正/补偿梯度所带来的误差
* Loss修正 - 最朴素的是进行正则化,加一个奇形怪状的项来拉动
* 比较高一点的是一系列Differentiable的方法,把Loss导到各个部分来协助优化
* 更多learnable parameter - LQNet
* 对STE进行修正- [Binary Neural Networks: A Survey](https://arxiv.org/abs/2004.03333)
* 存在的问题: severe information loss & Discontinuality
* Explainable解释
* some filters are redundnant(play a similar role)
* 有这样的操作,去衡量某一层是否重要-将它从fp换成0/1,看acc decay
* traitioinal understanding for the layer sensitivity, is keeping the 1st and -1st layer with full precision
* beneficial for understanding the structure of NN
* ensemble to help training the BNN, but may face over-fitting
* 分类 naive binarization / optimized binarization
* 后者针对quantize error或者是loss function做了一些额外的操作
* Preliminary
* ![](https://github.com/A-suozhang/MyPicBed/raw/master//img/20200603142600.png)
* BP: STE with clamp(-1,1)
* Category
* ![](https://github.com/A-suozhang/MyPicBed/raw/master//img/20200603142842.png)
* 几个方向
* Minimize Q Error
* Introducing Scale Factor(XNOR)
* Hash Map with Scaling Factor(BWHN)
* Train with Quantized Gradient, improved XNOR(DoReFa)
* More Channels, imporved XNOR(WRPN)
* Group and gradually Quantize
* Recursive approximation (HOPQ-Higher-order residue quantization)
* not 1 step quantization, linear combination of each recursive step
* Similar linear combination of binary masks - ABCNet
* 2 Step Quantization
1. only quantize activation with learnable function
2. then fix function and quantize weight (learn scale factor)
* PACT
* learnable upper bound
* LQ-Net (Learnt Quantization)
* Jointly trainig the NN and the Quantizer
* Surrport arbitary bit
* XNORNet++
* Fuse w/a scaling factor together to support arbitary bits
* Improved Loss func
* 主要focus在如何去globally适配binarization(对应上面的Q-error的就是更focus在局部上)
* LAB - Loss-aware-binarization (quasi-Newton )
* Problems
* degeneration
* saturation
* gradient mismatch
* INQ (Incremental Network Quantization)
* 逐渐把Group扩大
* Apprentice Network
* DBNN - Distillation BinaryNN
* Reduce Gradient Error
* 修正STE
* 最朴素的方式会导致在[-1,1]之间的参数不会被更新(?)
* 核心方式是Soft,Adjustable,SignFunction(然后不用STE)
* ![](https://github.com/A-suozhang/MyPicBed/raw/master//img/20200605144853.png)
* Bi-real: approx sign function(一个二次函数)
* CBCB - Circulant BNN
* BNN+
* HWGQ - Half Gaussian Quantization
* Differentiable Soft Quantization## Others
* Fitting Weight or Activation?
* 从二值化网络开始,很多analytical的求解方式往往是在参数分布高斯的假设前提下,最小化量化后参数与量化前参数的MSE
* 后续的如[这篇](https://arxiv.org/abs/1907.05686),认为参数可以变化,需要保存的是每一层的输出Activation
* 把WA压缩到能放到片上能够显著提升硬件设计性能(显而易见)
* 这可能也是BNN系列的比较繁荣的原因
* Huge Batch Size可以帮助Binary的训练(原理上对低比特同理?)
* Rounding Methods - Neareset/Stochastic/Biased Rounding
* Stochastic Rounding很关键,可以用到更低比特的信息
* ~~但是对硬件设计不太友好~~
* 也是[TernGrad](https://papers.nips.cc/paper/6749-terngrad-ternary-gradients-to-reduce-communication-in-distributed-deep-learning)之所以能work的原因
* 从模型压缩角度看,硬件部署训练的几大难点(和不太elegant的地方)
* 依赖大batch对存储要求高
* 随机rounding,硬件难实现### 经典的建模方式
* 如[Post-Training 4bit](https://arxiv.org/abs/1810.05723)一文,各种Analytical求解的方法都是将量化过程抽象为一个优化问题
* (一般都假设输入的分布为Gauss),最小化量化前后参数的MSE### Stochastic Roudning Related
* Stochastic Rounding对保证Convergence重要
* 但是WRPN作者文中表示自己利用了Full-Precision Grad所以不关键
* Scalable Methods for 8-bit Training of Neural Networks也提到了
* Mixed Precision Training With 8-bit Floating Point文中对RNE和Stochastic Rounding做了一个对比
* 比如imagenet Res18,问题不大,但是Res50会明显Overfiting
* 作者归因为Noisy Gradient错误指引了模型的前进方向,weight被错误更新了
* 同时发现L2范数会激烈增大,导致了Gradient变得更加Noisy,从而Vicious Circle
* 结论是**RNE**的方法对Gradient的Quantization Noise不是很有效
* 认为stochastic rounding参考了被丢掉bit的信息,更稳定### BN Related
* WAGE首先提出BN是训练瓶颈
* [L1BN (Linear BN) ](https://arxiv.org/pdf/1802.09769.pdf)
* 将BN所有的操作都转化为线性
* [RangeBN](https://arxiv.org/pdf/1802.09769.pdf)
* 将BN的Var转化为一个|Max-Min|*(1/sqrt(2ln(n)))
* 假设BN的running mean/var已经稳定
* 然后把他当作常数来计算# Genre
> 这一部分列举了很多文章,但是很多并没有阅读过
## Binary及其延申(极低比特)
> 从一开始的BNN延申开来的一系列work,基本都是利用了二值之后乘法变bitwise,对硬件部署采用非传统运算单元。
* [BNN](https://arxiv.org/abs/1602.02830)
* [BinaryConnect](https://arxiv.org/abs/1511.00363)
* [TernaryNet(TWN)](https://arxiv.org/abs/1605.04711)
* [XNorNet](https://arxiv.org/pdf/1603.05279.pdf)
* [ABCNet](https://arxiv.org/abs/1711.11294)
* [WRPN-Intel-ICLR2018](https://openreview.net/pdf?id=B1ZvaaeAZ)
* [DoReFaNet](https://arxiv.org/pdf/1606.06160.pdf)
* [TTQ(Trained Ternary Quantization)](https://arxiv.org/pdf/1612.01064.pdf)
* [Simultaneously Optimizing Weight and Quantizer of Ternary Neural Network using Truncated Gaussian Approximation](https://arxiv.org/abs/1810.01018)
* [Training Competitive Binary Neural Networks from Scratch](https://arxiv.org/abs/1812.01965)
* ~~最后这篇文章放在这里只是为了告诉大家这个领域到了2019年还在蓬勃发展~~## 量化方法(低比特)
> 使用相对较为传统的比特数(4,6,8),在具体量化方式,以及训练方式入手
* [Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference](https://arxiv.org/abs/1712.05877)
* [PACT](https://arxiv.org/abs/1805.06085)
* [Incremental Quantization](https://ieeexplore.ieee.org/document/4476718/)
* [Model compression via distillation and quantization](https://arxiv.org/abs/1802.05668)
* [Training Deep Neural Networks with 8-bit Floating Point Numbers](https://papers.nips.cc/paper/7994-training-deep-neural-networks-with-8-bit-floating-point-numbers.pdf)## 理论分析
* [Training Quantized Nets: A Deeper Understanding](https://arxiv.org/abs/1706.02379)
* [Towards The Limits Of Network Quantization](https://openreview.net/pdf?id=rJ8uNptgl)
* [Accumulation bit-width Scaling](https://arxiv.org/abs/1901.06588)
* [Per-Tensor-Quantization of BackProp](https://arxiv.org/abs/1812.11732)
* [Understanding Straight-Through Estimator in Training Activation Quantized Neural Nets](https://arxiv.org/abs/1903.05662)
* [An Empirical study of Binary Neural Networks' Optimisation](https://openreview.net/pdf?id=rJfUCoR5KX)
* [Scalable Methods for 8-bit Training of Neural Networks | Part 5](https://arxiv.org/abs/1805.11046)## ~~奇技淫巧~~
* [ernGrad: Ternary Gradients to Reduce Communication in Distributed Deep Learning](https://papers.nips.cc/paper/6749-terngrad-ternary-gradients-to-reduce-communication-in-distributed-deep-learning)
* [Deep Learning with Low Precision by Half-wave Gaussian Quantization](https://arxiv.org/abs/1702.00953)
* [Learning low-precision neural networks without Straight-Through Estimator (STE)](https://arxiv.org/abs/1903.01061)
* [SWALP: Stochastic Weight Averaging in Low-Precision Training](https://arxiv.org/abs/1904.11943)
* [Analysis Of Quantized MOdels-ICLR2019](https://openreview.net/forum?id=ryM_IoAqYX)
* [Training Quantized Network with Auxiliary Gradient Module](https://arxiv.org/abs/1903.11236)
* [Learning to Quantize Deep Networks by Optimizing Quantization Intervals with Task Loss](https://arxiv.org/abs/1808.05779)
* [ReLeQ: An Automatic Reinforcement Learning Approach for Deep Quantization of Neural Networks](http://www-users.cselabs.umn.edu/classes/Spring-2019/csci8980/papers/releq.pdf)
* [And the Bit Goes Down: Revisiting the Quantization of Neural Networks](https://arxiv.org/abs/1907.05686)
* [Ternary MobileNets via Per-Layer Hybrid Filter Banks](https://arxiv.org/abs/1911.01028)
* [Effective Training of Convolutional Neural Networks with Low-bitwidth Weights and Activations](https://arxiv.org/abs/1908.04680)
* [MoBiNet: A Mobile Binary Network for Image Classification](https://arxiv.org/abs/1907.12629)
* [Multi-Precision Quantized Neural Networks via Encoding Decomposition of -1 and +1](https://arxiv.org/abs/1905.13389)# Docs
> 看一下大公司主流的压缩工具都提供了什么功能
## Tensorflow Lite
> tf.contrib.quantize & [Tensorflow Lite](https://www.tensorflow.org/lite/guide)
* 提供了一个[Post-Training-Quantize的工具](https://www.tensorflow.org/lite/performance/post_training_quantization)
* 看上去是很直接的Quantize没有用到什么技巧,标准数据格式之间的转变(float16/int8) **没有自定义的数据格式**
* 文档中直接写到 ```If you want Higher Performance, Use Quantize-aware Training```
* 没有Finetune的过程?
* [Quantize-Aware Training](https://github.com/tensorflow/tensorflow/tree/r1.14/tensorflow/contrib/quantize)
* 只支持部分网络结构(合理)以卷积为主,还有一些RNN
* 这里默认Fold了Conv和BN(我们所说的MergeConvBNs)
* 有一个TOCO(Tf Lite Optimizing Converter)工具可以直接将训练好的FrozenGraph转化为真正的定点模型## PyTorch
> [Quantization Tool](https://pytorch.org/docs/stable/quantization.html?highlight=quantize)
* **QNNPack**
* 支持PerTensor和PerChannel的量化,采用带zeropoint的rounding
* Quantize Aware Training at ```torch.nn.qat torch.nn.intrinsic.qat```
* 提供了很多Observer
* 支持Symmetrical和Asymmetrical的量化---
# Groups (Low-Bit Training)
* Intel (AIPG)
* KAIST (Korea)
* IBM
* Kaust (迪拜的一个学校...)# TODO
* 实现对BN的魔改
* 首先研究它的定点Beahaviour
* ~~还需要再过一过文献中,统计以下目前对BN的认识(除了不能定点之外,大家都是怎么处理的)~~
* 实现Stochastic rounding
* 实现PACT
* 对目前的几种Range方法实现分组
* 对WA做Clamp实现,作为超参数加入
* (?) 能否从TernaryNet这一系列中提取出一些可以参考的优化点(比如训练出一个Range)# References
* [Awesome-Model-Compression](https://github.com/csyhhu/Awesome-Deep-Neural-Network-Compression/blob/master/Paper/PaperByConference.md)
* [Blog N0.0](https://blueardour.github.io/2019/04/29/model-compression-summary.html)
* [TensorBoard Lite Doc](https://www.tensorflow.org/lite/performance/post_training_quantization)
* [Distiller Doc](https://nervanasystems.github.io/distiller/quantization.html)