CNN/Deep_Compression/模型定点化路线图.md
首先定点化的setting分好几种,主要如下所示 (w代表weight,a代表activation,g代表gradient)
最近两年的目前有13篇直接相关的论文,截止到2016年7月
w定点,a浮点
Resiliency of Deep Neural Networks under Quantization [Wongyong Sung, Sungho Shin, 2016.01.07, ICLR2016] {5bit在CIFAR10上恢复正确率}
Fixed Point Quantization of Deep Convolutional Networks [Darryl D.Lin, Sachin S.Talathi, 2016.06.02] {每层定点化策略不同,解析解求出}
w+a定点
{ImageNet上8bit-8bit掉0.9%,AlexNet}
{结合硬件的trick可以在ImageNet上4-10bit}
{ImageNet上8bit-8bit掉1%,AlexNet}
fine-tune整个网络
w定点,a+g浮点
{2bit即三值网络在CIFAR10上恢复正确率}
w+a定点,g浮点
{每层定点化策略不同,解析解求出,CIFAR10上fine-tune后4bit-4bit掉1.32%}
w+a+g定点
Overcoming Challenges in Fixed Point Training of Deep Convolutional Networks [Darryl D.Lin, Sachin S. Talathi, Qualcomm Research,2016.07.08] {无随机rounding,ImageNet上4bit-16bit-16bit掉7.2%,a和g再小就不收敛}
DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients [Shuchang Zhou, Zekun Ni, 2016.06.20] {1bit-2bit-4bit, 第一层和最后一层没有量化,ImageNet上掉5.2%}
fine-tune最高几层
w+a+g定点
{无随机rounding,ImageNet上4bit-4bit-4bit掉23.3%}
分阶段地从低层到高层fine-tune网络
w+a+g定点
{无随机rounding,ImageNet上4bit-4bit-4bit Top5掉11.5%}
w定点,a+g浮点
二值网络
{CIFAR10上8.27%, state-of-art}
{ImageNet上39.2%,掉2.8%, AlexNet}
三值网络
{ImageNet掉2.3%, ResNet-18B}
w+a定点,g浮点
二值网络
{CIFAR10上10.15%}
{ImageNet上55.8%, 掉12.4%}
w+a+g定点
{ 随机rounding技巧,CIFAR10上16bit+16bit+16bit复现正确率}
Training deep neural networks with low precision multiplications [Matthieu Courbariaux, Yoshua Bengio, ICLR 2015 Workshop]
{ CIFAR10上10bit+10bit+12bit复现正确率 }
DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients [Shuchang Zhou, Zekun Ni, 2016.06.20]
{1bit-2bit-4bit, 第一层和最后一层没有量化,ImageNet上掉8.8%}
{1bit-2bit-6bit,ImageNet上超过DoReFa 0.33%}