Quantizaiton on weight is part of model compression issue.
This paper proposes a
After the training stage, the means of weights densely cluster at three locations in M=3 case. The histogram showes that the distribution over weights is "multiple spike-and-slab". Also, the locations of spikes are asymmetric about zero.
The quantized loss is the difference between quantized weights and full-precision weights, and it converges fastly.
"train_densenet.py" is the main file that runs adaptive quantization method on deep Bayesian DenseNet with CIFAR10.
"train_nonsymLeNet.py" is the main file that runs adaptive quantization method on deep LeNet with MNIST.
"layers.py" is the module of Bayesian convolutional layer and fully connected layer.
- Hardware:
- CPU: Intel Core i7-4930K @3.40 GHz
- RAM: 64 GB DDR3-1600
- GPU: GeForce GTX 1080ti
- pytorch
- Dataset
- MNIST
- CIFAR10
- CIFAR100