@inproceedings{5053d9cfa02e4d7a9b818848dceb336e,
title = "Adaptive weight compression for memory-efficient neural networks",
abstract = "Neural networks generally require significant memory capacity/bandwidth to store/access a large number of synaptic weights. This paper presents an application of JPEG image encoding to compress the weights by exploiting the spatial locality and smoothness of the weight matrix. To minimize the loss of accuracy due to JPEG encoding, we propose to adaptively control the quantization factor of the JPEG algorithm depending on the error-sensitivity (gradient) of each weight. With the adaptive compression technique, the weight blocks with higher sensitivity are compressed less for higher accuracy. The adaptive compression reduces memory requirement, which in turn results in higher performance and lower energy of neural network hardware. The simulation for inference hardware for multilayer perceptron with the MNIST dataset shows up to 42X compression with less than 1\% loss of recognition accuracy, resulting in 3X higher effective memory bandwidth and ∼19X lower system energy.",
keywords = "Compression, JPEG, Memory-efficient, MLP, Neural network, Weight",
author = "Ko, \{Jong Hwan\} and Duckhwan Kim and Taesik Na and Jaeha Kung and Saibal Mukhopadhyay",
note = "Publisher Copyright: {\textcopyright} 2017 IEEE.; 20th Design, Automation and Test in Europe, DATE 2017 ; Conference date: 27-03-2017 Through 31-03-2017",
year = "2017",
month = may,
day = "11",
doi = "10.23919/DATE.2017.7926982",
language = "English",
series = "Proceedings of the 2017 Design, Automation and Test in Europe, DATE 2017",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "199--204",
booktitle = "Proceedings of the 2017 Design, Automation and Test in Europe, DATE 2017",
}