Implications of Deep Compression with Complex Neural Networks

Main Article Content

Lily Young
James Richrdson York
Byeong Kil Lee

Abstract

Deep learning and neural networks have become increasingly popular in the area of artificial intelligence. These models have the capability to solve complex problems, such as image recognition or language processing. However, the memory utilization and power consumption of these networks can be very large for many applications. This has led to research into techniques to compress the size of these models while retaining accuracy and performance. One of the compression techniques is the deep compression three-stage pipeline, including pruning, trained quantization, and Huffman coding. In this paper, we apply the principles of deep compression to multiple complex networks in order to compare the effectiveness of deep compression in terms of compression ratio and the quality of the compressed network. While the deep compression pipeline is effectively working for CNN and RNN models to reduce the network size with small performance degradation, it is not properly working for more complicated networks such as GAN. In our GAN experiments, performance degradation is too much from the compression. For complex neural networks, careful analysis should be done for discovering which parameters allow a GAN to be compressed without loss in output quality.

Downloads

Download data is not yet available.

Article Details

How to Cite
[1]
Lily Young, James Richrdson York, and Byeong Kil Lee, “Implications of Deep Compression with Complex Neural Networks”, IJSCE, vol. 13, no. 3, pp. 1–6, Jul. 2023, doi: 10.35940/ijsce.C3613.0713323.
Section
Articles

How to Cite

[1]
Lily Young, James Richrdson York, and Byeong Kil Lee, “Implications of Deep Compression with Complex Neural Networks”, IJSCE, vol. 13, no. 3, pp. 1–6, Jul. 2023, doi: 10.35940/ijsce.C3613.0713323.

References

S. Han, H, Mao, and W J. Dally, “Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization, and Huffman Coding”, (ICLR) 2016

J.Luo, et al., “ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression”, IEEE International Conference on Computer Vision, 2017.

H.Li, et al., “Pruning Filter for Efficient ConvNets”, International Conference in Learning Represenations (ICLR), 2017.

Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770-778

“Trim insignificant weights | TensorFlow Model Optimization,” https://www.tensorflow.org/model_optimization/guide/pruning (accessed Dec. 12, 2022).

“Quantization aware training in Keras example | TensorFlow Model Optimization,” https://www.tensorflow.org/model_optimization/guide/quantization/training_example (accessed Dec. 12, 2022).

“RNN, LSTM & GRU,” dProgrammer lopez, Apr. 06, 2019. http://dprogrammer.org/rnn-lstm-gru

Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2020. Generative adversarial networks. Commun. ACM 63, 11 (November 2020), 139–144.