TY - JOUR
T1 - A survey on GAN acceleration using memory compression techniques
AU - Tantawy, Dina
AU - Zahran, Mohamed
AU - Wassal, Amr
N1 - Funding Information:
No funding was provided for this work.
Publisher Copyright:
© 2021, The Author(s).
PY - 2021/12
Y1 - 2021/12
N2 - Since its invention, generative adversarial networks (GANs) have shown outstanding results in many applications. GANs are powerful, yet resource-hungry deep learning models. The main difference between GANs and ordinary deep learning models is the nature of their output and training instability. For example, GANs output can be a whole image versus other models detecting objects or classifying images. Thus, the architecture and numeric precision of the network affect the quality and speed of the solution. Hence, accelerating GANs is pivotal. Data transfer is considered the main source of energy consumption, that is why memory compression is a very efficient technique to accelerate and optimize GANs. Two main types of memory compression exist: lossless and lossy ones. Lossless compression techniques are general among all models; thus, we will focus in this paper on lossy techniques. Lossy compression techniques are further classified into (a) pruning, (b) knowledge distillation, (c) low-rank factorization, (d) lowering numeric precision, and (e) encoding. In this paper, we survey lossy compression techniques for CNN-based GANs. Our findings showed the superiority of knowledge distillation over pruning alone and the gaps in the research field that needs to be explored like encoding and different combination of compression techniques.
AB - Since its invention, generative adversarial networks (GANs) have shown outstanding results in many applications. GANs are powerful, yet resource-hungry deep learning models. The main difference between GANs and ordinary deep learning models is the nature of their output and training instability. For example, GANs output can be a whole image versus other models detecting objects or classifying images. Thus, the architecture and numeric precision of the network affect the quality and speed of the solution. Hence, accelerating GANs is pivotal. Data transfer is considered the main source of energy consumption, that is why memory compression is a very efficient technique to accelerate and optimize GANs. Two main types of memory compression exist: lossless and lossy ones. Lossless compression techniques are general among all models; thus, we will focus in this paper on lossy techniques. Lossy compression techniques are further classified into (a) pruning, (b) knowledge distillation, (c) low-rank factorization, (d) lowering numeric precision, and (e) encoding. In this paper, we survey lossy compression techniques for CNN-based GANs. Our findings showed the superiority of knowledge distillation over pruning alone and the gaps in the research field that needs to be explored like encoding and different combination of compression techniques.
KW - Acceleration
KW - Compression
KW - Generative adversarial networks
KW - Knowledge distillation
KW - Optimization
KW - Pruning
KW - Quantization
KW - Survey
UR - http://www.scopus.com/inward/record.url?scp=85121419754&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85121419754&partnerID=8YFLogxK
U2 - 10.1186/s44147-021-00045-5
DO - 10.1186/s44147-021-00045-5
M3 - Review article
AN - SCOPUS:85121419754
SN - 1110-1903
VL - 68
JO - Journal of Engineering and Applied Science
JF - Journal of Engineering and Applied Science
IS - 1
M1 - 47
ER -