TY - GEN
T1 - Gabor is Enough
T2 - 14th IEEE Image, Video, and Multidimensional Signal Processing Workshop, IVMSP 2022
AU - Janjusevic, Nikola
AU - Khalilian-Gourtani, Amirhossein
AU - Wang, Yao
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Image processing neural networks, natural and artificial, have a long history with orientation-selectivity, often described mathematically as Gabor filters. Gabor-like filters have been observed in the early layers of CNN classifiers and even throughout low-level image processing networks. In this work, we take this observation to the extreme and explicitly constrain the filters of a natural-image denoising CNN to be learned 2D real Gabor filters. Surprisingly, we find that the proposed network (GDLNet) can achieve near state-of-the-art denoising performance amongst popular fully convolutional neural networks, with only a fraction of the learned parameters. We further verify that this parameterization maintains the noiselevel generalization (training vs. inference mismatch) characteristics of the base network, and investigate the contribution of individual Gabor filter parameters to the performance of the denoiser. We present positive findings for the interpretation of dictionary learning networks as performing accelerated sparse-coding via the importance of untied learned scale parameters between network layers. Our network's success suggests that representations used by low-level image processing CNNs can be as simple and interpretable as Gabor filterbanks.
AB - Image processing neural networks, natural and artificial, have a long history with orientation-selectivity, often described mathematically as Gabor filters. Gabor-like filters have been observed in the early layers of CNN classifiers and even throughout low-level image processing networks. In this work, we take this observation to the extreme and explicitly constrain the filters of a natural-image denoising CNN to be learned 2D real Gabor filters. Surprisingly, we find that the proposed network (GDLNet) can achieve near state-of-the-art denoising performance amongst popular fully convolutional neural networks, with only a fraction of the learned parameters. We further verify that this parameterization maintains the noiselevel generalization (training vs. inference mismatch) characteristics of the base network, and investigate the contribution of individual Gabor filter parameters to the performance of the denoiser. We present positive findings for the interpretation of dictionary learning networks as performing accelerated sparse-coding via the importance of untied learned scale parameters between network layers. Our network's success suggests that representations used by low-level image processing CNNs can be as simple and interpretable as Gabor filterbanks.
UR - http://www.scopus.com/inward/record.url?scp=85131346488&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85131346488&partnerID=8YFLogxK
U2 - 10.1109/IVMSP54334.2022.9816313
DO - 10.1109/IVMSP54334.2022.9816313
M3 - Conference contribution
AN - SCOPUS:85131346488
T3 - IVMSP 2022 - 2022 IEEE 14th Image, Video, and Multidimensional Signal Processing Workshop
BT - IVMSP 2022 - 2022 IEEE 14th Image, Video, and Multidimensional Signal Processing Workshop
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 26 June 2022 through 29 June 2022
ER -