TY - JOUR
T1 - Classifier-agnostic saliency map extraction
AU - Zolna, Konrad
AU - Geras, Krzysztof J.
AU - Cho, Kyunghyun
N1 - Funding Information:
Konrad Zolna is supported by the National Science Center, Poland (2017/27/N/ST6/00828, 2018/28/T/ST6/00211). Kyunghyun Chothanks support by AdeptMind, eBay, TenCent, NVIDIA and CIFAR. The authors would also like to thank Catriona C. Geras for correcting earlier versions of the manuscript.
Funding Information:
Konrad Zolna is supported by the National Science Center, Poland ( 2017/27/N/ST6/00828 , 2018/28/T/ST6/00211 ). Kyunghyun Chothanks support by AdeptMind , eBay , TenCent , NVIDIA and CIFAR . The authors would also like to thank Catriona C. Geras for correcting earlier versions of the manuscript.
Publisher Copyright:
© 2020
PY - 2020/7
Y1 - 2020/7
N2 - Currently available methods for extracting saliency maps identify parts of the input which are the most important to a specific fixed classifier. We show that this strong dependence on a given classifier hinders their performance. To address this problem, we propose classifier-agnostic saliency map extraction, which finds all parts of the image that any classifier could use, not just one given in advance. We observe that the proposed approach extracts higher quality saliency maps than prior work while being conceptually simple and easy to implement. The method sets the new state of the art result for localization task on the ImageNet data, outperforming all existing weakly-supervised localization techniques, despite not using the ground truth labels at the inference time. The code reproducing the results is available at https://github.com/kondiz/casme.
AB - Currently available methods for extracting saliency maps identify parts of the input which are the most important to a specific fixed classifier. We show that this strong dependence on a given classifier hinders their performance. To address this problem, we propose classifier-agnostic saliency map extraction, which finds all parts of the image that any classifier could use, not just one given in advance. We observe that the proposed approach extracts higher quality saliency maps than prior work while being conceptually simple and easy to implement. The method sets the new state of the art result for localization task on the ImageNet data, outperforming all existing weakly-supervised localization techniques, despite not using the ground truth labels at the inference time. The code reproducing the results is available at https://github.com/kondiz/casme.
KW - Convolutional neural networks
KW - Image classification
KW - Saliency map
KW - Weakly supervised localization
UR - http://www.scopus.com/inward/record.url?scp=85083900530&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85083900530&partnerID=8YFLogxK
U2 - 10.1016/j.cviu.2020.102969
DO - 10.1016/j.cviu.2020.102969
M3 - Article
AN - SCOPUS:85083900530
SN - 1077-3142
VL - 196
JO - Computer Vision and Image Understanding
JF - Computer Vision and Image Understanding
M1 - 102969
ER -