TY - JOUR
T1 - Understanding the Detrimental Class-level Effects of Data Augmentation
AU - Kirichenko, Polina
AU - Ibrahim, Mark
AU - Balestriero, Randall
AU - Bouchacourt, Diane
AU - Vedantam, Ramakrishna
AU - Firooz, Hamed
AU - Wilson, Andrew Gordon
N1 - Publisher Copyright:
© 2023 Neural information processing systems foundation. All rights reserved.
PY - 2023
Y1 - 2023
N2 - Data augmentation (DA) encodes invariance and provides implicit regularization critical to a model's performance in image classification tasks.However, while DA improves average accuracy, recent studies have shown that its impact can be highly class dependent: achieving optimal average accuracy comes at the cost of significantly hurting individual class accuracy by as much as 20% on ImageNet.There has been little progress in resolving class-level accuracy drops due to a limited understanding of these effects.In this work, we present a framework for understanding how DA interacts with class-level learning dynamics.Using higher-quality multi-label annotations on ImageNet, we systematically categorize the affected classes and find that the majority are inherently ambiguous, co-occur, or involve fine-grained distinctions, while DA controls the model's bias towards one of the closely related classes.While many of the previously reported performance drops are explained by multi-label annotations, our analysis of class confusions reveals other sources of accuracy degradation.We show that simple class-conditional augmentation strategies informed by our framework improve performance on the negatively affected classes.
AB - Data augmentation (DA) encodes invariance and provides implicit regularization critical to a model's performance in image classification tasks.However, while DA improves average accuracy, recent studies have shown that its impact can be highly class dependent: achieving optimal average accuracy comes at the cost of significantly hurting individual class accuracy by as much as 20% on ImageNet.There has been little progress in resolving class-level accuracy drops due to a limited understanding of these effects.In this work, we present a framework for understanding how DA interacts with class-level learning dynamics.Using higher-quality multi-label annotations on ImageNet, we systematically categorize the affected classes and find that the majority are inherently ambiguous, co-occur, or involve fine-grained distinctions, while DA controls the model's bias towards one of the closely related classes.While many of the previously reported performance drops are explained by multi-label annotations, our analysis of class confusions reveals other sources of accuracy degradation.We show that simple class-conditional augmentation strategies informed by our framework improve performance on the negatively affected classes.
UR - http://www.scopus.com/inward/record.url?scp=85191189679&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85191189679&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85191189679
SN - 1049-5258
VL - 36
JO - Advances in Neural Information Processing Systems
JF - Advances in Neural Information Processing Systems
T2 - 37th Conference on Neural Information Processing Systems, NeurIPS 2023
Y2 - 10 December 2023 through 16 December 2023
ER -