CONSISTENT COUNTERFACTUALS FOR DEEP MODELS

Emily Black, Zifan Wang, Anupam Datta, Matt Fredrikson

    Research output: Contribution to conferencePaperpeer-review

    Abstract

    Counterfactual examples are one of the most commonly-cited methods for explaining the predictions of machine learning models in key areas such as finance and medical diagnosis. Counterfactuals are often discussed under the assumption that the model on which they will be used is static, but in deployment models may be periodically retrained or fine-tuned. This paper studies the consistency of model prediction on counterfactual examples in deep networks under small changes to initial training conditions, such as weight initialization and leave-one-out variations in data, as often occurs during model deployment. We demonstrate experimentally that counterfactual examples for deep models are often inconsistent across such small changes, and that increasing the cost of the counterfactual, a stability-enhancing mitigation suggested by prior work in the context of simpler models, is not a reliable heuristic in deep networks. Rather, our analysis shows that a model's Lipschitz continuity around the counterfactual, along with confidence of its prediction, is key to its consistency across related models. To this end, we propose Stable Neighbor Search as a way to generate more consistent counterfactual explanations, and illustrate the effectiveness of this approach on several benchmark datasets.

    Original languageEnglish (US)
    StatePublished - 2022
    Event10th International Conference on Learning Representations, ICLR 2022 - Virtual, Online
    Duration: Apr 25 2022Apr 29 2022

    Conference

    Conference10th International Conference on Learning Representations, ICLR 2022
    CityVirtual, Online
    Period4/25/224/29/22

    ASJC Scopus subject areas

    • Language and Linguistics
    • Computer Science Applications
    • Education
    • Linguistics and Language

    Fingerprint

    Dive into the research topics of 'CONSISTENT COUNTERFACTUALS FOR DEEP MODELS'. Together they form a unique fingerprint.

    Cite this