PlumX Metrics
Embed PlumX Metrics

Contrasting Explanations for Understanding and Regularizing Model Adaptations

Neural Processing Letters, ISSN: 1573-773X, Vol: 55, Issue: 5, Page: 5273-5297
2023
  • 3
    Citations
  • 0
    Usage
  • 19
    Captures
  • 0
    Mentions
  • 0
    Social Media
Metric Options:   Counts1 Year3 Year

Metrics Details

  • Citations
    3
    • Citation Indexes
      3
  • Captures
    19

Article Description

Many of today’s decision making systems deployed in the real world are not static—they are changing and adapting over time, a phenomenon known as model adaptation takes place. Because of their wide reaching influence and potentially serious consequences, the need for transparency and interpretability of AI-based decision making systems is widely accepted and thus have been worked on extensively—e.g. a very prominent class of explanations are contrasting explanations which try to mimic human explanations. However, usually, explanation methods assume a static system that has to be explained. Explaining non-static systems is still an open research question, which poses the challenge how to explain model differences, adaptations and changes. In this contribution, we propose and (empirically) evaluate a general framework for explaining model adaptations and differences by contrasting explanations. We also propose a method for automatically finding regions in data space that are affected by a given model adaptation—i.e. regions where the internal reasoning of the other (e.g. adapted) model changed—and thus should be explained. Finally, we also propose a regularization for model adaptations to ensure that the internal reasoning of the adapted model does not change in an unwanted way.

Bibliographic Details

Provide Feedback

Have ideas for a new metric? Would you like to see something else here?Let us know