Validation: Counterfactual Explanation Robustness to Adversarial Noise

ValidatedEL 6/8TF 6/8

Innovation Maturity

Evidence Level:6/8Explicitly Described
Timeframe:6/8Short Term (6-12 mo)

Evidence: The FCA builds on several published methods (CECAS, DCMP, etc.) that are explicitly described in literature, but the integrated architecture itself is a novel combination not yet deployed.

Timeframe: Integrating existing components and validating robustness can be achieved within 6–12 months of focused development.

7.1 Identify the Objective

The central research challenge is to develop counterfactual explanation (CE) mechanisms that remain faithful, actionable, and interpretable when subjected to adversarial perturbations—both input‑level noise and model‑level shifts. Existing CE methods exhibit brittleness: perturbations that flip a model’s prediction are often treated as noisy artifacts rather than actionable changes, leading to misleading explanations and compromised user trust. Our objective is to bridge the gap between the optimization goals of adversarial attacks and the human‑interpretable, causally grounded requirements of counterfactual explanations in multi‑agent, adversarial settings.

7.3 Ideate/Innovate

We propose a Frontier CE Architecture (FCA) that integrates four complementary innovations:

  1. Causally‑Guided Adversarial Steering (CECAS‑style)
    Employ a causal graph learned from domain data to steer adversarial perturbations only along edges that preserve causal consistency. This prevents unintended alterations that violate domain semantics, as demonstrated in CECAS [1][2].

  2. Diffusion‑Constrained Manifold Projection (ACE‑DMP)
    Use a denoising diffusion probabilistic model (DDPM) to project raw adversarial perturbations onto the data manifold before evaluation. The filtering function (F_{\tau}) ensures high‑frequency artifacts are removed while retaining the semantic direction of the perturbation [3] .

  3. Multi‑Modal Adversarial Recourse Module (MARM)
    Extend CE to images, text, and graph data simultaneously by generating adversarial examples that respect cross‑modal causal constraints. This is essential for multi‑agent coordination where agents share heterogeneous observations.

  4. Robust Recourse Optimizer with Lp‑Bounded Model Change (RO‑Lp)
    Incorporate an optimization framework that bounds model changes in the (\ell_p) sense [4][5], ensuring that the CE remains valid even when the underlying model undergoes adversarial or data‑poisoning updates.

The FCA pipeline first learns a causal graph (or uses an expert‑defined one), then uses diffusion‑based on‑manifold projection to generate candidate counterfactuals, and finally optimizes for minimal action cost under an (\ell_p) model‑change constraint. The final CE is evaluated against a held‑out robustness oracle that simulates potential adversarial model variations.

Independent Validation

Causal‑Guided Adversarial Steering

causal graph steering adversarial perturbations causal consistencyCECAS causal steering adversarial robustnesscausal edge perturbation prevention spurious correlationcausal consistency adversarial example generationdomain semantics preserving adversarial steering
Causal‑guided adversarial steering seeks to exploit the causal structure of multimodal representations so that perturbations are both efficient and semantically coherent. In vision‑language‑action (VLA) models, the SAGA framework demonstrates that targeting high‑attention regions with sparse, patch‑wise perturbations yields attack success rates comparable to or exceeding dense‑patch methods while preserving visual plausibility [v4266]. This attention‑guided strategy aligns with the observation that attention scores correlate positively with loss sensitivity, enabling a more focused use of the perturbation budget.Building on this, a Cognitive Perturbation Protocol introduces user‑bias simulations during training, which are distilled into a lightweight Evidence Critic that scores documents for evidential strength. The critic learns to steer the model toward correct outputs even when queries are adversarially perturbed [v1211]. This causal intervention approach mirrors the Residual Semantic Steering (RSS) framework, which disentangles physical affordance from semantic intent by employing Monte Carlo syntactic integration, thereby mitigating the “modality collapse” that causes VLA agents to overfit to specific linguistic cues [v8528].A key challenge for these methods is the stability of the underlying representational geometry. Recent work provides a metric that predicts steering success a priori by measuring the geometric stability of linear directions assumed by representation‑engineering techniques [v17005]. When this stability is low, steering vectors become unreliable across contexts or model updates, limiting the practical impact of causal‑guided attacks. Cross‑modal preference steering further illustrates the power of joint visual‑textual perturbations, achieving higher manipulation success under realistic attacker capabilities than single‑modal attacks [v15838]. Together, these studies underscore that effective causal‑guided adversarial steering requires both attention‑aware perturbation design and robust, causally interpretable representations.

Diffusion‑Constrained Manifold Projection

denoising diffusion probabilistic model manifold projection counterfactualsDDPM data manifold filtering high‑frequency artifactsdiffusion‑based projection counterfactual fidelityACE‑DMP diffusion constrained counterfactual generationsemantic direction diffusion counterfactuals
Diffusion‑constrained manifold projection (DCMP) is a framework that leverages denoising diffusion probabilistic models (DDPMs) to generate counterfactual or edited images that remain on the underlying data manifold. By iteratively denoising a perturbed sample, the diffusion process implicitly enforces that the final output is a realistic data point, thereby avoiding the off‑manifold artifacts that plague naïve gradient‑based perturbations. This approach has been formalized in visual counterfactual explainer (VCE) pipelines, where the DDPM is used as a generative prior that guides the search for plausible counterfactuals while suppressing gradients that do not align with the manifold [v12930].The manifold constraint not only improves visual plausibility but also mitigates on‑manifold spurious function variations. By projecting the gradient through the decoder stack, DCMP removes components of the model’s decision surface that are orthogonal to the data manifold, leading to counterfactuals that are both minimal and semantically meaningful. Recent work on inverse problems has shown that adding a manifold penalty to the diffusion objective yields higher fidelity reconstructions and reduces hallucinations, especially in high‑dimensional image spaces [v2830].In medical imaging, DCMP has been applied to generate healthy counterfactuals for lesion analysis. A typical pipeline first constructs a healthy reference image via inpainting, then optimizes a latent diffusion objective that balances fidelity to the original and similarity to the healthy reference. The resulting counterfactuals preserve anatomical context while removing pathological features, enabling interpretable model explanations and data augmentation for scarce clinical datasets [v15368]. Similar strategies have been used for histopathology, where diffusion autoencoders produce realistic tissue edits that expose classifier decision boundaries [v16089].Implementing DCMP requires careful tuning of the diffusion schedule and guidance strength. The standard DDPM forward–reverse process is computationally intensive, but recent fast samplers (e.g., DDIM, DPM‑Solver) reduce the number of denoising steps while maintaining manifold adherence [v14059]. Consequently, DCMP offers a principled, scalable method for producing high‑quality counterfactuals that respect the intrinsic structure of complex image domains.

Multi‑Modal Adversarial Recourse Module

multi‑modal counterfactual explanation images text graphcross‑modal causal constraints adversarial recourseMARM multi‑modal adversarial example generationheterogeneous observation counterfactuals multi‑agentvision‑language graph counterfactual robustness
Multi‑modal adversarial recourse modules aim to combine robust, explainable, and clinically actionable outputs from vision‑language models (VLMs) with downstream decision‑support pipelines. Recent work on VLM defenses shows that parameter‑efficient adversarial training (e.g., AdvPT, APT) can harden cross‑modal embeddings while keeping inference latency low, and that a cross‑modal consistency loss further improves robustness to multimodal perturbations [v9141]. These techniques provide a foundation for a recourse module that can generate counterfactual explanations that remain valid even under adversarial manipulation.Explainability is critical in medical settings, where a VLM’s diagnostic prediction must be interpretable to clinicians and patients. An integrated explainable‑AI component that produces visual heatmaps and textual rationales, and that can embed the resulting report into an electronic health record via HL7/FHIR standards, has been demonstrated in recent radiology‑AI systems [v16245]. Coupling such a module with adversarially robust embeddings ensures that the explanations themselves are not easily spoofed, thereby preserving trust.Counterfactual recourse requires that the model can identify minimal, clinically plausible changes to multimodal inputs that would alter a prediction. Recent research proposes adaptive adversarial training that dynamically adjusts difficulty based on model state, and introduces contrastive loss regularization to enforce a structured latent space that supports counterfactual reasoning [v11082]. By aligning visual and textual modalities in a shared space, the module can generate coherent “what‑if” scenarios that respect both image‑based pathology and textual clinical context.Finally, the module must be evaluated against a suite of multimodal adversarial attacks, including prompt‑injection and cross‑modal consistency violations. Benchmarking frameworks such as CARLA and RAG‑Anything provide a standardized testbed for measuring robustness and interpretability across modalities [v15921]. Integrating these benchmarks into the development cycle allows continuous validation of both the adversarial defenses and the recourse generation logic, ensuring that the system remains reliable in real‑world clinical deployments.

Robust Recourse Optimizer with Lp‑Bounded Model Change

Lp bounded model change counterfactual optimizerrobust recourse optimization Lp norm model driftmodel change constraint counterfactual validityadversarial training poisoning Lp bounded recoursedistribution shift robust counterfactual Lp
Robust counterfactual recourse that remains valid under model updates is a growing research frontier. Recent work has formalised the problem as a min‑max optimisation over a bounded uncertainty set in parameter space, typically measured by an \(L_{p}\) norm. For generalized linear models, Kayastha et al. derived an optimal algorithm that reduces the non‑convex robust recourse problem to a tractable collection of convex sub‑problems, achieving substantial cost savings compared with naïve \(L_{\infty}\)‑based methods and with existing heuristic generators [v6294]. Their empirical studies on real‑world datasets show that the algorithm can lower the price of recourse by orders of magnitude while preserving proximity and feasibility.Theoretical guarantees for robustness have also been extended beyond linear models. A recent framework introduces a “naturally‑occurring” model‑change abstraction that allows arbitrary parameter shifts as long as prediction changes on the data manifold are bounded. This relaxation captures realistic scenarios where models drift in high‑dimensional parameter space yet maintain similar decision boundaries. The authors provide probabilistic robustness guarantees for any model class, and demonstrate that their robust recourse construction remains valid under such natural changes [v1977]. These results bridge the gap between worst‑case adversarial bounds and more realistic, data‑driven model evolution.Robustness metrics are essential for evaluating and comparing methods. A recent study proposes a multiplicity‑based robustness score that quantifies the fraction of counterfactuals that stay valid across a set of perturbed models. The score, ranging from 0 to 1, is computed by sampling models within a prescribed \(L_{p}\) radius and checking counterfactual feasibility. Experiments on benchmark tabular datasets show that robust generators achieve higher scores than conventional approaches, confirming the practical relevance of the metric [v8791]. Together, these advances establish a coherent pipeline: a formal robustness definition, an efficient algorithm for optimal recourse under \(L_{p}\) constraints, and a principled evaluation metric that captures real‑world model drift.

FCA Pipeline: Causal Graph + Diffusion Projection

FCA pipeline causal graph learning counterfactual generationcausal graph diffusion projection minimal action costcounterfactual optimization Lp model change FCAFCA counterfactual pipeline evaluation robustness oracleadversarial model variation counterfactual pipeline
The FCA Pipeline proposes a two‑stage workflow that first learns a causal graph from observational data and then projects counterfactual scenarios through a diffusion model. The causal discovery step leverages fast, graph‑free techniques such as FCI and GAC to identify admissible mediators and proxies while preserving differential privacy, thereby enabling per‑instance counterfactual consistency (SCC) without requiring a full structural causal model ([v13179], 0ffcc068918df33). By separating discovery from inference, the pipeline mitigates the risk of overfitting to spurious correlations and supports robust fairness audits that focus on individual‑level stability rather than group parity.Diffusion projection is employed to generate realistic counterfactual samples conditioned on the learned causal structure. Recent work on graph‑aware diffusion models shows that incorporating GNN‑based message passing can preserve local dependencies while allowing global perturbations, which is essential for faithfully simulating interventions [v5831]. The CCAGNN architecture demonstrates how dual‑encoder GNNs can jointly estimate causal and non‑causal feature effects, providing a principled way to embed counterfactual constraints into the diffusion process [v7542]. However, diffusion models remain computationally intensive, and their training stability can degrade when the underlying graph is large or highly connected.Topological ordering and directed graph policy optimization (DGPO) offer a complementary strategy to enforce causal directionality in the diffusion step. By imposing an upper‑triangular adjacency structure and positional encodings that respect node ordering, DGPO reduces the search space for valid interventions and improves interpretability of the generated counterfactuals [v7081]. This approach also facilitates efficient inference on edge‑directed graphs, which is critical for real‑time decision support in high‑stakes domains such as healthcare and finance.Overall, the FCA Pipeline’s modular design—causal graph discovery, privacy‑preserving feature selection, and diffusion‑based counterfactual generation—offers a scalable framework for individual‑level fairness and robustness. Future work should focus on integrating approximate inference techniques for large‑scale graphs, developing lightweight diffusion backbones that maintain fidelity, and establishing standardized evaluation suites that jointly assess causal consistency, privacy guarantees, and computational efficiency.

Robustness Oracle Evaluation

robustness oracle adversarial model simulation counterfactualsworst‑case scenario counterfactual evaluation oraclerobustness oracle sanity‑check protocols counterfactualadversarial model variants evaluation counterfactualoracle‑based counterfactual robustness assessment
Robustness oracle evaluation seeks to replace the elusive “ground‑truth” oracle that many AI systems lack with a reproducible, model‑agnostic proxy. Metamorphic testing provides a principled way to do this by checking that a model’s output transforms consistently under known input manipulations (e.g., image rotation or synonym replacement) and that invariant logical properties hold across perturbations. This approach is especially valuable for non‑deterministic generative models where a single correct answer is unavailable. [v3453]A practical instantiation of an oracle is the in‑the‑loop gain evaluation, which treats the user as a surrogate oracle and measures the improvement in model performance rather than relying on subjective feedback. By quantifying the percentage of the performance gap closed between a baseline and a corrected model, this method avoids logical fallacies inherent in human‑based studies and yields fully reproducible results. [v10859]Oracle distillation further refines robustness assessment by training a separate classifier to mimic the decision strategy of the target model. Because the distilled oracle is trained from scratch, it is immune to weight‑specific adversarial attacks that would otherwise transfer to the original model. The resulting “gain” metric normalizes across baselines of varying difficulty, providing a fair comparison of robustness improvements. [v5423]The effectiveness of counterfactual (CF) oracles depends on the number of labeled CF examples. Empirical studies show that the constraint‑feasibility score rises sharply with additional labeled inputs, reaching about 80 % feasibility with 100 labels, while the generation time per CF example decreases as batch size grows. These findings highlight the trade‑off between labeling effort and oracle reliability, and suggest that generative CF methods can offer computational advantages over search‑based baselines. [v12247]Finally, robustness evaluation must be coupled with bias and fairness audits. Counterfactual testing—creating prompt pairs that differ only in a protected attribute—provides a transparent, legally defensible way to detect discriminatory behavior. When combined with automated bias‑detection tools, this approach ensures that an oracle’s predictions remain equitable across demographic groups. [v12560]

FCA vs Conventional Counterfactual Methods

FCA causal integrity counterfactual superioritymanifold fidelity counterfactual diffusion advantagemulti‑modal robustness counterfactual comparisonmodel drift resilience counterfactual FCAscalable evaluation counterfactual robustness oracle
Fairness‑centric counterfactual analysis (FCA) explicitly embeds outcome‑parity or equal‑opportunity constraints into the generation of counterfactuals, ensuring that the synthetic “what‑if” scenarios respect protected‑group fairness metrics. Conventional counterfactual methods, by contrast, focus primarily on three desiderata—validity, proximity, and plausibility—without regard to how the counterfactuals may shift risk or benefit across demographic slices. FCA therefore offers a principled way to audit and correct bias in downstream decisions, but it also demands individual‑level causal models that are often unavailable in aggregate or high‑dimensional settings. [v16482]A key vulnerability of standard counterfactual explanations is their susceptibility to data‑poisoning attacks. By subtly corrupting a small subset of training examples, an adversary can inflate the cost of recourse or force the model to produce implausible counterfactuals, thereby undermining user trust. FCA’s fairness constraints can mitigate some of these effects by penalizing counterfactuals that disproportionately alter protected‑group outcomes, but the underlying model still needs to be robust to poisoning. Recent work demonstrates that both local and global poisoning can significantly degrade counterfactual reliability, highlighting the need for integrated robustness checks. [v12056]Fine‑grained counterfactual explanation frameworks have emerged to reconcile the tension between validity and plausibility. By operating in a disentangled latent space and weighting component contributions via Shapley‑based saliency partitions, these methods generate counterfactuals that alter only semantically meaningful features while preserving the data manifold. Such granularity not only improves interpretability but also reduces the likelihood of generating counterfactuals that violate domain constraints, a common failure mode in conventional approaches. [v12981]In terms of computational overhead, FCA typically incurs additional cost due to the optimization of fairness constraints and the requirement for causal graph estimation. Conventional counterfactual generators, especially those based on diffusion models or gradient‑based search, can be deployed more efficiently but may produce counterfactuals that are less actionable or ethically sound. Recent comparative studies show that fine‑tuned diffusion‑based counterfactuals can match FCA’s fidelity while remaining scalable to large datasets, suggesting a hybrid strategy that leverages the strengths of both paradigms. [v12977][v12899]

7.4 Justification

The proposed FCA surpasses conventional CE methods for several reasons:

In sum, FCA aligns the optimization objective of adversarial robustness with the interpretability and actionability demands of counterfactual explanations, thereby advancing the frontier of trustworthy, coordinated AI systems in adversarial environments.

Appendix A: Validation References

[v1211]Beyond Semantic Relevance: Counterfactual Risk Minimization for Robust Retrieval-Augmented Generation
https://arxiv.org/abs/2605.01302
[v1977]Counterfactual Explanations with Probabilistic Guarantees on their Robustness to Model Change
https://arxiv.org/abs/2408.04842
[v2830]Controllable Stylistic Text Generation with Train-Time Attribute-Regularized Diffusion
https://arxiv.org/abs/2510.06386
[v3453]Artificial Intelligence (AI) is becoming a crucial part of almost every industry.
https://www.validaitor.com/post/understanding-the-basics-of-ai-testing
[v4266] Fugu-MT 論文翻訳(概要): When and Where to Attack?
https://fugumt.com/fugumt/paper_check/2602.04356v1
[v5423]Visual Disentangled Diffusion Autoencoders: Scalable Counterfactual Generation for Foundation Models
https://doi.org/10.48550/arXiv.2601.21851
[v5831]Generative artificial intelligence in diabetes healthcare
https://doi.org/10.1016/j.isci.2025.113051
[v6294]Recourse provides individuals who received undesirable labels (e.g., denied a loan) from algorithmic decision-making systems with a minimum-cost improvement suggestion to achieve the desired outcome.
https://arxiv.org/html/2509.21293v1
[v7081]DSSA-TCN: Exploiting adaptive sparse attention and diffusion graph convolutions in temporal convolutional networks for traffic flow forecasting
https://doi.org/10.1371/journal.pone.0336787
[v7542]Optimizing Graph Causal Classification Models: Estimating Causal Effects and Addressing Confounders
https://arxiv.org/abs/2602.17941
[v8528]Stable Language Guidance for Vision-Language-Action Models
https://arxiv.org/abs/2601.04052
[v8791]ElliCE: Efficient and Provably Robust Algorithmic Recourse via the Rashomon Sets
https://arxiv.org/abs/2602.07674
[v9141]NutVLM: A Self-Adaptive Defense Framework against Full-Dimension Attacks for Vision Language Models in Autonomous Driving
https://arxiv.org/abs/2602.13293
[v10859]Towards desiderata-driven design of visual counterfactual explainers
https://doi.org/10.1016/j.patcog.2025.112811
[v11082]Cross-Modal Attention Analysis and Optimization in Vision-Language Models: A Study on Visual Reliability
https://arxiv.org/abs/2604.17217
[v12056]The effect of data poisoning on counterfactual explanations
https://doi.org/10.1016/j.inffus.2026.104237
[v12247]Preserving Causal Constraints in Counterfactual Explanations for Machine Learning Classifiers
https://arxiv.org/abs/1912.03277
[v12560] GitHub - erwanlemerrer/awesome-audit-algorithms: A curated list of algorithms and papers for auditing black-box algorithms.
https://github.com/erwanlemerrer/awesome-audit-algorithms
[v12899]Data science: a natural ecosystem
https://doi.org/10.1016/j.inffus.2025.104113
[v12930]Towards desiderata-driven design of visual counterfactual explainers
https://doi.org/10.1016/j.patcog.2025.112811
[v12977]Protein Counterfactuals via Diffusion-Guided Latent Optimization
https://arxiv.org/abs/2603.10811
[v12981]Towards Fine-Grained Interpretability: Counterfactual Explanations for Misclassification with Saliency Partition
https://doi.org/10.1109/cvpr52734.2025.02797
[v13179]Toward Individual Fairness Without Centralized Data: Selective Counterfactual Consistency for Vertical Federated Learning
https://arxiv.org/abs/2605.07117
[v14059]12.6.2025 Paper discussion: InstaSHAP: Interpretable Additive Models Explain Shapley Values Instantly.
http://tml.cs.uni-tuebingen.de/teaching/tml_graduate_seminar/past_tml_graduate_seminar.php
[v15368]"Learnings from Paying Artists Royalties for AI-Generated Art: A Retrospective on Tess.Design, Our Attempt to Make an Ethical, Artist-Friendly AI Marketplace.
https://gwern.net/doc/ai/nn/diffusion/index
[v15838]4 Oct 202566B23F41159AB61353DF219B4E3FE4ADarXiv:2510.03612v1[cs.AI]User query: "Find a Thriller Movie"
https://doi.org/10.48550/arxiv.2510.03612
[v15921]This week in deep learning, we bring you Tensorflow Similarity, faster quantized inference with XNNPACK, the world's first 5G and AI enabled drone platform and a paper on transformer-based 3D dance g
https://www.deeplearningweekly.com/p/deep-learning-weekly-issue-215
[v16089]Generative Image Layer Decomposition with Visual Effects
https://doi.org/10.1109/cvpr52734.2025.00716
[v16245]AI-Based System and Method for Generating Enhanced Radiology Reports
https://ppubs.uspto.gov/pubwebapp/external.html?q=(20260128138).pn
[v16482]FASE : A Fairness-Aware Spatiotemporal Event Graph Framework for Predictive Policing
https://arxiv.org/abs/2604.18644
[v17005]The Geometric Canary: Predicting Steerability and Detecting Drift via Representational Stability
https://arxiv.org/abs/2604.17698

Appendix: Cited Sources

1
Counterfactual Visual Explanation via Causally-Guided Adversarial Steering 2025-09-29
Abstract: Recent work on counterfactual visual explanations has contributed to making artificial intelligence models more explainable by providing visual perturbation to flip the prediction. However, these approaches neglect the causal relationships and the spurious correlations behind the image generation process, which often leads to unintended alterations in the counterfactual images and renders the explanations with limited quality. To address this challenge, we introduce a novel framework C...
2
Counterfactual Visual Explanation via Causally-Guided Adversarial Steering 2025-07-13
Recent work on counterfactual visual explanations has contributed to making artificial intelligence models more explainable by providing visual perturbation to flip the prediction. However, these approaches neglect the causal relationships and the spurious correlations behind the image generation process, which often leads to unintended alterations in the counterfactual images and renders the explanations with limited quality. To address this challenge, we introduce a novel framework CECAS, whic...
3
Diffusion Counterfactuals for Image Regressors 2025-12-31
Adversarial Counterfactual Explanations (ACE) generate counterfactual images by optimizing adversarial perturbations in the image space while filtering high-frequency and out-of-distribution artifacts using a diffusion model. More specifically, consider L class (x, y) as a function that quantifies the match between a sample x and a class y, typically the cross-entropy loss, which we aim to minimize.Consider a filtering function F that constrains a counterfactual x ' to the data manifold of the t...
4
Optimal Robust Recourse with L p -Bounded Model Change 2025-12-31
Our Contributions and Results Our main goal is to understand the true price of recourse for more restricted adversarial model changes.In particular, we measure model changes by bounding the L p norm of the difference between initial and changed models, where p 1 but p = .We provide a new algorithm that provably computes the optimal robust recourse for generalized linear models for this type of model change. The key insight in the design of our algorithm is the observation that the optimal soluti...
5
Recourse provides individuals who received undesirable labels (e.g., denied a loan) from algorithmic decision-making systems with a minimum-cost improvement suggestion to achieve the desired outcome. 2026-04-20
Our main goal is to understand the true price of recourse for more restricted adversarial model changes. In particular, we measure model changes by bounding the LpL^{p} norm of the difference between initial and changed models, where p 1p\geq 1 but p peq\infty. We provide a new algorithm that provably computes the optimal robust recourse for generalized linear models for this type of model change. The key insight in the design of our algorithm is the observation that the optimal solution of the...
6
Counterfactual explanations and adversarial attacks have a related goal: flipping output labels with minimal perturbations regardless of their characteristics. 2026-03-17
Counterfactual explanations and adversarial attacks have a related goal: flipping output labels with minimal perturbations regardless of their characteristics. Yet, adversarial attacks cannot be used directly in a counterfactual explanation perspective, as such perturbations are perceived as noise and not as actionable and understandable image modifications....
7
Adversarial Counterfactual Visual Explanations 2023-03-16
Yet, adversarial attacks cannot be used directly in a counterfactual explanation perspective, as such perturbations are perceived as noise and not as actionable and understandable image modifications. (2023)...
8
Towards desiderata-driven design of visual counterfactual explainers 2026-05-07
This can be e.g. the inclusion or removal of object parts, but also more intricate changes in image quality or color, that may not be accessible with other explanation techniques such as feature attribution.Another advantage of counterfactuals is that they are inherently actionable, e.g.together with a human in the loop, counterfactuals provide an implicit data augmentation scheme that can serve to address a model's missing invariances or reliance on spurious correlations .Mathematically, the se...
10
The effect of data poisoning on counterfactual explanations 2026-05-07
We demonstrate that state-of-the-art counterfactual generation methods and toolboxes are vulnerable to such data poisoning. Introduction Nowadays, many Artificial Intelligence (AI-) and Machine Learning (ML-) based systems are deployed in the real world [Zhao et al., 2023;Ho et al., 2022].These systems show an impressive performance but are still not perfecte.g.failures, issues of fairness, and vulnerability to data poisoning can cause harm when applied in the real world....
11
In November 2023, Mount Sinai Health System deployed an explainable AI diagnostic system across its network of 8 hospitals serving 7.4 million patients annually in New York, addressing critical trust 2026-04-23
However, saliency methods face faithfulness challenges: generated visualizations may not accurately reflect true model behavior due to saturation effects, adversarial perturbations, and implementation choices that produce visually appealing but technically incorrect attributions. Research from Google analyzing 47,000 Grad-CAM explanations found that 23% highlighted regions provably irrelevant to model predictions (determined through ablation studies zeroing out highlighted regions without changi...