Poster "explainable ai" Papers

44 papers found

$\mathcal{X}^2$-DFD: A framework for e$\mathcal{X}$plainable and e$\mathcal{X}$tendable Deepfake Detection

Yize Chen, Zhiyuan Yan, Guangliang Cheng et al.

NEURIPS 2025

Advancing Interpretability of CLIP Representations with Concept Surrogate Model

Nhat Hoang-Xuan, Xiyuan Wei, Wanli Xing et al.

NEURIPS 2025

AI2TALE: An Innovative Information Theory-based Approach for Learning to Localize Phishing Attacks

Van Nguyen, Tingmin Wu, Xingliang YUAN et al.

ICLR 2025
2
citations

AIGI-Holmes: Towards Explainable and Generalizable AI-Generated Image Detection via Multimodal Large Language Models

Ziyin Zhou, Yunpeng Luo, Yuanchen Wu et al.

ICCV 2025arXiv:2507.02664
13
citations

A Unified, Resilient, and Explainable Adversarial Patch Detector

Vishesh Kumar, Akshay Agarwal

CVPR 2025
2
citations

Contimask: Explaining Irregular Time Series via Perturbations in Continuous Time

Max Moebus, Björn Braun, Christian Holz

NEURIPS 2025

Data-centric Prediction Explanation via Kernelized Stein Discrepancy

Mahtab Sarvmaili, Hassan Sajjad, Ga Wu

ICLR 2025arXiv:2403.15576
2
citations

Derivative-Free Diffusion Manifold-Constrained Gradient for Unified XAI

Won Jun Kim, Hyungjin Chung, Jaemin Kim et al.

CVPR 2025arXiv:2411.15265
2
citations

Explainable Reinforcement Learning from Human Feedback to Improve Alignment

Shicheng Liu, Siyuan Xu, Wenjie Qiu et al.

NEURIPS 2025arXiv:2512.13837

Explainably Safe Reinforcement Learning

Sabine Rieder, Stefan Pranger, Debraj Chakraborty et al.

NEURIPS 2025

FakeShield: Explainable Image Forgery Detection and Localization via Multi-modal Large Language Models

Zhipei Xu, Xuanyu Zhang, Runyi Li et al.

ICLR 2025arXiv:2410.02761
77
citations

F-Fidelity: A Robust Framework for Faithfulness Evaluation of Explainable AI

Xu Zheng, Farhad Shirani, Zhuomin Chen et al.

ICLR 2025arXiv:2410.02970
15
citations

Fish-Vista: A Multi-Purpose Dataset for Understanding & Identification of Traits from Images

Kazi Sajeed Mehrab, M. Maruf, Arka Daw et al.

CVPR 2025arXiv:2407.08027
8
citations

HEIE: MLLM-Based Hierarchical Explainable AIGC Image Implausibility Evaluator

Fan Yang, Ru Zhen, Jianing Wang et al.

CVPR 2025arXiv:2411.17261
11
citations

Interpreting Language Reward Models via Contrastive Explanations

Junqi Jiang, Tom Bewley, Saumitra Mishra et al.

ICLR 2025arXiv:2411.16502
5
citations

LeapFactual: Reliable Visual Counterfactual Explanation Using Conditional Flow Matching

Zhuo Cao, Xuan Zhao, Lena Krieger et al.

NEURIPS 2025arXiv:2510.14623
1
citations

LOKI: A Comprehensive Synthetic Data Detection Benchmark using Large Multimodal Models

Junyan Ye, Baichuan Zhou, Zilong Huang et al.

ICLR 2025arXiv:2410.09732
30
citations

Minimizing False-Positive Attributions in Explanations of Non-Linear Models

Anders Gjølbye, Stefan Haufe, Lars Kai Hansen

NEURIPS 2025arXiv:2505.11210
1
citations

Mol-LLaMA: Towards General Understanding of Molecules in Large Molecular Language Model

Dongki Kim, Wonbin Lee, Sung Ju Hwang

NEURIPS 2025arXiv:2502.13449
12
citations

On Logic-based Self-Explainable Graph Neural Networks

Alessio Ragno, Marc Plantevit, Céline Robardet

NEURIPS 2025

PathFinder: A Multi-Modal Multi-Agent System for Medical Diagnostic Decision-Making Applied to Histopathology

Fatemeh Ghezloo, Saygin Seyfioglu, Rustin Soraki et al.

ICCV 2025arXiv:2502.08916
20
citations

RadZero: Similarity-Based Cross-Attention for Explainable Vision-Language Alignment in Chest X-ray with Zero-Shot Multi-Task Capability

Jonggwon Park, Byungmu Yoon, Soobum Kim et al.

NEURIPS 2025arXiv:2504.07416
1
citations

Reconsidering Faithfulness in Regular, Self-Explainable and Domain Invariant GNNs

Steve Azzolin, Antonio Longa, Stefano Teso et al.

ICLR 2025arXiv:2406.15156
5
citations

Regression-adjusted Monte Carlo Estimators for Shapley Values and Probabilistic Values

R. Teal Witter, Yurong Liu, Christopher Musco

NEURIPS 2025arXiv:2506.11849
4
citations

Representational Difference Explanations

Neehar Kondapaneni, Oisin Mac Aodha, Pietro Perona

NEURIPS 2025arXiv:2505.23917

Scalable, Explainable and Provably Robust Anomaly Detection with One-Step Flow Matching

Zhong Li, Qi Huang, Yuxuan Zhu et al.

NEURIPS 2025arXiv:2510.18328

Seeing Through Deepfakes: A Human-Inspired Framework for Multi-Face Detection

Juan Hu, Shaojing Fan, Terence Sim

ICCV 2025arXiv:2507.14807
1
citations

Smoothed Differentiation Efficiently Mitigates Shattered Gradients in Explanations

Adrian Hill, Neal McKee, Johannes Maeß et al.

NEURIPS 2025

Sound Logical Explanations for Mean Aggregation Graph Neural Networks

Matthew Morris, Ian Horrocks

NEURIPS 2025arXiv:2511.11593
1
citations

Start Smart: Leveraging Gradients For Enhancing Mask-based XAI Methods

Buelent Uendes, Shujian Yu, Mark Hoogendoorn

ICLR 2025

Towards Synergistic Path-based Explanations for Knowledge Graph Completion: Exploration and Evaluation

Tengfei Ma, Xiang song, Wen Tao et al.

ICLR 2025
3
citations

VERA: Explainable Video Anomaly Detection via Verbalized Learning of Vision-Language Models

Muchao Ye, Weiyang Liu, Pan He

CVPR 2025arXiv:2412.01095
10
citations

Attribution-based Explanations that Provide Recourse Cannot be Robust

Hidde Fokkema, Rianne de Heide, Tim van Erven

ICML 2024arXiv:2205.15834
22
citations

Counterfactual Metarules for Local and Global Recourse

Tom Bewley, Salim I. Amoukou, Saumitra Mishra et al.

ICML 2024arXiv:2405.18875
4
citations

EiG-Search: Generating Edge-Induced Subgraphs for GNN Explanation in Linear Time

Shengyao Lu, Bang Liu, Keith Mills et al.

ICML 2024arXiv:2405.01762
7
citations

Generating In-Distribution Proxy Graphs for Explaining Graph Neural Networks

Zhuomin Chen, Jiaxing Zhang, Jingchao Ni et al.

ICML 2024arXiv:2402.02036
7
citations

Good Teachers Explain: Explanation-Enhanced Knowledge Distillation

Amin Parchami, Moritz Böhle, Sukrut Rao et al.

ECCV 2024arXiv:2402.03119
19
citations

Graph Neural Network Explanations are Fragile

Jiate Li, Meng Pang, Yun Dong et al.

ICML 2024arXiv:2406.03193
18
citations

Layer-Wise Relevance Propagation with Conservation Property for ResNet

Seitaro Otsuki, Tsumugi Iida, Félix Doublet et al.

ECCV 2024arXiv:2407.09115
10
citations

Manifold Integrated Gradients: Riemannian Geometry for Feature Attribution

Eslam Zaher, Maciej Trzaskowski, Quan Nguyen et al.

ICML 2024arXiv:2405.09800
9
citations

On Gradient-like Explanation under a Black-box Setting: When Black-box Explanations Become as Good as White-box

Yi Cai, Gerhard Wunder

ICML 2024arXiv:2308.09381
3
citations

Position: Do Not Explain Vision Models Without Context

Paulina Tomaszewska, Przemyslaw Biecek

ICML 2024arXiv:2404.18316
1
citations

Probabilistic Conceptual Explainers: Trustworthy Conceptual Explanations for Vision Foundation Models

Hengyi Wang, Shiwei Tan, Hao Wang

ICML 2024arXiv:2406.12649
9
citations

Token Transformation Matters: Towards Faithful Post-hoc Explanation for Vision Transformer

Junyi Wu, Bin Duan, Weitai Kang et al.

CVPR 2024arXiv:2403.14552
16
citations