"explainable ai" Papers

68 papers found • Page 1 of 2

$\mathcal{X}^2$-DFD: A framework for e$\mathcal{X}$plainable and e$\mathcal{X}$tendable Deepfake Detection

Yize Chen, Zhiyuan Yan, Guangliang Cheng et al.

NEURIPS 2025

Advancing Interpretability of CLIP Representations with Concept Surrogate Model

Nhat Hoang-Xuan, Xiyuan Wei, Wanli Xing et al.

NEURIPS 2025

AI2TALE: An Innovative Information Theory-based Approach for Learning to Localize Phishing Attacks

Van Nguyen, Tingmin Wu, Xingliang YUAN et al.

ICLR 2025
2
citations

AIGI-Holmes: Towards Explainable and Generalizable AI-Generated Image Detection via Multimodal Large Language Models

Ziyin Zhou, Yunpeng Luo, Yuanchen Wu et al.

ICCV 2025arXiv:2507.02664
13
citations

AudioGenX: Explainability on Text-to-Audio Generative Models

Hyunju Kang, Geonhee Han, Yoonjae Jeong et al.

AAAI 2025paperarXiv:2502.00459

A Unified, Resilient, and Explainable Adversarial Patch Detector

Vishesh Kumar, Akshay Agarwal

CVPR 2025
2
citations

Constructing Fair Latent Space for Intersection of Fairness and Explainability

Hyungjun Joo, Hyeonggeun Han, Sehwan Kim et al.

AAAI 2025paperarXiv:2412.17523
2
citations

Contimask: Explaining Irregular Time Series via Perturbations in Continuous Time

Max Moebus, Björn Braun, Christian Holz

NEURIPS 2025

Data-centric Prediction Explanation via Kernelized Stein Discrepancy

Mahtab Sarvmaili, Hassan Sajjad, Ga Wu

ICLR 2025arXiv:2403.15576
2
citations

Derivative-Free Diffusion Manifold-Constrained Gradient for Unified XAI

Won Jun Kim, Hyungjin Chung, Jaemin Kim et al.

CVPR 2025arXiv:2411.15265
2
citations

Disentangled Concepts Speak Louder Than Words: Explainable Video Action Recognition

Jongseo Lee, Wooil Lee, Gyeong-Moon Park et al.

NEURIPS 2025spotlightarXiv:2511.03725

Explainable Reinforcement Learning from Human Feedback to Improve Alignment

Shicheng Liu, Siyuan Xu, Wenjie Qiu et al.

NEURIPS 2025arXiv:2512.13837

Explainably Safe Reinforcement Learning

Sabine Rieder, Stefan Pranger, Debraj Chakraborty et al.

NEURIPS 2025

Explaining Decisions of Agents in Mixed-Motive Games

Maayan Orner, Oleg Maksimov, Akiva Kleinerman et al.

AAAI 2025paperarXiv:2407.15255
4
citations

Exploiting Symmetries in MUS Computation

Ignace Bleukx, Hélène Verhaeghe, Bart Bogaerts et al.

AAAI 2025paperarXiv:2412.13606
1
citations

FakeShield: Explainable Image Forgery Detection and Localization via Multi-modal Large Language Models

Zhipei Xu, Xuanyu Zhang, Runyi Li et al.

ICLR 2025arXiv:2410.02761
77
citations

F-Fidelity: A Robust Framework for Faithfulness Evaluation of Explainable AI

Xu Zheng, Farhad Shirani, Zhuomin Chen et al.

ICLR 2025arXiv:2410.02970
15
citations

Fish-Vista: A Multi-Purpose Dataset for Understanding & Identification of Traits from Images

Kazi Sajeed Mehrab, M. Maruf, Arka Daw et al.

CVPR 2025arXiv:2407.08027
8
citations

HEIE: MLLM-Based Hierarchical Explainable AIGC Image Implausibility Evaluator

Fan Yang, Ru Zhen, Jianing Wang et al.

CVPR 2025arXiv:2411.17261
11
citations

Higher Order Structures for Graph Explanations

Akshit Sinha, Sreeram Vennam, Charu Sharma et al.

AAAI 2025paperarXiv:2406.03253
4
citations

Interpreting Language Reward Models via Contrastive Explanations

Junqi Jiang, Tom Bewley, Saumitra Mishra et al.

ICLR 2025arXiv:2411.16502
5
citations

LeapFactual: Reliable Visual Counterfactual Explanation Using Conditional Flow Matching

Zhuo Cao, Xuan Zhao, Lena Krieger et al.

NEURIPS 2025arXiv:2510.14623
1
citations

LOKI: A Comprehensive Synthetic Data Detection Benchmark using Large Multimodal Models

Junyan Ye, Baichuan Zhou, Zilong Huang et al.

ICLR 2025arXiv:2410.09732
30
citations

Minimizing False-Positive Attributions in Explanations of Non-Linear Models

Anders Gjølbye, Stefan Haufe, Lars Kai Hansen

NEURIPS 2025arXiv:2505.11210
1
citations

Mol-LLaMA: Towards General Understanding of Molecules in Large Molecular Language Model

Dongki Kim, Wonbin Lee, Sung Ju Hwang

NEURIPS 2025arXiv:2502.13449
12
citations

NOMATTERXAI: Generating “No Matter What” Alterfactual Examples for Explaining Black-Box Text Classification Models

Tuc Van Nguyen, James Michels, Hua Shen et al.

AAAI 2025paperarXiv:2408.10528
1
citations

On Logic-based Self-Explainable Graph Neural Networks

Alessio Ragno, Marc Plantevit, Céline Robardet

NEURIPS 2025

PathFinder: A Multi-Modal Multi-Agent System for Medical Diagnostic Decision-Making Applied to Histopathology

Fatemeh Ghezloo, Saygin Seyfioglu, Rustin Soraki et al.

ICCV 2025arXiv:2502.08916
20
citations

PrivateXR: Defending Privacy Attacks in Extended Reality Through Explainable AI-Guided Differential Privacy

Ripan Kumar Kundu, Istiak Ahmed, Khaza Anuarul Hoque

ISMAR 2025paperarXiv:2512.16851

Provable Gradient Editing of Deep Neural Networks

Zhe Tao, Aditya V Thakur

NEURIPS 2025spotlight
1
citations

RadZero: Similarity-Based Cross-Attention for Explainable Vision-Language Alignment in Chest X-ray with Zero-Shot Multi-Task Capability

Jonggwon Park, Byungmu Yoon, Soobum Kim et al.

NEURIPS 2025arXiv:2504.07416
1
citations

Reconsidering Faithfulness in Regular, Self-Explainable and Domain Invariant GNNs

Steve Azzolin, Antonio Longa, Stefano Teso et al.

ICLR 2025arXiv:2406.15156
5
citations

Regression-adjusted Monte Carlo Estimators for Shapley Values and Probabilistic Values

R. Teal Witter, Yurong Liu, Christopher Musco

NEURIPS 2025arXiv:2506.11849
4
citations

Representational Difference Explanations

Neehar Kondapaneni, Oisin Mac Aodha, Pietro Perona

NEURIPS 2025arXiv:2505.23917

Scalable, Explainable and Provably Robust Anomaly Detection with One-Step Flow Matching

Zhong Li, Qi Huang, Yuxuan Zhu et al.

NEURIPS 2025arXiv:2510.18328

Seeing Through Deepfakes: A Human-Inspired Framework for Multi-Face Detection

Juan Hu, Shaojing Fan, Terence Sim

ICCV 2025arXiv:2507.14807
1
citations

SHAP values via sparse Fourier representation

Ali Gorji, Andisheh Amrollahi, Andreas Krause

NEURIPS 2025spotlightarXiv:2410.06300
3
citations

Smoothed Differentiation Efficiently Mitigates Shattered Gradients in Explanations

Adrian Hill, Neal McKee, Johannes Maeß et al.

NEURIPS 2025

Sound Logical Explanations for Mean Aggregation Graph Neural Networks

Matthew Morris, Ian Horrocks

NEURIPS 2025arXiv:2511.11593
1
citations

Start Smart: Leveraging Gradients For Enhancing Mask-based XAI Methods

Buelent Uendes, Shujian Yu, Mark Hoogendoorn

ICLR 2025

Towards Synergistic Path-based Explanations for Knowledge Graph Completion: Exploration and Evaluation

Tengfei Ma, Xiang song, Wen Tao et al.

ICLR 2025
3
citations

Towards Unifying Evaluation of Counterfactual Explanations: Leveraging Large Language Models for Human-Centric Assessments

Marharyta Domnich, Julius Välja, Rasmus Moorits Veski et al.

AAAI 2025paperarXiv:2410.21131
8
citations

Understanding Emotional Body Expressions via Large Language Models

Haifeng Lu, Jiuyi Chen, Feng Liang et al.

AAAI 2025paperarXiv:2412.12581
11
citations

Understanding Individual Agent Importance in Multi-Agent System via Counterfactual Reasoning

Jianming Chen, Yawen Wang, Junjie Wang et al.

AAAI 2025paperarXiv:2412.15619
7
citations

VERA: Explainable Video Anomaly Detection via Verbalized Learning of Vision-Language Models

Muchao Ye, Weiyang Liu, Pan He

CVPR 2025arXiv:2412.01095
10
citations

Accelerating the Global Aggregation of Local Explanations

Alon Mor, Yonatan Belinkov, Benny Kimelfeld

AAAI 2024paperarXiv:2312.07991
7
citations

Attribution-based Explanations that Provide Recourse Cannot be Robust

Hidde Fokkema, Rianne de Heide, Tim van Erven

ICML 2024arXiv:2205.15834
22
citations

Beyond TreeSHAP: Efficient Computation of Any-Order Shapley Interactions for Tree Ensembles

Maximilian Muschalik, Fabian Fumagalli, Barbara Hammer et al.

AAAI 2024paperarXiv:2401.12069
33
citations

CGS-Mask: Making Time Series Predictions Intuitive for All

Feng Lu, Wei Li, Yifei Sun et al.

AAAI 2024paperarXiv:2312.09513
2
citations

Counterfactual Metarules for Local and Global Recourse

Tom Bewley, Salim I. Amoukou, Saumitra Mishra et al.

ICML 2024arXiv:2405.18875
4
citations
PreviousNext