"adversarial machine learning" Papers

11 papers found

CAMH: Advancing Model Hijacking Attack in Machine Learning

Xing He, Jiahao Chen, Yuwen Pu et al.

AAAI 2025paperarXiv:2408.13741

Concept-ROT: Poisoning Concepts in Large Language Models with Model Editing

Keltin Grimes, Marco Christiani, David Shriver et al.

ICLR 2025arXiv:2412.13341
6
citations

Meme Trojan: Backdoor Attacks Against Hateful Meme Detection via Cross-Modal Triggers

Ruofei Wang, Hongzhan Lin, Ziyuan Luo et al.

AAAI 2025paperarXiv:2412.15503
3
citations

PoisonedEye: Knowledge Poisoning Attack on Retrieval-Augmented Generation based Large Vision-Language Models

Chenyang Zhang, Xiaoyu Zhang, Jian Lou et al.

ICML 2025
3
citations

PSBD: Prediction Shift Uncertainty Unlocks Backdoor Detection

Wei Li, Pin-Yu Chen, Sijia Liu et al.

CVPR 2025arXiv:2406.05826
4
citations

The Implicit Bias of Structured State Space Models Can Be Poisoned With Clean Labels

Yonatan Slutzky, ‪Yotam Alexander‬‏, Noam Razin et al.

NEURIPS 2025spotlightarXiv:2410.10473
2
citations

Where the Devil Hides: Deepfake Detectors Can No Longer Be Trusted

Shuaiwei Yuan, Junyu Dong, Yuezun Li

CVPR 2025arXiv:2505.08255
2
citations

Your Scale Factors are My Weapon: Targeted Bit-Flip Attacks on Vision Transformers via Scale Factor Manipulation

Jialai Wang, Yuxiao Wu, Weiye Xu et al.

CVPR 2025
3
citations

BadRL: Sparse Targeted Backdoor Attack against Reinforcement Learning

Jing Cui, Yufei Han, Yuzhe Ma et al.

AAAI 2024paperarXiv:2312.12585
26
citations

Energy-based Backdoor Defense without Task-Specific Samples and Model Retraining

Yudong Gao, Honglong Chen, Peng Sun et al.

ICML 2024

Uniformly Stable Algorithms for Adversarial Training and Beyond

Jiancong Xiao, Jiawei Zhang, Zhi-Quan Luo et al.

ICML 2024arXiv:2405.01817
2
citations