"adversarial machine learning" Papers
11 papers found
Conference
CAMH: Advancing Model Hijacking Attack in Machine Learning
Xing He, Jiahao Chen, Yuwen Pu et al.
AAAI 2025paperarXiv:2408.13741
Concept-ROT: Poisoning Concepts in Large Language Models with Model Editing
Keltin Grimes, Marco Christiani, David Shriver et al.
ICLR 2025arXiv:2412.13341
6
citations
Meme Trojan: Backdoor Attacks Against Hateful Meme Detection via Cross-Modal Triggers
Ruofei Wang, Hongzhan Lin, Ziyuan Luo et al.
AAAI 2025paperarXiv:2412.15503
3
citations
PoisonedEye: Knowledge Poisoning Attack on Retrieval-Augmented Generation based Large Vision-Language Models
Chenyang Zhang, Xiaoyu Zhang, Jian Lou et al.
ICML 2025
3
citations
PSBD: Prediction Shift Uncertainty Unlocks Backdoor Detection
Wei Li, Pin-Yu Chen, Sijia Liu et al.
CVPR 2025arXiv:2406.05826
4
citations
The Implicit Bias of Structured State Space Models Can Be Poisoned With Clean Labels
Yonatan Slutzky, Yotam Alexander, Noam Razin et al.
NEURIPS 2025spotlightarXiv:2410.10473
2
citations
Where the Devil Hides: Deepfake Detectors Can No Longer Be Trusted
Shuaiwei Yuan, Junyu Dong, Yuezun Li
CVPR 2025arXiv:2505.08255
2
citations
Your Scale Factors are My Weapon: Targeted Bit-Flip Attacks on Vision Transformers via Scale Factor Manipulation
Jialai Wang, Yuxiao Wu, Weiye Xu et al.
CVPR 2025
3
citations
BadRL: Sparse Targeted Backdoor Attack against Reinforcement Learning
Jing Cui, Yufei Han, Yuzhe Ma et al.
AAAI 2024paperarXiv:2312.12585
26
citations
Energy-based Backdoor Defense without Task-Specific Samples and Model Retraining
Yudong Gao, Honglong Chen, Peng Sun et al.
ICML 2024
Uniformly Stable Algorithms for Adversarial Training and Beyond
Jiancong Xiao, Jiawei Zhang, Zhi-Quan Luo et al.
ICML 2024arXiv:2405.01817
2
citations