"adversarial attacks" Papers

115 papers found • Page 1 of 3

A²RNet: Adversarial Attack Resilient Network for Robust Infrared and Visible Image Fusion

Jiawei Li, Hongwei Yu, Jiansheng Chen et al.

AAAI 2025paperarXiv:2412.09954
3
citations

Adversarial Attacks against Closed-Source MLLMs via Feature Optimal Alignment

Xiaojun Jia, Sensen Gao, Simeng Qin et al.

NEURIPS 2025arXiv:2505.21494
18
citations

Adversarial Attacks on Data Attribution

Xinhe Wang, Pingbang Hu, Junwei Deng et al.

ICLR 2025arXiv:2409.05657
1
citations

Adversarial Attacks on Event-Based Pedestrian Detectors: A Physical Approach

Guixu Lin, Muyao Niu, Qingtian Zhu et al.

AAAI 2025paperarXiv:2503.00377
4
citations

Adversarial Robustness of Discriminative Self-Supervised Learning in Vision

Ömer Veysel Çağatan, Ömer TAL, M. Emre Gursoy

ICCV 2025arXiv:2503.06361

Adversary Aware Optimization for Robust Defense

Daniel Wesego, Pedram Rooshenas

NEURIPS 2025

AIM: Additional Image Guided Generation of Transferable Adversarial Attacks

Teng Li, Xingjun Ma, Yu-Gang Jiang

AAAI 2025paperarXiv:2501.01106
4
citations

A New Adversarial Perspective for LiDAR-based 3D Object Detection

Shijun Zheng, Weiquan Liu, Yu Guo et al.

AAAI 2025paperarXiv:2412.13017
3
citations

Autonomous LLM-Enhanced Adversarial Attack for Text-to-Motion

Honglei Miao, Fan Ma, Ruijie Quan et al.

AAAI 2025paperarXiv:2408.00352
8
citations

Bits Leaked per Query: Information-Theoretic Bounds for Adversarial Attacks on LLMs

Masahiro Kaneko, Timothy Baldwin

NEURIPS 2025spotlightarXiv:2510.17000

Boosting Adversarial Transferability with Spatial Adversarial Alignment

Zhaoyu Chen, HaiJing Guo, Kaixun Jiang et al.

NEURIPS 2025arXiv:2501.01015
1
citations

Bridging Symmetry and Robustness: On the Role of Equivariance in Enhancing Adversarial Robustness

Longwei Wang, Ifrat Ikhtear Uddin, Prof. KC Santosh (PhD) et al.

NEURIPS 2025spotlightarXiv:2510.16171
2
citations

Confidence Elicitation: A New Attack Vector for Large Language Models

Brian Formento, Chuan Sheng Foo, See-Kiong Ng

ICLR 2025arXiv:2502.04643
2
citations

Democratic Training Against Universal Adversarial Perturbations

Bing Sun, Jun Sun, Wei Zhao

ICLR 2025arXiv:2502.05542
1
citations

DepthVanish: Optimizing Adversarial Interval Structures for Stereo-Depth-Invisible Patches

Yun Xing, Yue Cao, Nhat Chung et al.

NEURIPS 2025arXiv:2506.16690

Detecting Adversarial Data Using Perturbation Forgery

Qian Wang, Chen Li, Yuchen Luo et al.

CVPR 2025arXiv:2405.16226
3
citations

DIA: The Adversarial Exposure of Deterministic Inversion in Diffusion Models

SeungHoo Hong, GeonHo Son, Juhun Lee et al.

ICCV 2025arXiv:2510.00778

Dynamical Low-Rank Compression of Neural Networks with Robustness under Adversarial Attacks

Steffen Schotthöfer, Lexie Yang, Stefan Schnake

NEURIPS 2025oralarXiv:2505.08022
6
citations

Endowing Visual Reprogramming with Adversarial Robustness

Shengjie Zhou, Xin Cheng, Haiyang Xu et al.

ICLR 2025
2
citations

Enhancing Graph Classification Robustness with Singular Pooling

Sofiane Ennadir, Oleg Smirnov, Yassine ABBAHADDOU et al.

NEURIPS 2025arXiv:2510.22643

Exploring Visual Vulnerabilities via Multi-Loss Adversarial Search for Jailbreaking Vision-Language Models

Shuyang Hao, Bryan Hooi, Jun Liu et al.

CVPR 2025arXiv:2411.18000
6
citations

Fit the Distribution: Cross-Image/Prompt Adversarial Attacks on Multimodal Large Language Models

Hai Yan, Haijian Ma, Xiaowen Cai et al.

NEURIPS 2025

Fortifying Time Series: DTW-Certified Robust Anomaly Detection

Shijie Liu, Tansu Alpcan, Christopher Leckie et al.

NEURIPS 2025oral

GSBA$^K$: $top$-$K$ Geometric Score-based Black-box Attack

Md Farhamdur Reza, Richeng Jin, Tianfu Wu et al.

ICLR 2025arXiv:2503.12827
3
citations

Hypergraph Attacks via Injecting Homogeneous Nodes into Elite Hyperedges

Meixia He, Peican Zhu, Keke Tang et al.

AAAI 2025paperarXiv:2412.18365
6
citations

Instant Adversarial Purification with Adversarial Consistency Distillation

Chun Tong Lei, Hon Ming Yam, Zhongliang Guo et al.

CVPR 2025arXiv:2408.17064
13
citations

IPAD: Inverse Prompt for AI Detection - A Robust and Interpretable LLM-Generated Text Detector

Zheng CHEN, Yushi Feng, Jisheng Dang et al.

NEURIPS 2025arXiv:2502.15902

Jailbreaking as a Reward Misspecification Problem

Zhihui Xie, Jiahui Gao, Lei Li et al.

ICLR 2025arXiv:2406.14393
11
citations

Jailbreaking Multimodal Large Language Models via Shuffle Inconsistency

Shiji Zhao, Ranjie Duan, Fengxiang Wang et al.

ICCV 2025arXiv:2501.04931
30
citations

Keeping an Eye on LLM Unlearning: The Hidden Risk and Remedy

Jie Ren, Zhenwei Dai, Xianfeng Tang et al.

NEURIPS 2025arXiv:2506.00359
7
citations

LARGO: Latent Adversarial Reflection through Gradient Optimization for Jailbreaking LLMs

Ran Li, Hao Wang, Chengzhi Mao

NEURIPS 2025arXiv:2505.10838
4
citations

LiD-FL: Towards List-Decodable Federated Learning

Hong Liu, Liren Shan, Han Bao et al.

AAAI 2025paperarXiv:2408.04963

MIP against Agent: Malicious Image Patches Hijacking Multimodal OS Agents

Lukas Aichberger, Alasdair Paren, Guohao Li et al.

NEURIPS 2025arXiv:2503.10809
10
citations

MOS-Attack: A Scalable Multi-objective Adversarial Attack Framework

Ping Guo, Cheng Gong, Fei Liu et al.

CVPR 2025arXiv:2501.07251

Non-Adaptive Adversarial Face Generation

Sunpill Kim, Seunghun Paik, Chanwoo Hwang et al.

NEURIPS 2025arXiv:2507.12107
1
citations

NoPain: No-box Point Cloud Attack via Optimal Transport Singular Boundary

Zezeng Li, Xiaoyu Du, Na Lei et al.

CVPR 2025arXiv:2503.00063
5
citations

NumbOD: A Spatial-Frequency Fusion Attack Against Object Detectors

Ziqi Zhou, Bowen Li, Yufei Song et al.

AAAI 2025paperarXiv:2412.16955
15
citations

On the Alignment between Fairness and Accuracy: from the Perspective of Adversarial Robustness

Junyi Chai, Taeuk Jang, Jing Gao et al.

ICML 2025

On the Stability of Graph Convolutional Neural Networks: A Probabilistic Perspective

Ning Zhang, Henry Kenlay, Li Zhang et al.

NEURIPS 2025arXiv:2506.01213

Pixel Is Not a Barrier: An Effective Evasion Attack for Pixel-Domain Diffusion Models

Chun-Yen Shih, Li-Xuan Peng, Jia-Wei Liao et al.

AAAI 2025paperarXiv:2408.11810
2
citations

Prompt2Perturb (P2P): Text-Guided Diffusion-Based Adversarial Attack on Breast Ultrasound Images

Yasamin Medghalchi, Moein Heidari, Clayton Allard et al.

CVPR 2025arXiv:2412.09910
4
citations

RAT: Adversarial Attacks on Deep Reinforcement Agents for Targeted Behaviors

Fengshuo Bai, Runze Liu, Yali Du et al.

AAAI 2025paperarXiv:2412.10713
12
citations

Rationalizing and Augmenting Dynamic Graph Neural Networks

Guibin Zhang, Yiyan Qi, Ziyang Cheng et al.

ICLR 2025oral

Robust LLM safeguarding via refusal feature adversarial training

Lei Yu, Virginie Do, Karen Hambardzumyan et al.

ICLR 2025arXiv:2409.20089
45
citations

Robust SuperAlignment: Weak-to-Strong Robustness Generalization for Vision-Language Models

Junhao Dong, Cong Zhang, Xinghua Qu et al.

NEURIPS 2025spotlight

R-TPT: Improving Adversarial Robustness of Vision-Language Models through Test-Time Prompt Tuning

Lijun Sheng, Jian Liang, Zilei Wang et al.

CVPR 2025arXiv:2504.11195
15
citations

Safe RLHF-V: Safe Reinforcement Learning from Multi-modal Human Feedback

Jiaming Ji, Xinyu Chen, Rui Pan et al.

NEURIPS 2025arXiv:2503.17682
9
citations

SECA: Semantically Equivalent and Coherent Attacks for Eliciting LLM Hallucinations

Buyun Liang, Liangzu Peng, Jinqi Luo et al.

NEURIPS 2025arXiv:2510.04398

Stochastic Regret Guarantees for Online Zeroth- and First-Order Bilevel Optimization

Parvin Nazari, Bojian Hou, Davoud Ataee Tarzanagh et al.

NEURIPS 2025arXiv:2511.01126
2
citations

TAROT: Towards Essentially Domain-Invariant Robustness with Theoretical Justification

Dongyoon Yang, Jihu Lee, Yongdai Kim

CVPR 2025arXiv:2505.06580
1
citations
PreviousNext