"adversarial robustness" Papers

129 papers found • Page 2 of 3

Resolution Attack: Exploiting Image Compression to Deceive Deep Neural Networks

Wangjia Yu, Xiaomeng Fu, Qiao Li et al.

ICLR 2025

Robust Conformal Prediction with a Single Binary Certificate

Soroush H. Zargarbashi, Aleksandar Bojchevski

ICLR 2025arXiv:2503.05239
3
citations

Robust Contextual Pricing

Anupam Gupta, Guru Guruganesh, Renato Leme et al.

NEURIPS 2025

Robust Feature Learning for Multi-Index Models in High Dimensions

Alireza Mousavi-Hosseini, Adel Javanmard, Murat A Erdogdu

ICLR 2025arXiv:2410.16449
5
citations

Robust SuperAlignment: Weak-to-Strong Robustness Generalization for Vision-Language Models

Junhao Dong, Cong Zhang, Xinghua Qu et al.

NEURIPS 2025spotlight

R-TPT: Improving Adversarial Robustness of Vision-Language Models through Test-Time Prompt Tuning

Lijun Sheng, Jian Liang, Zilei Wang et al.

CVPR 2025arXiv:2504.11195
15
citations

Support is All You Need for Certified VAE Training

Changming Xu, Debangshu Banerjee, Deepak Vasisht et al.

ICLR 2025arXiv:2504.11831

Synergy Between the Strong and the Weak: Spiking Neural Networks are Inherently Self-Distillers

Yongqi Ding, Lin Zuo, Mengmeng Jing et al.

NEURIPS 2025oralarXiv:2510.07924

Towards Adversarially Robust Dataset Distillation by Curvature Regularization

Eric Xue, Yijiang Li, Haoyang Liu et al.

AAAI 2025paperarXiv:2403.10045
18
citations

Towards Adversarial Robustness via Debiased High-Confidence Logit Alignment

Kejia Zhang, Juanjuan Weng, Zhiming Luo et al.

ICCV 2025arXiv:2408.06079
2
citations

Towards Robust Knowledge Unlearning: An Adversarial Framework for Assessing and Improving Unlearning Robustness in Large Language Models

Hongbang Yuan, Zhuoran Jin, Pengfei Cao et al.

AAAI 2025paperarXiv:2408.10682
25
citations

Two is Better than One: Efficient Ensemble Defense for Robust and Compact Models

Yoojin Jung, Byung Cheol Song

CVPR 2025arXiv:2504.04747
1
citations

UIBDiffusion: Universal Imperceptible Backdoor Attack for Diffusion Models

Yuning Han, Bingyin Zhao, Rui Chu et al.

CVPR 2025highlightarXiv:2412.11441
6
citations

Understanding and Improving Adversarial Robustness of Neural Probabilistic Circuits

Weixin Chen, Han Zhao

NEURIPS 2025arXiv:2509.20549

WMCopier: Forging Invisible Watermarks on Arbitrary Images

Ziping Dong, Chao Shuai, Zhongjie Ba et al.

NEURIPS 2025

Your Text Encoder Can Be An Object-Level Watermarking Controller

Naresh Kumar Devulapally, Mingzhen Huang, Vishal Asnani et al.

ICCV 2025arXiv:2503.11945

Zero-cost Proxy for Adversarial Robustness Evaluation

Yuqi Feng, Yuwei Ou, Jiahao Fan et al.

ICLR 2025
1
citations

Adaptive Hierarchical Certification for Segmentation using Randomized Smoothing

Alaa Anani, Tobias Lorenz, Bernt Schiele et al.

ICML 2024arXiv:2402.08400
2
citations

Adversarial Attacks on Combinatorial Multi-Armed Bandits

Rishab Balasubramanian, Jiawei Li, Tadepalli Prasad et al.

ICML 2024arXiv:2310.05308
3
citations

Adversarially Robust Deep Multi-View Clustering: A Novel Attack and Defense Framework

Haonan Huang, Guoxu Zhou, Yanghang Zheng et al.

ICML 2024

Adversarially Robust Distillation by Reducing the Student-Teacher Variance Gap

Junhao Dong, Piotr Koniusz, Junxi Chen et al.

ECCV 2024
10
citations

Adversarially Robust Hypothesis Transfer Learning

Yunjuan Wang, Raman Arora

ICML 2024

Adversarial Prompt Tuning for Vision-Language Models

Jiaming Zhang, Xingjun Ma, Xin Wang et al.

ECCV 2024arXiv:2311.11261
34
citations

Adversarial Robustification via Text-to-Image Diffusion Models

Daewon Choi, Jongheon Jeong, Huiwon Jang et al.

ECCV 2024arXiv:2407.18658
2
citations

Adversarial Robustness Limits via Scaling-Law and Human-Alignment Studies

Brian Bartoldson, James Diffenderfer, Konstantinos Parasyris et al.

ICML 2024arXiv:2404.09349
37
citations

Attack-free Evaluating and Enhancing Adversarial Robustness on Categorical Data

Yujun Zhou, Yufei Han, Haomin Zhuang et al.

ICML 2024

BadCLIP: Trigger-Aware Prompt Learning for Backdoor Attacks on CLIP

Jiawang Bai, Kuofeng Gao, Shaobo Min et al.

CVPR 2024arXiv:2311.16194
68
citations

BadPart: Unified Black-box Adversarial Patch Attacks against Pixel-wise Regression Tasks

Zhiyuan Cheng, Zhaoyi Liu, Tengda Guo et al.

ICML 2024arXiv:2404.00924
10
citations

Better Safe than Sorry: Pre-training CLIP against Targeted Data Poisoning and Backdoor Attacks

Wenhan Yang, Jingdong Gao, Baharan Mirzasoleiman

ICML 2024arXiv:2310.05862
18
citations

Be Your Own Neighborhood: Detecting Adversarial Examples by the Neighborhood Relations Built on Self-Supervised Learning

Zhiyuan He, Yijun Yang, Pin-Yu Chen et al.

ICML 2024arXiv:2209.00005
10
citations

Boosting Adversarial Training via Fisher-Rao Norm-based Regularization

Xiangyu Yin, Wenjie Ruan

CVPR 2024arXiv:2403.17520
13
citations

Breaking the Barrier: Enhanced Utility and Robustness in Smoothed DRL Agents

Chung-En Sun, Sicun Gao, Lily Weng

ICML 2024arXiv:2406.18062
6
citations

Can Implicit Bias Imply Adversarial Robustness?

Hancheng Min, Rene Vidal

ICML 2024arXiv:2405.15942
5
citations

Catastrophic Overfitting: A Potential Blessing in Disguise

MN Zhao, Lihe Zhang, Yuqiu Kong et al.

ECCV 2024arXiv:2402.18211
1
citations

Causality Based Front-door Defense Against Backdoor Attack on Language Models

Yiran Liu, Xiaoang Xu, Zhiyi Hou et al.

ICML 2024

Certifiably Robust Image Watermark

Zhengyuan Jiang, Moyang Guo, Yuepeng Hu et al.

ECCV 2024arXiv:2407.04086
9
citations

Characterizing Model Robustness via Natural Input Gradients

Adrian Rodriguez-Munoz, Tongzhou Wang, Antonio Torralba

ECCV 2024arXiv:2409.20139
2
citations

Collapse-Aware Triplet Decoupling for Adversarially Robust Image Retrieval

Qiwei Tian, Chenhao Lin, Zhengyu Zhao et al.

ICML 2024arXiv:2312.07364
3
citations

Comparing the Robustness of Modern No-Reference Image- and Video-Quality Metrics to Adversarial Attacks

Anastasia Antsiferova, Khaled Abud, Aleksandr Gushchin et al.

AAAI 2024paperarXiv:2310.06958
19
citations

Compositional Curvature Bounds for Deep Neural Networks

Taha Entesari, Sina Sharifi, Mahyar Fazlyab

ICML 2024arXiv:2406.05119
1
citations

Consistent Adversarially Robust Linear Classification: Non-Parametric Setting

Elvis Dohmatob

ICML 2024

Coupling Graph Neural Networks with Fractional Order Continuous Dynamics: A Robustness Study

Qiyu Kang, Kai Zhao, Yang Song et al.

AAAI 2024paperarXiv:2401.04331
10
citations

DataFreeShield: Defending Adversarial Attacks without Training Data

Hyeyoon Lee, Kanghyun Choi, Dain Kwon et al.

ICML 2024arXiv:2406.15635
1
citations

Defense Against Adversarial Attacks on No-Reference Image Quality Models with Gradient Norm Regularization

Yujia Liu, Chenxi Yang, Dingquan Li et al.

CVPR 2024arXiv:2403.11397
12
citations

DiffuseMix: Label-Preserving Data Augmentation with Diffusion Models

Khawar Islam, Muhammad Zaigham Zaheer, Arif Mahmood et al.

CVPR 2024arXiv:2405.14881
90
citations

DiG-IN: Diffusion Guidance for Investigating Networks - Uncovering Classifier Differences Neuron Visualisations and Visual Counterfactual Explanations

Maximilian Augustin, Yannic Neuhaus, Matthias Hein

CVPR 2024arXiv:2311.17833
9
citations

Enhancing Adversarial Robustness in SNNs with Sparse Gradients

Yujia Liu, Tong Bu, Ding Jianhao et al.

ICML 2024arXiv:2405.20355
14
citations

Et Tu Certifications: Robustness Certificates Yield Better Adversarial Examples

Andrew C. Cullen, Shijie Liu, Paul Montague et al.

ICML 2024arXiv:2302.04379
3
citations

Extending Adversarial Attacks to Produce Adversarial Class Probability Distributions

Jon Vadillo, Roberto Santana, Jose A Lozano

ICML 2024arXiv:2004.06383
1
citations

Geometry-Aware Instrumental Variable Regression

Heiner Kremer, Bernhard Schölkopf

ICML 2024arXiv:2405.11633