α
Research
Alpha Leak
Conferences
Topics
Top Authors
Rankings
Browse All
EN
中
Home
/
Authors
/
Nicholas Carlini
Nicholas Carlini
21
papers
7,023
total citations
papers (21)
FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence
NEURIPS 2020
arXiv
4,370
citations
On Adaptive Attacks to Adversarial Example Defenses
NEURIPS 2020
arXiv
917
citations
Measuring Robustness to Natural Distribution Shifts in Image Classification
NEURIPS 2020
arXiv
642
citations
Are aligned neural networks adversarially aligned?
NEURIPS 2023
arXiv
320
citations
Counterfactual Memorization in Neural Language Models
NEURIPS 2023
arXiv
170
citations
Stealing part of a production language model
ICML 2024
arXiv
145
citations
The Privacy Onion Effect: Memorization is Relative
NEURIPS 2022
arXiv
141
citations
Handcrafted Backdoors in Deep Neural Networks
NEURIPS 2022
arXiv
89
citations
Students Parrot Their Teachers: Membership Inference on Model Distillation
NEURIPS 2023
arXiv
40
citations
Persistent Pre-training Poisoning of LLMs
ICLR 2025
arXiv
38
citations
Indicators of Attack Failure: Debugging and Improving Optimization of Adversarial Examples
NEURIPS 2022
arXiv
35
citations
Adversarial Perturbations Cannot Reliably Protect Artists From Generative AI
ICLR 2025
arXiv
35
citations
Increasing Confidence in Adversarial Robustness Evaluations
NEURIPS 2022
arXiv
22
citations
Measuring Non-Adversarial Reproduction of Training Data in Large Language Models
ICLR 2025
arXiv
13
citations
Effective Robustness against Natural Distribution Shifts for Models with Different Training Data
NEURIPS 2023
arXiv
12
citations
Exploring and Mitigating Adversarial Manipulation of Voting-Based Leaderboards
ICML 2025
arXiv
12
citations
Initialization Matters for Adversarial Transfer Learning
CVPR 2024
arXiv
11
citations
AutoAdvExBench: Benchmarking Autonomous Exploitation of Adversarial Example Defenses
ICML 2025
arXiv
8
citations
Position: In-House Evaluation Is Not Enough. Towards Robust Third-Party Evaluation and Flaw Disclosure for General-Purpose AI
ICML 2025
2
citations
IF-Guide: Influence Function-Guided Detoxification of LLMs
NEURIPS 2025
arXiv
1
citations
Position: Considerations for Differentially Private Learning with Large-Scale Public Pretraining
ICML 2024
0
citations