Prompt Risk Control: A Rigorous Framework for Responsible Deployment of Large Language Models

12citations
arXiv:2311.13628
12
citations
#1221
in ICLR 2024
of 2297 papers
6
Top Authors
4
Data Points

Abstract

With the explosion of the zero-shot capabilities of (and thus interest in) pre-trained large language models, there has come accompanying interest in how best to prompt a language model to perform a given task. While it may be tempting to choose a prompt based on empirical results on a validation set, this can lead to a deployment where an unexpectedly high loss occurs. To mitigate this prospect, we propose a lightweight framework, Prompt Risk Control, for selecting a prompt based on rigorous upper bounds on families of informative risk measures. We provide and compare different methods for producing bounds on a diverse set of risk metrics like mean, CVaR, and the Gini coefficient of the loss distribution. In addition, we extend the underlying statistical bounding techniques to accommodate the possibility of distribution shifts in deployment. Extensive experiments on high-impact applications like chatbots, medical question answering, and news summarization highlight why such a framework is necessary to reduce exposure to the worst outcomes.

Citation History

Jan 28, 2026
11
Feb 13, 2026
12+1
Feb 13, 2026
12
Feb 13, 2026
12