Quantile Risk Control: A Flexible Framework for Bounding the Probability of High-Loss Predictions

December 27, 2022 ยท Entered Twilight ยท ๐Ÿ› International Conference on Learning Representations

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: .gitignore, LICENSE, MANIFEST.in, Makefile, README.md, data, environment.yml, models, output, paper_experiments.sh, requirements.txt, scripts, setup.py, var_control

Authors Jake C. Snell, Thomas P. Zollo, Zhun Deng, Toniann Pitassi, Richard Zemel arXiv ID 2212.13629 Category cs.LG: Machine Learning Cross-listed stat.ML Citations 12 Venue International Conference on Learning Representations Repository https://github.com/jakesnell/quantile-risk-control โญ 2 Last Checked 1 month ago
Abstract
Rigorous guarantees about the performance of predictive algorithms are necessary in order to ensure their responsible use. Previous work has largely focused on bounding the expected loss of a predictor, but this is not sufficient in many risk-sensitive applications where the distribution of errors is important. In this work, we propose a flexible framework to produce a family of bounds on quantiles of the loss distribution incurred by a predictor. Our method takes advantage of the order statistics of the observed loss values rather than relying on the sample mean alone. We show that a quantile is an informative way of quantifying predictive performance, and that our framework applies to a variety of quantile-based metrics, each targeting important subsets of the data distribution. We analyze the theoretical properties of our proposed method and demonstrate its ability to rigorously control loss quantiles on several real-world datasets.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning