An open API service indexing awesome lists of open source software.

https://github.com/msesia/conditional-conformal-pvalues

Conditional calibration of conformal p-values for outlier detection.
https://github.com/msesia/conditional-conformal-pvalues

conformal-prediction false-discovery-rate machine-learning outlier-detection statistics

Last synced: 3 months ago
JSON representation

Conditional calibration of conformal p-values for outlier detection.

Awesome Lists containing this project

README

          

# Testing for Outliers with Conformal p-values

We study the construction of p-values for nonparametric outlier detection, taking a multiple-testing perspective. The framework is that of conformal prediction, which wraps around any machine-learning algorithm to provide finite-sample guarantees regarding the validity of predictions for future testpoints. In this setting, existing methods can compute p-values that are marginally valid but mutually dependent for different future test points.

This repository contains a software implementation and guided examples for the methodology developed in the [accompanying paper](https://arxiv.org/abs/2104.08279), which provides a new method to compute p-values that are both conditionally valid and independent of each other for different future test points, thus allowing multiple testing with stronger stronger type-I error guarantees.
```
"Testing for Outliers with Conformal p-values"
Stephen Bates, Emmanuel Candes, Lihua Lei, Yaniv Romano, and Matteo Sesia.
accepted in Annals of Statistics (2022)
arXiv pre-print: https://arxiv.org/abs/2104.08279
```