Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
173
result(s) for
"conformal prediction"
Sort by:
Nonparametric predictive distributions based on conformal prediction
by
Manokhin, Valery
,
Min-ge, Xie
,
Vovk, Vladimir
in
Distribution functions
,
Least squares
,
Nonparametric statistics
2019
This paper applies conformal prediction to derive predictive distributions that are valid under a nonparametric assumption. Namely, we introduce and explore predictive distribution functions that always satisfy a natural property of validity in terms of guaranteed coverage for IID observations. The focus is on a prediction algorithm that we call the Least Squares Prediction Machine (LSPM). The LSPM generalizes the classical Dempster–Hill predictive distributions to nonparametric regression problems. If the standard parametric assumptions for Least Squares linear regression hold, the LSPM is as efficient as the Dempster–Hill procedure, in a natural sense. And if those parametric assumptions fail, the LSPM is still valid, provided the observations are IID.
Journal Article
Majority vote ensembles of conformal predictors
2019
We study majority vote ensembles of \\[\\varepsilon \\]-valid conformal predictors (CP). We show that the prediction set \\[\\varGamma ^\\eta \\] produced as the majority vote among the prediction sets \\[\\varGamma ^\\varepsilon _i\\] of k independent \\[\\varepsilon \\]-valid CPs is also valid, for some significance level \\[\\eta \\]; we provide a method to compute \\[\\varepsilon \\] to achieve a desired \\[\\eta \\]. We further indicate an error upper bound for an ensemble of correlated CPs, and derive a value \\[\\varepsilon \\] for which such an ensemble guarantees \\[\\eta \\] conservative validity. We evaluate empirically our findings, and compare them with alternative strategies for combining CPs’ predictions.
Journal Article
Automatic face recognition with well-calibrated confidence measures
by
Papadopoulos, Harris
,
Eliades, Charalambos
,
Král, Pavel
in
Annotations
,
Classifiers
,
Face recognition
2019
Automatic face recognition (AFR) has gained the attention of many institutes and researchers in the past two decades due to its wide range of applications. This attention resulted in the development of a variety of techniques for the particular task with a high recognition accuracy when the environment is well-controlled. In the case of moderately controlled or fully uncontrolled environments however, the performance of most techniques is dramatically reduced due to the much higher difficulty of the task. As a result, the provision of some kind of indication of the likelihood of a recognition being correct is a desirable property of AFR techniques in many applications, such as the detection of wanted persons or the automatic annotation of photographs. This work investigates the application of the conformal prediction (CP) framework for extending the output of AFR techniques with well-calibrated measures of confidence. In particular we combine CP with one classifier based on patterns of oriented edge magnitudes descriptors, one classifier based on scale invariant feature transform descriptors, and a weighted combination of the similarities computed by the two. We examine and compare the performance of five nonconformity measures for the particular task in terms of their accuracy and informational efficiency.
Journal Article
Classification Under Partial Reject Options
2024
In many applications there is ambiguity about which (if any) of a finite number
N
of hypotheses that best fits an observation. It is of interest then to possibly output a whole set of categories, that is, a scenario where the size of the classified set of categories ranges from 0 to
N
. Empty sets correspond to an outlier, sets of size 1 represent a firm decision that singles out one hypothesis, sets of size
N
correspond to a rejection to classify, whereas sets of sizes
2
,
…
,
N
-
1
represent a partial rejection to classify, where some hypotheses are excluded from further analysis. In this paper, we review and unify several proposed methods of Bayesian set-valued classification, where the objective is to find the optimal Bayesian classifier that maximizes the expected reward. We study a large class of reward functions with rewards for sets that include the true category, whereas additive or multiplicative penalties are incurred for sets depending on their size. For models with one homogeneous block of hypotheses, we provide general expressions for the accompanying Bayesian classifier, several of which extend previous results in the literature. Then, we derive novel results for the more general setting when hypotheses are partitioned into blocks, where ambiguity within and between blocks are of different severity. We also discuss how well-known methods of classification, such as conformal prediction, indifference zones, and hierarchical classification, fit into our framework. Finally, set-valued classification is illustrated using an ornithological data set, with taxa partitioned into blocks and parameters estimated using MCMC. The associated reward function’s tuning parameters are chosen through cross-validation.
Journal Article
Criteria of efficiency for set-valued classification
by
Petej, Ivan
,
Fedorova, Valentina
,
Nouretdinov, Ilia
in
Artificial Intelligence
,
Classification
,
Complex Systems
2017
We study optimal conformity measures for various criteria of efficiency of set-valued classification in an idealised setting. This leads to an important class of criteria of efficiency that we call probabilistic and argue for; it turns out that the most standard criteria of efficiency used in literature on conformal prediction are not probabilistic unless the problem of classification is binary. We consider both unconditional and label-conditional conformal prediction.
Journal Article
Aleatoric and epistemic uncertainty in machine learning: an introduction to concepts and methods
2021
The notion of uncertainty is of major importance in machine learning and constitutes a key element of machine learning methodology. In line with the statistical tradition, uncertainty has long been perceived as almost synonymous with standard probability and probabilistic predictions. Yet, due to the steadily increasing relevance of machine learning for practical applications and related issues such as safety requirements, new problems and challenges have recently been identified by machine learning scholars, and these problems may call for new methodological developments. In particular, this includes the importance of distinguishing between (at least) two different types of uncertainty, often referred to as aleatoric and epistemic. In this paper, we provide an introduction to the topic of uncertainty in machine learning as well as an overview of attempts so far at handling uncertainty in general and formalizing this distinction in particular.
Journal Article
Conformal prediction under feedback covariate shift for biomolecular design
2022
Many applications of machine-learning methods involve an iterative protocol in which data are collected, a model is trained, and then outputs of that model are used to choose what data to consider next. For example, a data-driven approach for designing proteins is to train a regression model to predict the fitness of protein sequences and then use it to propose new sequences believed to exhibit greater fitness than observed in the training data. Since validating designed sequences in the wet laboratory is typically costly, it is important to quantify the uncertainty in the model’s predictions. This is challenging because of a characteristic type of distribution shift between the training and test data that arises in the design setting—one in which the training and test data are statistically dependent, as the latter is chosen based on the former. Consequently, the model’s error on the test data—that is, the designed sequences—has an unknown and possibly complex relationship with its error on the training data. We introduce a method to construct confidence sets for predictions in such settings, which account for the dependence between the training and test data. The confidence sets we construct have finite-sample guarantees that hold for any regression model, even when it is used to choose the test-time input distribution. As a motivating use case, we use real datasets to demonstrate how our method quantifies uncertainty for the predicted fitness of designed proteins and can therefore be used to select design algorithms that achieve acceptable tradeoffs between high predicted fitness and low predictive uncertainty.
Journal Article
Distribution‐free prediction bands for non‐parametric regression
2014
We study distribution‐free, non‐parametric prediction bands with a focus on their finite sample behaviour. First we investigate and develop different notions of finite sample coverage guarantees. Then we give a new prediction band by combining the idea of ‘conformal prediction’ with non‐parametric conditional density estimation. The proposed estimator, called COPS (conformal optimized prediction set), always has a finite sample guarantee. Under regularity conditions the estimator converges to an oracle band at a minimax optimal rate. A fast approximation algorithm and a data‐driven method for selecting the bandwidth are developed. The method is illustrated in simulated and real data examples.
Journal Article
Valid prediction intervals for regression problems
by
Dewolf, Nicolas
,
Waegeman, Willem
,
Baets, Bernard De
in
Bayesian analysis
,
Calibration
,
Comparative analysis
2023
Over the last few decades, various methods have been proposed for estimating prediction intervals in regression settings, including Bayesian methods, ensemble methods, direct interval estimation methods and conformal prediction methods. An important issue is the validity and calibration of these methods: the generated prediction intervals should have a predefined coverage level, without being overly conservative. So far, no study has analysed this issue whilst simultaneously considering these four classes of methods. In this independent comparative study, we review the above four classes of methods from a conceptual and experimental point of view in the i.i.d. setting. Results on benchmark data sets from various domains highlight large fluctuations in performance from one data set to another. These observations can be attributed to the violation of certain assumptions that are inherent to some classes of methods. We illustrate how conformal prediction can be used as a general calibration procedure for methods that deliver poor results without a calibration step.
Journal Article
Conformal load prediction with transductive graph autoencoders
2025
Predicting edge weights on graphs has various applications, from transportation systems to social networks. This paper describes a Graph Neural Network (GNN) approach for edge weight prediction with guaranteed coverage. We leverage conformal prediction to calibrate the GNN outputs and produce valid prediction intervals. We handle data heteroscedasticity through error reweighting and Conformalized Quantile Regression (CQR). We compare the performance of our method against baseline techniques on real-world transportation datasets. Our approach has better coverage and efficiency than all baselines and showcases robustness and adaptability.
Journal Article