Know When to Abstain: Optimal Selective Classification with Likelihood Ratios
Overview
Overall Novelty Assessment
The paper proposes a Neyman–Pearson framework for selective classification under covariate shift, introducing two distance-based selector methods (∆-MDS and ∆-KNN) and providing comprehensive empirical evaluation. It resides in the 'Likelihood Ratio and Neyman-Pearson Approaches' leaf, which contains only two papers total. This represents a relatively sparse research direction within the broader taxonomy of fifty papers across thirty-six topics, suggesting the theoretical grounding of selective classification in classical statistical decision theory remains underexplored compared to post-hoc confidence methods or OOD detection approaches.
The taxonomy reveals neighboring work in sibling leaves under 'Theoretical Foundations and Optimal Selection Mechanisms,' including minimax analysis using transfer exponents and unified rejection frameworks across multiple loss functions. The paper's emphasis on likelihood ratios distinguishes it from the more crowded 'Post-Hoc Selection Strategies' branch (six papers across three leaves) and 'Selective Classification with Out-of-Distribution Detection' branch (thirteen papers across four leaves). The scope note for this leaf explicitly excludes heuristic confidence-based methods, positioning the work as a principled alternative to empirical baselines that dominate neighboring branches.
Among thirty candidates examined through limited semantic search, none clearly refuted the three main contributions. The Neyman–Pearson framework contribution examined ten candidates with zero refutable matches, as did the two novel distance-based methods and the comprehensive evaluation under covariate shift. This absence of overlapping prior work within the examined scope suggests the specific combination of classical statistical optimality theory with modern selective classification under distribution shift has received limited direct attention, though the search scale means potentially relevant work outside the top-thirty semantic matches may exist.
The analysis reflects a bounded literature search rather than exhaustive coverage of the field. The sparse population of the taxonomy leaf and lack of refutable candidates among thirty examined papers suggest novelty within the explored scope, though the limited search depth means the assessment cannot definitively rule out related work in adjacent statistical learning theory or domain adaptation literature not captured by semantic similarity to the paper's abstract and introduction.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors apply the classical Neyman–Pearson lemma from statistics to selective classification, showing that the optimal selection function is a likelihood ratio test between correct and incorrect predictions. This theoretical framework provides principled guidance for designing selector functions in modern deep networks.
The authors propose ∆-MDS and ∆-KNN, which are modified versions of Mahalanobis distance and k-nearest neighbors methods that explicitly estimate separate distributions for correctly and incorrectly classified samples. They also introduce a linear combination strategy to merge distance-based and logit-based scores, all motivated by the Neyman–Pearson framework.
The authors provide extensive experiments on covariate shift scenarios across vision (ImageNet variants) and language (Amazon Reviews) tasks, evaluating both vision-language models like CLIP and supervised classifiers. This addresses an underexplored setting in selective classification where input distributions change while label spaces remain fixed.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
[2] Model-free selective inference under covariate shift via weighted conformal p-values PDF
Contribution Analysis
Detailed comparisons for each claimed contribution
Neyman–Pearson framework for optimal selective classification
The authors apply the classical Neyman–Pearson lemma from statistics to selective classification, showing that the optimal selection function is a likelihood ratio test between correct and incorrect predictions. This theoretical framework provides principled guidance for designing selector functions in modern deep networks.
[70] Revamping Conformal Selection With Optimal Power: A Neyman--Pearson Perspective PDF
[71] Sequential analysis: Hypothesis testing and changepoint detection PDF
[72] Density Ratio Estimation and Neyman Pearson Classification with Missing Data PDF
[73] Weighted joint LRTs for cooperative spectrum sensing using K-means clustering PDF
[74] Online anomaly detection in the Neyman-Pearson hypothesis testing framework PDF
[75] Optimal Decision Rules for Composite Binary Hypothesis Testing under Neyman-Pearson Framework PDF
[76] Hybrid Fusion Combining Palmprint and Palm Vein for Large-Scale Palm-Based Recognition PDF
[77] Explicit Abstention Knobs for Predictable Reliability in Video Question Answering PDF
[78] On the validity of the likelihood ratio and maximum likelihood methods PDF
[79] Adaptive Robust and Nonparametric Procedures with Application to Communications, Radar, Sonar and Array Signal Processing PDF
Two novel distance-based selector methods: ∆-MDS and ∆-KNN
The authors propose ∆-MDS and ∆-KNN, which are modified versions of Mahalanobis distance and k-nearest neighbors methods that explicitly estimate separate distributions for correctly and incorrectly classified samples. They also introduce a linear combination strategy to merge distance-based and logit-based scores, all motivated by the Neyman–Pearson framework.
[60] Feature Selection based on Mahalanobis Distance for Early Parkinson Disease Classification PDF
[61] Maturity status classification of papaya fruits based on machine learning and transfer learning approach PDF
[62] A novel anomaly detection method based on adaptive Mahalanobis-squared distance and one-class kNN rule for structural health monitoring under environmental ⦠PDF
[63] Imbalance data classification using local mahalanobis distance learning based on nearest neighbor PDF
[64] Quantum K-nearest neighbors classification algorithm based on Mahalanobis distance PDF
[65] An empirical study of distance metrics for k-nearest neighbor algorithm PDF
[66] Spatial outlier detection on discrete GNSS velocity fields using robust Mahalanobis-distance-based unsupervised classification PDF
[67] A novel anomaly detection method based on adaptive Mahalanobis-squared distance and one-class kNN rule for structural health monitoring under environmental effects PDF
[68] Learning a Mahalanobis distance metric for data clustering and classification PDF
[69] Mahalanobis distance based multivariate outlier detection to improve performance of hypertension prediction PDF
Comprehensive evaluation under covariate shift
The authors provide extensive experiments on covariate shift scenarios across vision (ImageNet variants) and language (Amazon Reviews) tasks, evaluating both vision-language models like CLIP and supervised classifiers. This addresses an underexplored setting in selective classification where input distributions change while label spaces remain fixed.