Scalable Oversight via Partitioned Human Supervision
Overview
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors introduce a framework that exploits partitioned human expertise to collect complementary labels (indicating incorrect options) at scale for superhuman tasks. This enables evaluation and training of AI systems without requiring full ground truth or comprehensive expert verification.
The authors derive an unbiased linear correction estimator for accuracy using only complementary labels, analyze its variance properties, and propose two mixture estimators (inverse-variance weighted and maximum-likelihood) that combine ordinary and complementary labels with finite-sample deviation guarantees.
The authors empirically validate that their estimators enable both evaluation of large language models without ground truth and training of agentic AI systems by using complementary labels as fitness signals in agent search pipelines, demonstrating improved downstream performance.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
Contribution Analysis
Detailed comparisons for each claimed contribution
Scalable oversight framework via partitioned human supervision using complementary labels
The authors introduce a framework that exploits partitioned human expertise to collect complementary labels (indicating incorrect options) at scale for superhuman tasks. This enables evaluation and training of AI systems without requiring full ground truth or comprehensive expert verification.
[51] Federated benchmarking of medical artificial intelligence with MedPerf PDF
[52] Beyond Manual Annotation: A Human-AI Collaborative Framework for Medical Image Segmentation Using Only âBetter or Worseâ Expert Feedback PDF
[53] A Human-Centric Assessment Framework for AI PDF
Unbiased estimator of top-1 accuracy from complementary labels with variance analysis and mixture estimators
The authors derive an unbiased linear correction estimator for accuracy using only complementary labels, analyze its variance properties, and propose two mixture estimators (inverse-variance weighted and maximum-likelihood) that combine ordinary and complementary labels with finite-sample deviation guarantees.
[54] Unbiased Risk Estimators Can Mislead: A Case Study of Learning with Complementary Labels PDF
Demonstration of evaluation and agentic training using complementary labels
The authors empirically validate that their estimators enable both evaluation of large language models without ground truth and training of agentic AI systems by using complementary labels as fitness signals in agent search pipelines, demonstrating improved downstream performance.