DUET: Optimizing Training Data Mixtures via Coarse, Noisy Feedback from Unseen Evaluation Tasks
Overview
Overall Novelty Assessment
The paper introduces DUET, a global-to-local algorithm that optimizes training data mixtures using Bayesian optimization guided by coarse, noisy feedback from unseen evaluation tasks. It resides in the 'Bayesian Optimization with Task Feedback' leaf, which contains only two papers including this one. This is a relatively sparse research direction within the broader 'Feedback-Driven Data Mixture Optimization' branch, suggesting the specific combination of Bayesian optimization with coarse task feedback for mixture optimization remains underexplored compared to reweighting-based or predictive modeling approaches.
The taxonomy reveals several neighboring directions: 'Gradient-Based Feedback Alignment' uses online gradient signals rather than Bayesian search, while 'Adversarial and Agent-Based Feedback' employs self-learning agents. The sibling paper in the same leaf shares the Bayesian optimization framework but may differ in feedback granularity or data selection mechanisms. Adjacent branches like 'Reweighting-Based Data Mixture Optimization' (e.g., DoReMi) adjust domain weights without iterative feedback loops, and 'Predictive Modeling' approaches extrapolate performance from small-scale experiments rather than deploying models iteratively. DUET's interleaving of data selection with Bayesian optimization distinguishes it from these single-pass or purely predictive methods.
Among 24 candidates examined, no contribution was clearly refuted. The DUET algorithm itself was assessed against 4 candidates with no refutable overlap; the theoretical convergence analysis examined 10 candidates with no prior work providing equivalent regret bounds; and the problem formulation for unseen tasks reviewed 10 candidates without finding direct precedents. This suggests that within the limited search scope—top-K semantic matches plus citation expansion—the specific integration of coarse feedback, Bayesian optimization, and convergence guarantees appears novel. However, the modest candidate pool means the analysis cannot rule out relevant work outside this sample.
Given the sparse taxonomy leaf and absence of refutable candidates in the examined set, the work appears to occupy a relatively unexplored niche. The limited search scope (24 papers) and the narrow sibling set (one other paper) indicate that while the approach seems novel within the sampled literature, a more exhaustive review—especially of adjacent optimization and active learning communities—would be necessary to fully assess originality. The convergence analysis and coarse-feedback formulation stand out as potentially distinctive contributions based on the available evidence.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors introduce DUET, an algorithm that combines data selection methods with Bayesian optimization in an iterative manner to optimize training data mixtures using only coarse, noisy feedback from unseen evaluation tasks, without requiring fine-grained data information.
The authors provide a theoretical analysis demonstrating that DUET converges to the optimal training data mixture by analyzing the algorithm's attained cumulative regret under the Bayesian optimization framework, proving convergence without requiring detailed evaluation task data.
The authors formalize a new problem setting where practitioners lack fine-grained information about evaluation task data but can iteratively gather coarse performance feedback, addressing a gap between traditional domain adaptation and domain generalization approaches.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
[1] DUET: Optimizing Training Data Mixtures via Feedback from Unseen Evaluation Tasks PDF
Contribution Analysis
Detailed comparisons for each claimed contribution
DUET algorithm for optimizing training data mixtures via coarse feedback
The authors introduce DUET, an algorithm that combines data selection methods with Bayesian optimization in an iterative manner to optimize training data mixtures using only coarse, noisy feedback from unseen evaluation tasks, without requiring fine-grained data information.
[1] DUET: Optimizing Training Data Mixtures via Feedback from Unseen Evaluation Tasks PDF
[69] BPR: Bayesian personalized ranking from implicit feedback PDF
[70] Interactive Training: Feedback-Driven Neural Network Optimization PDF
[71] Relevance feedback using generalized Bayesian framework with region-based optimization learning PDF
Theoretical convergence analysis via cumulative regret
The authors provide a theoretical analysis demonstrating that DUET converges to the optimal training data mixture by analyzing the algorithm's attained cumulative regret under the Bayesian optimization framework, proving convergence without requiring detailed evaluation task data.
[1] DUET: Optimizing Training Data Mixtures via Feedback from Unseen Evaluation Tasks PDF
[51] Convergence Guarantees for Adaptive Bayesian Quadrature Methods PDF
[52] Convergence-Guaranteed Parametric Bayesian Distributed Cooperative Localization PDF
[53] Mixed-variable Bayesian optimization PDF
[54] Bayesian Optimization with Exponential Convergence PDF
[55] Think global and act local: Bayesian optimisation over high-dimensional categorical and mixed search spaces PDF
[56] A hybrid optimization algorithm with Bayesian inference for probabilistic model updating PDF
[57] An empirical Bayesian strategy for solving the simultaneous sparse approximation problem PDF
[58] Blending Data and Knowledge for Process Industrial Modeling Under Riemannian Preconditioned Bayesian Framework PDF
[59] Bayesian maximum entropy and data fusion for processing qualitative data: theory and application for crowdsourced cropland occurrences in Ethiopia PDF
Novel problem formulation for unseen evaluation tasks
The authors formalize a new problem setting where practitioners lack fine-grained information about evaluation task data but can iteratively gather coarse performance feedback, addressing a gap between traditional domain adaptation and domain generalization approaches.