Derandomized Online-to-Non-convex Conversion for Stochastic Weakly Convex Optimization
Overview
Overall Novelty Assessment
The paper proposes a deterministic variant of online-to-non-convex conversion specifically for weakly convex functions, achieving optimal oracle complexity without random interpolation or scaling. It occupies the sole leaf within the Deterministic Online-to-Non-convex Conversion Methods branch, with no sibling papers in the taxonomy. This positioning suggests the paper addresses a relatively sparse research direction, as the taxonomy contains only three total papers across three distinct branches, indicating limited prior exploration of deterministic conversion frameworks in this setting.
The taxonomy reveals two neighboring branches: Randomized Online-to-Non-convex Conversion Methods and Robust Conversion Under Heavy-Tailed Noise. The randomized branch handles general non-smooth non-convex objectives using stochastic techniques, while the robust branch addresses heavy-tailed gradient noise through clipping mechanisms. The paper diverges from these directions by eliminating randomness entirely and restricting attention to weakly convex functions, trading generality for deterministic guarantees. The scope notes clarify that weak convexity assumptions distinguish this work from general non-smooth methods, while the exclusion of heavy-tailed settings separates it from robust variants.
Among fourteen candidates examined, the core derandomization contribution shows no clear refutation across four candidates reviewed. However, the periodically restarted OGD variant with momentum-restarted SGDM reveals one refutable candidate among six examined, and the regularized Goldstein stationarity criterion shows one refutable candidate among four examined. These statistics suggest the algorithmic framework appears relatively novel within the limited search scope, though specific technical components encounter more substantial prior work. The modest candidate pool indicates the analysis captures nearby semantic matches rather than exhaustive field coverage.
Based on the limited literature search of fourteen candidates, the derandomization framework for weakly convex functions appears to occupy underexplored territory, though certain algorithmic and stationarity components show overlap with existing work. The sparse taxonomy structure and absence of sibling papers reinforce this impression, though the small search scope precludes definitive claims about field-wide novelty. The analysis primarily reflects proximity to top semantic matches rather than comprehensive prior art coverage.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors develop a deterministic variant of the online-to-non-convex conversion (O2NC) framework specifically for weakly convex optimization. This derandomized algorithm eliminates the random interpolation or scaling steps required in prior O2NC methods while maintaining optimal oracle complexity and resembling standard SGDM more closely.
The authors introduce a novel periodically restarted online gradient descent procedure that resets momentum updates after fixed iteration periods. This variant corresponds to a momentum-restarted version of SGDM and enables more aggressive updates without explicit ball constraints, demonstrated to be effective for training ResNet and ViT networks.
The authors define a new regularized version of Goldstein stationarity that simultaneously controls the scale of convex hulls of subgradients and proximity of underlying subsets. This criterion obviates the need for explicit tiny-ball constraints in the online learning module, allowing for potentially more progressive increment updates.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
Contribution Analysis
Detailed comparisons for each claimed contribution
Derandomized online-to-non-convex conversion for weakly convex functions
The authors develop a deterministic variant of the online-to-non-convex conversion (O2NC) framework specifically for weakly convex optimization. This derandomized algorithm eliminates the random interpolation or scaling steps required in prior O2NC methods while maintaining optimal oracle complexity and resembling standard SGDM more closely.
[13] The cost of nonconvexity in deterministic nonsmooth optimization PDF
[14] Turnstile leverage score sampling with applications PDF
[15] A tight lower bound for online convex optimization with switching costs PDF
[16] Technical Report: Reactive Navigation in Partially Known Non-Convex Environments PDF
Periodically restarted OGD variant with momentum-restarted SGDM
The authors introduce a novel periodically restarted online gradient descent procedure that resets momentum updates after fixed iteration periods. This variant corresponds to a momentum-restarted version of SGDM and enables more aggressive updates without explicit ball constraints, demonstrated to be effective for training ResNet and ViT networks.
[9] Proximal gradient algorithm with momentum and flexible parameter restart for nonconvex optimization PDF
[7] Momentum-Based Multi-Agent Approaches to Distributed Nonconvex Optimization PDF
[8] A momentum accelerated stochastic method and its application on policy search problems PDF
[10] Single-loop stochastic smoothing algorithms for nonsmooth Riemannian optimization PDF
[11] Augmenting Optimizers with NARS: The NARS Learning Rate Scheduler PDF
[12] MoXCo: How I learned to stop exploring and love my local minima? PDF
Regularized Goldstein stationarity criterion
The authors define a new regularized version of Goldstein stationarity that simultaneously controls the scale of convex hulls of subgradients and proximity of underlying subsets. This criterion obviates the need for explicit tiny-ball constraints in the online learning module, allowing for potentially more progressive increment updates.