Corner Gradient Descent
Overview
Overall Novelty Assessment
The paper proposes generalized stationary SGD algorithms with infinite memory that achieve convergence rates up to O(t^{-2ζ}) on infinite-dimensional quadratic problems with power-law spectra. It resides in the 'Generalized Stationary SGD with Infinite Memory' leaf under 'Momentum and Acceleration Techniques', where it is currently the sole paper. This placement indicates a sparse research direction within the broader momentum-based acceleration landscape, suggesting the infinite-memory perspective on SGD acceleration is relatively unexplored compared to finite-memory momentum methods or explicit preconditioning approaches.
The taxonomy reveals neighboring work in adjacent leaves: Heavy Ball Momentum Under Gradient Noise addresses anisotropic noise effects, Nesterov Acceleration for Stochastic Settings focuses on interpolation-based guarantees, and Krylov Subspace Acceleration projects gradients onto subspaces. The paper's contour-based framework and corner algorithms diverge from these by framing acceleration through complex-plane geometry rather than classical momentum schedules or subspace projections. The exclude_note clarifies that finite-memory momentum methods belong elsewhere, positioning this work as fundamentally distinct in its use of infinite weighted gradient history to handle power-law spectral conditions.
Among 23 candidates examined across three contributions, none were found to clearly refute the proposed ideas. The contour view examined 10 candidates with no refutations, corner algorithms with acceleration factor θ examined 3 candidates with none refuting, and finite-memory approximations examined 10 candidates with no overlapping prior work identified. This suggests that within the limited search scope—focused on top semantic matches and citation expansion—the specific combination of contour-based analysis, corner geometry for acceleration, and infinite-memory stationary schemes appears novel. The absence of sibling papers in the same taxonomy leaf reinforces that this particular framing has not been extensively studied.
The analysis covers a targeted literature search rather than an exhaustive survey of all optimization methods. The taxonomy structure shows the field has many alternative acceleration strategies (preconditioning, finite-memory momentum, learning rate schedules), but the infinite-memory stationary perspective with complex-plane contours represents a distinct angle. The limited candidate pool and sparse taxonomy leaf suggest this direction is relatively underexplored, though the search scope does not rule out related work in adjacent mathematical optimization communities not captured by the semantic search.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors establish that stationary SGD algorithms with arbitrary linear memory can be represented as contours in the complex plane through a response map Ψ. This geometric viewpoint allows algorithm design to be reformulated as contour design, where the loss evolution is determined by the contour properties.
The authors introduce corner algorithms based on contours with corners having external angle θπ (1 < θ < 2). They prove these algorithms accelerate convergence from O(t−ζ) to O(t−θζ) and establish a complete phase diagram showing the maximum achievable acceleration θmax = min(2, ν, 2/(ζ+1/ν)) in the signal-dominated regime.
The authors show that ideal corner algorithms, which require infinite memory, can be efficiently approximated by practical finite-memory algorithms. This is achieved using fast rational approximations of power functions with error bounds of O(e−c√M), making the theoretical corner algorithms implementable in practice.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
Contribution Analysis
Detailed comparisons for each claimed contribution
Contour view of generalized (S)GD algorithms
The authors establish that stationary SGD algorithms with arbitrary linear memory can be represented as contours in the complex plane through a response map Ψ. This geometric viewpoint allows algorithm design to be reformulated as contour design, where the loss evolution is determined by the contour properties.
[37] Adaptive orthogonal gradient descent algorithm for fully complex-valued neural networks PDF
[38] Multi-depth computer-generated hologram based on stochastic gradient descent algorithm with weighted complex loss function and masked diffraction PDF
[39] Accelerating Smooth Games by Manipulating Spectral Shapes PDF
[40] The saddle-point method in differential privacy PDF
[41] Load Distribution Optimization of Steel Storage Rack Based on Genetic Algorithm PDF
[42] Complex s-plane modeling and 2d characterization of the stochastic scattering loss in symmetric dielectric slab waveguides exhibiting ergodic surface-roughness with ⦠PDF
[43] Active contour model in deep learning era: A revise and review PDF
[44] An extension of the steepest descent method for contour integrals PDF
[45] Natural Gradient Descent of Complex-Valued Neural Networks Invariant under Rotations PDF
[46] " Brownian strings": Segmenting images with stochastically deformable contours PDF
Corner algorithms with acceleration factor θ
The authors introduce corner algorithms based on contours with corners having external angle θπ (1 < θ < 2). They prove these algorithms accelerate convergence from O(t−ζ) to O(t−θζ) and establish a complete phase diagram showing the maximum achievable acceleration θmax = min(2, ν, 2/(ζ+1/ν)) in the signal-dominated regime.
Finite-memory approximation of corner algorithms
The authors show that ideal corner algorithms, which require infinite memory, can be efficiently approximated by practical finite-memory algorithms. This is achieved using fast rational approximations of power functions with error bounds of O(e−c√M), making the theoretical corner algorithms implementable in practice.