A Law of Data Reconstruction for Random Features (And Beyond)
Overview
Overall Novelty Assessment
The paper establishes a 'law of data reconstruction' for random features models, showing that training data can be recovered when the number of parameters p exceeds dn (data dimensionality times sample count). It resides in the 'Gradient-Based Theoretical Analysis' leaf under 'Theoretical Foundations and Identifiability', alongside one sibling paper (Gradient Reconstruction Provably). This leaf is notably sparse, containing only two papers within a broader taxonomy of 50 works, suggesting the paper targets a relatively underexplored theoretical niche focused on formal identifiability conditions rather than empirical attack development.
The taxonomy reveals that most reconstruction research concentrates on practical attack methods (Direct Parameter-Based Reconstruction, Gradient Inversion Attacks) and defense mechanisms (Differential Privacy Bounds), with fewer works establishing theoretical foundations. Neighboring leaves include 'Implicit Bias and Margin Maximization' (four papers exploiting gradient descent properties) and 'Bayesian and Probabilistic Frameworks' (one paper on inverse problems). The paper's focus on random features and parameter-dimensionality thresholds distinguishes it from these adjacent directions, which emphasize optimization dynamics or statistical inference rather than capacity-based reconstruction limits.
Among 30 candidates examined, the theoretical contributions (law of data reconstruction, Theorems 1 and 2) show no clear refutation across 10 candidates each, suggesting limited prior work on this specific threshold phenomenon. However, the optimization method for reconstruction faces overlap: 2 of 10 candidates appear refutable, indicating that algorithmic approaches to parameter-based data recovery have received more attention. This pattern aligns with the taxonomy structure, where attack algorithms dominate the literature while theoretical characterizations of reconstruction regimes remain less developed.
Based on this limited search scope (top-30 semantic matches), the theoretical framing around the dn threshold appears relatively novel, though the optimization method connects to a more established algorithmic literature. The analysis does not cover exhaustive citation networks or domain-specific venues, so additional related work may exist outside the examined candidates. The sparse population of the theoretical analysis leaf suggests the paper addresses a gap in formal reconstruction theory.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors establish a theoretical threshold showing that entire training datasets can be reconstructed from random features models when the number of parameters p exceeds dn, where d is data dimensionality and n is the number of training samples. This reveals a fundamental law governing when data reconstruction becomes feasible.
The authors prove that when the feature representations of training samples lie in the span of reconstructed sample features under sufficient over-parameterization, the reconstructed samples must be close to original training data. Theorem 2 extends this to show reconstructed samples are distinct, ruling out duplicates.
The authors develop a practical reconstruction algorithm based on minimizing the projection of trained parameters onto the orthogonal complement of reconstructed feature space. They demonstrate this method successfully recovers training data across random features, two-layer networks, and deep residual networks when p exceeds dn.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
[8] Reconstructing Training Data from Model Gradient, Provably PDF
Contribution Analysis
Detailed comparisons for each claimed contribution
Law of data reconstruction for random features
The authors establish a theoretical threshold showing that entire training datasets can be reconstructed from random features models when the number of parameters p exceeds dn, where d is data dimensionality and n is the number of training samples. This reveals a fundamental law governing when data reconstruction becomes feasible.
[51] Robust training under label noise by over-parameterization PDF
[52] Model repair: Robust recovery of over-parameterized statistical models PDF
[53] Recovery of training data from overparameterized autoencoders: An inverse problem perspective PDF
[54] Invertible kernel PCA with random fourier features PDF
[55] Overparameterization improves stylegan inversion PDF
[56] Noise Regularizes Over-parameterized Rank One Matrix Recovery, Provably PDF
[57] Regularized Random Fourier Features and Finite Element Reconstruction for Operator Learning in Sobolev Space PDF
[58] More is Less: Inducing Sparsity via Overparameterization PDF
[59] Wind field reconstruction with adaptive random Fourier features. PDF
[60] Overparameterized Neural Networks Implement Associative Memory PDF
Theoretical characterization via Theorems 1 and 2
The authors prove that when the feature representations of training samples lie in the span of reconstructed sample features under sufficient over-parameterization, the reconstructed samples must be close to original training data. Theorem 2 extends this to show reconstructed samples are distinct, ruling out duplicates.
[61] Hybrid two-stage reconstruction of multiscale subsurface flow with physics-informed residual connected neural operator PDF
[62] Generalized transfer subspace learning through low-rank constraint PDF
[63] Secure split learning against property inference, data reconstruction, and feature space hijacking attacks PDF
[64] Graph regularized feature selection with data reconstruction PDF
[65] Adaptive surface reconstruction with multiscale convolutional kernels PDF
[66] Bayesian and related methods in image reconstruction from incomplete data PDF
[67] Data reconstruction coverage based on graph signal processing for wireless sensor networks PDF
[68] SAR target recognition with limited training data based on angular rotation generative network PDF
[69] Frustratingly Easy Feature Reconstruction for Out-of-Distribution Detection PDF
[70] Kernel PCA and de-noising in feature spaces PDF
Optimization method for data reconstruction
The authors develop a practical reconstruction algorithm based on minimizing the projection of trained parameters onto the orthogonal complement of reconstructed feature space. They demonstrate this method successfully recovers training data across random features, two-layer networks, and deep residual networks when p exceeds dn.