DeepAFL: Deep Analytic Federated Learning

ICLR 2026 Conference SubmissionAnonymous Authors
Analytic LearningFederated LearningContinual LearningLifelong LearningIncremental LearningRepresentation LearningData Heterogeneity
Abstract:

Federated Learning (FL) is a popular distributed learning paradigm to break down data silo. Traditional FL approaches largely rely on gradient-based updates, facing significant issues about heterogeneity, scalability, convergence, and overhead, etc. Recently, some analytic-learning-based work has attempted to handle these issues by eliminating gradient-based updates via analytical (i.e., closed-form) solutions. Despite achieving superior invariance to data heterogeneity, these approaches are fundamentally limited by their single-layer linear model with a frozen pre-trained backbone. As a result, they can only achieve suboptimal performance due to their lack of representation learning capabilities. In this paper, to enable representable analytic models while preserving the ideal invariance to data heterogeneity for FL, we propose our Deep Analytic Federated Learning approach, named DeepAFL. Drawing inspiration from the great success of ResNet in gradient-based learning, we design gradient-free residual blocks in our DeepAFL with analytical solutions. We further introduce an efficient layer-wise protocol for training our deep analytic models layer by layer in FL through least squares. Both theoretical analyses and empirical evaluations validate our DeepAFL's superior performance with its dual advantages in heterogeneity invariance and representation learning, outperforming state-of-the-art baselines by up to 5.68%-8.42% across three benchmark datasets. The related codes will be made open-sourced upon the acceptance of this paper.

Disclaimer
This report is AI-GENERATED using Large Language Models and WisPaper (A scholar search engine). It analyzes academic papers' tasks and contributions against retrieved prior work. While this system identifies POTENTIAL overlaps and novel directions, ITS COVERAGE IS NOT EXHAUSTIVE AND JUDGMENTS ARE APPROXIMATE. These results are intended to assist human reviewers and SHOULD NOT be relied upon as a definitive verdict on novelty.
NOTE that some papers exist in multiple, slightly different versions (e.g., with different titles or URLs). The system may retrieve several versions of the same underlying work. The current automated pipeline does not reliably align or distinguish these cases, so human reviewers will need to disambiguate them manually.
If you have any questions, please contact: mingzhang23@m.fudan.edu.cn

Overview

Overall Novelty Assessment

The paper proposes DeepAFL, a deep analytic federated learning approach that extends single-layer analytic models to multi-layer architectures with representation learning capabilities. It resides in the 'Analytic and Gradient-Free Federated Learning' leaf of the taxonomy, which currently contains only this paper as a sibling. This indicates a sparse research direction within the broader federated learning landscape, where gradient-based methods dominate. The taxonomy shows fifty papers across thirty-six topics, with most concentrated in personalized learning, privacy mechanisms, and domain-specific applications, making this analytic approach a relatively isolated niche.

The taxonomy reveals that neighboring branches focus on gradient-based personalization (e.g., representation-classifier decoupling with eight papers), contrastive self-supervised methods (four papers), and privacy-preserving techniques (six papers across three subcategories). DeepAFL diverges fundamentally by eliminating iterative optimization entirely, contrasting with methods like Feature Alignment and Classifier Collaboration or Meta-Learning Personalized that rely on gradient updates. The scope note for this leaf explicitly excludes gradient-based methods, positioning DeepAFL as a complementary paradigm rather than an incremental refinement of existing iterative approaches. Its connection to domain-specific applications remains unclear from the taxonomy structure.

Among thirty candidates examined, the DeepAFL framework itself shows no clear refutation (zero of ten candidates), suggesting novelty in the overall system design. However, gradient-free residual blocks and layer-wise least-squares training each face one refutable candidate among ten examined. The limited search scope means these statistics reflect top-thirty semantic matches, not exhaustive coverage. The contribution-level analysis indicates that while the integrated approach appears novel, individual technical components (residual blocks, layer-wise protocols) may have precedents in the examined literature. The sparse sibling count in the taxonomy leaf supports the impression of a relatively unexplored direction.

Based on the limited search scope of thirty candidates, the work appears to occupy a genuinely sparse research area within federated learning. The taxonomy structure confirms that analytic methods represent a minority approach compared to gradient-based personalization and privacy-preserving techniques. However, the analysis cannot rule out relevant prior work outside the top-thirty semantic matches or in adjacent fields like distributed optimization or kernel methods. The contribution-level statistics suggest moderate novelty, with the system-level integration appearing more distinctive than individual components.

Taxonomy

Core-task Taxonomy Papers
50
3
Claimed Contributions
30
Contribution Candidate Papers Compared
2
Refutable Paper

Research Landscape Overview

Core task: deep analytic federated learning with representation learning capabilities. The field encompasses a diverse set of approaches that address how distributed clients can collaboratively learn meaningful representations while respecting privacy, heterogeneity, and communication constraints. The taxonomy reveals several major branches: unsupervised and self-supervised methods (e.g., Federated Unsupervised Representation[1], Pretrained Contrastive[5]) that learn features without labels; vertical federated learning (Vertical Unsupervised[2]) where different feature sets reside at different parties; personalized federated learning via representation decoupling (Personalization Layers[8], Common Representation Personalized[36]) that separates shared and local knowledge; meta-learning and matching-based personalization (Meta-Learning Personalized[9], Matched Averaging[7]); classifier calibration and regularization (Classifier Calibration[26], Feature Alignment Classifier[4]) to handle data heterogeneity; domain generalization and adaptation (Domain Generalization[17], Domain Bias Elimination[33]); multimodal and vision-language learning (Multimodal IoT[14], Vision-Language Grounding[22]); privacy-preserving mechanisms (Soteria[12], Privacy Survey[28]); security and robustness (Backdoor Attack[38], Robust Aggregation[39]); domain-specific applications spanning healthcare, IoT, and beyond (Medical Image Analysis[37], Network Anomaly Detection[32]); and analytic or gradient-free methods that bypass iterative optimization. A particularly active line of work explores how to decouple shared representations from personalized components, balancing global knowledge transfer with local adaptation under non-IID data. Another contrasting theme is the tension between privacy-preserving techniques and the need for rich feature sharing, with methods like Pravfed[3] and Learning-Outcome Privacy[25] proposing novel trade-offs. DeepAFL[0] sits squarely within the analytic and gradient-free branch, offering a closed-form solution that avoids the iterative gradient descent common in most federated approaches. Compared to gradient-based personalization methods such as Feature Distribution Adaptation[13] or DualFed[27], DeepAFL[0] emphasizes computational efficiency and theoretical tractability, potentially sacrificing some representational flexibility for faster convergence and reduced communication overhead. This positions it as a complementary direction to the dominant iterative paradigm, appealing to scenarios where analytic solutions can yield competitive performance with lower resource demands.

Claimed Contributions

DeepAFL: Deep Analytic Federated Learning approach

The authors introduce DeepAFL, a novel federated learning method that enables gradient-free deep representation learning while maintaining invariance to data heterogeneity. This approach addresses the fundamental limitation of existing analytic-learning-based FL methods that rely on single-layer linear models.

10 retrieved papers
Gradient-free residual blocks with analytical solutions

The authors design residual blocks inspired by ResNet that can be trained without gradients using closed-form solutions. These blocks incorporate random projections, activation functions, and learnable transformations to enable deep representation learning in the analytic learning framework.

10 retrieved papers
Can Refute
Layer-wise training protocol via least squares

The authors develop a layer-wise training protocol that allows deep analytic models to be trained efficiently in federated settings using least squares optimization. This protocol enables clients to perform lightweight forward-propagation computations while the server aggregates global models layer by layer.

10 retrieved papers
Can Refute

Core Task Comparisons

Comparisons with papers in the same taxonomy category

Within the taxonomy built over the current TopK core-task papers, the original paper is assigned to a leaf with no direct siblings and no cousin branches under the same grandparent topic. In this retrieved landscape, it appears structurally isolated, which is one partial signal of novelty, but still constrained by search coverage and taxonomy granularity.

Contribution Analysis

Detailed comparisons for each claimed contribution

Contribution

DeepAFL: Deep Analytic Federated Learning approach

The authors introduce DeepAFL, a novel federated learning method that enables gradient-free deep representation learning while maintaining invariance to data heterogeneity. This approach addresses the fundamental limitation of existing analytic-learning-based FL methods that rely on single-layer linear models.

Contribution

Gradient-free residual blocks with analytical solutions

The authors design residual blocks inspired by ResNet that can be trained without gradients using closed-form solutions. These blocks incorporate random projections, activation functions, and learnable transformations to enable deep representation learning in the analytic learning framework.

Contribution

Layer-wise training protocol via least squares

The authors develop a layer-wise training protocol that allows deep analytic models to be trained efficiently in federated settings using least squares optimization. This protocol enables clients to perform lightweight forward-propagation computations while the server aggregates global models layer by layer.