Two-Way Is Better Than One: Bidirectional Alignment with Cycle Consistency for Exemplar-Free Class-Incremental Learning

ICLR 2026 Conference SubmissionAnonymous Authors
continual learningexemplar freeexemplar free class incremental learningclass incremental learningexemplar-free
Abstract:

Continual learning (CL) seeks models that acquire new skills without erasing prior knowledge. In exemplar-free class-incremental learning (EFCIL), this challenge is amplified because past data cannot be stored, making representation drift for old classes particularly harmful. Prototype-based EFCIL is attractive for its efficiency, yet prototypes drift as the embedding space evolves; thus, projection-based drift compensation has become a popular remedy. We show, however, that existing one-directional projections introduce systematic bias: they either retroactively distort the current feature geometry or align past classes only locally, leaving cycle inconsistencies that accumulate across tasks. We introduce bidirectional projector alignment during training: two maps, old\tonew and new\toold, are trained during each new task with stop-gradient gating and a cycle-consistency objective so that transport and representation co-evolve. Analytically, we prove that the cycle loss contracts the singular spectrum toward unity in whitened space and that improved transport of class means/covariances yields smaller perturbations of classification log-odds, preserving old-class decisions and directly mitigating catastrophic forgetting. Empirically, across standard EFCIL benchmarks, our method achieves unprecedented reductions in forgetting while maintaining very high accuracy on new tasks, consistently outperforming state-of-the-art approaches.

Disclaimer
This report is AI-GENERATED using Large Language Models and WisPaper (A scholar search engine). It analyzes academic papers' tasks and contributions against retrieved prior work. While this system identifies POTENTIAL overlaps and novel directions, ITS COVERAGE IS NOT EXHAUSTIVE AND JUDGMENTS ARE APPROXIMATE. These results are intended to assist human reviewers and SHOULD NOT be relied upon as a definitive verdict on novelty.
NOTE that some papers exist in multiple, slightly different versions (e.g., with different titles or URLs). The system may retrieve several versions of the same underlying work. The current automated pipeline does not reliably align or distinguish these cases, so human reviewers will need to disambiguate them manually.
If you have any questions, please contact: mingzhang23@m.fudan.edu.cn

Overview

Overall Novelty Assessment

The paper introduces bidirectional projector alignment with cycle consistency to address prototype drift in exemplar-free class-incremental learning. It resides in the 'Bidirectional and Cycle-Consistent Alignment' leaf under 'Prototype Drift Compensation Mechanisms,' where it is currently the sole occupant among 40 papers across the taxonomy. This isolation suggests the specific combination of bidirectional projection and cycle consistency during training represents a relatively unexplored niche within the broader drift compensation landscape, which includes four other subcategories addressing prototype shift through alternative mechanisms.

The taxonomy reveals substantial activity in neighboring directions. 'Adaptive Prototype Correction and Reconstruction' contains four papers exploring density-based reinforcement and topology correction, while 'Semantic Shift Estimation and Dual-Projection' includes two works on dual-projection frameworks without cycle constraints. The 'Prototype Calibration and Refinement Strategies' branch offers complementary post-hoc adjustment methods, and 'Pseudo-Replay and Synthetic Data Generation' provides an alternative paradigm through feature regeneration. The paper's bidirectional approach bridges alignment-based drift correction with geometric consistency enforcement, occupying conceptual space between unidirectional projection methods and calibration strategies that lack mutual alignment guarantees.

Among 11 candidates examined, the bidirectional alignment contribution shows one refutable candidate from six examined, while the geometry-preserving transport mechanism found no refutations among five candidates. The theoretical analysis contribution was not tested against prior work. The limited search scope—11 total candidates rather than an exhaustive survey—means these statistics reflect top-K semantic matches and citation expansion, not comprehensive coverage. The bidirectional alignment's single refutation suggests some overlap exists within the examined subset, while the transport mechanism appears more distinctive among the candidates reviewed.

Based on the 11-candidate search, the work appears to occupy a sparse research direction with limited direct precedent in its specific technical approach. The taxonomy structure confirms that while drift compensation is a crowded area overall, the particular combination of bidirectional projection and cycle consistency during training has minimal representation. However, the analysis cannot rule out relevant work outside the examined candidate set, and the single refutation for the core contribution warrants careful examination of overlap boundaries.

Taxonomy

Core-task Taxonomy Papers
40
3
Claimed Contributions
11
Contribution Candidate Papers Compared
1
Refutable Paper

Research Landscape Overview

Core task: exemplar-free class-incremental learning with prototype drift compensation. The field addresses the challenge of learning new classes sequentially without storing past exemplars, while mitigating the drift that occurs when class prototypes shift as the feature extractor adapts to new data. The taxonomy reveals a rich landscape organized around ten major branches. Prototype Drift Compensation Mechanisms focus on alignment strategies that correct or stabilize prototypes across incremental steps, including bidirectional and cycle-consistent methods like Bidirectional Alignment Cycle[0]. Prototype Calibration and Refinement Strategies emphasize post-hoc adjustments using techniques such as Training-Free Prototype Calibration[1] and Adaptive Prototype Correction[2]. Feature Representation Stabilization aims to preserve old-class discriminability through regularization and consolidation, while Pseudo-Replay and Synthetic Data Generation leverage generative models to reconstruct old-class information. Additional branches cover Specialized Architectural and Modular Approaches, Heterogeneity-Aware and Distribution-Based Methods like Distribution-Aware Knowledge Prototyping[10], Prototype Evolution and Reminiscence Mechanisms such as Prototype Reminiscence[33], Domain-Specific extensions including Traffic Sign Incremental[3], and strategies for Old Class Reconstruction like Resurrecting Old Classes[12]. A central tension across these branches involves balancing stability and plasticity: methods must prevent catastrophic forgetting of old classes while accommodating new ones. Within Prototype Drift Compensation Mechanisms, Bidirectional Alignment Cycle[0] pursues cycle-consistent alignment to ensure forward and backward prototype consistency, contrasting with simpler unidirectional corrections found in works like Learnable Drift Compensation[5] or Semantic Drift Compensation[13]. This bidirectional approach sits at the intersection of alignment-based drift correction and feature-space consistency enforcement, sharing conceptual ground with calibration strategies such as Adaptive Prototype Correction[2] but emphasizing mutual alignment rather than one-way adjustment. Meanwhile, methods in Pseudo-Replay branches like Retrospective Feature Synthesis[15] and Dual-Consistency Model Inversion[18] tackle drift indirectly by regenerating old-class features, offering a complementary perspective. The original work's focus on cycle consistency positions it as a principled solution to drift, addressing a core challenge that resonates across multiple branches while offering a distinct geometric perspective on prototype alignment.

Claimed Contributions

Bidirectional projector alignment with cycle consistency during training

The authors propose learning two projection maps simultaneously during each task—one from old to new feature space and one from new to old—using stop-gradient operations and a cycle-consistency loss. This approach allows transport and representation to evolve together, addressing limitations of prior two-stage, one-directional methods.

6 retrieved papers
Can Refute
Theoretical analysis of cycle loss and decision stability

The authors provide theoretical guarantees showing that minimizing the cycle loss contracts singular values toward one in whitened feature space, and that reduced alignment and cycle errors lead to tighter bounds on classification log-odds perturbations, thereby preserving decisions on old classes.

0 retrieved papers
Geometry-preserving transport mechanism for drift mitigation

The method maintains the geometric structure of old-class representations even as the feature extractor adapts to new tasks, which reduces the tendency to favor recent classes and improves retention of previously learned knowledge.

5 retrieved papers

Core Task Comparisons

Comparisons with papers in the same taxonomy category

Within the taxonomy built over the current TopK core-task papers, the original paper is assigned to a leaf with no direct siblings and no cousin branches under the same grandparent topic. In this retrieved landscape, it appears structurally isolated, which is one partial signal of novelty, but still constrained by search coverage and taxonomy granularity.

Contribution Analysis

Detailed comparisons for each claimed contribution

Contribution

Bidirectional projector alignment with cycle consistency during training

The authors propose learning two projection maps simultaneously during each task—one from old to new feature space and one from new to old—using stop-gradient operations and a cycle-consistency loss. This approach allows transport and representation to evolve together, addressing limitations of prior two-stage, one-directional methods.

Contribution

Theoretical analysis of cycle loss and decision stability

The authors provide theoretical guarantees showing that minimizing the cycle loss contracts singular values toward one in whitened feature space, and that reduced alignment and cycle errors lead to tighter bounds on classification log-odds perturbations, thereby preserving decisions on old classes.

Contribution

Geometry-preserving transport mechanism for drift mitigation

The method maintains the geometric structure of old-class representations even as the feature extractor adapts to new tasks, which reduces the tendency to favor recent classes and improves retention of previously learned knowledge.

Two-Way Is Better Than One: Bidirectional Alignment with Cycle Consistency for Exemplar-Free Class-Incremental Learning | Novelty Validation