LS-Merge: Merging Language Models in Latent Space

ICLR 2026 Conference SubmissionAnonymous Authors
LS-MergeLLM merginglatent spaceweight space learning
Abstract:

Model merging in weight space is an efficient way to reuse pretrained models, but existing methods typically assume matching architectures or sizes, making heterogeneous merges brittle or infeasible. We address this limitation by encoding model weights into a smooth latent space, enabling cross-architecture operations, and performing the merge in the latent space before decoding back to weights. This approach faces two major challenges. First, LLMs contain billions of parameters, which makes latent encoding computationally demanding. Second, using high compression ratios often hinders the encoder’s ability to generalize to unseen weights. We tackle these issues with a transformer-based variational autoencoder (VAE) trained in a two-stage compression curriculum with structured layer-aware chunking: the model first learns a high-capacity latent representation and then distills to a compact code, improving both stability and out-of-distribution generalization. To align heterogeneous models, we introduce a dimensionality-matching projection that allows interpolation between models of different sizes. Empirically, latent-space interpolation is consistently more robust than direct weight-space averaging and yields stronger downstream performance when merging models of different sizes. Together, these components provide a scalable, architecture-agnostic recipe for model merging.

Disclaimer
This report is AI-GENERATED using Large Language Models and WisPaper (A scholar search engine). It analyzes academic papers' tasks and contributions against retrieved prior work. While this system identifies POTENTIAL overlaps and novel directions, ITS COVERAGE IS NOT EXHAUSTIVE AND JUDGMENTS ARE APPROXIMATE. These results are intended to assist human reviewers and SHOULD NOT be relied upon as a definitive verdict on novelty.
NOTE that some papers exist in multiple, slightly different versions (e.g., with different titles or URLs). The system may retrieve several versions of the same underlying work. The current automated pipeline does not reliably align or distinguish these cases, so human reviewers will need to disambiguate them manually.
If you have any questions, please contact: mingzhang23@m.fudan.edu.cn

Overview

Overall Novelty Assessment

The paper proposes LS-Merge, a framework that encodes model weights into a latent space using a transformer-based VAE, enabling cross-architecture merging operations before decoding back to weights. This work resides in the 'Latent-Space and Embedding-Based Merging' leaf, which contains only three papers including the original. This is a relatively sparse research direction within the broader taxonomy of 50 papers across 22 leaf nodes, suggesting that latent-space encoding approaches for heterogeneous model merging remain an emerging area compared to more established techniques like direct weight interpolation or ensemble methods.

The taxonomy reveals that this work sits within 'Parameter-Space Merging and Alignment Techniques', adjacent to 'Weight-Space Interpolation and Coefficient Optimization' (3 papers) and 'Layer-Level Integration and Permutation' (2 papers). These neighboring leaves focus on direct parameter manipulation without latent encoding, highlighting a methodological divergence. The broader taxonomy also includes 'Knowledge Transfer and Ensemble Collaboration' (5 papers) and 'Mixture-of-Experts Architectures' (3 papers), which preserve model independence rather than merging parameters. The scope notes clarify that latent-space methods explicitly exclude direct weight averaging and ensemble approaches, positioning this work as a distinct strategy for achieving heterogeneous integration through learned representations.

Among 29 candidates examined, the contribution-level analysis reveals mixed novelty signals. The core LS-Merge framework (Contribution 1) examined 10 candidates with 1 appearing to provide overlapping prior work. The dimensionality-matching projection and optimal transport alignment (Contribution 2) examined 9 candidates with 2 potentially refuting papers. The two-stage compression curriculum with layer-aware chunking (Contribution 3) examined 10 candidates with none clearly refuting it, suggesting this training strategy may be more novel within the limited search scope. These statistics indicate that while the overall latent-space merging concept has some precedent, specific technical components—particularly the compression curriculum—appear less explored in the examined literature.

Based on the limited search of 29 semantically similar papers, the work appears to occupy a relatively sparse research direction with modest prior overlap. The taxonomy structure confirms that latent-space encoding for heterogeneous merging is less crowded than direct weight-space methods or ensemble approaches. However, the analysis does not cover exhaustive literature search or systematic review of all related compression and alignment techniques, leaving open the possibility of additional relevant work outside the top-K semantic matches examined.

Taxonomy

Core-task Taxonomy Papers
50
3
Claimed Contributions
29
Contribution Candidate Papers Compared
3
Refutable Paper

Research Landscape Overview

Core task: merging language models with heterogeneous architectures. The field has evolved to address the challenge of combining models that differ in structure, training objectives, or domain specialization. The taxonomy reveals several complementary strategies: Parameter-Space Merging and Alignment Techniques focus on direct weight manipulation and embedding-based fusion methods such as those explored in LS-Merge[0] and Mergenet[1]; Knowledge Transfer and Ensemble Collaboration emphasize collaborative inference and distillation approaches like Ensemble Heterogeneous LLMs[5] and Mixture-of-Agents[9]; Mixture-of-Experts Architectures provide modular frameworks for routing across heterogeneous components; while Multimodal and Cross-Domain Model Integration tackles the broader challenge of fusing models trained on different modalities or tasks. Domain-Specific Merging Applications and Optimization and Evaluation Frameworks address practical deployment and benchmarking concerns, with supporting methods in auxiliary techniques rounding out the landscape. A particularly active line of work centers on latent-space and embedding-based merging, where models are aligned through learned transformations rather than naive parameter averaging. LS-Merge[0] exemplifies this direction by operating in a shared latent space to bridge architectural differences, positioning itself alongside Mergenet[1] which also emphasizes learned alignment mechanisms, and Knowledge Fusion LLMs[2] which explores fusion at the representation level. These approaches contrast with ensemble methods like Mixture-of-Agents[9] that preserve model independence during inference, and with mixture-of-experts frameworks such as CoMoE[42] that introduce explicit gating. A recurring theme across branches is the trade-off between integration depth—whether to merge parameters directly, align intermediate representations, or coordinate outputs—and the preservation of specialized capabilities. Open questions include how to efficiently search the space of possible merges, as addressed by Evolutionary Model Merging[16], and how to evaluate merged models across diverse benchmarks without retraining from scratch.

Claimed Contributions

LS-Merge framework for merging LLMs in latent space

The authors introduce LS-Merge, a framework that encodes model weights into a smooth latent space using a transformer-based variational autoencoder, performs merging operations in this latent space, and decodes back to weights. This approach enables both homogeneous and heterogeneous model merging without requiring architectural alignment.

10 retrieved papers
Can Refute
Dimensionality-matching projection and OT-based alignment for heterogeneous merging

The authors develop a method combining proportional dimensionality mapping with Optimal Transport alignment to enable merging of models with mismatched architectures (different depths or widths). This addresses the geometric incompatibility of latent distributions from heterogeneous models by registering their manifolds before interpolation.

9 retrieved papers
Can Refute
Two-stage compression curriculum with layer-aware chunking

The authors propose a training strategy that first learns a high-capacity latent representation using a deterministic autoencoder, then enables the KL term to structure the latent space. This curriculum, combined with layer-aware chunking of weight tensors, improves stability and out-of-distribution generalization when encoding LLM weights with heavy-tailed distributions.

10 retrieved papers

Core Task Comparisons

Comparisons with papers in the same taxonomy category

Contribution Analysis

Detailed comparisons for each claimed contribution

Contribution

LS-Merge framework for merging LLMs in latent space

The authors introduce LS-Merge, a framework that encodes model weights into a smooth latent space using a transformer-based variational autoencoder, performs merging operations in this latent space, and decodes back to weights. This approach enables both homogeneous and heterogeneous model merging without requiring architectural alignment.

Contribution

Dimensionality-matching projection and OT-based alignment for heterogeneous merging

The authors develop a method combining proportional dimensionality mapping with Optimal Transport alignment to enable merging of models with mismatched architectures (different depths or widths). This addresses the geometric incompatibility of latent distributions from heterogeneous models by registering their manifolds before interpolation.

Contribution

Two-stage compression curriculum with layer-aware chunking

The authors propose a training strategy that first learns a high-capacity latent representation using a deterministic autoencoder, then enables the KL term to structure the latent space. This curriculum, combined with layer-aware chunking of weight tensors, improves stability and out-of-distribution generalization when encoding LLM weights with heavy-tailed distributions.