← Back to Papers

Token-Level LLM Collaboration via FusionRoute

Nuoya Xiong, Yuhang Zhou, Hanqing Zeng, Zhaorun Chen, Furong Huang, Shuchao Bi, Lizhu Zhang, Zhuokai Zhao
arXiv.org | 2026
FusionRoute is a token-level multi-LLM collaboration framework that jointly selects the best domain-expert LLM at each decoding step AND contributes a trainable complementary logit to correct/refine the selected expert's next-token distribution. This dual mechanism theoretically and empirically overcomes the fundamental limitations of pure expert-selection routing.

Problem Statement

Deploying a single large general-purpose LLM strong across all domains is prohibitively expensive, while smaller domain-specialized models fail to generalize. Existing token-level collaboration methods rely solely on fixed expert outputs, which is theoretically shown to be insufficient for realizing the optimal decoding policy without unrealistic global coverage assumptions. There is a need for an efficient, principled method to combine specialized models that retains their strengths while compensating for their weaknesses.

Key Novelty

  • Theoretical proof that pure expert-only token-level routing is fundamentally limited and cannot recover the optimal decoding policy without strong global coverage assumptions, motivating the need for a complementary generator
  • A lightweight router that simultaneously performs expert selection AND generates a complementary logit added to the chosen expert's output at each token step, expanding the effective policy class
  • Demonstration that the augmented FusionRoute framework can recover optimal value functions under mild conditions, with empirical validation across multiple model families (Llama-3, Gemma-2) and diverse tasks

Evaluation Highlights

  • FusionRoute outperforms sequence-level collaboration, token-level collaboration, model merging, and direct fine-tuning baselines across mathematical reasoning, code generation, and instruction following benchmarks on both Llama-3 and Gemma-2 families
  • FusionRoute remains competitive with individual domain experts on their respective specialized tasks while generalizing across domains, demonstrating it does not sacrifice specialization for breadth

Breakthrough Assessment

7/10 FusionRoute makes a meaningful theoretical contribution by formally characterizing the failure mode of pure routing and proposes a principled fix with strong empirical backing across multiple model families and task types; however, it is an incremental architectural advance in the growing LLM collaboration/routing space rather than a paradigm shift.

Methodology

  1. Train a lightweight router model that, at each autoregressive decoding step, scores available domain-expert LLMs and selects the most suitable one based on the current context
  2. The same router simultaneously generates a complementary logit vector representing corrections or refinements not captured by the selected expert's output distribution
  3. The final next-token distribution is computed by adding the router's complementary logit to the selected expert's logit, then sampling; the router is trained end-to-end to minimize task loss across domains

System Components

Lightweight Router

A small trainable model that operates at each token decoding step to both select the best domain expert and produce a complementary logit correction signal

Expert LLMs

A pool of domain-specialized smaller LLMs (e.g., math, code, instruction-following experts) whose logits are queried at each step based on router selection

Logit Addition Mechanism

The complementary logit from the router is added directly to the selected expert's next-token logit distribution, enabling fine-grained correction and recovery of distributions not representable by any single expert

Theoretical Coverage Analysis

Formal analysis proving that expert-only routing cannot realize optimal decoding without global coverage, motivating and justifying the complementary generator component

Results

Benchmark/Method Best Single Expert This Paper (FusionRoute) Delta
Mathematical Reasoning Domain expert (strong in-domain) Outperforms all baselines Positive improvement over collaboration baselines
Code Generation Domain expert (strong in-domain) Outperforms all baselines Positive improvement over collaboration baselines
Instruction Following Domain expert (strong in-domain) Outperforms all baselines Positive improvement over collaboration baselines
vs. Token-level Collaboration Prior SOTA token routing FusionRoute superior Consistent gains across families
vs. Model Merging Merged model baseline FusionRoute superior Consistent gains across families

Key Takeaways

  • Pure token-level routing that only selects among experts is theoretically insufficient; practitioners building LLM routing systems should include a trainable correction mechanism (e.g., logit augmentation) to avoid fundamental expressivity limitations
  • FusionRoute's lightweight router adds minimal overhead while enabling effective multi-domain generalization, making it a practical alternative to expensive monolithic large models or brittle model merging approaches
  • The framework is model-family agnostic (validated on both Llama-3 and Gemma-2), suggesting it can be applied broadly to combine any set of specialized LLMs without retraining the experts themselves

Abstract

Large language models (LLMs) exhibit strengths across diverse domains. However, achieving strong performance across these domains with a single general-purpose model typically requires scaling to sizes that are prohibitively expensive to train and deploy. On the other hand, while smaller domain-specialized models are much more efficient, they struggle to generalize beyond their training distributions. To address this dilemma, we propose FusionRoute, a robust and effective token-level multi-LLM collaboration framework in which a lightweight router simultaneously (i) selects the most suitable expert at each decoding step and (ii) contributes a complementary logit that refines or corrects the selected expert's next-token distribution via logit addition. Unlike existing token-level collaboration methods that rely solely on fixed expert outputs, we provide a theoretical analysis showing that pure expert-only routing is fundamentally limited: unless strong global coverage assumptions hold, it cannot in general realize the optimal decoding policy. By augmenting expert selection with a trainable complementary generator, FusionRoute expands the effective policy class and enables recovery of optimal value functions under mild conditions. Empirically, across both Llama-3 and Gemma-2 families and diverse benchmarks spanning mathematical reasoning, code generation, and instruction following, FusionRoute outperforms both sequence- and token-level collaboration, model merging, and direct fine-tuning, while remaining competitive with domain experts on their respective tasks.

Generated on 2026-03-02 using Claude