Abstract
A two-stage prompt routing architecture uses graph-based clustering and mixture-of-experts learning to efficiently select optimal language models for queries, achieving superior performance at reduced computational cost.
Prompt routing dynamically selects the most appropriate large language model from a pool of candidates for each query, optimizing performance while managing costs. As model pools scale to include dozens of frontier models with narrow performance gaps, existing approaches face significant challenges: manually defined task taxonomies cannot capture fine-grained capability distinctions, while monolithic routers struggle to differentiate subtle differences across diverse tasks. We propose a two-stage routing architecture that addresses these limitations through automated fine-grained task discovery and task-aware quality estimation. Our first stage employs graph-based clustering to discover latent task types and trains a classifier to assign prompts to discovered tasks. The second stage uses a mixture-of-experts architecture with task-specific prediction heads for specialized quality estimates. At inference, we aggregate predictions from both stages to balance task-level stability with prompt-specific adaptability. Evaluated on 10 benchmarks with 11 frontier models, our method consistently outperforms existing baselines and surpasses the strongest individual model while incurring less than half its cost.
Community
Code and model release soon
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- RouteMoA: Dynamic Routing without Pre-Inference Boosts Efficient Mixture-of-Agents (2026)
- Dynamic Model Routing and Cascading for Efficient LLM Inference: A Survey (2026)
- Effective LoRA Adapter Routing using Task Representations (2026)
- Towards Fair and Comprehensive Evaluation of Routers in Collaborative LLM Systems (2026)
- Expert Pyramid Tuning: Efficient Parameter Fine-Tuning for Expertise-Driven Task Allocation (2026)
- CoMoL: Efficient Mixture of LoRA Experts via Dynamic Core Space Merging (2026)
- MMR-Bench: A Comprehensive Benchmark for Multimodal LLM Routing (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper