AI Glossary

LLM Router

A system that automatically selects the best LLM for each query based on complexity, cost, and capability requirements.

Overview

An LLM router intelligently directs incoming queries to the most appropriate language model based on factors like query complexity, required capabilities, cost constraints, and latency requirements. Simple questions go to smaller, cheaper models, while complex queries are routed to more capable (and expensive) models.

Benefits

LLM routing can reduce API costs by 50-80% while maintaining quality by avoiding the use of expensive frontier models for simple tasks. Approaches include trained classifier models that predict query difficulty, rule-based routing using keyword detection, and cascade systems that try cheaper models first and escalate if quality is insufficient.

← Back to AI Glossary

Last updated: March 5, 2026