Plano is an AI-native proxy and data plane for agentic apps — with built-in orchestration, safety, observability, and smart LLM routing so you stay focused on your agents core logic.
-
Updated
Mar 6, 2026 - Rust
Plano is an AI-native proxy and data plane for agentic apps — with built-in orchestration, safety, observability, and smart LLM routing so you stay focused on your agents core logic.
A curated list of Awesome-LLM-Ensemble papers for the survey "Harnessing Multiple Large Language Models: A Survey on LLM Ensemble"
High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model discovery across local and remote inference backends.
RouterArena: An open framework for evaluating LLM routers with standardized datasets, metrics, an automated framework, and a live leaderboard.
Intelligent model routing for OpenClaw with quota prediction, task classification, and automatic optimization
MCP AI Bridge — smart multi-model routing for Antigravity. Route tasks to the best AI model automatically.
3-Tier hybrid AI router that orchestrates FunctionGemma-270M on-device and Gemini 2.5 Flash Lite in the cloud for 99% function-calling accuracy at 548ms avg latency. Built at the Cactus × Google DeepMind Hackathon.
Compose, train and test fast LLM routers
An applied AI system using LLM routing, hybrid retrieval, and structured positive/negative reasoning for decision support.
Production-ready AI Agent Template optimized for Azure
NeurIPS 2025 paper "MESS+: Dynamically Learned Inference-Time LLM Routing in Model Zoos with Service Level Guarantees"
Unified interface server for various LLM providers with OpenAI API format
Hybrid AI routing: LOCAL Ollama + CLOUD GitHub Copilot
A neural multi-armed bandit framework for routing prompts to the most suitable LLM in a multi-agent system.
InsightBox - Hệ thống chống đứt gãy tri thức & Trợ lý học tập Đa luồng (Edge-Cloud Hybrid AI)
ARCHIVED: See openclaw-tactician for the active version
Enrutador inteligente de LLMs para el IndesIAhack 2025 que selecciona el mejor modelo por consulta para minimizar coste y consumo energético manteniendo la calidad.
Smart LLM routing that sends each query to the cheapest model that can handle it well. Cuts LLM costs by 60-75% without sacrificing quality. Includes a feedback loop that improves routing decisions over time.
Route multi-agent tasks to the right AI model — automatically
AURIXA: Real-time conversational AI orchestration platform. Multi-tenant, modular microservices with pluggable LLM providers (OpenAI, Claude, Gemini, local models), agent orchestration, RAG, safety guardrails, and voice streaming.
Add a description, image, and links to the llm-routing topic page so that developers can more easily learn about it.
To associate your repository with the llm-routing topic, visit your repo's landing page and select "manage topics."