SmarterRouter: An intelligent LLM gateway and VRAM-aware router for Ollama, llama.cpp, and OpenAI. Features semantic caching, model profiling, and automatic failover for local AI labs.
-
Updated
Mar 4, 2026 - Python
SmarterRouter: An intelligent LLM gateway and VRAM-aware router for Ollama, llama.cpp, and OpenAI. Features semantic caching, model profiling, and automatic failover for local AI labs.
Dropbox and pastebin for AI agents. Open and free. Useful for caching contexts and bridging multi-agent workflows
Add a description, image, and links to the ai-cache topic page so that developers can more easily learn about it.
To associate your repository with the ai-cache topic, visit your repo's landing page and select "manage topics."