π©
A lightweight Go-based proxy that lets you run multiple local LLMs simultaneously and switch between them on demand. Works with any OpenAI/Anthropic-compatible server (llama.cpp, vllm, etc.), making it perfect for developers experimenting with different models in their AI workflows.