What is it?
Hellas Gate is a developer-first LLM model router with three unique features:
- Smart Routing: dynamically choose models and providers
- Virtual Models: easily integrate into your editors and tools
- Local Compute: access local compute from anywhere; think "tailscale for your local LLM"
Read on for more details, or see the Quickstart to get started.
Smart Routing
Gateway's smart routing lets you dynamically choose which models and providers to use based on a number of preferences:
- Cost
- Latency
- Geographical location
- Model (e.g. Llama)
- Provider (e.g., OpenAI, Anthropic, etc.)
- ... and more.
You can also route by intent. For example, you can ask for the best model for writing code.
Virtual Models
Virtual models are like aliases for a router.
For example, you could create an autocomplete model which picks the lowest-latency model on any provider.
Then you can configure your editor to use this model for autocompletion.
Local Compute
If you run an OpenAI-compatible API locally (like vLLM), you can make it available to your own account anywhere in the world. Think of this as "tailscale for your local LLM".
See local compute for setup instructions.