Skip to content
#

inference-gateway

Here are 17 public repositories matching this topic...

Engine-agnostic LLM gateway in Rust. Full OpenAI & Anthropic API compatibility across SGLang, vLLM, TRT-LLM, OpenAI, Gemini & more. Industry-first gRPC pipeline, KV cache-aware routing, chat history, tokenization caching, Responses API, embeddings, WASM plugins, MCP, and multi-tenant auth.

  • Updated Apr 4, 2026
  • Rust

Core Elixir primitives for building reliable self-hosted inference clients, provider adapters, transport boundaries, and operational controls for private AI runtimes across local, edge, and dedicated infrastructure.

  • Updated Apr 2, 2026
  • Elixir

Improve this page

Add a description, image, and links to the inference-gateway topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the inference-gateway topic, visit your repo's landing page and select "manage topics."

Learn more