•1 min read
Building a Local LLM Gateway on a Mac Mini
A small OpenAI-compatible proxy in front of Ollama — Bearer auth, model routing, streaming — so no cluster pod has to hold a vendor API key.
1 post tagged with ollama. This index aggregates both frontmatter tags: entries and inline #ollama mentions.
A small OpenAI-compatible proxy in front of Ollama — Bearer auth, model routing, streaming — so no cluster pod has to hold a vendor API key.