OpenResponses.Adapters.Ollama (OpenResponses v0.1.1)

View Source

Adapter for Ollama — run any open-weight model locally.

Ollama exposes an OpenAI-compatible chat endpoint. No API key is required. GPU latency dominates for local models; run Ollama on the same host or a fast private network.

Configuration

No API key needed. Override the base URL if Ollama is running on a different host:

config :open_responses, :provider_config, %{
  ollama: [
    base_url: "http://gpu-server:11434"
  ]
}

Routing

Selected automatically for common local model name prefixes:

config :open_responses, :routing, %{
  ~r/^llama|^mistral|^phi|^qwen/ => OpenResponses.Adapters.Ollama
}

Add your own model name patterns to the routing table as needed.