Ollama
Configure Ollama for local model inference — no API key, OpenAI-compatible pass-through.
The ollama provider connects to a local Ollama instance. Ollama exposes an OpenAI-compatible API, so requests are forwarded as-is.
Configuration
[providers.ollama]
kind = "ollama"
models = ["llama3.2", "mistral"]The default base URL is http://localhost:11434/v1. Override it if Ollama runs on a different host:
[providers.ollama]
kind = "ollama"
base_url = "http://192.168.1.100:11434/v1"
models = ["llama3.2"]No API key is needed for local Ollama.
Usage
Start Ollama, pull a model, then send requests through CrabLLM:
ollama pull llama3.2
crabllm --config crabllm.tomlcurl http://localhost:8080/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "llama3.2",
"messages": [{"role": "user", "content": "Hello!"}]
}'Supported endpoints
- Chat completions (streaming and non-streaming)
- Embeddings (if supported by the Ollama model)