Ollama (Local Models)
Use Ollama to run open-source models locally with RadarOS. No API key required—ideal for development, privacy-sensitive workloads, and cost-free experimentation.Setup
Install Ollama
Download and install Ollama from ollama.ai. Start the Ollama service (it runs on
http://localhost:11434 by default).Factory
The Ollama model name (e.g.,
llama3.1, codellama, mistral).Optional configuration. See Config below.
Config
Ollama server URL. Defaults to
http://localhost:11434. Use this for remote Ollama instances or custom ports.No API key needed
No API key needed
Ollama runs locally and does not require an API key. Just ensure the Ollama service is running.
Example
Popular Models
| Model | Use Case |
|---|---|
llama3.1 | General purpose, strong all-around performance |
codellama | Code generation and understanding |
mistral | Fast, efficient, good for chat |
mixtral | Mixture of experts, higher capability |
phi3 | Small, fast, good for edge devices |
Discover more models
Discover more models
Run
ollama list to see installed models. Browse ollama.com/library for the full catalog.Multi-Modal Support
Ollama supports image input for vision-capable models likellava, bakllava, and llama3.2-vision.
Images
Pass images as base64 data inContentPart[]:
Unsupported: Audio & Files
Audio and file inputs are not supported by Ollama. If passed, the provider logs a warning and skips them.Tool Calling
Ollama supports function calling with select models. Enable tools on your agent as usual — RadarOS handles the tool call protocol automatically:Models with Tool Support
| Model | Tool Calling |
|---|---|
llama3.1 | Yes |
llama3.2 | Yes |
mistral | Yes |
mixtral | Yes |
codellama | No |
phi3 | No |
llava | No (vision only) |
Performance Tips
GPU Acceleration
Ollama automatically uses GPU when available. Check GPU usage with:Context Size
By default, Ollama uses a 2048-token context window. For agents with long conversations or large tool results, increase it:Model Selection Guidelines
| Scenario | Recommended Model |
|---|---|
| General chat + tools | llama3.1 or llama3.1:70b |
| Code generation | codellama:13b or codellama:34b |
| Vision tasks | llava:13b or llama3.2-vision |
| Fast responses (edge) | phi3:mini or llama3.2:1b |
| Complex reasoning | mixtral:8x7b or llama3.1:70b |