Ollama
General AI model hosting.
- Source: https://github.com/ollama/ollama
- License: MIT
- Alternatives: vLLM, LocalAI, LM Studio
Pick Your LLM
In ollama_entrypoint.sh, uncomment the LLM you want to preload.
For GPU setup, see Proxmox GPU.
Pre-download Models
Connect to running container
podman exec -it ollama sh
Pull a model
ollama pull gpt-oss:120b
# ollama pull qwen3:8b
Create a Larger Context Model
For 1000x better results with coding tools:
Create a Modelfile
FROM gpt-oss:20b
PARAMETER num_ctx 32000
Build the variant
ollama create gpt-oss:20b-32k -f Modelfile
Opencode Integration
Ollama works with opencode for AI-assisted coding.
# Install (avoid 1.3.2)
npm install -g opencode-ai@1.2.26
# Optional: disable autoupdate checks
# echo 'export OPENCODE_DISABLE_AUTOUPDATE=true' >> ~/.bashrc
# Language servers
npm install -g typescript-language-server typescript
npm install -g pyright
npm install -g vscode-langservers-extracted
Copy opencode.json.example_ to ~/.config/opencode/opencode.json and reference your model variant.