Skip to main content

Ollama

General AI model hosting.

Pick Your LLM

In ollama_entrypoint.sh, uncomment the LLM you want to preload.

For GPU setup, see Proxmox GPU.

Pre-download Models

Connect to running container

podman exec -it ollama sh

Pull a model

ollama pull gpt-oss:120b
# ollama pull qwen3:8b

Create a Larger Context Model

For 1000x better results with coding tools:

Create a Modelfile

FROM gpt-oss:20b
PARAMETER num_ctx 32000

Build the variant

ollama create gpt-oss:20b-32k -f Modelfile

Opencode Integration

Ollama works with opencode for AI-assisted coding.

# Install (avoid 1.3.2)
npm install -g opencode-ai@1.2.26
# Optional: disable autoupdate checks
# echo 'export OPENCODE_DISABLE_AUTOUPDATE=true' >> ~/.bashrc

# Language servers
npm install -g typescript-language-server typescript
npm install -g pyright
npm install -g vscode-langservers-extracted

Copy opencode.json.example_ to ~/.config/opencode/opencode.json and reference your model variant.