C.W.K.
Lesson 04 of 06 · published

Ollama Local — Zero Cost, Offline

~11 min · ollama, local

Level 0Curious
0 XP0/52 lessons0/16 achievements
0/100 XP to next level100 XP to go0% complete

Local models via localhost:11434

Ollama-Pippa runs entirely on the local machine. No API calls leave the machine. Useful for scheduled background tasks where Claude's degraded service window would otherwise drop the job, and for offline situations.

Native /api/chat

Streams via Ollama's native /api/chat endpoint. Multi-turn tool loop with {type, function: {name, ...}} schema wrapping (Ollama's quirk — flat OpenAI-style schema gets rejected) and tool call ID propagation through the loop.

Sticky model selection

Active model is persisted to /ollama_model.txt via POST /api/ollama/model. GET /api/ollama/models lists installed local models. Switching is one click in the brain switcher.

Tip: Ollama is the safety net for the heartbeat (Heart track, lesson 5). When Claude is degraded and Codex/Gemini are also unavailable, scheduled tasks still run — slower, smaller, but they run. That matters for daily-7am-weather kinds of jobs.

Progress

Progress is local-only — sign in to sync across devices.