curl -fsSL https://openclaw.ai/install.sh | bash
curl -fsSL https://ollama.com/install.sh | sh
ollama run gpt-oss:20b
openclaw uninstall
{
"models": {
"providers": {
"ollama": {
"baseUrl": "http://localhost:11434/v1",
"apiKey": "ollama-local",
"api": "openai-completions",
"models": [
{
"id": "gpt-oss:20b",
"name": "gpt-oss:20b",
"reasoning": false,
"input": [
"text"
],
"cost": {
"input": 0,
"output": 0,
"cacheRead": 0,
"cacheWrite": 0
},
"contextWindow": 131072,
"maxTokens": 8192
}
]
}
}
},
"agents": {
"defaults": {
"model": {
"primary": "ollama/gpt-oss:20b"
},
"workspace": "/Users/abhi/.openclaw/workspace",
"maxConcurrent": 4,
"subagents": {
"maxConcurrent": 8
}
}
},
"tools": {
"web": {
"search": {
"enabled": false
},
"fetch": {
"enabled": true
}
}
}
}openclaw onboard
Steps explained in the video.
Prompt injections, for one. Frontier models are hardened against this in training, whilst local models tend to rely more on system prompts,which is less secure.