curl -fsSL https://openclaw.ai/install.sh | bash
curl -fsSL https://ollama.com/install.sh | sh
ollama run gpt-oss:20b
openclaw uninstall
{
"models": {
"providers": {
"ollama": {
"baseUrl": "http://localhost:11434/v1",
"apiKey": "ollama-local",
"api": "openai-completions",
"models": [
{
"id": "gpt-oss:20b",
"name": "gpt-oss:20b",
"reasoning": false,
"input": [
"text"
],
"cost": {
"input": 0,
"output": 0,
"cacheRead": 0,
"cacheWrite": 0
},
"contextWindow": 131072,
"maxTokens": 8192
}
]
}
}
},
"agents": {
"defaults": {
"model": {
"primary": "ollama/gpt-oss:20b"
},
"workspace": "/Users/abhi/.openclaw/workspace",
"maxConcurrent": 4,
"subagents": {
"maxConcurrent": 8
}
}
},
"tools": {
"web": {
"search": {
"enabled": false
},
"fetch": {
"enabled": true
}
}
}
}openclaw onboard
Steps explained in the video.
i dont think local model like gpt-oss:20b para will work for all the needs like we wont get that experience like all those when we use claude,gpt models but somehow secutity wise its ok but performence wise we need to search alternative on local machines can run.