@@ -671,9 +671,9 @@ You can configure opencode to use local models through [llama.cpp's](https://git
"name": "Qwen3-Coder: a3b-30b (local)"
}
},
- "limit": {
- "context": 128000,
- "output": 65536
+ "limit": {
+ "context": 128000,
+ "output": 65536