Original Reddit post

So I was triing to use ollama for use opencode as VS estention Opencode works fine with the BigPickle but if i try to use for example with qwen2.5-coder:7b i cannot make the simpler task that give me no problem with BigPickle like : “Make a dir called testdirectory” I get this as response: { name: todo list, arguments: { todos: [ { content: Create a file named TEST.TXT, priority: low, status: pending } ] } } I was following this tutorial https://www.youtube.com/watch?v=RIvM-8Wg640&t= this is the opencode.json { “$schema”: “https://opencode.ai/config.json”, “provider”: { “ollama”: { “models”: { “qwen2.5-coder:7b”: { “name”: “qwen2.5-coder:7b” } }, “name”: “Ollama (local)”, “npm”: “@ai-sdk/openai-compatible”, “options”: { “baseURL”: “http://localhost:11434/v1” } } } } There is anything i can do to fix it? someone suggest to use lmstudio but this really work? anyone tested it? Claudecode will fix it? submitted by /u/DiscoverFolle

Originally posted by u/DiscoverFolle on r/ClaudeCode