What is best code editor for local LLM deployment (LM Studio, llama.cpp) as of May 2026?
Posted by jingtianli@reddit | LocalLLaMA | View on Reddit | 25 comments
Hello folks
What is best code editor for local LLM deployment (LM Studio, llama.cpp)?
I wish to test my LM studio + Qwen 3.6 27B and Gemma 4 31B with a legit local code editor. I want it to have the same user experience like cursor (sub agent support, auto find linter error, or auto complete and so on). But i could not find any yet.
Vscode Continue plugin works but not fully works, I dont know the problem is occurs with LM studio or what, large context of code often got cutted off for some reason.
Is there such tool existed? I heard Void was popular once but they paused that project unfortunately. Is there free opensourced one or paid tool that allow user to have the same experience as cursor or codex but for local LLM deployment?
NO Claude Code please no CLI interface, sometimes I want to chat with LLM with screenshoted image, so native windows support is must not a commandline
Thank you in advance!
Alan_Silva_TI@reddit
RooCode was the best option, but they’ve announced that they’re abandoning the project (check their subreddit).
It originally started as a fork of Cline, and they even recommended that users switch back to it. Cline works quite well and has some useful features.
By the way, both are VS Code extensions, so they should be right up your alley.
jingtianli@reddit (OP)
Thank you kindly!
OneSlash137@reddit
These models are ass for coding or development.
It’s probably getting caught in tool call loops because it constantly drops prompt prefill forcing the entire conversation to preprocess from the beginning including all the agents tool parameters etc. then while all that is chugging the tools just keep requiring the same thing over and over until they time out.
If you could manage to get that solved and run it full speed, it won’t give you anything reliable, well built, best practice, scalable, production ready.
LLMs for professional home development is not here yet.
false79@reddit
^-- I'd take that advice with a grain of salt. As a professional developer, there are gains to be made by offloading repetitive work to an LLM if you provide it with a sufficient amount of context, smaller achievable goals, let its reasoning connect the dots faster than a human would, free up time to do other things.
I find a common pattern among those who are frustrated with local LLMs for coding is not really knowing how to activate the parameters needed to produce the output they need. Cloud models do a lot of the heavy lifting but there are several here who have figured out you can get just as much done if you understand the limits of the model and how it does not read your mind with a zero shot prompt.
OneSlash137@reddit
Wrong.
false79@reddit
I dunno if you look over post history but there are fair amount people down voting your lack of expertise in this area.
I am sure if you knew what you were talking about, it would be very different.
__JockY__@reddit
Sounds like someone tried coding on an 8GB card with an IQ2_XXS on Cherry Studio and wondered why it doesn’t work.
OneSlash137@reddit
Try again.
stormy1one@reddit
Sounds like you have a configuration issue. Qwen3.6-27B is extremely capable when configured correctly. It’s not going to zero shot anything complex, but will absolutely be able to produce reliable code for production use. Tool calling was one of the main things they addressed in 3.6. Performance is excellent in vLLM v19.1 — all works having at least two instances - one for orchestration managing the other for coding.
OneSlash137@reddit
Wrong. I’m an enterprise architect. You don’t know best practice or production code.. you don’t know enough to know it’s bad. Because LLMs won’t tell you they are bad at something, they’ll just cinfidently say the code works.
And maybe it does, but most likely it’s doing really stupid things like making 5 separate database queries when only 1 is necessary.
Those are the types of things that fly under the radar until 10,000 people are using your app at once and the prod database explodes.
So, yeah. It isn’t a config issue.
Levipl@reddit
This because things are just too fresh today. It’ll get there but you have to work within the limitations for now.
Ok_Hope_4007@reddit
I use the jetbrains IDEs (Pycharm/Rider/We storm) and the cline plugin. Some people use Kilo Code as we'll.
bnightstars@reddit
Switch to VSCode-Insiders and use Copilot :?
dwrz@reddit
This may not be what you want to hear, but I use these models with Emacs and
gpteland they work great.bobby-chan@reddit
And if you want agentic, `gptel-agent`.
ea_man@reddit
If you want "work out of the box" exp with QWEN you gotta use Qwencode, learn to like it.
StardockEngineer@reddit
Have you tried good ol' Github Copilot?
super_g_sharp@reddit
Roo code all day long.
I've tried continue.dev , goose
I switch to it today and it's JUST like cursor. I think it had 1 bad tool call all day. And it indexes your code. Context management was fantastic
samandiriel@reddit
Roo has switched to a saas only model and won't be supporting their plugin past may 15th, unfortunately.
I passed on roo as a result and am trying out kilo code instead
jingtianli@reddit (OP)
Yes I just tried this is awesome!
miklosp@reddit
https://docs.roocode.com/sunset
super_g_sharp@reddit
Also look for the club-3090 GitHub. He has recipies for vllm that are amazing.
michaelsoft__binbows@reddit
You haven't learned how to send screenshots into CLI harnesses yet? Are you serious?
different_tom@reddit
Who's a grumpy little fella?
Usual-Carrot6352@reddit
Hermes