Complete beginner to Agentic coding, is Qwen3.6-27B + pi.dev the right starting point or should I be looking elsewhere?
Posted by SarcasticBaka@reddit | LocalLLaMA | View on Reddit | 33 comments
Hello fellow members of this lovely community,
Let me start by saying that I’m about as far from a professional developer as it gets. I’m a hobbyist whose entire coding experience consists of building various Python/VBA tools and simple JavaScript web apps mostly using VS Code. So far, my approach to using AI for coding has basically been copying and pasting sections of my code into ChatGPT and asking for changes or additions as needed.
Since small local models seem to have improved quite a bit for coding, I decided to dip my toes into this whole “agentic coding” space I’ve been hearing about. Hardware-wise, I have a measly 2080 Ti with 22 GB of VRAM, in which I managed to fit Unsloth’s Qwen3.6-27B-UD-Q4_K_XL with 128k context at q8_0 KV using the parameters below, while getting around 20–22 tok/s.
"qwen3.6-27b-coder":
cmd: |
${llama_server}
--host 0.0.0.0 --port ${PORT} -ngl 999 -fa on --jinja --no-mmap -cram 2048 --no-warmup -np 1
--model ${host_model_dir}/Qwen3.6-27B/Qwen3.6-27B-UD-Q4_K_XL.gguf
--mmproj ${host_model_dir}/Qwen3.6-27B/mmproj-F16-Qwen3.6-27B.gguf
--no-mmproj-offload
--spec-type ngram-mod
--spec-ngram-size-n 24
--draft-min 12
--draft-max 48
--ctx-size 131072
--cache-type-k q8_0
--cache-type-v q8_0
--temp 0.6
--presence-penalty 0.0
--repeat-penalty 1.0
--min-p 0.0
--top-k 20
--top-p 0.95
--fit off
--reasoning on
--reasoning-budget -1
--chat-template-kwargs '{"enable_thinking":true}'
--chat-template-kwargs '{"preserve_thinking":true}'
While searching for a coding agent that fits my setup, I saw PI being recommended quite a bit for being fast and lightweight. I installed it, hooked it up with Qwen3.6, and so far so good.
The issue I’m running into is that PI feels like a very barebones “DIY” type of agent. I’m sure that’s great if you know what you’re doing, but as a complete beginner to CLI-based coding agents, I’m honestly a bit lost on how to use it effectively or what a good workflow even looks like.
So I have a few questions for you more knowledgeable folks:
-
Should I stick with PI and just go through the documentation until I’m more comfortable? Or would it make more sense to switch to something more “batteries included” like Opencode, Qwencode, etc.? Alternatively, should I just stick with VS Code and use an extension that connects to a local LLM?
-
Regarding my model choice: is 128k context and ~20 tok/s actually usable for coding, or would I be better off switching to a 35B MoE model with CPU offload for higher speed and/or context?
-
Any recommended optimizations for my llama-server parameters?
-
Lastly, I’m running into an issue with PI where, even though reasoning is enabled on the llama-server side, the model doesn’t seem to “think” based on my initial tests. The thinking_level setting in PI is also set to off, and I can’t seem to change it.
Thanks in advance for any help or guidance.
StardockEngineer@reddit
Brand new and going to use pi? Ooof. Dangerous.
Solexe@reddit
would you rather recommend opencode? or something else?
SarcasticBaka@reddit (OP)
That's what I'm gathering from all the replies so far haha. I'm glad I asked.
ai_guy_nerd@reddit
Qwen 3.6 27B is a solid choice for a start, especially with 22GB VRAM. The model's ability to follow complex instructions makes it great for agentic workflows. Using pi.dev helps bridge the gap between the model and the actual execution.
Depending on the goals, it might be worth looking into frameworks like CrewAI or AutoGen if the focus is on multi-agent coordination. For a more streamlined experience with a local setup, OpenClaw is another interesting way to handle autonomous pipelines. Stick with the current setup for a bit to understand the basics of tool use and loops before jumping into heavier frameworks.
arthor@reddit
why is every 2nd post suddenly pumping pi.dev on this sub
SarcasticBaka@reddit (OP)
Where in my post am I "pumping" pi.dev?
En-tro-py@reddit
Begging the question in your title...
pi.dev definitely seems to have a lot of 'hype' posts pushing it lately, so it's a bit of a red flag that a 'complete beginner' already is going with it...
SarcasticBaka@reddit (OP)
I guess this is a chicken and egg scenario because seeing all the recent posts in this sub is what pushed me to try pi in the first place and thus post about it.
En-tro-py@reddit
No guardrails or sandbox, doesn't seem like the thing a 'complete beginner' should be getting recommended or recommending.
Hence why it's seems like hype driven not actual organic growth.
edeltoaster@reddit
I'm no beginner and on my very first testcall it tried to self-improve by implementing a webfetch using Python and the BeautifulSoup package... which it started to install by itself without asking. I am hesitating if I should install pi again with guardrails added and use it again - becaus it really could be great for local models - or if I just leave it be for the next nice thing.
SarcasticBaka@reddit (OP)
Fair enough. I definitely see your point.
vr_fanboy@reddit
because is great, i keep throwing it tasks and it nails them in the same way as sonnet would. "Hey my langfuse does not work in dev-ml, you have mcp access go fix it". 30 seconds later i go to the browser and langfuse is running fine, error report generated to be reviewed later. (by another pi instance) Try to do this with Claude Code using a local LLM, after first prompt you have 30% token usage reached (128k context available) then go check the trazability, you have TONS of bloat inside the CC prompt request, it needs to keep tabs with the corpo guard rails. Is great to see PI context growing slow and steady while it works, it can running for 15 mins easy without losing itself.
En-tro-py@reddit
Username is on point... Fanboy can't stand even the lightest criticism...
What part of that is a 'complete beginner' doing?
chimph@reddit
a simple question about new tech is begging?
chimph@reddit
Seem people don’t understand the flywheel effect of shiny new things. It’s not suspicious.
sine120@reddit
A couple youtubers found it recently and people are liking it. It's a better experience than opencode if you have slow PP.
Interesting-Print366@reddit
Depends on what machine you are using. If you have enough vram and using gpus like rtx series use opencode it would be much better for you. But if you are using SFF workstation with unified ram. Pi would be better but 27b would be still very much slow
Flylink2@reddit
I am a newbe like you in all of this ! I have been using Vscodium with Cline extension (and Ollama behind it but I think you are better with llama.cpp, I am thinking about switching to that as well...
I tried to play with that and really enjoyed ! You can Plan & Act with different model. Since you plan with the tool, you see how it thinks and you allow him to do what you want.
I spent hrs with Gemini Pro trying to find the best local/opensource way to do agent locally and it's the best I found... Would be curious to have your feedback or find is there is anything better around :)
Old-Sherbert-4495@reddit
i would suggest opencode, the desktop version. you would switch between that and vscode.
sine120@reddit
OpenCode is worth trying but probably not the best fit. It's a very good agent, but the system prompt is huge. If your PP speeds are sub 1k, you'll be waiting tens of seconds for your response to even start. Pi's prompt is way smaller as it has way fewer tools. I've used both and for his system I'd lean Pi
Ok-Internal9317@reddit
Same goes for Hermes Agent, inital prompt is too insane and confusing as well maybe? I saw sometimes qwen brute force write code through terminal command although there is also write_code api skill.
mr_Owner@reddit
Pi dev seems too loose and risky, haven't tried but if your local llm sucks at instructions following then i would still go for kilocode or cline as vscode extension.
ibishitl@reddit
The thing with pi.dev is it doens't use too many tokens for the system prompt wich make it lightweight for local models, I'm liking my experience with it
Just ask it to build things for itself, something like "build a skill/extension to ask for permissions before making any dangerous change" idk, I don't do that but it is recommended
sasquatch3277@reddit
20tps is gonna be ass in opencode
opencode will have you running out of the box
Ask yourself if you really want to build the tool to build your project (pi)
use the moe and get double speed and context instead of waiting 10 minutes for PP and CoT
If accuracy is bad and you want to optimize things go to dense + pi
hurdurdur7@reddit
Is this post written by an AI bot?
SarcasticBaka@reddit (OP)
No?
hurdurdur7@reddit
Because it sure sounds like it is. at one moment you are hesitant, as if appearing not to know what you are doing, at next moment you are dishing out fine grained params that make a lot of sense ... and pretty much flawless grammar and structuring.
SarcasticBaka@reddit (OP)
Okay well I can somewhat understand where you're coming from. I'm not at all familiar with agentic coding as I said in my OP but I am very familiar with llama.cpp and its various parameters.
ea_man@reddit
That depends on what you wanna do and with how much hand holding:
Edit a file in vscodium: extension
If you just wanna modify a few files in an existing proj: Aider
Have the sf do all planning and implementing and testing: Qwencode / Opencode
Same but build your own from little: Pi code
Graphic editor / reports / eyecandy: antigravity
gtrak@reddit
You're not clear about your goals so I suggest just tinkering for a while and trying different things.
SarcasticBaka@reddit (OP)
My main goal is to familiarize myself with agentic coding as a whole.
MoodyPurples@reddit
In that case it would probably make sense to try a handful of different harnesses to see what you like. If you find something that feels like a perfect fit, great! Otherwise, you’ll know what you want to add to Pi
gtrak@reddit
To make an analogy to graph search, you want a strategy like BFS (breadth-first-search), not DFS (depth-first-search). Map out the entire space of things you can try, then just dabble and try something new for a while, don't go deep until you have more specific goals and the lay of the land.