I've been trying to work towards "Level 5" AI-assisted development for a year. I'm not quite there, but this is the flow that's working for me.

Posted by rgeade@reddit | ExperiencedDevs | View on Reddit | 29 comments

To be clear this is not "vibe coding", but a pragmatic system for managing coding agents at scale. It's the structured process I've been working on since I first started using claude code nearly a year ago.

The thing that made the biggest difference wasn't the AI model I was using (although I've primarily used Opus 4.6 for dev with Codex on reviews), it was the specs. Specifically writing a real per-task spec and then running it through a separate agent that's prompted to poke holes in it, NOT validate it. Adversarial spec review sounds dramatic but it's just a second agent trying to break the plan before you build anything.

The process is ever changing, but the current flow I landed on:

strategy → phases → sprints → tasks → per-task spec → adversarial spec review → build → adversarial code review → staging → prod

The adversarial spec review was actually the last thing I added. I was doing adversarial code reviews when Codex came out and pitted a Codex agent against Opus. This caught far more issues earlier in the process, and has substantially reduced by review cycle count during code review.

Is anyone else landing on something similar? I'm by no means at "dark factory" yet, but I spend far more time speccing than I have in my entire career.

I also wrote a longer form of the above to share the details of my flow, but I'm still looking at ways to improve the process. Ultimately, it becomes a function of token cost; and at some point the review cycles can double or triple that.