• Aurornis 10 hours ago

    > Steve Yegge said he feels "sorry for people" who merely "use Cursor, ask it questions sometimes, review its code really carefully, and then check it in."

    Steve Yegge is building a multi-agent orchestration system. This is him trying to FOMO listeners into using his project.

    From what I've observed, the people trying to use herds of agents to work on different things at the same time are just using tokens as fast as possible because they think more tokens means more progress. As you scale up the sub-agents you spend so much time managing the herd and trying to backtrack when things go wrong that you would have been better off handling it serially with yourself in the loop.

    If you don't have someone else paying the bill for unlimited token usage it's going to be a very expensive experiment.

  • politelemon 10 hours ago

    Having gone through his interview just now, his advice and experience seems centered around Vibe coding new applications and not really reflective of the reality of the industry.

    > But I feel sorry for people who are good engineers – or who used to be – and they use Cursor, ask it questions sometimes, review its code really carefully, and then check it in. And I’m like: ‘dude, you’re going to get fired [because you are not keeping up with modern tools] and you’re one of the best engineers I know!’”

    I would certainly take a careful person over the likes of yegge who seems to be neither pragmatic, nor an engineer.

    • linkregister 9 hours ago

      Yegge became famous from his blog recounting his hiring as a software engineer at Google in the early 2010s. He has been an engineer for a long time.

      However, the implication that someone failing to use an experimental technology is falling behind is hyperbole.

      • enraged_camel 8 hours ago

        >> I would certainly take a careful person over the likes of yegge who seems to be neither pragmatic, nor an engineer.

        What utter nonsense. Yegge has been a programmer for longer than some people on this board have been alive, has worked on a lot of interesting and massively challenging projects and generously shared what he has learned with the community. Questioning his engineering chops is both laughable and absurd.

        • swordsith 5 hours ago

          the buck on engineer status in my opinion stops when someone becomes a crypto scammer.

      • avaer 9 hours ago

        I think people who run 15 agents to write a piece of software could probably use 1 or 2 and a better multi-page prompt and have the same results for a fraction of the cost.

        Especially with the latest models which pack quite a long and meaningful horizon into a single session, if you prompt diligently for what exactly you want it to do. Modern agentic coding spins up its own sub-agents when it makes sense to parallelize.

        It's just not as sexy as typing a sentence and letting your AI bill go BRR (and then talking about it).

        I'd like to see some actual results with a meaningful benchmark of software output that shows that agent orchestrators accomplish any meaningful improvement in the state of the art of software engineering, other than spending more tokens.

        Maybe it's time to dredge up the Mythical Man-Month?

        • esperent 8 hours ago

          Steve is basically an Instagram influencer for coders.

          He'll say whatever he can to stay in the spotlight, try to make you feel bad, that you're doing things wrong, that he invented things like agent orchestration when in fact he's just a loudmouth.

          Ignore him and his stupid gastown and get on with your life.

          • jolux 10 hours ago

            No point. Claude Code with skills and subagents is plenty. If they would stop breaking it constantly it would be fine.

            The bottleneck has not been how quickly you can generate reasonable code for a good while now. It’s how quickly you can integrate and deploy it and how much operational toil it causes. On any team > 1, that’s going to rely on getting a lot of people to work together effectively too, and it turns out that’s a completely different problem with different solutions.

            • fooster 10 hours ago

              What if you could remove that toil.

            • dolebirchwood 9 hours ago

              I don't know what kind of work he's doing that doesn't require actually reading the code to ensure it's appropriately maintainable, but more power to him. I actually like knowing what the hell my code is doing and that it conforms to my standards before committing it. I'll accept his condolences.

              • wasmainiac 9 hours ago

                Same, seems completely irresponsible.

                • utopiah 9 hours ago

                  We don't have time for safety, or security, or accuracy, or even understandability anymore. We need to move fast! /s

                • d4rkp4ttern 3 hours ago

                  I think there’s a level beyond 8: not reviewing AI-generated code.

                  There’s a lot of discussion about whether to let AI write most of your code (which at least in some circles is largely settled by now), but when I see hype-posts about “AI is writing almost all of our code”, the top question I’m curious about is, how much of the AI-written code are they reviewing ?

                  • 0xbadcafebee 7 hours ago

                    No I'm not, but not because I don't want to. To safely use an AI agent, it needs a ton of safety guardrails that (afaict) are difficult to set up. A lot of the safety guardrails we need don't even exist yet.

                    I'm working on all that currently. Trying to set up local systems to do practical and secure orchestrated AI work, without over-reliance on proprietary systems and platforms. Turns out it's a buttload of work. Yegge's own project (Gas Town) is a real world attempt to build just the agent part, and still many more parts are needed. It's so complicated, I don't think any open source solution is going to become dominant, because there's too much to integrate. The company that perfects this is going to be the next GitHub and Heroku rolled into one.

                    I get why people question all this. It's a completely different way of working that flies in the face of every best practice and common-sense lesson you learn as a software developer. But once you wrap your head around it, it makes total sense. You don't need to read code to know a system works and is reliable. You don't need to manually inspect the quality of things if there's other ways to establish trust. Work gets done a lot faster with automation, ironically with fewer errors. You can use cutting-edge technology to improve safety and performance, and ship faster.

                    These aren't crazy hypothetical ideals - what I just described is modern auto manufacturing. If it's safe enough for a car, it's safe enough for a web app.

                    • hrishikesh-s 8 hours ago

                      Yes, I'm using an agent orchestrator to write code. In fact, a couple of days before Anthropic introduced agent teams, I built a custom tool for myself inside emacs: https://github.com/hrishikeshs/magnus

                      I basically cycle through prompts and approve/deny/guide agents while looking at the buffer and thinking traces as text scrolls through. It has changed my life :)

                      • lubujackson 9 hours ago

                        I think Yegge needs to keep up with the tech a bit more. Cursor has gotten quite powerful - it's plan mode now seems about on par with Claude Code, producing Mermaid charts and detailed multi-phase plans that pretty much just work. I also noticed their debug mode will now come up with several thesises (thesi?), create some sort of debugging harness and logging system, test each thesis, tear down the debugging logic and present a solution. I have no idea when that happened, but it helped solve a tricky frontend race condition for me a day or two ago.

                        I still like Claude, but man does it suck down tokens.

                        • tbrownaw 9 hours ago

                          Sometimes I tell the AI to change something, sometimes I just do it myself. Sometimes I start to do it and then the magic tab-complete guesses well enough that I can just tab through the rest of it.

                          Sometimes the magic tab-complete insists on something silly and repeatedly gets in the way.

                          Sometimes I tell the AI to do something, and then have to back out the whole thing and do it right myself. Sometimes it's only a little wrong, and I can accept the result and then tweak it a bit. Sometimes it's a little wrong in a way that's easy to tell it to fix.

                          • mlaretallack 9 hours ago

                            Not the best way to do it, but I use xfce, multiple workspaces, each with there own version of AWS Kiro, and each kiro has its own project I am working on. This allows me to "switch context" easier between each project to check how the agents are getting on. Kiro also notifies me when an agent wants somthing. Usually I keep it to about 4 projects at a time, just to keep the context switching down.

                            • eshaham78 7 hours ago

                              The token cost is real, but I think the core problem isn't orchestration per se - it's context management. When you run multiple agents, each one needs a coherent view of the codebase. The real bottleneck isn't the number of agents, it's how you slice ownership.

                              The best pattern I've seen: give each agent clear module boundaries (not file-level, but architectural). Think micro-services even in a monolith. Then each agent owns its domain end-to-end.

                              Also: dry-run everything. Let agents propose changes, then review before applying. It's like having a senior dev who says 'wait, let me think about this' before every commit. Works surprisingly well.

                              • johnfn 9 hours ago

                                I am unfortunately in level 8. God help me. But honestly building an agent orchestrator is a really fun problem. It's like building an IDE and then using that IDE to build itself. Or building a programming language and then coding in that language! But with an entirely new host of different and interesting problems.

                                • tiku 9 hours ago

                                  When orchestrating you need to have a damn good plan / requirements. And then I'm typing or thinking a lot beforehand. And at the end it's never 100% what you want.

                                  That is why I'm going back to per function/small scope ai questions.

                                  • _sinelaw_ 9 hours ago

                                    I did when just starting on a new project, it was working well when I had many new components to implement. But as the project matured and stabilized every new feature is cross-cutting and it's impossible to parallelize the work without running into conflicts (design conflicts, where two agents would add similar overlapping mechanisms, and also the usual code conflicts, touching the same files). Also, with project maturity I'm much more concerned about keeping it stable and correct, which is hard to do with parallel agents running amok.

                                    • johnfn 9 hours ago

                                      I find if you just ask the agents to resolve the conflicts they do a pretty great job. It's even better if you can feed them all the context while resolving the conflict.

                                      • _sinelaw_ 8 hours ago

                                        The harder problem is conflicting design choices, or duplicating similar infra. It means I need to be much more involved in steering individual agents and planning up front (waterfall style), which limits the parallelism further

                                    • andy_ppp 10 hours ago

                                      I think people should figure out what works for them rather than letting people on the internet gate-keep what is good. Everything is about personal choices and refining your own taste. I would not be happy being unable to understand everything deeply so having a million agents all doing stuff would just cause me a load of stress even if I could churn stuff out more quickly.

                                      • Glyptodon 10 hours ago

                                        The stumbling block we have is spinning up separate environments for every agent so they have isolation for their branches. I think this is solveable, but we aren't trying to solve it ourselves. In practice it means we aren't doing a lot of agent supervision.

                                        • SkyPuncher 8 hours ago

                                          Git worktrees essentially solve this. It essentially copies your repo to a new folder

                                          • tbrownaw 9 hours ago

                                            That sounds like an excellent match for containers.

                                          • wasmainiac 9 hours ago

                                            > At my work, this wouldn't fly

                                            How does one even review the code from multiple agents. The quality imo is still to low to just let run on its own.

                                            • dsifry 9 hours ago

                                              I have been helping people get onboarded with Claude Code and the orchestrator I wrote called Metaswarm [1) and the response has been way beyond my expectations.

                                              But don't take my word for it, try it out for yourself, it is MIT licensed, and you can create new projects with it or add it to an existing project.

                                              [1] https://github.com/dsifry/metaswarm

                                              • woutr_be 8 hours ago

                                                I would love to experience this, but I'm only at the level were I occasionally open ChatGPT or Claude, asked it a question, and then get frustrated because it can't even give me a straight answer, or makes incorrect assumptions.

                                                I can't even imagine having multiple agents write code that somehow works.

                                                • freakynit 8 hours ago

                                                  Same here. I've tried agent integrations in VS Code and also have agentic CLIs installed (Claude Code, Gemini cli). But honestly, I still find it more reliable, and often faster, to just ask focused questions, let it generate a method or two, review the output, and copy-paste it into my project. Rinse and repeat. Kind of like how we used to do in the good old days an year back.

                                                  For now at least, the full agent workflows feel less efficient and more headache-inducing than being helpful.

                                                  And agentic swarms: that's marketing bs.. at least for now.

                                                • slopinthebag 10 hours ago

                                                  No, I don't even use agents to generate code most of the time. I mainly use the inline assistant to modify or fill out blocks of code, and agents sometimes for refactors, asking questions, search, debugging, generating documentation etc.

                                                  I feel bad for Yegge.

                                                  • lmeyerov 9 hours ago

                                                    I stopped manually writing code 6-9mo ago, and am generating high-quality code on the dimensions we care about like GPU perf benchmarks, internal & industry conformance standards test suites, evals benchmarks, lint/type checkers, etc. It's not perfect code - there are clear AI slop tell tales that review cycles still let linger - but it's doing more ambitious things than we'd do on most dimensions like capability, quality, and volume. We're solving years-old GPU bugs that we had given up on as mere mortals.

                                                    And yes, we build our own orchestrator tech, both as our product (not vibes coding but vibes investigating), and more relevant here, our internal tooling. For example, otel & evals increasingly drive our AI coding loops rather than people. Codex and claude code are great agentic coding harnesses, so our 'custom orchestration' work is more about more intelligently using them in richer pipelines, like the above eval-driven loop. They've been pretty steadily adding features like parallel subagents that work in teams, and hookable enough to do most tricks, that I don't feel the need to use others. We're busy enough adapting on our own!

                                                    • pdyc 9 hours ago

                                                      i tried but it didn't worked for me. Now i use agents as editors for fully formed solution so slightly better editor than typing.

                                                      • nprateem 9 hours ago

                                                        There's important stuff to review, 10-20% (eg overall architecture, use of existing utilities/patterns), and there's the specifics of the client code.

                                                        My reviews pick out the first and gloss over the latter. They take a few minutes. So I run multiple distinct tasks across agents in antigravity, so there's less chance of conflict. This is on 500k+ line codebase. I'm amazed by the complexity of changes it can handle.

                                                        But I agree with his take. Old fashioned programming is dead. Now I do the work of a team of 3 or 4 people each day: AI speed but also no meetings, no discussions, no friction.

                                                        • gimmeslop 8 hours ago

                                                          I couldn’t ship 1.5 million lines of code daily without orchestrated agents.

                                                          • dboreham 10 hours ago

                                                            People lie. Let's see a video of them doing this, or logs of the sessions, and the generated code, so we can judge for ourselves.

                                                            • petesergeant 10 hours ago

                                                              "Claude writes, Codex reviews" has shown huge promise as a pattern for me, so I wrote a Dockerfile and some instructions on how to make that happen for agents, and ended up with https://github.com/pjlsergeant/moarcode

                                                              I am spending most of my day in this harness. It has rough edges for sure, but it means I trust the code coming out much more than I did just Claude.

                                                              • joshuaisaact 9 hours ago

                                                                I don't think you need two separate models for this - I get similarly good results re-prompting with Claude. Well, not re-prompting, I just have a skill that wipes the context then gets Claude to review the current PR and make improvements before I review it.

                                                                • neumann 9 hours ago

                                                                  I tried to opposite because claude was not coding as well as codex some additional modules for my codebase and codex could. Then I tried to get claude to read and critique and it got so many fundamentals wrong I was wondering if I am using the wrong model.

                                                                • whattheheckheck 10 hours ago

                                                                  Vscode agent mode is pretty slick

                                                                  • bitwize 10 hours ago

                                                                    We're not there yet, but it's going to happen. Given the nature of the application I'm working on, I wouldn't be surprised if the entire headcount of the engineering department were reduced to around five or so in a year or two.