Gas town and the like all basically sound to me like, "AI is amazing! Ok, actually it isn't very good, but maybe we can just use more AI with our AI and then it'll be good!"
And I'm not surprised at all to learn that this path took us to a "Maintenance Manager Checker Agent." I wonder what he'll call the inevitable Maintenance Manager Checker Agent Checker Agent?
Maybe I've been in this game too long, but I've encountered managers that think like this before. "We don't need expensive, brilliant, developers, we just need good processes for the cheap inexperienced developers to follow." I think what keeps this idea alive is that it sort of works for simple CRUD apps and other essentially "solved" problems. At least until the app needs to become more than just a simple CRUD app
For years we had people trying to make voice agents, like Iron Man's Jarvis, a thing. You had people super bought into the idea that if you could talk to your computer and say "Jarvis, book me a flight from New York to Hawaii" and it would just do it just like the movies, that was the future, that was sci-fi, it was awesome.
But it turns out that voice sucks as a user interface. The only time people use voice controls is when they can't use other controls, i.e. while driving. Nobody is voluntarily booking a flight with their Alexa. There's a reason every society on the planet shifted from primarily phone calls to texting once the technology was available!
It's similar with vibe coding. People like Yegge are extremely bought into the idea of being a hyperpowered coder, sitting in a dimly lit basement in front of 8 computer screens, commanding an army of agents with English, sipping coffee between barking out orders. "Agent 1, refactor that method to be more efficient. Agent 5, tighten up the graphics on level 3!"
Whether or not it's effective or better than regular software development is secondary, if it's a concern at all. The purpose is the process. It's the future. It's sci-fi. It's awesome.
AI is an incredible tool and we're still discovering the right way to use it, but boy, "Gas Town" is not it.
This is confusing. Voice is not a UI, it's an input device. When I call my bank and have to input some numbers into the automated system, I prefer to say them than to type them. The phone menu system is the UI, fingers or voice are two different input modes for the same UI.
The problem with alexa booking tickets is not the use of my voice but that there are a lot of decisions (comparison shopping, seat selection etc) to be made. Alexa can't read my mind to make the trade-offs I would make, although it could ask me 10 zillion questions. The difference between voice/ears and fingers/eyes is the bandwidth of information transfer, but also the availability of the tools. Hands and eyes may be busy as in your car example, but they are also busy if I'm carrying a toddler around the house or can't be bothered to reach into my pocket or am already using my phone for something else (game, video etc). So voice is a good option for many tasks. And LLMs/agents do have the potential to make more tasks (simple ones, not booking tickets) accessible to voice since "AI as UI" is where it holds the most potential IMHO. And that's great because we need all the help we can get to avoid taking our phones out of our pockets and getting sucked into random tangents like HN comment threads just bc we wanted to check the weather
"Agent 1, refactor that method to be more efficient. Agent 5, tighten up the graphics on level 3!"
I'm not sure its even that, his description of his role in this is:
"You are a Product Manager, and Gas Town is an Idea Compiler. You just make up features, design them, file the implementation plans, and then sling the work around to your polecats and crew. Opus 4.5 can handle any reasonably sized task, so your job is to make tasks for it. That’s it."
And he says he isn't reviewing the code, he lets agents review each others code from look of it. I am interested to see the specs/feature definitions he's giving them, that seems to be one interesting part of his flow.
Yeah maybe the refactoring was a bad example because it implies looking at the code. It's more like "Agent 1, change the color of this widget. Agent 9, add a red dot whenever there's a new message. Agent 67, send out a marketing email blast advertising our new feature."
Assuming both agents are using the same model, what could the reviewer agent add of value to the agent writing the code? It feels like "thinking mode" but with extra steps, and more chance of getting them stuck in a loop trying to overcorrect some small inane detail.
"I implemented a formula for Jeffrey Emanuel’s “Rule of Five”, which is the observation that if you make an LLM review something five times, with different focus areas each time though, it generates superior outcomes and artifacts. So you can take any workflow, cook it with the Rule of Five, and it will make each step get reviewed 4 times (the implementation counts as the first review)."
And I guess more generally, there is a level of non-determinism in there anyway.
> Nobody is voluntarily booking a flight with their Alexa.
Rich people use voice because they have disposable income and they don't care if a flight is $800 or $4,000. They are likely buying business/first class anyways.
Tony Stark certainly doesn't care.
Elon Musk certainly uses voice to talk to his management team to book his flights.
The average person doesn't have the privilege of using voice because it doesn't have enough fuck-you-money to not care for prices.
As someone who's friends with executive assistants: rich people use executive assistants (humans) because they are busy and/or value their time more than money and don't want to bother with the details. None of them are using voice assistants.
> Tony Stark certainly doesn't care. Elon Musk certainly uses voice to talk to his management team to book his flights.
Delegating to a human isn't the same as using a voice assistant, this should be obvious, unless you believe that managers are doing all the real work and every IC is a brainless minion. Maybe far in the future when there's AGI, but certainly not today.
> The average person doesn't have the privilege of using voice because it doesn't have enough fuck-you-money to not care for prices.
You can order crap off Amazon for the same price as you would through the website with your Alexa right now, but Amazon themselves have admitted approximately 0% of people actually do this which is why the entire division ended up a minor disaster. It's just a shitty interface in the same way that booking a flight through voice is a shitty interface.
Rich people will literally just talk to their executive assistants and just ask what they want. They may use phone calls, voice mails, emails, and text. But you'd be crazy to argue that they never use just voice with their IRL assistants.
Your point is that voice is a terrible interface to get something done.
My point is that some people have the privilege to use voice to get something done.
Some companies are just trying to remove the human who is taking the voice command and replacing with AI.
You're completely missing the distinction between "using a human brain (by telling a person what to do)" [yes this includes via voice] and "using a machine via voice interface."
Still the same. "Hey look, I got these crappy developers (LLMs) to actually produce working code! This is a game-changer!" When the working code is a very small, limited thing.
I don't know, your talking about an incredibly talented engineer saying:
"In the past week, just prompting, and inspecting the code to provide guidance from time to time, in a few hours I did the following four tasks, in hours instead of weeks"
Its up to you to decide how to behave, but I can't see any reasons to completely dismiss this. It ends with good guidance what to do if you can't replicate though.
The fact that he's an extremely talented developer actually supports my overall understanding that AI producing code is way over hyped. Sure, a master of his field can get a boost out of it, after spending (unaccounted for) time learning how best to coax good code out of it. Neat?
Any sort of evidence! I see none! It's not really a new thing to have no evidence of productivity gains when it comes to software development tools. Some feel like vim is a huge productivity boost and some don't. Some believe rust is amazing, some hate it. It's really hard to measure these things.
Here’s a professional developer who built a product used across the planet daily saying they built a feature for said product, and then asked AI to build the same feature based on the design doc: the AI succeeded and did the same work in minutes: https://antirez.com/news/158
You could always test this yourself. Draw up a design doc for a problem, and implement it. Then ask an AI to do the same thing. Compare your time.
If you add enough agents, you basically recreate the structure of human teams again. And the lessons from mythical man month etc. start applying. Large companies/teams don't seem to produce higher quality software than small ones, in fact they usually seem to be worse.
At best the notion of "subagents" today seems to be a hack to work around context length limits.
This is a good point, but with ai it’s a little different, because both your process and ai are getting better. You build processes that can aspirationally support inferior AIs, while at the same time AIs themselves improve and meet you half way. This thought does not help my mental well being unfortunately.
I think in the end people will realise AI is not a silver bullet that will solve all problems and make all software engineers obsolete. Instead it will just become an extra tool in our toolbelt right alongside LSP, linters/fixers, unit test frameworks, well though out text editors and IDE's, etc.
When the bubble has burst in a few years, the managers will have moved on to the next fad.
It's definitely helpful for search and summarisation.
In terms of prototyping, I can see the benefits but they're negated by the absurd amount of work it takes to get the code into more maintainable form.
I guess you can just do really heavy review throughout, but then you lose a lot of the speed gains.
That being said, putting a textual interface over everything is super cool, and will definitely have large downstream impacts, but probably not enough to justify all the spending.
And I'm not surprised at all to learn that this path took us to a "Maintenance Manager Checker Agent." I wonder what he'll call the inevitable Maintenance Manager Checker Agent Checker Agent?
Maybe I've been in this game too long, but I've encountered managers that think like this before. "We don't need expensive, brilliant, developers, we just need good processes for the cheap inexperienced developers to follow." I think what keeps this idea alive is that it sort of works for simple CRUD apps and other essentially "solved" problems. At least until the app needs to become more than just a simple CRUD app