scuff3d
3 months ago
The core of the entire argument is that the $150/hour is based on a developers ability to physically write code, which is not true. Having something that can generate code reliabily (which these things can barely do even with an expert at the wheel) doesn't address any of the actual hard problems we deal with on a daily basis.
Plus running AI tools is going to get much more expensive. The current prices aren't sustainable long term and they don't have any viable path to reducing costs. If anything the cost of operations for the big company are going to get worse. They're in the "get 'em hooked" stage of the drug deal.
bigiain
3 months ago
> Having something that can generate code reliabily (which these things can barely do even with an expert at the wheel) doesn't address any of the actual hard problems we deal with on a daily basis.
Not understanding that is something I've been seeing management repeatedly doing for decades.
This article reads like all the things I discovered and the mistakes the company I worked for made learning how to outsource software development back in the late 90s and early 2000s. The only difference is this is using AI to generate the code instead of lower paid developers from developing nations. And, just like software outsourcing as an industry created practices and working styles to maximise profit to outsourcing companies, anyone who builds their business relying on OpenAI/Anthropic/Google/Meta/whoever - is going to need to address the risk of their chosen AI tool vendor ramping up the costs of using the tools to extract all; the value of the apparent cost savings.
This bit matches exactly with my experience:
"The trouble comes in that most people don't know what code needs to be created to solve their problem, for any but the most trivial problems. Who does know what code would be needed to solve complex problems? Currently that's only known by software developers, development managers and product managers, three job classifications that are going to be merging rapidly."
We found that assuming the people you employ as "developers" weren't actually also doing the dev management and product management roles was wrong. At least for our business where there were 6 or 8 devs who all understood the business goals and existing codebase and technology. When we eventually;y got successful outsourced development working was after we realised that writing code from lists of tasks/requirements was way less than 50% of what our in-house development team had been doing for years. We ended up saving a lot of money on that 30 or 40% of the work, but 60 or 70% of the higher level _understanding the business and tech stack_ work still needed to be done by people who understood the whole business and had a vested interest in the business succeeding.
Terr_
3 months ago
> assuming the people you employ as "developers" weren't actually also doing the dev management and product management roles was wrong.
Also in the mix: Stuff involving B2B customers and integrating systems, where "developer" blurs a bit with sales-engineer or consultant, ex:
* Be on a call to ask significant questions (grounded in reading the code) to determine what the customer's real problem is.
* Help craft diplomatic but accurate explanations of what's going wrong.
* Explain what bugs or changes you can do for them versus which parts are fully on their end, sometimes with "here's what I think your engineers should consider doing" advice.
Not always fun, but sometimes enlightened self-interest means I'd rather spend an 1 hour being "one of our developers" in a customer-meeting, as opposed to 6 hours discovering everything is actually working as-intended and the customer just misunderstood what feature they were using.
simonw
3 months ago
Completely agree on your first point: software development is so much more than writing code. LLMs are a threat to programmers for whom the job is 8 hours a day of writing code to detailed specifications provided by other people. I can't remember any point in my own career where I worked with people who got to do that.
There's a great example of that in the linked post itself:
> Let's build a property-based testing suite. It should create Java classes at random using the entire range of available Java features. These random classes should be checked to see whether they produce valid parse trees, satisfying a variety of invariants.
Knowing what that means is worth $150/hour even if you don't type a single line of code to implement it yourself!
And to be fair, the author makes that point themselves later on:
> Agentic AI means that anything you know to code can be coded very rapidly. Read that sentence carefully. If you know just what code needs to be created to solve an issue you want, the angels will grant you that code at the cost of a prompt or two. The trouble comes in that most people don't know what code needs to be created to solve their problem, for any but the most trivial problems.
On your second point: I wouldn't recommend betting against costs continuing to fall. The cost reduction trend has been reliable over the past three years.
In 2022 the best available models was GPT-3 text-davinci-003 at $60/million input tokens.
GPT-5 today is $1.25/million input tokens - 48x cheaper for a massively more capable model.
... and we already know it can be even cheaper. Kimi K2 came out two weeks ago benchmarking close to (possibly even above) GPT-5 and can be run at an even lower cost.
I'm willing to bet there are still significantly more optimizations to be discovered, and prices will continue to drop - at least on a per-token basis.
We're beginning to find more expensive ways to use the models though. Coding Agents like Claude Code and Codex CLI can churn through tokens.
scuff3d
3 months ago
I get your point, but I don't think the pricing is long term viable. We're in the burn everything to the ground to earn market share phase. Once things start to stabilize and there is no more user growth, they'll start putting the screws to the users.
I said the same thing about Netflix in 2015 and Gamepass in 2020. It might have taken a while but eventually it happened. And they're gonna have to raise prices higher and faster at some point.
Nevermark
3 months ago
Netflix can't sell you 48 hours of video watching, no matter how many videos they make, or how good they get.
Gamepass, same thing.
Those are completely incomparable businesses.
jimbokun
3 months ago
Netflix prices went up a little bit but not very much.
handoflixue
3 months ago
Netflix was initially $7/mo in 2007. Currently it goes for $18/mo in 2025.
150% increase in 18 years is about triple the rate of inflation
I'd call that more than "a little bit", but I'd agree that if LLMs go the same way it probably doesn't change the equation much at all.
Plus they do still have an $8/mo ad-supported plan. After adjusting for inflation, that's actually cheaper than the original!
scuff3d
3 months ago
I think it's gonna be a lot worse with LLMs. Mainly because they're substantially under charging, I don't think the cost of operating is going to drop much, and the workflows are only going to get more token hungry.
The incentives here are also fucking atrocious. They aren't incentivised to make the model as good as possible. It's in their best interest to tune so it's good enough to not drive you off, but bad enough to push you to spend more.
brazukadev
3 months ago
> I think it's gonna be a lot worse with LLMs
It wont: if/when LLMs start to get too expensive, people will just migrate to open models, run it local, etc. I see no scenario where we are held hostage by the main providers.
simonw
3 months ago
I was a lot more worried about this when only OpenAI and Anthropic had truly great models - but now we also have Google and five different Chinese AI labs who are releasing open weight models that are in the same ballpark as the OpenAI and Anthropic ones. I think we'll be fine.
handoflixue
3 months ago
Very much agreed - right now someone might hold a few months of competitive lead, but open models catch up fast. Plus the lack of any real vendor lock-in means there's just not room for extortionate pricing.
bigiain
3 months ago
> In 2022 the best available models was GPT-3 text-davinci-003 at $60/million input tokens.
>GPT-5 today is $1.25/million input tokens - 48x cheaper for a massively more capable model.
Yes - but.
GPT-5 and all the other modern "reasoning models" and tools burn through way more tokens to answer the same prompts.
As you said:
> We're beginning to find more expensive ways to use the models though. Coding Agents like Claude Code and Codex CLI can churn through tokens.
Right now, it feels that "frontier models" costs to use are staying the same as they've been for the entire ~5 year history of the current LLM/AI industry. But older models these days are comparably effectively free.
I'm wondering when/if there'll be a asymptotic flattening, where new frontier models are insignificantly better that older ones, and running some model off Huggingface on a reasonably specced up Mac Mini or gaming PC will provide AI coding assistance at basically electricity and hardware depreciation prices?
simonw
3 months ago
That really is the most interesting question for me: when will it be possible to run a model that is good enough to drive Claude Code or Codex CLI on consumer hardware?
gpt-oss-120b fits on a $4000 NVIDIA Spark and can be used by Codex - it's OK but still nowhere near the bigger ones: https://til.simonwillison.net/llms/codex-spark-gpt-oss
But... MiniMax M2 benchmarks close to Sonnet 4 and is 230B - too big for one Spark but can run on a $10,000 Mac Studio.
And Kimi K2 runs on two Mac Studios ($20,000).
So we are getting closer.
CamperBob2
3 months ago
Also, at some point the Blackwell-generation DGX Station is supposed to ship with 768 GB of unified memory. It will presumably come with a high five-figure price tag, and it should be able to run most open-source models with little need to trade off quality for speed.
Trouble is, there's not even much hype surrounding the launch yet, much less shipping hardware. Which seems kind of ominous.
lelanthran
3 months ago
> Knowing what that means is worth $150/hour even if you don't type a single line of code to implement it yourself!
Yeah but eventually there wont be enough people who actually do know all that.
The hope amongst the proponents is that by the time that happens they wont need anyone who knows that because SOTA will have replaced those people too.
hastamelo
3 months ago
even if token cost and usage increases, it's far far away from the cost of a developer - $10k+/month
jimbokun
3 months ago
To be fair the author does point to the many parts of software development that remain excluding the writing of code.
atleastoptimal
3 months ago
Open source AI is getting cheaper and cheaper. Model companies run inference at a profit, the lack of profitability from AI companies is just due to them putting all their capital into training the next generation of models.
sameermanek
3 months ago
Seems like openAI spent ~7B on inference in first half of the year. Their revenue was 4B in the same time. They spent another 5B on rnd to deprecate their own, and rest of industry's assets.
scuff3d
3 months ago
I'd love to see the financial data on that.
ManuelKiessling
3 months ago
Model companies running inference at a profit — do you have a good source for this?
winter_blue
3 months ago
> They're in the "get 'em hooked" stage of the drug deal.
You're implying that people are selling inference at below cost right now. That's certainly not true for most third-party inference providers. I doubt API pricing at Anthropic or OpenAI is being solid below cost either.
The only place where you get what you're talking about are the fixed price plans OpenAI, Anthropic, Cursor, etc. sell.
scuff3d
3 months ago
OpenAI is claiming they'll post 74 billion in loses through 2028. Anthropic is on course to lose 3 billion by the end of this year, they lost 5 billion last year.
As far as I can tell the inference provider landscape is a fucking mess, and I can't find any decent financial information on any of the ones I tried. So unless you have something showing those companies are profitable I'm not buy it.
Nevermark
3 months ago
Their big spend isn't inference. It is the training, which they can pull back on at any time.
Inference itself will keep getting cheaper.
scuff3d
3 months ago
Even if you eliminate all of OpenAIs other costs besides inference they're still in the red. And they can't just stop training new models. That's like saying Honda can just quit designing new cars. They technically could, but it would destroy their business.
They have one method of monitization right now, and there is no clear evidence that their costs are suddenly going to decrease anytime soon. Despite claims to the contrary, no one has actually provided any evidence of a pathway to those costs magically cutting in half over the next few years.
The entire industry is being propped up by insane over investment and an obsession with growth at all costs. Investments will dry up sooner or later, and you can't grow forever.
Nevermark
3 months ago
Inference keeps getting cheaper, so "it isn't cheap enough yet" isn't an issue. Even with zero efficiency innovations from here, cost per instruction is the most deflationary commodity of all time.
So how was that ever going to be a problem?
The optimal choice for marginal costs, which will naturally drop on their own, at the beginning of a new tech cycle is to run in the red. It would be a sign of gross incompetence if they were fine tuning those costs already.
Training spend is the giant expense. And either training costs are unsustainable, and training spend will hit a pause, or it is not unsustainable and training spend will continue.
So, which is it?
Critical point: The majority of their costs are not required to serve the highest level of capability they have achieved at any given time.
That is unusual. In the sense that it is an exceptionally healthy cost control structure. Note that not even open source offers a cost advantage, for training or inference.
austin-cheney
3 months ago
I so completely want to agree with you. If this were anything other than JavaScript agreeing with your comment would be simple.
I wrote JavaScript in the corporate world for 15 years. Here is the reality:
* Almost nobody wants to do it. The people that get paid for it don't want to do it. They just want to get paid. The result is that everybody who does get paid for it completely sucks. Complete garbage, at least at work. There a lot of amazing people writing JavaScript, just not at work, and why would they try harder. Delivering quality at work far outside the bell curve just results in hostility aside for some very rare exceptions. My exception was when I was doing A/B testing for a major .com.
* Since everybody in the corporate JavaScript world completely sucks every major project eventually fails from a business perspective or stalls into lifeless maintenance mode. It just gets too expensive to maintain 5+ years later or too fragile to pivot to the next business demand. So, it has to get refactored or rebuilt. Sometime that means hoping the next generation framework is ready, and the business is willing to train people on it, and willing to go through growing pains. More often this means calling in outside parties who can do it correctly the first time. Its not about scale. Its about the ability to actually build something original and justify every hour productively. I was on both sides of that fence.
* The reason why the corporate overlords hire outside parties to fix problems from internal teams isn't just about talent. Keep in mind it's tremendously expensive. Yes, those people are capable of producing something that doesn't suck and do so faster. The bigger issue is that they will always deliver reliably, because they are executing under a contract with a work performance statement. The internal teams do not have a contract performance definition that will kill their careers or terminates their incomes. They just have to hope the business remains financial solvent so they don't get caught in a mass layoff. This breeds a lot of entitlement and false expectations that seem to grow on each other.
So, yes, in this case it really is about the ability to write code physically. Yes, you need to juggle client nonsense and have soft skills too, but those are layered on top of just being able to write the code. When your options are limited to a bunch of 0s that depend on copy/paste from predefined framework templates you need somebody who can actually justify their existence in a very practical solutions delivery way.
scuff3d
3 months ago
That might be a problem in the frontend/JavaScript world, but I think people forget the industry is a lot bigger than web dev. I've managed to go my whole career without ever having to write a single line of JavaScript.
But to be fair your point, and the original authors, isn't a bad one. And I think someone else in this thread said it too. If you're only skill is typing out code against a very narrow spec that someone else did all the work to figure out, you're probably in trouble.
beefnugs
3 months ago
Yeah this is rough... so assuming this blog post is mostly true:
Why would the ai model makers charge less than $149/hr ?
Why hasn't outsourcing attacked equal chunks of that $150/hr all these years now?
If companies dont realize that if an employee is required 1 hour per week, they still need a full salary that covers all the rising costs of housing/food/health/necessity... then most knowledge workers just die. Even more so in up and coming countries, its just massive world war and suffering if we dont change capitalism somehow. Why would AI employees keep working toward such destruction and death?
What if the 80/20 problem is more reality, machine learning and LLMs can never get that last 20 percent working right, but now nobody knows how to finish the last chunks? Seems more like the last 10% of coders that dont die, should be charging hundreds of times more $/hr. Even this doesnt fix the problem because the knowledgeable will die off and nobody hired juniors for years.
You think housing prices can sustain their value in this nonsense? Old people who are still in charge of everything and own everything will destroy all of this before it starts affecting their assets
_wdfs
3 months ago
This is why I think the article whiffs on something important, in the section where he asks "If we can code quickly and cheaply, what is the new constraint?"
Energy. The new constraint is energy.
XenophileJKO
3 months ago
I think the new constraint is actually decisions.