rkozik1989
4 hours ago
This article is really only useful if LLMs are actually able to close the gap from where they are now to where they want to be in a reasonable amount of time. There are plenty of historical examples of technologies where the last few milestones are nearly impossible to achieve: hypersonic/supersonic travel, nuclear waste disposal, curing cancer, error-free language translation, etc. All of which have had periods of great immediate success, but development/research always gets stuck in the mud (sometimes for decades) because the level complexity to complete the race is exponentially higher than it was at the start.
Not saying you should disregard today's AI advancements, I think some level of preparedness is a necessity, but to go all in on the idea that deep learning will power us to true AGI is a gamble. We've dumped billions of dollars and countless hours of research into developing a cancer cure for decades but we still don't have a cure.
dakiol
3 hours ago
In software we are always 90% there. Is that 10% the part that gives us jobs. I don’t see LLMs that different from, let’s say, the time compilers or high level languages appeared.
bpt3
3 hours ago
Until LLMs become as reliable as compilers, this isn't a meaningful comparison IMO.
To put it in your words, I don't think LLMs get us 90% of the way there because they get us almost 100% of the way there sometimes, and other times less than 0%.
Jensson
2 hours ago
Compilers reliably solves 90% of your problems, LLM unreliably solves 100% of your problems.
So yeah, very different.
bluefirebrand
an hour ago
If it's not reliable then the problem is not solved
You've just moved the problem from "I can't solve this" to "I can't trust if the LLM solved this properly"
nomel
an hour ago
There is a level of reliability that is sufficient, as proven by us humans, the existence of issue trackers, and the entire industry that is software QA.
And, further, the existence of offshore, low quality, contractors that are in such frequent use.
xp84
36 minutes ago
Precisely. The code I would get from that type of contractor had a similar reliability as what I generate today with nothing but the $20 a month level of AI stuff. Of course, we have the option of making the AI rewrite it in 2 minutes or so to fix its mistakes without waiting for it to be day there again.
AI replacing outsourcing and (sadly) junior SWEs is more likely than it just eliminating coding jobs across the board. Lord help them when our generation of senior SWEs retires, though!
bluefirebrand
34 minutes ago
> AI replacing outsourcing and (sadly) junior SWEs is more likely than it just eliminating coding jobs across the board. Lord help them when our generation of senior SWEs retires, though
Not them
It's on current software devs to make sure this doesn't happen! People in senior positions need to be loud and aggressive about telling the money people that we cannot rely on AI to do this work!
Every time you shrug and say "yeah the LLM does ok junior level work" you are part of the goddamn problem
xp84
a minute ago
What problem? It's not my problem if my employer is screwed after my generation retires. That's the shareholders or owners' problem. The people making 2-10x my salary upstream of me in the org chart are being paid that presumably because they have such greater wisdom and foresight than I do. If I'm the CTO or have very significant equity, maybe I'll talk about restarting hiring of juniors. Otherwise I'll just sit and wait for the desperate consulting offers. It'll be like the COBOL boom in the late 90s.
Note: That isn't my retirement plan, but it'll just be a fun source of extra money if I'm right.
CamperBob2
an hour ago
If it makes money, the problem is solved. At least from the perspective of the people with the money.
Less cynically, it doesn't matter whether some code was written by a human or an LLM -- it still has to be tested and accepted. That responsibility ultimately must end up on a human's desk.
derefr
3 hours ago
I would argue that "augmented programming" (as the article terms it) both is and isn't analogous to the other things you mentioned.
"Augmented programming" can be used to refer to a fully-general-purpose tool that one always programs with/through, akin in its ubiquity to the choice to use an IDE or a high-level language. And in that sense, I think your analogies make sense.
But "augmented programming" can also be used to refer to use of LLMs under constrained problem domains, where the problem already can be 100% solved with current technology. Your analogies fall apart here.
A better analogy that covers both of these cases, might be something like grid-scale power storage. We don't have any fully-general grid-scale power storage technologies that we could e.g. stick in front of every individual windmill or solar farm, regardless of context. But we do have domain-constrained grid-scale power storage technologies that work today to buffer power in specific contexts. Pumped hydroelectric storage is slow and huge and only really reasonable in terms of CapEx in places you're free to convert an existing hilltop into a reservoir, but provides tons of capacity where it can be deployed; battery-storage power stations are far too high-OpEx to scale to meet full grid needs, but work great for demand smoothing to loosen the design ramp-rate tolerances for upstream power stations built after the battery-storage station is in place; etc. Each thing has trade-offs that make it inapplicable to general use, but perfect for certain uses.
I would argue that "augmented programming" is in exactly that position: not something you expect to be using 100% of the time you're programming; but something where there are already very specific problems that are constrained-enough that we can design agentive systems that have been empirically observed to solve those problems 100% of the time.
derektank
2 hours ago
There are no technical hurdles remaining with respect to nuclear waste disposal. The only obstacle is social opposition
golemotron
an hour ago
..and regulation. The same with supersonic.
BinaryIgor
4 hours ago
100%; Exactly as you've pointed out, some technologies - or their "last" milestones - might never arrive or could be way further into the future than people initially anticipated.
bgroat
4 hours ago
We're 90%... we're almost half way there!
HumblyTossed
3 hours ago
It costs 10% to get 90% of the way there. Nobody ever wants to spend the remaking 90% to get us all the way there.
strbean
3 hours ago
I think parent was making an (apt) reference to Old School RuneScape, where level 92 represent half of the total XP needed to reach the max level of 99.
germandiago
4 hours ago
Exactly this.
jrecyclebin
3 hours ago
I'm already not going back to the way things were before LLMs. This is fortunately not a technology where you have to go all-in. Having it generate tests and classes, solve painful typing errors and help me brainstorm interfaces is already life-changing.
xp84
28 minutes ago
I am in a similar place, I think, to you. Caveat: I don't spend a lot of my day-to-day coding anyway, so that helps, but I've never even tried Cursor or Windsurf. I just poke copilot to write whole functions, or ask ChatGPT for things that seem like they'd be tedious or error-prone for me. Then I spend 3-5 minutes tying those things together and test. It saves about 80% of the time, but I still end up knowing exactly how the code works because I wrote most of the function signatures and reviewed all the code.
I know in the very right circumstances the "all-in" way could be faster, but it's already significant that I can do 5x as many coding tasks or do one in a fifth the time. Even if it never ever improves at all.
CuriouslyC
3 hours ago
LLMs are noisy channels. There's some P(correct|context). You can increase the reliability of noisy channels to an arbitrary epsilon using codes. The simplest example of this in action is the majority decoding logic, which maps 1:1 to parallel LLM implementation and solution debate among parallel implementers. You can implement more advanced codes but it requires being able to decompose structured LLM output and have some sort of correctness oracle in most cases.