Animats
9 months ago
I'm tired of LLMs.
Enough billions of dollars have been spent on LLMs that a reasonably good picture of what they can and can't do has emerged. They're really good at some things, terrible at others, and prone to doing something totally wrong some fraction of the time. That last limits their usefulness. They can't safely be in charge of anything important.
If someone doesn't soon figure out how to get a confidence metric out of an LLM, we're headed for another "AI Winter". Although at a much higher level than last time. It will still be a billion dollar industry, but not a trillion dollar one.
At some point, the market for LLM-generated blithering should be saturated. Somebody has to read the stuff. Although you can task another system to summarize and rank it. How much of "AI" is generating content to be read by Google's search engine? This may be a bigger energy drain than Bitcoin mining.
datahack
9 months ago
It’s probably generally irrelevant what they can do today, or what you’ve seen so far.
This is conceptually essentially Moore’s law, but about every 5.5 months. That’s the only thing that matters at this stage.
I watched everyone make the same arguments about the general Internet, and then the Web, then mobile, then Bitcoin. It’s just a toy. It’s not that useful. Is this supposed to be the revolution? It uses too much power. It won’t scale. The technology is a dead end.
The general pattern of improvement to technology has been radically to the upside at an increasing pace parabolically for decades and there’s nothing indicating that this is a break in the pattern. In fact it’s setting up to be an order of magnitude greater impact than the Internet was. At a minimum, I don’t expect it to be smaller.
Looking at early telegraphs doesn’t predict the iPhone, etc.
Optimism is warranted here until it isn’t.
YeGoblynQueenne
9 months ago
>> Looking at early telegraphs doesn’t predict the iPhone, etc.
The problem with this line of argument is that LLMs are not new technology, rather they are the latest evolution of statistical language modelling, a technology that we've had at least since Shannon's time [1]. We are way, way past the telegraph era, and well into the age of large telephony switches handling millions of calls a second.
Does that mean we've reached the end of the curve? Personally, I have no idea, but if you're going to argue we're at the beginning of things, that's just not right.
________________
[1] In "A Mathematical Theory of Communication", where he introduces what we today know as information theory, Shannon gives as an example of an application a process that generates a string of words in natural English according to the probability of the next letter in a word, or the next word in a sentence. See Section 3 "The Series of Approximations to English":
https://people.math.harvard.edu/~ctm/home/text/others/shanno...
Note: Published 1948.
bburnett44
9 months ago
I think we can pretty safely say bitcoin was a dead end other than for buying drugs, enabling ransomware payments, or financial speculation.
Show me an average person who has bought something real w bitcoin (who couldn’t have bought it for less complexity/transaction cost using a bank) and I’ll change my mind
jiggawatts
9 months ago
Speaking of the iPhone, I just ugpraded to the 16 Pro because I want to try out the new Apple Intelligence features.
As soon as I saw integrated voice+text LLM demos, my first thought was that this was precisely the technology needed to make assistants like Siri not total garbage.
Sure, Apple's version 1.0 will have a lot of rough edges, but they'll be smoothed out.
In a few versions it'll be like something out of Star Trek.
"Computer, schedule an appointment with my Doctor. No, not that one, the other one... yeah... for the foot thing. Any time tomorrow. Oh thanks, I forgot about that, make that for 2pm."
Try that with Siri now.
In a few years, this will be how you talk to your phone.
Or... maybe next month. We're about to find out.
newaccountman2
9 months ago
I am generally on your side of this debate, but Bitcoin is a reference that is in favor of the opposite position. Crypto is/was all hype. It's a speculative investment, that's all atm.
runeks
9 months ago
> I watched everyone make the same arguments about the general Internet, and then the Web, then mobile, then Bitcoin.
You’re conveniently forgetting all the things that followed the same trajectory as LLMs and then died out.
Melenahill
9 months ago
Hello everyone! With immense joy in my heart, I want to take a moment to express my heartfelt gratitude to an incredible lottery spell psychic, Priest Ray. For years, I played the lottery daily, hoping for a rewarding outcome, but despite my efforts and the various tips I tried, success seemed elusive. Then, everything changed when I discovered Priest Ray. After requesting a lottery spell, he cast it for me and provided me with the lucky winning numbers. I can't believe it—I am now a proud lottery winner of $3,000,000! I felt compelled to share my experience with all of you who have been tirelessly trying to win the lottery. There truly is a simpler and more effective way to achieve your dreams. If you've ever been searching for a way to increase your chances of winning, I encourage you to reach out via email: psychicspellshrine@gmail.com
otabdeveloper4
9 months ago
> ...about the general Internet, and then the Web, then mobile, then Bitcoin. It’s just a toy. It’s not that useful.
Well, they're not wrong. They are not that useful toys.
(Yes, the "Web" included.)
mhowland
9 months ago
"They're really good at some things, terrible at others, and prone to doing something totally wrong some fraction of the time."
I agree 100% with this sentiment, but, it also is a decent description of individual humans.
This is what processes and control systems/controls are for. These are evolving at a slower pace than the LLMs themselves at the moment so we're looking to the LLM to be its own control. I don't think it will be any better than the average human is at being their own control, but by no means does that mean it's not a solvable problem.
latexr
9 months ago
> I agree 100% with this sentiment, but, it also is a decent description of individual humans.
But you can understand individual humans and learn which are trustworthy for what. If I want a specific piece of information, I have people in my life that I know I can consult to get an answer that will most likely be correct and that person will be able to give me an accurate assessment of their certainty and they know how to accurately confirm their knowledge and they’ll let me know later if it turns out they were wrong or the information changed and…
None of that is true with LLMs. I never know if I can trust the output, unless I’m already an expert on the subject. Which kind of defeats the purpose. Which isn’t to say they’re never helpful, but in my experience they waste my time more often than they save it, and at an environmental/energy cost I don’t personally find acceptable.
Gazoche
9 months ago
> I agree 100% with this sentiment, but, it also is a decent description of individual humans.
But humans can be held accountable, LLMs cannot.
If I pay a human expert to compile a report on something and they decide to randomly make up facts, that's malpractice and there could be serious consequences for them.
If I pay OpenAI to do the same thing and the model hallucinates nonsense, OpenAI can just shrug it and say "oh that's just a limitation of current LLMs".
linsomniac
9 months ago
>also is a decent description of individual humans
A friend of mine was moving from software development into managing devs. He told me: "They often don't do things the way or to the quality I'd like, but 10 of them just get so much more done than I could on my own." This was him coming to terms with letting go of some control, and switching to "guiding the results" rather than direct control.
The LLMs are a lot like this.
YeGoblynQueenne
9 months ago
>> I agree 100% with this sentiment, but, it also is a decent description of individual humans.
Why would that be a good thing? The big thing with computers is that they are reliable in ways that humans simply can't ever be. Why is it suddenly a success to make them just as unreliable as humans?
Too
9 months ago
You missed quoting the next sentence about providing confidence metric.
Humans may be wrong a lot but at least the vast majority will have the decency to say “I don’t know”, “I’m not sure”, “give me some time to think”, “my best guess is”. In contrast to most LLMs today that in full confidence just spews out more hallucinations.
jajko
9 months ago
I'll keep buying (and paying premium) for dumber things. Cars are a prime example, I want it to be dumb as fuck, offline, letting me decide what to do. At least next 2 decades, and thats achievable. After that I couldnt care less, I'll probably be a bad driver at that point anyway so switch may make sense. I want dumb beautiful mechanival wristwatch.
I am not ocd-riddled insecure man trying to subconsiously immitate much of the crowd, in any form of fasion. If that will make me an outlier, so be it, a happier one.
I suspect new branch of artisanal human-mind-made trademark is just behind the corner, maybe niche but it will find its audience. Beautiful imperfections, clear clunky biases and all that.
mattgreenrocks
9 months ago
What dumb cars are you looking at?
spencerchubb
9 months ago
LLMs have been improving exponentially for a few years. let's at least wait until exponential improvements slow down to make a judgement about their potential
bloppe
9 months ago
They have been improving a lot, but that improvement is already plateauing and all the fundamental problems have not disappeared. AI needs another architectural breakthrough to keep up the pace of advancement.
COAGULOPATH
9 months ago
In some domains (math and code), progress is still very fast. In others it has slowed or arguably stopped.
We see little progress in "soft" skills like creative writing. EQBench is a benchmark that tests LLM ability to write stories, narratives, and poems. The winning models are mostly tiny Gemma finetunes with single-digit parameter counts. Huge foundation models with hundreds of billions of parameters (Claude 3 Opus, Llama 3.1 405B, GPT4) are nowhere near the top. (Yes, I know Gemma is a pruned Gemini). Fine-tuning > model size, which implies we don't have a path to "superhuman" creative writing (if that even exists). Unlike model size, fine-tuning can't be scaled indefinitely: once you've squeezed all the juice out of a model, what then?
OpenAI's new o1 model exhibits amazing progress in reasoning, math, and coding. Yet its writing is worse than GPT4-o's (as backed by EQBench and OpenAI's own research).
I'd also mention political persuasion (since people seem concerned about LLM-generated propaganda). In June, some researchers tested LLM ability to change the minds of human subjects on issues like privatization and assisted suicide. Tiny models are unpersuasive, as expected. But once a model is large enough to generate coherent sentences, persuasiveness kinda...stops. All large models are about equally persuasive. No runaway scaling laws are evident here.
This picture is uncertain due to instruction tuning. We don't really know what abilities LLMs "truly" possess, because they've been crippled to act as harmless, helpful chatbots. But we now have an open-source GPT-4-sized pretrained model to play with (Llama-3.1 405B base). People are doing interesting things with it, but it's not setting the world on fire.
9cb14c1ec0
9 months ago
I can't think of any exponential improvements that have happened recently.
rifty
9 months ago
I don’t think you should expect exponential growth towards greater correctness past good enough for any given domain of knowledge it is able to mirror. It is reliant on human generated material, and so rate limited by the number of humans able to generate the quality increase you need - which decreases in availability as you expect higher quality. I also don’t believe greater correctness for any given thing is an open ended question that allows for experientially exponential improvements.
Though maybe you are just using exponential figuratively in place of meaning rapid and significant development and investment.
bamboozled
9 months ago
Do you know what exponential means? They might be getting getting but it hardly seems exponential at this stage.
__loam
9 months ago
Funnily enough, bitcoin mining still uses at least about 3x more power that AI at the moment, while providing less value imo. AI power use is also dwarfed by other industries even in computing. We should still consider whether it's worth it, but most research and development on LLMs in corporate right now seems to be focused on making them more efficient, and therefore both cheaper and less power intensive, to run. There's also stuff like Apple intelligence that is moving it out to edge devices with much more efficient chips.
I'm still a big critic of AI generally but they're definitely not as bad as crypto which is shocking.
illiac786
9 months ago
Do you have a nice reference for this? I could really use something like this, this topic comes up a lot in my social circle.
Ferret7446
9 months ago
How do you measure the value of bitcoin, if not by its market cap? Do you interview everyone and ask them how much they're willing to pay for a service that allows them to transfer money digitally without institutional oversight/bureaucracy?
latexr
9 months ago
> They're really good at some things, terrible at others, and prone to doing something totally wrong some fraction of the time. (…) They can't safely be in charge of anything important.
Agreed. If everyone understood that and operated under that assumption, it wouldn’t be that much of an issue. Alas, these guessing machines are marketed as all-knowing oracles that can already solve half of humanity’s problems and a significant number of people treat them as being right every time, even in instances where they’re provably wrong.
seandoe
9 months ago
Totally agree on the confidence metric. The way chatbots spew complete falsities in such a confident tone is really disheartening. I want to use AI more but I don't feel I can trust it at all. If I can't trust it and have to search for other resources to verify it's claims, the value is really diminished.
naming_the_user
9 months ago
Is it even possible in principle for an LLM to produce a confidence interval given that in a lot of cases the input is essentially untrusted?
What comes to mind is - I consider myself an intelligent being capable of recognising my limits - but if you put my brain in a vat and taught me a new field of science, I could quite easily make claims about it that were completely incorrect if your teaching was incorrect because I have no actual real world experience to match it up to.
theamk
9 months ago
Right, and that's why "years of experience" matters in humans. You will be giving incorrect answers, but as long as you get feedback, you will improve, or at least calibrate your confidence meter.
This is not the case with current model - they are forever stuck at junior level, and they won't improve no matter how much you correct them.
I know humans like that too. I don't ask them questions that I need good answers too.
wrycoder
9 months ago
Just wait until they get saturated with subtle (and not so subtle) advertising. Then, you'll really hate them.
rldjbpin
9 months ago
LLMs are to AI what BTC is to blockchain, let me explain.
blockchain and no-trust decentralization has so much promise, but grifters all go for what got done first and can be squeezed money out of. same is happening with LLMs, as a lot of current AI work started with text first.
they might still lowkey be necessary evils because without them there would not have been so much money or attention flowing in this way.
agubelu
9 months ago
> blockchain and no-trust decentralization has so much promise
I've been hearing this for the past 5 years, yet nothing of practical use based on blockchains has materialized yet.
Jommi
9 months ago
you dont think an open finance network that's accessible for anyone with an internet connection is useful?
your westernness is showing
go ask SA or Africa how useful it is that they arent restricted by insane dictatorial capital controls anymore
throw45678943
9 months ago
Indeed. Decentralised currency is at least a technology that can power the individual at times, rather than say governments, big corps, etc especially in certain countries. Yes it didn't change as much as was marketed but I don't see that as a bad thing. Its still a "tool" that people can use, in some cases to enable use cases they couldn't do or didn't have the freedom to do before.
AI, given its requirements for large computation and money, and its ability to make easily available intelligence to certain groups, IMO has a real potential to do the exact opposite - take away power from individuals especially if they are middle class or below. In the wrong hands it can definitely destroy openness and freedom.
Even if it is "Open" AI, for most of society their ability to offer labor and intelligence/brain power is the only thing they can offer to gain wealth and sustenance - making it a commodity tilts the power scales. If it changes even a slice of what it is marketed at; there are real risks for current society. Even if it increases production of certain goods, it won't increase production of the goods the ultra wealthy tend to hold (physical capital, land, etc) making them as a proportion even more wealthy. This is especially true if AI doesn't end up working in the physical realm quick enough. The benefits seem more like novelties to most individuals that they could do without where to large corps and ultra wealthy individuals the the benefits IMO are much more obvious with AI (e.g. we finally don't need workers). Surveillance, control, persuasion, propaganda, mass uselessness of most of the population, medical advances for the ultra wealthy, weapons, etc can now be done at almost infinite scale and with great detail. If it ever gets to the point of obsoleting human intelligence would be a very interesting adjustment period for humanity.
The flaw isn't the technology; its the likely use of it by humans and their nature. Not saying LLMs are there yet or even if they are the architecture to do this, but agentic behaviour and running corporations (as OpenAI makes its goal on their presentation slides to be) seem to be a way to rid many of the need for other people in general (to help produce, manage, invent and control). That could be a good or bad thing, depending on how we manage it but one thing it wouldn't be would be simple.
bschmidt1
9 months ago
I love how people are like "there's no use case" and there's already products on shelves. I see AI art everywhere, AI writing, customer support - already happened. You guys are naysaying something that already happened people already replaced jobs with LLMs and already profit due to AI. There are already startups with users where you provide a OPENAI_API_KEY, or customers where you provide theirs.
If you can't see how this tech is useful Idk what to tell you, you have no imagination AND aren't looking around you at the products, marketing, etc. that already exists. These takes remind me of the luddites of ~2012 who were still doubting the Internet in general.
lmm
9 months ago
> I see AI art everywhere, AI writing, customer support - already happened.
Is any of it adding value though? I can see that AI has made it easier to do SEO spam and make an excuse for your lack of customer support, just like IVR systems before it. But I don't believe those added any real value (they may have generated profits for their makers, but I think that was a zero- or negative-sum trade). Put it this way: is AI being used to generate anything that people are actually happy to receive?
fragmede
9 months ago
> But I don't believe those added any real value (they may have generated profits for their makers, but I think that was a zero- or negative-sum trade).
Okay, so some people are making money with it, but no true value was added, eh?
lmm
9 months ago
Do new scams create value? No, even though they make money for some people. The same with speculative ventures that don't pan out. You can only say something's added value when it's been positive sum overall, not just allowed some people to take a profit at the expense of others.
bschmidt3
9 months ago
[flagged]
anon7725
9 months ago
There is a difference between “being useful” and living up to galactic-scale hype.
bschmidt1
9 months ago
[flagged]
dang
9 months ago
You've continued to break the site guidelines, not just with this account but with others like https://news.ycombinator.com/item?id=41681416, and ignored our requests to stop.
Between that and the personally abusive emails you've been sending, it's clear that you don't want to use HN as intended, so I've banned the accounts.
lukev
9 months ago
The utility of LLMs clearly exists (I'm building a product on this premise, so I'm not uninterested!)
But hype also exists. How closely they are matched is not yet clear.
But your comment seems to indicate that the "pro-tech" position is automatically the best. This is _not_ true, as cryptocurrency has already demonstrated.
bschmidt3
9 months ago
Funny thing is you are actually the pro [corporate] tech one, not on the side of freedom. Furthermore nobody said anything about crypto - you are seriously grasping at straws. You have said nothing about the products on shelves (billions of dollars in the industry already) argument, only presented crypto as an argument which has nothing to do with the conversation.
bschmidt1
9 months ago
> This is _not_ true, as cryptocurrency has already demonstrated.
His whole argument against AI is basically the anti-tech stance: "Well crypto failed that means AI will too" It's coming from a place of disdain for technology. That's your typical Hacker News commenter. This site is like Fox News in 2008 - some of the dumbest people alive
lukev
9 months ago
Not at all! I am broadly speaking very pro-tech.
What I am against is the “if it’s tech it’s good” mindset that seems to have infected far too many. I mention crypto because it’s the largest example of tech that is not good for the world.
anon7725
9 months ago
You’re certainly able to sus out everything about my worldview, knowledge level and intentions from my one-sentence comment.
The only thing that LLMs are at risk of subverting is the livelihood of millions of people. AI is a capital intensifier, so the rich will get richer as it sees more uptake.
About copyright - yeah, I’m quite concerned for my friends who are writers and artists.
> You'll get left behind with these takes, it's not smart. If you don't care about advancing technology or society then have fun being a luddite, but you're on the wrong side of history.
FWIW I work in applied research on LLMs.
bschmidt3
9 months ago
[flagged]
user
9 months ago