Controlling AI's Growing Energy Needs

37 pointsposted a month ago
by pseudolus

64 Comments

jazzyjackson

a month ago

Controlling [unjustified futurist hype's] Growing Energy Needs

Haven't seen a case study yet where a company outside of spammers and scammers increased their productivity using LLMs to do any kind of classification or unstructured-data-to-structured-data translation.

I've tried having it build up a table of contact information from email signatures, llama3 hallucinates phone numbers when the one I wanted was in the prompt, would have been better off with regex.

Looking up film developer recipes with Perplexity Pro/GPT4o, the answer comes back with a confident 1:119 ratio of HC110 for my P33 film, when I click through the source it pulled those numbers from a completely different film stock, since naturally the forum thread drifted from discussing one brand to another midway through conversation and doesn't matter how big your context window is, LLMs cant keep associations straight.

I'm going to keep trying to find uses for these garbage generators because the appearance of omniscience is so seductive, but I'm more cynical by the day, the investments and nuke fast tracking seems to be coming straight out of faith that AGI is just around the corner.

(Yea yea I know, I'm holding it wrong)

devjab

a month ago

Our engineers feed a LLM engineering plans as a quality control. Not as a replacement of any of the regular process, simply as an additional step. It discovered what would’ve been a fairly serious error in some component design for one of our energy plants that nobody else had caught. Similarity it “fact checked” one of our subcontractors on something with cable sizes that I honestly don’t understand. Finding a better option than what the subcontractor was suggesting. I say “fact checked” because you shouldn’t think of it as that since everything it outputs need to be actual fact checked.

I mostly view LLMs as advanced auto-complete for what I do and haven’t been very impressed by them, but some of the examples our engineers show us are wild.

jazzyjackson

25 days ago

Thanks. I actually do have a workflow that would benefit from feeding an LLM the data corresponding to 4 'stages' of an order in varying formats. Having a computer flag inconsistencies sounds feasible.

llm_trw

a month ago

I just wrote a multilinear algebra research paper using natural language to generate all the latex. This used to take me several hour of tedious mind numbing boiler plate. I finished in in 10 minutes today.

People who aren't on board with llms as the biggest driver of growth since internet search were the type of people who still thought the yellow pages were safe from Google in 2004.

bertylicious

25 days ago

How do you generate "all the latex" in a latex document with "natural language"? What specifically did you generate? The whole research paper? What "boilerplate" are you talking about?

llm_trw

25 days ago

Are you familiar with LaTeX?

I was writing a lot of rather complex augmented matrices.

Instead of having to manually type them in with all the formatting done by hand I literally gave ChatGPT a python nested list and told it what to subscript, what to bold, etc.

In previous papers I have done this by hand if there's only a few matrices that I need to show, or by writing a script to auto-generate them in whatever format I need if there will be many.

This time I just talked to a chat bot and got it working in minutes.

Then it turns out I made a mistake at the start. The fix was just to tell it what the error was and propagate the changes. It did that flawlessly.

bertylicious

25 days ago

Yes, I'm familiar with latex. That's why I was asking.

Using an LLM to reformat/restructure your matrix data makes sense to me and this was the missing detail from your original post.

I would be too paranoid about hallucinations and opt for a script still, but that might just be personal preference.

llm_trw

25 days ago

The question is: how is more likely to make a stupid mistake and not spot it? Me or an LLM? My money is on me.

Yoric

24 days ago

With a simple script? I'd bet on the LLM making more mistakes.

user

a month ago

[deleted]

add-sub-mul-div

a month ago

> I'm going to keep trying to find uses for these garbage generators

Is it really worth your time? Sometimes the simplest explanation is the right one. If you've already mastered your craft past a certain point and care about the quality of your work, this will only drag you down.

yallpendantools

25 days ago

This reminded me of a conversation I recently had at work. Someone was turning over a project to me so he was explaining the context and history of the technical design decisions they've gone through. One stage of the project was basically linting through someone else's source code, evaluating possible code branches.

"So, we want to know all the possible states this could go through. We gave that task to OpenAI with the following prompt."

He shows me a sample prompt they would've sent to OpenAI. It started with the premise "You are a senior Ruby developer with ten years experience. Read through this source file and give me a list of all the code paths which involves global variables".

"Really?" I said, amused. I knew ahead of time that this project they were turning over is LLM-heavy. I have misgivings about LLMs but I try to keep an open mind so it was really impressive that it was used at this stage. "How well did it work?"

"It didn't. Instead we parsed it to a syntax tree and just performed static code analysis."

(Usual disclaimer that details of the anecdote were changed/obfuscated but the spirit of the interaction has been preserved.)

ta12653421

a month ago

For all tech people, LLM are an absolute boon: it brings productivity by 10x (at least!) onto the next level, esp. for develops.

I developed an financial servic application with 1MB+ source of code (not including 3rd party libs) within 3 month - thats more than 1 million keystrokes that created a system that works perfectly fine and produces exactly the results as specified. (for comparison: my masterthesis had around 110.000 letters, and i wrote it over a couple of weeks, while the app is slightly more complicated: its a complex system where each part relies on each another, while the masterthesis is just a dumb document in which i could write in theory everything and nobody would care/check if the descript is just bs)

arp242

25 days ago

> I developed an financial servic application with 1MB+ source of code (not including 3rd party libs) within 3 month - thats more than 1 million keystrokes that created a system that works perfectly fine and produces exactly the results as specified

How do you know it "works fine"? How do you know there aren't tons of caveats as soon as some boundary conditions are exceeded? Certainly not because you actually understand it all.

Just yesterday we had "Engineers do not get to make startup mistakes when they build ledgers".[1]

I sure hope it's just for your personal use because it would be completely irresponsible otherwise, to the point of criminal neglect. This is a Knightmare[2] waiting to happen and I can only hope it will only your own money that's effected and that you won't ruin anyone's lives with this.

[1]: https://news.ycombinator.com/item?id=42269227

[2]: https://dougseven.com/2014/04/17/knightmare-a-devops-caution...

ta12653421

25 days ago

Interesting why i get downvoted for appraising LLM and clapping-up that it grows my productivity? Very interesting...

Good one: I read the article about this Fintech yesterday as well :) To your questions: It works because it can see it? (For sure, every software has some bugs which may be never enconutered, but in this case its a fairly simple application, doint exactly 4 functions for me)

The link to the Knight story, thanks for pointing out on that, interesting read! Well, they are some magnitudes above what we are doing here with our own trading infrastructure, apart from that: we are using our own money on our own risk.

ptero

25 days ago

You are almost certainly downvoted primarly for the tone, not content. "For all the tech people" is very likely to bring downvotes. If you wrote it as a personal data point it would be fine.

Also, advocating for a general availability fintech app with 1MB of llm-generated code makes me shudder. I want my finances (x-rays, car software, etc.) to not be done that way.

If you had been clear that you made your own trading infra for your own money -- good luck! I have zero problems with this. My 2c.

ta12653421

25 days ago

Well, with "all tech people" i actually meant ">everyone working in tech (or like the HN community<", and i was serious about that: I havent met one person in the business on the tech side who didnt profit from it?

Hm, because of which wording do you imply "general availability"? Interesting take! :)

consp

25 days ago

> I developed an financial servic application with 1MB+ source of code (not including 3rd party libs) within 3 month

The timeframe + loc + financial sector seems .... Off (I'm not touching that).

ta12653421

25 days ago

Wanna join me for a code/QA flight, i'm open for external audits :)

But actually you are somehow right: Without LLM this wouldnt have been possible, e.g. The platform uses Hibernate - for this, you need to create your entity class, a mapping class and a SQL script, and these things have to be somehow in line to each another to make it work; before LLM, you had to type those files.

Its just about automating those things which you had to type earlier & timecostly on your own.

yallpendantools

25 days ago

> The platform uses Hibernate - for this, you need to create your entity class, a mapping class and a SQL script, and these things have to be somehow in line to each another to make it work; before LLM, you had to type those files.

Do you really need an LLM for that? Sure LLMs can do that but they're overkill, no?

My internship and first proper job (2011-2012) was at a consulting firm that used Hibernate. They managed to auto-generate all these things with an Excel macro. I'm willing to bet some JetBrains plugin can do that too without LLMs.

ta12653421

25 days ago

This was just a very simple example which came to my mind: you could also use active mapping etc. (though i prefer the XML files), its just one example out of many: i see LLM not as a "problem solver" but as a helper/sparringpartner

jncfhnb

a month ago

I’d be sweating furiously if told I was to inherit a new app to maintain and reading the text above

llm_trw

25 days ago

I'd be looking for a new job.

People like OP are the reason why there is so much skepticism about the new generation of AI.

>I just copy pasted all the answers I found on stack overflow into our live trading database at it works like a charm.

-- OP from a previous wave of computing.

ta12653421

25 days ago

No, actually not:

The thing is - i have 20+ years of IT & SWdev experience, i'm an old guy in your eyes: If you can check and QA, what the LLM gives you, you are very-pretty save.

Also, just sharing some of your current thoughts usually leads to new ideas when applying LLM; its very inspiring because they usually point you out to new contexts & ideas.

llm_trw

25 days ago

Again, that's also true for copy/pasting anything from stack overflow.

ta12653421

25 days ago

by far not! :-D

on stack overflow, i do not get "context-adjacent" ideas & inspirations - thats the difference.

ta12653421

25 days ago

Hahah, i heard that sub-tone :)

Well, the thing is, the application does only a few things and it is very well structured, i'd say - i could line out the architecture with a few sentences. For sure: To understand how the internal mechanics are working, you need to have some background knowhow on several layers.

bertylicious

25 days ago

Do you mind revealing the name of your "financial servic application" so we can stay clear of it?

ta12653421

25 days ago

Its a proprietary trading app which is used for our own activities only:

For a "standard user" its too complicated, its not made for the retail world. Would be tricky to sell this, esp. as it works only with one specific broker currently.

But if i make a retail version, i will post it here :)

shrubble

a month ago

It seems more than ridiculous that this is considered a serious issue. Datacenter space, generators (to cover the power needs of the datacenter) and power cost real money and thus there is plenty of incentive to reduce it.

Already we are seeing both hardware alternatives and software optimizations; human ingenuity will eventually take care of it.

croes

a month ago

Eventually.

And if not we accelerate climate change because we built fast solutions for the energy problems instead of sustainable ones.

XorNot

25 days ago

The solution to climate change is not reducing energy usage. That can, at best, delay it. The solution is to generate energy without emitting CO2. We already have the technology to do this.

The conflation of supposed moral virtues versus actual, practical solutions is a huge problem with messaging in this space: there are a lot of people who don't see climate change as a problem to be solved, but as the stepping stone to bringing down capitalism or some other goal they want to use it as a bandwagon for.

croes

24 days ago

>The solution is to generate energy without emitting CO2.

These take time and emit CO2 themselves at least in the bulding phase.

And LLMs like GPT won't help us fighting climate change so we raise our energy consumption without later benefit on that matter.

Maybe we first should buld the energy sources and then use them for convenience tools.

Or do you want to explain our children we fucked up because we "needed" AI to write emails that are read by AI to generate a response? Much of AIs use is still just for unnecessary things just more convenient. It's still unclear if they have a productivity benefit or just changed the way we spend the same amount of time.

Intel's study found no such benefit and suggests teaching the people how to use LLMs means how to tell the machine what it should do.

Sounds like programming with extra steps and more ambiguity.

XorNot

24 days ago

No matter how you slice it though, not building LLMs doesn't change the underlying equation: you're emitting CO2 to make electricity. Every new human being born once they turn on electric light is. Every new industry. Every bit of economic growth, basically.

Any action which isn't the reduction of CO2 to generate electricity, or elimination, is just a delaying tactic. Even substantial population reduction would buy you what, maybe a few hundred years before atmospheric CO2 hits the same levels again?

There is exactly one problem to solve with anthropogenic climate change and it's CO2 emissions. Any action which isn't directly attacking that issue is a waste of time and political resources, and demonstrably*has been a complete waste of time and effort - every victory was won on the back of superior technology (i.e. efficiency didn't just happen - LEDs happened and they're just plain better and cheaper).

Yoric

24 days ago

Delay is good. Delay is necessary. While we have (some bricks of) the technology to generate/store/consume energy without emitting CO2, we haven't nearly rolled it out and we need time to do so. So, yes, by all means, any month we win by decreasing energy usage might save countless lives.

> The conflation of supposed moral virtues versus actual, practical solutions is a huge problem with messaging in this space: there are a lot of people who don't see climate change as a problem to be solved, but as the stepping stone to bringing down capitalism or some other goal they want to use it as a bandwagon for.

Fair point.

ben_w

25 days ago

On the vendor's side, there's incentive to reduce the costs per token, but also an incentive to sell more tokens.

On the consumer's side, it's more about what the alternative would be, and the alternative is getting a human to do it.

Right now the general quality of the LLMs is good enough to be a tool, but not good enough to fully replace humans using it — I've seen GPT-4 called an intern, I'd agree having seen student code not too long before ChatGPT came out, and you don't want to let students do a whole project unsupervised either. Even though o1 is better, I think it's still at the "interesting, but not good enough to leave unsupervised" level*.

For tasks where the AI is competent enough to be interesting for whatever task, even if it took 10 kW to run the compute, at $0.1/kWh that's often cheaper than offshoring to a poor country.

As the world's entire electricity supply is 2 TW, which is 250 watts per capita, I think it's not at all implausible that enough people decide that the output of various AI (not just LLMs, also vision, audio, robots etc.) are valuable enough to demand enough of that electricity to make the prices rise.

* and I'm not just saying that because when it gets that good none of us will be employable any more

llm_trw

a month ago

The number one predictor of human wealth is energy consumption. Since the stagnation of the west in the 1980s we have seen the same stagnation in energy consumption there.

jncfhnb

25 days ago

In what way has western energy consumption stagnated since the 1980s

consp

25 days ago

It hasn't, it has been linear for a while and is not exponential by the looks of it. Still climbing though.

doctorpangloss

a month ago

The reason people talk about this is because carbon credit trades are similar to minting shitcoins:

- You have to create demand for something totally subjective like, "Does this reduce carbon emissions?"

- There is a large liquid market for carbon credits.

- You have to turn something nonfungible like various carbon sequestration technologies and schemes into something fungible like tons of CO2 emissions.

- So you go and "mint" a "shitcoin," lets say literal chicken shit being processed "in a way" that makes it "make less carbon." Then some BS little certifying agency says that it does do that, whatever that means, and someone does some lab test once, that shows something, and then "AI Algorithms" are used to extrapolate the impact. Then you go to JP Morgan who needs various shitcoin-like things like this minted by people like you, and JP Morgan bundles it all up in a trade where Microsoft, who decided that they need to offset all datacenter emissions, will become the literal bagholder of chicken shit.

Is there much of a difference between science trade journals speculating about so and so carbon reduction scheme and crypto trade journals speculating about so and so utility token? They are similar energies. I would love for there to be utility in the other side of the carbon credit schemes, but it looks like bullshit more often than not.

dr_dshiv

25 days ago

You make it sound like banks tolerate fraud. A financial cleanup is coming to the carbon markets, because that’s the only way serious money will flow.

user

a month ago

[deleted]

vfclists

25 days ago

Shouldn't the title of the article be "Research on low power alternatives to LLMs" rather than "Controlling AI's ..."

naveen99

25 days ago

Controlling ?

Aren’t we try to increase fraction of global energy usage towards computing transformers as fast as possible ?

Desert solar anyone ?

avip

25 days ago

So ChatGPT-3 took ~1.3 GWh to train. And it provides some questionable value to people.

While you're at it, Bitcoin, a by definition no-value product, takes ~100 TWh to run annually.

Excuse the whataboutism.

benchmarkist

a month ago

No one knows where we are going, the aim of life has been forgotten, the end has been left behind. Man has set out at tremendous speed to go nowhere. - Jacques Ellul

deadbabe

a month ago

Why are we still stuck with computing technology that requires nuclear power plants worth of energy to power AI that can barely match the cognitive power of a human brain running on a cup of coffee and some calories?

xena

a month ago

Because the AI that can barely match the cognitive power of a human brain running on a cup of coffee and some calories doesn't sleep, eat, get sick, go on maternity/paternity leave, go on strike, want to be paid, want their pay to increase with the cost of living, or have pesky human rights that limit the value you can extract from them.

winocm

a month ago

This is increasingly reminding me of those dystopian science fiction films.

scott_w

a month ago

Would you say the same of any automated or mechanised process that massively reduces necessary human input for the above reasons?

winocm

a month ago

That is honestly how I actually see these things being used now, they can be powerful tools if used effectively.

The thought of people being replaced wholesale though…

scott_w

25 days ago

Tractors and combine harvesters replaced humans wholesale. Would you rather we go back to a world where all farming must be done by humans manually operating equipment? Countless advancements in manufacturing replaced humans wholesale in factories. Should we undo that?

deadbabe

25 days ago

I don’t think you understand. Can we create a computer the size of a bee’s brain and with the same intelligence and energy requirements? If we can’t, then I don’t think we’ll ever truly achieve an efficient general intelligence.

Teever

a month ago

You're not calculating the energy required for a human being to do computation. It requires food, clothing, recreation, temperature control, healthcare, and education and all sorts of things that require oodles more energy than a simply calculation of daily calorie consumption of a human being.

Yoric

25 days ago

True, but the alternative is letting people starve. I suggest that this should not be one of the goals of our society.

llm_trw

a month ago

The computing machines that the Manhattan project used were famously as fast as the average secretary doing the same calculations. Yet the computer age still happened. Ask yourself why.

m3kw9

a month ago

One is miniaturization and the other one is some sort of science breakthroughs to understand how the brain works.

monkaiju

a month ago

If it helps the AI doesnt 'match' humans at all, it does a completely different thing.

teitoklien

a month ago

Didn’t know there was a human brain running on a cup of coffee and some calories answering hundreds of millions of cognitive API requests across the world each minute and continuing to grow.

Tbh AI is already at the stage that its more effective and insightful on most subjects than a huge chunk of humans and their coffee and calorie sippin brains.

I don’t think this sort of sadist hot take on nascent revolutionary tech that’s becoming better exponentially each year, is productive or fits into HN.

WA

a month ago

Let’s see about the exponential part. Too early to judge. Some signs that the current capabilities are actually stagnating and progress is slowing down. The problem of determining when an LLM is wrong is fundamental and might be a show-stopper for exponential imorovements.