Tell HN: Is this AI doom loop? I thought o1-preview was amazing, until it wasn't

11 pointsposted 17 hours ago
by moomoo11

Item id: 41683399

13 Comments

proc0

16 hours ago

Right, I think it's a limitation of deep learning. Transformers and gigantic leaps in scaling have allowed AI models to reach impressive capabilities, but at its core it's limited by its training data. This is not how actual intelligence works. People don't need terabytes of information to learn how to communicate and reason.

For this reason the current AI trend will have a bigger impact on creative tasks, rather than critical technical ones. It's already great for generating art quickly, at least for the concepting phase, and other creative assets that can afford to be generic. Solving technical problems, on the other hand, requires reasoning beyond what can be extracted from training data.

We'll need a new paradigm of AI in order to have a chance at creating models that properly reason. Even without detail knowledge of the brain, we can safely speculate that the reason and language areas are extremely efficient compare to cutting edge LLMs, which means there are algorithms more complex and efficient than simple artificial neural connections that just sum weights with a bias.

mergisi

2 hours ago

AI can be frustrating when it falls short on complex tasks, especially with long or intricate code. To get the most out of it, you need to break down big problems into smaller pieces and choose the right AI for your needs—like OpenAI o1, Claude Sonnet. Also, leveraging dev tools like Cursor AI can help enhance productivity. For tasks like SQL generation, specialized tools like https://ai2sql.io/ work great. AI isn't perfect for everything, but when used selectively, it can still be super helpful.

eschneider

17 hours ago

Yes. You're expecting too much. Generative AI models don't "understand" your problem, they don't even "understand" how to program. They're just fitting whatever data they've seen to your input.

gwoolhurme

10 hours ago

To his defense, that is how it’s marketed. That this new model can reason.

loveparade

5 hours ago

Yeah, but "reason" is not a well-defined term. It means different things to different people in different contexts. It's just marketing speech. You can easily argue that all ML models, even those from 50 years ago, can reason to some extent.

gwoolhurme

4 hours ago

Fully agree that’s kind of my point though. It’s a very tall order for some people. Like the OP

jprete

16 hours ago

It never occurred to me before that chatbot randomness might have the same reward structure as a slot machine, but apparently it does. OpenAI got you to spend all your credits on seven attempts at this one problem. I'm not saying you're addicted, but I wonder about the people who absolutely gush over it.

resource0x

16 hours ago

It's time for "How many programmers does it take to screw in a lightbulb using AI" jokes.

muzani

14 hours ago

It might be just an OpenAI thing where they ramp up the power a week after the demo to get more subscriptions, then bring it down gradually. The forums now happily gaslight you into thinking it's a conspiracy theory, despite the evidence. It's easy to catch - just share an amazing input/output response with a friend, then try the exact same thing a month later.

It's one of the arguments for using open source AI even though it's still a little behind - at least when you're running it on your own system, you know if you're the problem.

solardev

15 hours ago

Does Claude work any better for you?

moomoo11

12 hours ago

I have tried Claude and I find that it is more "to the point", but still suffers from giving the wrong/unusable answers.

Both claude and chatgpt are good for rote tasks, but claude is definitely more succinct.

b20000

13 hours ago

these models are search engines with some interpolation thrown in

rvz

14 hours ago

> When people hype up that the AI solved something for them, I wonder were they lazy like me working with something complex, or were they lazy and didn't even try on something simple?

The truth they won't tell you is that they have likely invested in that AI tool and they are hyping it up with their VC friends that 'It works' even when they know it doesn't.

Each time I talk to the AI bros about these limitations, they retort to their whataboutisms with 'But humans hallucinate too!', 'The human brain is the same as an LLM' nonsense excuses.

LLMs do not 'understand' your problems nor can they reason about them. O1 is no different and instead of buying into the scam and prompting endlessly with garbage results and attempting to replace your co-worker, actual programmers can write the plan for the code themselves and solve it. Especially for unseen code or new changing syntax for it.

Whenever I see someone promoting another AI tool, I always see who invested and 9/10 of the time it is funded by VCs and ex-FAANG engineers yet again on the snake oil grift. (And they know it but will never admit it.)