miki123211
5 days ago
I think how the authors of this post think about "AI agents" is really interesting.
Most people think of agents like they think of human employees. They set up a limited number of agents to run in parallel (often just one), with each agent running in a loop and doing one task at a time. They're still in a world where you have a fixed (on the timescale of hours or days) number of employees, each employee can only do one thing at a time, and transferring tasks between employees is slow and costly.
LLMs don't really work like that. You effectively have an infinite number of agents that you can conjure out of thin air at any time. There's no cost advantage to performing LLM requests in series rather than in parallel.
If you realize this, the pattern of each agent fanning out and forking itself into as many sub-agents as are needed to fulfill the task becomes obvious. This is exactly what the authors have done.
I think a better way to think of agents is as "tasks" or "jobs", like those you might find in Celery or sidekik, and apply the learnings from those.
neom
5 days ago
For fun last month I decided to see if i could build a fully functional business of agents. It's 175 python files (employees) build up of internal roles within those files (tasks). So what I have is 175 employees who are able to pass output around each other, understand the work, complete the work, understand where to send the output. The whole system has the ability to do around 275 base processes (same as a business at > 100MM arr) I started on a Friday afternoon and slept a little bit and finished on Monday afternoon. After I had it running I sent it to a VC friend to show them and they sent back the deck of a startup that is in stealth with $25MM doing it the exact same way. With 1 month and a designer and an engineer, I could have it mvp functional for anyone to use ($40k?). Times are changing. Here is kinda how it looks: https://s.h4x.club/9ZuO4XQR / https://s.h4x.club/jkuB8ZED (I've evolved it a little since this, and if you're an engineer and look at my files and think, this guy is a moron: I know!:))
yusina
5 days ago
> understand the work
LLMs don't understand. It's mind-boggling to me that large parts of the tech industry think that.
Don't ascribe to them what they don't have. They are fantastic at faking understanding. Don't get me wrong, for many tasks, that's good enough. But there is a fundamental limit to what all this can do. Don't get fooled into believing there isn't.
motorest
5 days ago
> LLMs don't understand. It's mind-boggling to me that large parts of the tech industry think that.
I think you might be tied to a definition of "understanding" that doesn't really apply.
If you prompt a LLM with ambiguous instructions, it requests you to clarify (i.e., extend prompt to provide more context) and once you do the LLM outputs something that exactly meets the goals of the initial prompt, does it count as understanding?
If it walks like a duck and quacks like a duck, it's a duck,or something so close to a duck that we'd be better off calling it that.
acchow
5 days ago
> If you prompt a LLM with ambiguous instructions, it requests you to clarify (i.e., extend prompt to provide more context)
It does not understand that it needs clarification. This behavior is replicated pattern
Spivak
5 days ago
So you have two prompts, one is ambiguous and the second is the same prompt but with the ambiguity resolved.
In the first prompt the replicated pattern is to ask for clarification, in the second prompt the replicated pattern is to perform the work. The machine might understand nothing but does it matter when it responds appropriately to the different cases?
I don't really care whether it understands anything at all, I care that the machine behaves as though it did have understanding.
motorest
5 days ago
> So you have two prompts, one is ambiguous and the second is the same prompt but with the ambiguity resolved.
No. You have an initial prompt that is vague, and then you have another prompt that is more specific.
- "draw me an automobile"
- "here's a picture of an ambulance."
- "could you make it a convertible instead? Perhaps green."
- "ok, here's a picture of a jaguar e-type".
AlecSchueler
5 days ago
What is the difference? What would actual understanding look like?
intended
5 days ago
Your Question is an example of the difference.
Your question can be rephrased to “what would an actual difference look like.”
However, what you are asking underneath that, is a mix of “what is the difference” and “what is the PRACTICAL difference in terms of output”
Or in other words, if the output looks like what someone with understanding would say, how is it meaningfully different.
—-
Humans have a complex model of the world underlying their thinking. When I am explaining this to you, you are (hopefully) not just learning how to imitate my words. You are figuring out how to actually build a model of an LLM, that creates intuitions / predictions of its behavior.
In analogy terms, learning from this conversation, (understanding) is to create a bunch of LEGO blocks in your head, which you can then reuse and rebuild according to the rules of LEGO.
One of the intuitions is that humans can hallucinate, because they can have a version of reality in their head which they know is accurate and predicts physical reality, but they can be sick/ill and end up translating their sensory input as indicating a reality that doesn’t exist. OR they can lie.
Hallucinations are a good transition point to move back to LLMs, because LLMs cannot actually hallucinate, or lie. They are always “perceiving” their mathematical reality, and always faithfully producing outputs.
If we are to anthropomorphize it back to our starting point about “LLMs understand”, this means that even when LLMs “hallucinate” or “lie”, they are actually being faithful and honest, because they are not representing an alternate reality. They are actually precisely returning the values based on the previous values input into the system.
“LLMs understand” is misleading, and trojans in a concept of truth (therefore untruth) and other intuitions that are invalid.
—-
However, understanding this does not necessarily change how you use the LLMs 90% of the time, it just changes how you model them in your head, resulting in a higher match between observer reality and your predictive reality.
For an LLM this makes not difference, because its forecasting the next words the same way.
accCer
5 days ago
It depends on which human feedback was used to train the model. For humans, there are various communication models like the four-sides model. If the dataset has annotations for the specific facets of the communication model, then an LLM trained on this dataset will have specific probabilities that replicate that communication model. You may call this understanding what the prompter says, but it's just replication for me.
vinnymac
5 days ago
This isn’t a complete answer, but my short list for moving the tech many steps forward would be:
* replying with “I don’t know” a lot more often
* consistent responses based on the accessible corpus
* far fewer errors (hallucinations)
* being able to beat Pokémon reliably and in a decent time frame without any assistance or prior knowledge about the game or gaming in general (Gemini 2.5 Pro had too much help)
bandrami
5 days ago
The difference comes when it receives novel input
kenjackson
5 days ago
What is novel input?
spartanatreyu
4 days ago
Something it hasn't seen in it's training before.
Such as, trying to get an AI to create a design mockup of mastodon but it keeps falling over at laying out the page correctly: https://social.coop/@scottjenson/114593227688326501
intended
5 days ago
> If it walks like a duck and quacks like a duck, it's a duck,or something so close to a duck that we'd be better off calling it that.
Saying “LLMs match understanding well enough”, is to make the same core error if we were to say “rote learning is good enough” in a conversation about understanding a subject.
The issue is that they can pass the test(s), but they dont understand the work. This is the issue with a purely utilitarian measure of output.
baobun
5 days ago
I think most of us agree with Searle that a Chinese room does not understand Chinese.
anorwell
5 days ago
Nor does a neuron.
Argumentum ad populum, I have the impression that most computer scientists, at least, do not find Searle's argument at all convincing. Too many people for whom GEB was a formative book.
GoatInGrey
5 days ago
I don't believe the user meant "understand" in the classical biological and philosophical sense, or were otherwise attempting to anthropomorphize the systems. They were speaking from the practical experience of "this thing takes a somewhat ambiguous input with unique constraints and implements the ask more-or-less as intended".
squidbeak
5 days ago
They understand. Anything able to reason about any arbitrary request and form a plan tailored to that request understands well enough to qualify for the verb. The mechanism behind it may feel hollow or fake. But if its responses reliably show understanding, the LLM understands - by any ordinary measure.
intended
5 days ago
Rote learning is a term that exists which specifically punctures this output oriented measurement of understanding.
hayst4ck
5 days ago
Nearly every argument like this has the same fatal flaw, and it's generally not the critique of the AI, but the critique reflected back on to humans.
Humans also don't understand and are frequently faking understanding, which for many tasks is good enough. There are fundamental limits to what humans can do.
The AI of a few months ago before OpenAI's sycophancy was quite impressive, less so now which means it is being artificially stunted so more can be charged later. It means privately it is much better than what is public. I can't say it "understands," but I can say it outclasses many many humans. There are already numbers of tasks based around understanding where I would already choose an LLM over a human.
It's worth looking at bloom's taxonomy (https://en.wikipedia.org/wiki/Bloom%27s_taxonomy): In the 2001 revised edition of Bloom's taxonomy, the levels were renamed and reordered: Remember, Understand, Apply, Analyze, Evaluate, and Create. In my opinion it is at least human competitive for everything but create.
I used to be very bearish on AI, but if you haven't had a "wow" moment when using one, then I don't think you've tried to explore what it can do or tested it's limits with your own special expertise/domain knowledge, or if you have then I'm not sure we're using the same LLMs. Then compare that experience to normal people, not your peer groups. Compare an LLM to people into astrology, crystal healing, or homeopathy and ask which has more "understanding."
roryirvine
5 days ago
I do agree with you - but the big difference is that humans-who-are-faking-it tend to learn as they go so might, with a bit of effort, be expected to understand eventually.
Does that actually matter? Probably not for many everyday tasks...
yusina
5 days ago
Um, moving the goal post?
The claim was LLMs understand things.
The counter was, nope, they don't. They can fake it well though.
Your argument now is, well humans also often fake it. Kinda implying that it means it's ok to claim that LLMs have understanding?
They may outclass people in a bunch of things. That's great! My pocket calculator 20 years also did, and it's also great. Neither understands what they are doing though.
neom
5 days ago
It's fun to talk about, but personally he whole "understanding" debate is a red herring, imo what we actually care about when we talk about intelligence is the capacity and depth of: second order thinking, regardless of the underlying mechanism. I think personally key question isn't "do LLMs understand?" but, "can LLMs engage in second order thinking?" The answer seems to be yes - they can reason about reasoning, plan their approaches, critique their own outputs, and adapt their strategies, o1 has shown us that with RL and reasoning tokens you can include it in a single system, but our brains have multiple systems we can control and that can be combined in multiple ways at any given moment: emotions, feelings, thoughts combined into user space, 3 core systems input, memory, output. The nuances is in the fact that various reasons (nature + nurture), various humans appear to have varying levels of meta control over the multiple reasoning systems.
perching_aix
5 days ago
Why are you pretending to be participating in a debate? You mention things like "moving the goalpost", "counter[arguments]", and "arguments", as if you did anything more than just assert your opinion in the first place.
This is what you wrote:
> LLMs don't understand.
That's it. An assertion of opinion with nothing else included. I understand it sucks when people feel otherwise, but that's just kinda how this goes. And before you bring up how there were more sentences in your comment, I'd say they are squarely irrelevant, but sure, let's review those too:
> It's mind-boggling to me that large parts of the tech industry think that.
This is just a personal reporting of your own feelings. Zero argumentational value.
> Don't ascribe to them what they don't have.
A call for action, combined with the same assertion of opinion as before, just rehashed. Again, zero argumentational value.
> They are fantastic at faking understanding.
Opinion, loaded with the previous assertion of opinion. No value add.
> Don't get me wrong, for many tasks, that's good enough.
More opinion. Still no arguments or verifiable facts presented or referenced. Also a call for action.
> But there is a fundamental limit to what all this can do.
Opinion, and a vague one at that. Still nothing.
> Don't get fooled into believing there isn't.
Call for action + assertion of opinion again. Nope, still nothing.
It's pretty much the type of comment I wish would just get magically filtered out before it ever reached me. Zero substance, maximum emotion, and plenty of opportunities for people to misread your opinions as anything more than that.
Even within your own system of opinions, you provide zero additional clarification why you think what you think. There's literally nothing to counter, as strictly speaking you never actually ended up claiming anything. You just asserted your opinion, in its lonesome.
This is no way to discuss anything, let alone something you or others likely feel strongly about. I've had more engaging, higher quality, and generally more fruitful debates with the models you say don't understand, than anyone here so far could have possibly had with you. Please reconsider.
hayst4ck
5 days ago
> higher quality, and generally more fruitful debates with the models you say don't understand
My favorite thing about LLMs is that they can convincingly tell me why I'm wrong or how I could think about things differently, not for ideas on the order of sentences and paragraphs, but on the order of pages.
My second favorite thing is that it is amazingly good at deconstructing manipulative language and power tactics. It is scary good at developing manipulation strategies and inferring believable processes to achieve complex goals.
perching_aix
4 days ago
Had some success with that myself as well. Also found out about Claimify [0] recently, I should really get myself together and get a browser extension going one of these days. I think the quantized gemma3 models should be good enough for this, so it could remain all local too.
yusina
5 days ago
So, it is your opinion that the mere expression of opinion "without anything else" is not allowed in a discussion?
And if that is so, didn't you also "just" express an opinion? Would your own contribution to the discussion pass your own test?
You might have overlooked that I provided extensive arguments all around in this thread. Please reconsider.
perching_aix
4 days ago
> So, it is your opinion that the mere expression of opinion "without anything else" is not allowed in a discussion?
This is not what I said, no: I said that asserting your opinion over others' and then suddenly pretending to be in a debate is "not allowed" (read: is no way to have a proper discussion).
A mere expression of opinion would have been like this:
> [I believe] LLMs don't understand.
And sure, having to stick an explicit "I think / I believe" everywhere is annoying. But it became necessary, when all the other things you had to say continued to omit this magic phrase, and it became clearly intentionally not present, when you started talking as if you made any arguments of your own. Merely expressing your opinion is not what you did, even when reading it charitably. That's my problem.
> Would your own contribution to the discussion pass your own test?
And so yes, I believe it does.
> You might have overlooked that I provided extensive arguments all around in this thread. Please reconsider.
I did consider this. It cannot be established that the person whose comment you took a whole lot of issue with also considered those though, so why would I do so? And so, I didn't, and will not either. Should I change my mind, you'll see me in those subthreads later.
yusina
4 days ago
I challenge you to go through the history of your own posts and count how often you salt your statements of opinion with the magic dust of "I believe".
I did. You are not living up to the standard you are demanding of others (and which rarely anybody around here satisfies anyway).
Seems we are not getting anywhere. We can agree to disagree, which I'm fine with. Please refrain from personal attacks going forward, thank you.
perching_aix
4 days ago
> I challenge you to go through the history of your own posts and count how often you salt your statements of opinion with the magic dust of "I believe".
Challenge semi-accepted [0]. Looking through my entire comment history here so far on this wonderful forum (628 comments), there seem to be 179 hits for the word "think" and 21 for the word "believe". If we're being nice and assume these are all in separate comments, that would mean up to ~32% of my comments feature these words, and then only some portion of these will actually pertain to me guarding my own opinions with them. Still, feeling pretty chuffed about it if I'm honest, I think I'm doing pretty good.
For good measure, I also checked against your comment history of 100 comments. 2 counts of "believe", 9 counts of "think". Being nice here only yields us up to 11%, and focusing on expressions of opinion would only bring this down further.
That said, I think this is pretty dumb. [1]
> I did. You are not living up to the standard you are demanding of others (and which rarely anybody around here satisfies anyway).
Please do show me the numbers you got and your methodology. (And not from the research you're going to do after reading this comment - although if it's done actually proper, I'm interested in that too.)
> Seems we are not getting anywhere.
If only you put as much effort into actually considering what I wrote as you did into stalking my comment history or coming up with new fallacies and manipulation tactics, I think we would have.
Seriously:
- not being able to put it into words how you don't think LLMs understand is perfectly normal. You could have embraced this, but instead we're on like level 4 of you doubling down.
- sharing your opinion continues to be perfectly okay. Asserting your opinion over others continues to be super not okay.
- I (or others) don't need to be free of the faults that I described in order for these things to be faults. It's normal to make mistakes. It'd also be normal to just own them, but here I am, exporting my own comment history using the HN API, because you just can't acknowledge having been wrong and not defending it, even though reading between the lines you do seem to agree with basically everything I said, and are just trying to give me a rhetorical checkmate at this point.
> Please refrain from personal attacks going forward, thank you.
Tried my best. For real; I rewrote this like 6 times.
[0] You continue to heavily engage in manipulative language and fallacies, so I feel 100% uncompelled to honor your "challenge request" proper. I explicitly brought up several other criteria, such as a sentence presenting as an opinion when read in good faith, not being utilized as an accepted shared characterization when used in other sentences, and not being referred to as arguments elsewhere. What you describe as "statements of opinion with the magic dust of "I believe"" seem to intentionally gloss over these criteria, in what I can best describe as just a plain old strawman. So naturally, the challenge was as woefully weakly accepted as I possibly could.
[1] Obviously these statistics are completely bogus, since maybe you just don't offer your opinions much. Considering your performance here so far, this is pretty hard for me to believe, but it is entirely possible and I don't care to manually pore over 100 of your comments, sorry. If they are anything like the ones in this subthread here so far, I've already had more than enough. And if I went through the trouble of automating it ironically involving an LLM, I'd be doing a whole proper job of it at that point anyways, which would go against [0].
squidbeak
5 days ago
Excellently put.
zenburnmyface
5 days ago
meh. I feel this is just a linguistic shortcut, similar to how _trained_ biologists can talk about a species or organism evolving some trait. Of course the organism isn't _really_ evolving with any goal in mind, but that's clear to the speaker and audience. Whether or not LLMs understand (very unlikely), it's clear what we mean by an LLM "understanding": has the context + prior training to make reasonable predictions. But no one wants to write that each time.
yusina
5 days ago
That's an interesting take and in fact one I could get behind.
But I'm afraid that most folks using the term mean it more literally than you describe.
philistine
5 days ago
Exactly. The whole point of all the LLM companies is to get grandma to use it. If you say understand about a technology with the desired appeal of Facebook, then you’re talking to everyone and words matter extra hard.
GaggiX
5 days ago
They understand tho, it's different than how it's done in our brain but they solve task that would be impossible to do without understanding. I would even say that they can now reason through problems thanks to powerful reasoning models like Gemini 2.5 Pro and o3.
bobxmax
5 days ago
How do you know?
yusina
5 days ago
Extraordinary claim requires extraordinary proof. I don't know, but I'm also not the one claiming something.
(Besides, we know what LLMs do, and none of those things indicate understanding. Just statistics.)
shawabawa3
5 days ago
You can create a new game with new rules never seen before
You can explain this to an LLM
The LLM can then play the game following the rules
How can you say it hasn't understood the game?
intended
5 days ago
A Person who memorizes something by rote, can pass many tests. From a test and verifiability perspective, they cannot be distinguished from someone who understands a subject.
An LLM can pass many tests, it is indistinguishable from someone who understands the subject.
Indistinguishable does not imply that the processes followed match what a human is doing when it understands a subject.
I use this when I think of humans learning - humans learn the most when they are playing. They try new things, explore ideas and build a mental model of what they are playing with.
To understand something, is to have a mental model of that thing in ones head.
LLMs have models of symbol frequency, and with their compute, are able to pass most tests, simply because they are able to produce chains of symbols that build on each other.
However, similar to rote learning, they are able to pass tests. Not understand. The war is over the utilitarian point “LLMs are capable of passing most tests”, and the factual point “LLMs dont actually understand anything”.
This articulation of the utilitarian point is better than the lazier version which says “LLMs understand”, and this ends up anthropomorphizing a tool, and creating incorrect intuitions of how LLMs work, amongst other citizens and users.
yusina
5 days ago
The LLM is only capable of doing so if it has encountered something similar before as part of its training.
Claiming anything else requires a proof.
neom
5 days ago
the extraordinary claim would be that LLMs can only do things they've seen before exactly, given the compositional and emergent capabilities we observe. The evidence suggests they can generalize beyond their training in meaningful ways, even if imperfectly...if a human came out living but with a brain that had zero electrical activity, that would be extraordinary, we normally come out with a baseline of pre-programming. I sometimes think this debate happens because humans don't want to admit we're nothing more than LLMs programmed by nature and nurture, human seem to want to be especially special.
https://arxiv.org/abs/2206.07682
https://towardsdatascience.com/enhanced-large-language-model...
https://arxiv.org/abs/2308.00304
(and if MoRA is moving the goal posts, fine: RL/RT)
yusina
5 days ago
>if a human came out living but with a brain that had zero electrical activity, that would be extraordinary, we normally come out with a baseline of pre-programming.
That statement reveals deep deficiencies in your understanding of biological neural networks. "electrical activity" is very different from "pre-programming". Synapses fire all the time, no matter if meaningfully pre-programmed or not. In fact, electrical activity decreases over time in a human brain. So, if anything, programming over time reduces electrical activity (though there is no established causal link).
> I sometimes think this debate happens because humans don't want to admit we're nothing more than LLMs programmed by nature and nurture, human seem to want to be especially special.
It's not specific to humans. But indeed, we don't fully understand how brains of humans, apes, pigs, cats and other animals really work. We have some idea of synapses, but there is still a lot unclear. It's like thinking just because an internal combistion engine is made of atoms, and we mostly know how atom physics and chemistry work, that any body with this basic knowledge of atom physics can understand and even build an ICE. Good luck trying. It's similar with a brain. Yes, synapses play a role. But that doesn't mean a brain is "nothing more than an LLM".
neom
5 days ago
Neural activity begins around 6 weeks gestation, electrical patterns help establish basic neural circuits, activity dependent neural development shapes connectivity before any sensory input, critical periods where electrical activity literally sculpts brain architecture. Motor patterns get programmed before birth (why babies can suck, grasp, etc.), language processing areas develop structural biases before hearing language, visual cortex develops orientation maps before seeing anything, basic learning algorithms get "wired in" through developmental processes. If a human emerged, was able to function in the world, do things, but had zero electrical activity in the brain, that would be... normal? No: extraordinary.
Humans arrive out of the VJJ with innate neural architectures to be filled and developed - not literal blank slates, there is an OS. The electrical activity during development is literally the biological process that creates our "base programming." LLMs have architectural inductive biases (attention mechanisms, etc.), human brains have evolved architectural biases established through fetal development. We're both "pre-programmed" systems, just through different mechanisms.
Your response about "electrical activity decreases over time" is irrelevant - you weren't talking about adult brain activity, you were talking about the developmental process that creates our initial neural architecture.
tbh: I can't tell if you're engaging in good faith or not.
holoduke
5 days ago
Definition of understanding is based on connecting relations. If there is one thing a llm can do its connecting relations. So I am not sure why you say llms are not understanding.
neom
5 days ago
What is the limit my system will reach?
rzz3
5 days ago
Thats an interesting word to pick on. Understanding still means something here in a relative sense.
vasco
5 days ago
Asking a short question but in a serious way: so what?
yusina
5 days ago
You are asking why it is meaningful to use terms for what they mean instead of making up things?
Well, I prefer it that way, but the spirit of "AI" seems to go in another direction, and the leadership of US government also does, so maybe times are just changing.
acchow
5 days ago
> The whole system has the ability to do around 275 base processes
It’s incredibly easy to get LLMs to do a lot of stuff that seems convincing.
They are literally trained for plausibility.
robbomacrae
5 days ago
Is anyone else annoyed that VC's are out there sharing decks of startups in stealth with potential competitors? How often does this happen?
eterm
5 days ago
I would be annoyed along with you if I thought the post was true.
IncreasePosts
5 days ago
It's not a lie, it is just vibe posting
wslh
5 days ago
Your message doesn't make it clear what those 175 employees can realistically accomplish on their own.
For instance, you might have an SEO expert on the team, but that alone won't guarantee top search engine rankings. There are countless SEO professionals and tools (human or AI-powered), and even having the best one doesn't eliminate the underlying challenge: business competition. LLMs, like any other tool, don’t solve that fundamental problem.
neom
5 days ago
No employees accomplish anything on their own in the real world, all employees are part of a team. That's why I designed a business strategy and analysis layer (over half the system, in fact), with web tools and connections to all of the insights systems (like mix panel). I built the exact same thing I build at digitalocean but instead of humans I defined them with code, digitalocean runs just fine, so does my LLM system. The whole system I build is self learning, insight gathering and refinement. Competition is for losers, the best teams win via the best insights.
vasco
5 days ago
Why 175? Why not 5 billion employees? Why not 20000 companies in parallel? Why not simulate 5 earth's worth of history and setup a full universe of worlds full of startups?
This sounds like those guys in social media that one up each other with their bed times and end up saying they wake up every day at 2am to meditate and work out
neom
5 days ago
Because that was the scope of the project. When we got to 400 employees at DigitalOcean I noticed I thought it was really half that, original I just sat out to make the marketing and strategy team, but got bit carried away, the fp&a team was the only group I really struggled with, my cfo skills are very meh.
vasco
5 days ago
1 single agent with a good model is going to beat that approach every single time. The same way Whatsapp needed only 55 people (and probably the last hires were not needed for the outcome) to sell for $19B.
And other companies have existed for hundreds of years and had thousands of people work for them and never even made $100M.
neom
5 days ago
I'm confused what you're saying. There a loads of markets, loads of segments, loads of ways to do unit economics, yes, but business is business, it's prescriptive at it's core. I'm using a single model, it's just openai calls using the role function.
immibis
5 days ago
Does this experiment do anything useful or does it just soak up investor money? Not that there's anything wrong with the latter.
neom
5 days ago
The only investor is me. I build it on my own over a weekend, on my own. I just wanted to confirm it can be done therefore will exist, that is all. Personally, I decided not to peruse it because I am old and lazy and don't want to compete against a16z and sequoia funded adderall filled teenagers.
immibis
5 days ago
I meant the one that investors are paying for.
neom
5 days ago
It professes to be able to do the business side of business (not literal product or technology development) - they did not have any agents to code or design, mine didn't either but has agents that can call tools, I don't think mine can build a product, but I believe mine can build, operate, and grow a business around one, I presume theirs will be able to also.
literalAardvark
5 days ago
Engineers who would judge someone's frontier MVP like that are not even worth worrying about.
This is epic work. Would love to see more of it but I guess you're gonna take it the startup route since you have connections. Best of luck.
neom
5 days ago
Thanks!!! I decided not to build it, that space is already too busy, there is a startup with $25MM in stealth, who else is in stealth? On top of that, this method will get stale very very quickly, foundation model businesses are just too hard to work around right now, it's a silly way to do business. My magic is I've build a startup from scratch to over 400 people and watched what they do, it won't be long till that isn't worth much.
mucha
5 days ago
Cool. What goods/services does your business provide to customers?
neom
5 days ago
Goods and services are a byproduct of business, business is primarily concerned with systems and processes that facilitate value exchange, so my tool, can work with a user, to build a business, not a product or a service. If you bake cupcakes, my tool can get you 100 people at your door, it cannot open the door or provide the cakes.
iammrpayments
5 days ago
Sounds really interesting but I have no idea what’s the purpose of having 175 “employees” here? Maybe it is a smart way to sell the idea you’re going to replace 175 people if you buy the product? Could just buy chatgpt instead I guess, but a chatbot doesn’t sound as cool as 175 employees.
neom
5 days ago
I would love to know how to do it another way if you have any ideas, I'm sadly not experienced or intelligent enough to think of another way to do it.
jprokay13
5 days ago
I’ve been floating around a similar set of ideas and it’s been very fun (if not all that useful yet) to build Did you try taking it one step further where a “recruiter” has to hire the engineers after a screening process? I wonder if this could get you even better AI engineers
catlifeonmars
5 days ago
This really sounds like a “faster horse” scenario and totally misses the point of the GPs comment: why shackle yourself to modeling the way humans work?
neom
5 days ago
Humans modeled time, and morning and night are real.
catlifeonmars
4 days ago
Have you thought about what emergent properties might arise if you weren’t prescriptive about the organizational structure or task specialization?
neom
3 days ago
My thought is, it would be a dysfunctional system, like this: https://futurism.com/professors-company-ai-agents
I was just building a tool people can use to do the business side of product and technology. I wanted it to do basic tasks i KNOW a business needs at all scales, not pretend to do whatever "AI" thinks an employee does.
catlifeonmars
2 days ago
Ah I see. Basically a skeuomorphism for bizops?
neom
a day ago
fuckin' cool description, never would have put skeuomorphism there myself, but it's apt, nice.
londons_explore
5 days ago
> forking itself into as many sub-agents as are needed to fulfill the task
The forking is free. Running the sub-agents is linear cost, but the expensive bit is joining the agents responses back together again.
If a task has 6 subtasks and an agent is spawned for each, at some point some 'joiner' agent needs to parse and summarize the findings of the sub agents and feed it back to the parent. That step necessarily involves information loss, and uses more computation that a single linear agent design would not use.
neom
5 days ago
I designed something for a business and found I needed 4 major sub-systems (like a real business) - insight/data, cognition, meta cognition and execution, and if you don't define all 4, the system is junk.
motorest
5 days ago
> I designed something for a business and found I needed 4 major sub-systems (like a real business) - insight/data, cognition, meta cognition and execution, and if you don't define all 4, the system is junk.
Might it be just another realization of Conway's law?
https://en.wikipedia.org/wiki/Conway%27s_law
Might it be possible that the only reason you're assuming a system is junk is just that it doesn't resemble the systems you know and expect? There are so many ways to skin a cat, and certainly no business process represents the optimal process.
neom
5 days ago
I thought about this for 15 hours. First, I really appreciate how you wrote this comment. It's an extremely well composed comment. I also appreciate your use of the word might. Anyway, I suspect you are probably correct, and I hit the wall again because: I'm just too versed.
yusina
5 days ago
> You effectively have an infinite number of agents
You don't.
Sincerely, Your Electricity Bill
dwohnitmok
5 days ago
> If you realize this, the pattern of each agent fanning out and forking itself into as many sub-agents as are needed to fulfill the task becomes obvious.
And this is precisely how really bad things could happen:
https://www.lesswrong.com/posts/kpPnReyBC54KESiSn/optimality...
TimPC
5 days ago
The challenge with fan out is constructing a linear conversation that makes sense that captures previous history. In any context where the LLM needs that information linear loops often perform better than trying to splice together conversations from multiple parallel processes.
kposehn
5 days ago
This is similar to something we've been doing for a while. Instead of individual agents we are creating many iterations and sub-iterations of spawned agents that are largely autonomous. A lot of the human-centric paradigms just don't really apply to LLMs/AI but people are used to approaching them that way.
viraptor
5 days ago
> They set up a limited number of agents to run in parallel (often just one),
Most of what people use agents for daily can often be one-shotted though and even collating/rating 10 results would be costly.
If I had a harness for evaluating the results and VC level money, I'd be throwing an army at well defined experimental tasks as well.