fennecbutt
a day ago
I find this topic very interesting because it's something I've run into and mitigated ever since gpt3 was available.
Plenty of long, loooong and complex role plays, world building and tests to see if I could integrate dozens of different local models into a game project or similar.
All of the same issues there apply here for "agents" as well.
Very quickly learn that even current models are like distracted puppies. Larger models seem to be able to brute force their way through some of these problems but I wouldn't call that sustainable.
mberlove
7 hours ago
Sometimes it seems you can "remind" the more established models, and this will bring the context back into focus (just from personal experience) but why that would work, I can only guess.
What methods have you found to brute-force through the problem?