Plenty of long, loooong and complex role plays, world building and tests to see if I could integrate dozens of different local models into a game project or similar.
All of the same issues there apply here for "agents" as well.
Very quickly learn that even current models are like distracted puppies. Larger models seem to be able to brute force their way through some of these problems but I wouldn't call that sustainable.
What methods have you found to brute-force through the problem?
rorylaitila•7mo ago
fennecbutt•7mo ago
It's just beyond playing with more heads, specialised heads, kv caching etc it doesn't seem like anybody's figured out the next step here yet.
Attention is already pretty atrocious perf even with caching so additional context metadata would have to be implemented carefully.