Why do so many LLMs struggle with memory?
Why do so many LLMs struggle with memory?

Why do so many LLMs struggle with memory?

Hello! Hoping this is the right sub to ask. I like to play with AI models, mainly as chat bots. They're fun, they're very human-like, they are overall wayyyy beyond anything I would've expected even 10 years ago.

But their memory is atrocious. Various companies seem to be rolling out improvements, but it's still not good. Which seems bizarre to me. The entire chat history I have with the bot is probably a handful of kB, certainly not a super intensive thing to store or even to hold in RAM.

So, what gives? These bots can understand metaphor, make jokes, and pick up on implied meaning, but have the long-term memory of a concussed goldfish. It's exactly the opposite of what I would expect from a digital tool. It's fascinating. What's the reason for it, on the technical level?

submitted by /u/Blue-Jay27
[link] [comments]