Is this why they’re trying to obtain so damn much memory components? So they can ship the entire data relied upon by the statistical models to consumers?
I really didn’t think they could be more wasteful but clearly I was wrong, they want to eat up bandwidth like a fish drinks water. Or maybe they want to actually SHIP the DATABASE?
Here is what will change if your AI could remember better: nothing. The LLMs are statistical models. The next most likely token based on the sample data would rarely if ever change based on a new prompt being memorized. Normally, the models need to be retrained whenever a change is made to the data, if you could bypass that you will probably lose accuracy but even if you didn’t then you still won’t gain any tangible accuracy.
Is this why they’re trying to obtain so damn much memory components? So they can ship the entire data relied upon by the statistical models to consumers?
I really didn’t think they could be more wasteful but clearly I was wrong, they want to eat up bandwidth like a fish drinks water. Or maybe they want to actually SHIP the DATABASE?
Here is what will change if your AI could remember better: nothing. The LLMs are statistical models. The next most likely token based on the sample data would rarely if ever change based on a new prompt being memorized. Normally, the models need to be retrained whenever a change is made to the data, if you could bypass that you will probably lose accuracy but even if you didn’t then you still won’t gain any tangible accuracy.