Regardless, whatever memory engines people come up with, it's not in anyone's interest to have the memory layer sitting on Anthropic or Open AIs server. The memory layer should exist locally, with these external servers acting as nothing else but LLM request fulfillment.

Now, we'll never be able to educate most of the world on why they should seek out tools that handle the memory layer locally, and these big companies know that (the same way they knew most of the world would not fight back against data collection), but that is the big education that needs to spread diligently.

To put it another way, some games save your game state locally, some save it in the cloud. It's not much of a personal concern with games because what the fuck are you really going to learn from my Skyrim sessions? But the save state for my LLM convos? Yeah, that will stay on my computer, thank you very much for your offer.

Isn't the saved state still being sent as part of the prompt context with every prompt? The high token count is financially beneficial to the LLM vendor no matter where it's stored.

The saved state is sent on each prompt, yes. Those who are fully aware of this would seek a local memory agent and a local llm, or at the very least a provider that promises no-logging.

Every sacrifice we make for convenience will be financially beneficial to the vendor, so we need to factor them out of the equation. Engineered context does mean a lot more tokens, so it will be more business for the vendor, but the vendors know there is much more money in saving your thoughts.

Privacy-first intelligence requires these two things at the bare minimum:

1) Your thoughts stay on your device

2) At worst, your thoughts pass through a no-logging environment on the server. Memory cannot live here because any context saved to a db is basically just logging.

3) Or slightly worse, your local memory agent only sends some prompts to a no-logging server.

The first two things will never be offered by the current megacapitalist.

Finally, the developer community should not be adopting things like Claude memory because we know. We’re not ignorant of the implications compared to non-technical people. We know what this data looks like, where it’s saved, how it’s passed around, and what it could be used for. We absolutely know better.

This sounds similar to Proton's Lumo