The proposed design in this paper is bad, but the core of the idea is very interesting.
At a high-level, 90% of the complexity of their data retrieval system can be deleted by simply having attaching a `CLAUDE.md` file to every data store that is automatically kept up to date the agents can read.
High-throughput queries by an agent don't feel much different than high-throughput querying that large scale systems Instagram and Youtube need to service on a daily basis. Whatever works for 10M active users per second on IG would also work for 50 agents making 1M queries per second.
I can see a need for innovation in data store still. My little startup probably can't afford the same AWS bill than Meta but the tide would lift all boats, not just AI-specific use cases.