Are we sure people don't work that way? Almost all of us operate on instincts almost all of the time. We have guardrails, people who operate them are often committed to institutions. When we choose to do things, it is based on that static hardwiring. Our meta model later comes up with reasons why we did the things. Sometimes, but rarely, it is correct. The human brain is extremely heterogenous, modular even. Some of our modules function remarkably like a memory store fed back into a context window. Adding a meta model to an llm that is updated autonomously by an additional model that analyzes outcomes to upde this predictive meta model would quite likely result in the agent's models mistaking the meta model for a self. Much like we do.