'ability to reason' implies that LLMs are building a semantic model from their training data, whereas the simplest explanation for their behavior is that they are building a syntactic model (see Plato's Cave). Thus without new training they cannot 'learn', RAG or no RAG.

We have multiple threads of research demonstrating in-context learning, friend.

https://github.com/dqxiu/ICL_PaperList