Consider it an LLM cache. The result has already been cached so you don't have to generate it again.