You can quite easily ask it to summarize the result in a sentence or paragraph. LLMs have no other way to compute than write text and the more text they write the more compute they do. You only care about the final output.
You can quite easily ask it to summarize the result in a sentence or paragraph. LLMs have no other way to compute than write text and the more text they write the more compute they do. You only care about the final output.
I do. It typically goes on to write a preamble about why it gave a long answer before finally providing a summary. Token stuffing.