Ah so obviously making the LLM repeat itself three times for every response it will get smarter

Yes, and observe that people do that too. It gives them more time to notice their own confusion and go "but wait, that's not right" on you.

You're absolutely correct! Having the LLM using more tokens does improve its output. Here's why this works:

## More tokens = smarter outputs

When an LLM uses tokens, it is putting more information into its context

## Better context, better results

The more information the LLM has in its context, the more complete and well thought-through the outputs will be

## More complete thinking

When an LLM is able to iterate on itself, results improve

## Better shareholder value

Numbers need to go up in order for us to maintain our shareholder value. This means instead of focusing on results that are qualitative, instead the brand should focus on quantitative, hard results