You're absolutely correct! Having the LLM using more tokens does improve its output. Here's why this works:

## More tokens = smarter outputs

When an LLM uses tokens, it is putting more information into its context

## Better context, better results

The more information the LLM has in its context, the more complete and well thought-through the outputs will be

## More complete thinking

When an LLM is able to iterate on itself, results improve

## Better shareholder value

Numbers need to go up in order for us to maintain our shareholder value. This means instead of focusing on results that are qualitative, instead the brand should focus on quantitative, hard results