You're absolutely correct! Having the LLM using more tokens does improve its output. Here's why this works:
## More tokens = smarter outputs
When an LLM uses tokens, it is putting more information into its context
## Better context, better results
The more information the LLM has in its context, the more complete and well thought-through the outputs will be
## More complete thinking
When an LLM is able to iterate on itself, results improve
## Better shareholder value
Numbers need to go up in order for us to maintain our shareholder value. This means instead of focusing on results that are qualitative, instead the brand should focus on quantitative, hard results