Wouldn't this affect quality of output negatively?

Thanks to chain of thought, actually having the LLM be explicit in its output allows it to have more quality.

Chain of thought happens in the <think> tags, not the visible output.

Caveman only strips filler from what you see... the reasoning depth stays the same.

I found this visualisation pretty interesting - https://vectree.io/c/chain-of-thought-reasoning-how-llms-thi...