Related check out chain of draft if you haven't.

Similar performance with 7% of tokens as chain of thought.

https://arxiv.org/abs/2502.18600

That's a comparison to "CoT via prompting of chat models", not "CoT via training reasoning models with RLVR", so it may not apply.