Related check out chain of draft if you haven't.
Similar performance with 7% of tokens as chain of thought.
https://arxiv.org/abs/2502.18600
That's a comparison to "CoT via prompting of chat models", not "CoT via training reasoning models with RLVR", so it may not apply.
That's a comparison to "CoT via prompting of chat models", not "CoT via training reasoning models with RLVR", so it may not apply.