This is silly? The important metric is value per token, which is obviously increasing, and thus the relative token is getting cheaper because you need far less of them to produce anything of value.
Which then might lead to you using a lot more, because it offsets some other thing that costs even more still, like your time.
"Which is obviously increasing"
With the primary advancements over the past two years being Chain Of Thought which absolutely obliterates token counts in what world would the "per token" value of a model be going up...
If you are able to cogently explain how you would instruct GPT 3.5 with ANY amount of tokens to do what Sonnet 4 is able to do, I am sure there's a lot of wealthy people that would be very interested in having a talk with you.
You build a reasoning variant of GPT 3.5. Please deposit 1 billion dollars in my bank account for this advice.