Is their overall margin also about 60% too? Or something saner like 30%?

Their overall margin is negative.

No, it’s not. This is a dangerous perspective, usually held by engineers who think that accounting doesn’t matter and don’t understand it.

You MUST accrue the lifetime value of the assets against the capital expense (R&D in this case) to determine the answer to this question.

The company (until this announcement) had raised $17B and has a $14B revenue rate with 60% operating margin.

It is only negative on margin if you assume the prior 14B (e.g. Claude 4.6 plus whatever’s unreleased) will have no value in 24 months. In that case, well, they probably wasted money training.

If you think their growth rate will continue, then you must only believe the models have a useful 9 months or so life before they are break even.

Anthropic is, according to Dario, profitable on every model <<—- they have trained if you consider them individually. You would do best to think “will this pattern continue?”

What is the lifetime value of an individual pretraining run, and what is the cost to do it? Whether it is a net positive seems to still be an open question.

Actually there is a chart of answers to this question, because the frontier providers have been delivering new models for some time. The answer is that so far they have been net positive.

Sorry - if a model costs (say) 20B to train, lasts 12 months before it becomes obsolete, generates 2B/month revenue, but with 1B/month inference costs, then it has lost 8B.

Or are you suggesting that in fact each model comes out ahead over its lifespan, and all this extra cash is needed because the next model is so much more costly to train that it is sucking up all the profits from the current, but this is ok because revenue is expected to also scale?

I'm not suggesting, I'm saying that this is what has happened so far, full stop, based on multiple public statements from people like Dario.

Basically every model trained so far has made money for Anthropic and OpenAI. Well maybe not GPT4.5 - we liked you but we barely knew thee..

The cash spend is based on two beliefs a) this profitability will continue or improve, and b) scaling is real.

Therefore, rational actors are choosing to 2-10x their bets in sequence, seeing that the market keeps paying them more money for each step increase in quality, and believing that either lift off is possible or that the benefits from the next run will translate to increased real cash returns.

What's obscure to many is that these capital investments are happening time shifted from model income. Imagine a sequence of model training / deployments that started and finished sequenced: Pay $10m, make $40m. Pay $100m, make $400m. Pay $1bn, make $4bn. Pay $10bn, (we are here; expectation is: make $40bn).

If you did one of those per year, the company charts would look like: $30m in profits, $300m in profits, $3bn in profits. And in fact, if you do some sort of product-based accrual accounting, that's what you would see.

Pop quiz, if you spend in the first month your whole training budget, and the cycles all start in November, what would the cash basis statement look like for the same business model I just mentioned?

-$10m, -$60m, -$600m, $-6bn.. This is the same company with different accounting periods.

Back in reality, shortly into year 1, it was clear (or a hopeful dream) that the next step (-100 / +400) was likely, and so the company embarked on spending that money well ahead of the end of the revenue cycle for the first model. They then did it again and again. As a result naive journalists can convince engineers "they've never made money". Actually they've made money over and over and are making more and more money, and they are choosing to throw it all at the next rev.

Is it a good idea or not to do that is a question worth debating. But it's good to have a clear picture of the finances of these companies; it helps explain why they're getting the investment.