My opinion is that unless there is some insane breakthrough in power efficiency with video generation, or if energy costs go down to zero, there is no way such a thing actually becomes profitable at the scale of scrolling TikTok. It is far more power efficient (and cheaper) to have people post their own content.

I think you're overestimating how much power LLMs consume. Let's say one video pegs a top of the line Blackwell chip at 100% utilization for 10 minutes. I think a Blackwell chip (plus cooling and other data center overhead) is somewhere around 3000 watts when running 100%. So that's about 0.5 kilowatt-hours. I suspect this is a severe overestimate because there's probably a significant amount of batching that happens that cuts down on amortized power usage, and non-pro Sora 2 might be processed with weaker, smaller models, but I'm not very confident.

Data centers seem to have wholesale rates of around 4 cents per kilowatt-hour on the higher end.

This gets you 2 cents per video. If you're generating 50 million videos per day (an estimate on the higher side of how many TikTok videos are uploaded every day), that costs you a million dollars a day.

So if you entirely subsidized for free the entirety of all of TikTok's video generation just using LLMs, I don't think energy generation costs exceed 365 million a year (and I think this might be very severely estimating costs, but there are some large error bars here).

I'm pretty sure OpenAI (or any company) would be pretty happy to pay 365 million dollars a year for the soft social power of something like TikTok. Just the influence this buys in politics and social discourse would be worth the pricetag alone.

And that's of course leaving aside any form of monetization whatsoever (where in reality you'd likely be charging the heaviest users the most).

N.B. I'm also not sure it's actually more power efficient for users to post their own content in absolute terms. It seems not unlikely that the amount of energy it takes to produce, edit, and process a TikTok video exceeds half a kilowatt-hour. But maybe you're focused solely on the video hoster.

> It seems not unlikely that the amount of energy it takes to produce, edit, and process a TikTok video exceeds half a kilowatt-hour.

That would be really remarkable, considering the total power capacity of a phone battery is in the neighborhood of 0.015 kWh

Yeah I should clarify. This is a very vague estimate around "total energy spent for making a video you wouldn't otherwise do" which includes stuff like lighting, transportation, video transcoding on the server, script writing, actor coordination, etc. E.g. if someone drives somewhere solely to make a video they otherwise wouldn't, then it gets included.

I hedged as "not unlikely" because I'd need to think harder about the amortization of more energy expensive videos vs less energy expensive ones and how much energy you can actually attribute to a video vs the video solely being an activity that would be an add-on to something that would happen anyways.

But it's not just the energy expenditure of a phone.

(I also think that 0.5 kilowatt-hours is an overestimate of energy expenditure by potentially up to two orders of magnitude depending on how much batching is done, but my original comment did say 0.5 kWh).

[deleted]

> Let's say one video pegs a top of the line Blackwell chip at 100% utilization for 10 minutes.

Where do you get this from?

You didn't include the amortized cost of a Blackwell GPU, which is an order of magnitude larger expense than electricity.

Yeah that's fair (although the original comment was only talking about energy costs).

But this is kind of a worst case cost analysis. I fully expect that the average non-pro Sora 2 video has one to two orders of magnitude less GPU utilization than I listed here (because I think those video tokens are probably generated at a batch size of ~100 per batch).

I'm not sure you need a 'breakthrough', just many incremental improvements will do the trick.

We are also getting better at producing cheap power. For example thanks to intermittent sources like solar and wind, in many places electricity often becomes free in wholesale markets some times of the day.

AI generation (including video) currently takes at least a second, and users expect that delay. So that means inference is not latency sensitive and you can put these data centres anywhere in the world, wherever power is cheapest. Model training cares even less about latency.

At the moment, the hardware itself is too expensive (and nvidia has long backlogs), so people run them even when power is expensive. But you can easily imagine an alternative future where power never becomes cheaper than today (on average), but we have lots of AI data centres lying in wait around the world and only kicking into full gear when and where power is essentially free.

We are not getting better at producing cheaper power as the cost has increased per hour a lot over the last 50 years. But we are generating more power from different sources that are cleaner.

Power needs to be given away or people paid to take it is more of a function of limited storage abilities and limited ability to scale down rather then generating unlimited power. The free power is an issue with how the system is built (and the type of power source) rather than a sign of success. The same area has to import power at higher costs when the sun or wind isn't as powerful.

> Power needs to be given away or people paid to take it is more of a function of limited storage abilities and limited ability to scale down rather then generating unlimited power.

There's no need to scale down solar or wind power.

Yes, storage is another way to make money from power prices that differ over time.

> [...] the cost has increased per hour a lot over the last 50 years.

Some sources of power, like solar, have been dropping in price a lot recently.

There have been some big breakthroughs with hardware, though I'm on mobile and can't provide a link currently. I expect it to take done time to get into production though.

Also I suspect that this won't stay free very long. Silicon valley loves the model of starting free to build a user base and then monetizing more later