That AI will have to be significantly preferable to the baseline of open models running on cheap third-party inference providers, or even on-prem. This is a bit of a challenge for the big proprietary firms.

> the baseline of open models running on cheap third-party inference providers, or even on-prem. This is a bit of a challenge for the big proprietary firms.

It’s not a challenge at all.

To win, all you need is to starve your competitors of RAM.

RAM is the lifeblood of AI, without RAM, AI doesn’t work.

Assuming high bandwidth flash works out, RAM requirements should be drastically reduced as you'd keep the weights in much higher capacity flash.

> Sample HBF modules are expected in the second half of 2026, with the first AI inference hardware integrating the tech anticipated in early 2027.

https://www.tomshardware.com/tech-industry/sandisk-and-sk-hy...