I can only hope that Cerebras is able to keep their first party inference product going. It’s incredible to run a strong model at interactive latencies for whole results. Routinely less than seconds to product entire files / documents / outputs / …

https://cloud.cerebras.ai/