I just stopped my Groq API. Sad to see competition being eaten up by shitty Nvidia. I like their products but Jensen is an absolute mfer with deceitful marketing.

I literally said “oh no” out loud when I read the headline.

As a relative laymen in this hardware inference space, I am curious what exactly was Groq useful for vs. the typical hardware architecture? Or was this a “step in before they become more generally useful” situation for Nvidia/Groq?

Much faster responses, before this deal I thought it would be Google vs Groq for the superior tech with nvidea missing out.

was the API good

fast but furiously expensive

Honestly cerebras is good. I can recommend it. I talked to their team once on discord as a literal free user so that was something really nice personally

They were also the faster one compared to groq but they were always a little slow on adding new models compared to groq but not sure what changed right now.

Definitely recommend cerebras tho now that groq's been eaten up from inside basically

I had the feeling that Cerebras only supports smaller modes. Maybe something to do with their hardware arch? I never dove into it. I wanted to use Kimi K2 fast for coding and Groq was the only fast provider at the time

Cerebras currently has GLM4.6 on it, and will be getting GLM4.7 soon.