I just stopped my Groq API. Sad to see competition being eaten up by shitty Nvidia. I like their products but Jensen is an absolute mfer with deceitful marketing.
I just stopped my Groq API. Sad to see competition being eaten up by shitty Nvidia. I like their products but Jensen is an absolute mfer with deceitful marketing.
I literally said “oh no” out loud when I read the headline.
As a relative laymen in this hardware inference space, I am curious what exactly was Groq useful for vs. the typical hardware architecture? Or was this a “step in before they become more generally useful” situation for Nvidia/Groq?
Much faster responses, before this deal I thought it would be Google vs Groq for the superior tech with nvidea missing out.
was the API good
fast but furiously expensive
Honestly cerebras is good. I can recommend it. I talked to their team once on discord as a literal free user so that was something really nice personally
They were also the faster one compared to groq but they were always a little slow on adding new models compared to groq but not sure what changed right now.
Definitely recommend cerebras tho now that groq's been eaten up from inside basically
I had the feeling that Cerebras only supports smaller modes. Maybe something to do with their hardware arch? I never dove into it. I wanted to use Kimi K2 fast for coding and Groq was the only fast provider at the time
Cerebras currently has GLM4.6 on it, and will be getting GLM4.7 soon.