If China keeps releasing decent copies of SOTA models that only take 20% of the resources, then we may get some relief when those models become "good-enough"
If China keeps releasing decent copies of SOTA models that only take 20% of the resources, then we may get some relief when those models become "good-enough"
>copies of SOTA models that only take 20% of the resources
They might be 20% of the price (because they don't have to invest that much in training), but are probably not 20% of the resources (ie. inference), considering they take more tokens to do the same task, and have slower inference speeds.
https://x.com/scaling01/status/2050616057191072161
Even at 2x the tokens (max from that tweet), that makes them 40% of resources. Which is still only 40% of the resources.
I've been using deepseek and it's good enough for my personal use. It takes way more time/tokens/course-correcting to get things done, but I spend in a month what I spend in a day with opus 4.6