Do you think new LLMs are going to write better and better code? When all they are going to have is the slop generated by previous, worse models?

Yes. The models may have started from indiscriminate scraping, but people are undoubtedly working on refining the training data. Combined with the overall model capabilities, I suspect code quality will continue to go up.

What you're suggesting is a negative flywheel where quality spirals down, but I'm hoping it becomes a positive loop and the quality floor goes up. We had plenty of slop before LLMs, and not all LLM output is slop. Time will tell, but I think LLMs will continue to improve their coding abilities and push overall quality higher.

Let's agree to disagree.

What I see is that while LLMs can do real tasks, they often produce overengineered unmaintainable slop (plenty of examples where code can be reduced 10x to do the same). I hope this is not a base to continue training LLMs on.