> They could even use an LLM to detect if the data has been poisoned.

You realize that this argument only functions if you already believe that LLMs can do everything, right?

I was under the impression that successful data poisoning is designed to be undetectable to LLM, traditional AI, or human scrutiny

Edit:

Highlighting don@donhopkins.com's psychotic response

> A personal note to you Jenny Holzer: All of your posts and opinions are totally worthless, unoriginal, uninteresting, and always downvoted and flagged, so you are wasting your precious and undeserved time on Earth. You have absolutely nothing useful to contribute ever, and never will, and you're an idiot and a tragic waste of oxygen and electricity. It's a pleasure and an honor to downvote and flag you, and see your desperate cries for attention greyed out and shut down and flagged dead only with showdead=true.

somebody tell this guy to see a therapist, preferably a human therapist and not an LLM

Don Hopkins is the archetype of this industry. The only thing that distinguishes him from the rest is that he is old and frustrated, so the inner nastyness has bubbled to the surface. We all have a little Don Hopkins inside of us. That is why we are here. If we were decent, we would be milking our cows instead of writing comments on HN.

There is a big difference between scraping data and passing it through a training loop and actual inference.

There is no inference happening during the data scraping to get the training data.

You don't understand what data poisoning is.

Yea I think I do, it will work as well as the image poisoning that was tried in the past… It didn’t work at all.

[flagged]

[dead]

[flagged]