So they trained LLM's on a bunch of junk and then notice that it got worse? I don't understand how that's a surprising, or even interesting result?
replies(3):
just use a different model?
dont train it with bad data and just start a new session if your RAG muffins went off the rails?
what am I missing here