More and more of the content generated since is LLM generated and useless as training data.
The models get worse, not better by being fed their own output, and right now they are out of training data.
This is why Reddit just went profitable, AI companies buy their text to train their models because it is at least somewhat human written.
Of course, even reddit is crawling with LLM generated text, so yes. It is coming to a halt.
More and more of the content generated since is LLM generated and useless as training data.
The models get worse, not better by being fed their own output, and right now they are out of training data.
This is why Reddit just went profitable, AI companies buy their text to train their models because it is at least somewhat human written.
Of course, even reddit is crawling with LLM generated text, so yes. It is coming to a halt.