Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Is this paper wrong? - https://arxiv.org/abs/2311.09807


It shows that if you deliberately train LLMs against their own output in a loop you get problems. That's not what synthetic data training does.


I understand and appreciate your clarification. However would it not be the case some synthetic data strategies, if misapplied, can resemble the feedback loop scenario and thus risk model collapse?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: