Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

As if the language models currently would give a damn about copyright...




The problem is they have to hide their sources due to copyright. So they train on copyright data but must obscure it in the output. Thus they must hide the sources of truth making it impossible to fact check them directly and the reason that hallucinations are so common and unavoidable in the current pattern.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: