Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

And to give a concrete example, in my view it should be allowed to use any source code to train a model such that the model learns that code is bad or insecure or slow or otherwise undesirable. In other words, it should be allowed to train on anything as long as the model does NOT produce that training data verbatim.


Maybe you should update your view with 17 USC 106.

https://www.law.cornell.edu/uscode/text/17/106


What copyrightable elements of the original work persist in the model, if it is incapable of outputting them? I can derive a SHA-1 hash from a copyrighted image, and yet it would be absurd to call that a derivative work.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: