• thebestaquaman@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    edit-2
    1 year ago

    Of course, once the AI is trained, you can’t look at some arbitrary output and determine whether that specific output came due to some specific training data set. In principle, if some of your training data is found to violate copyrights you either have to compensate the copyright holder or re-train the model without that data set.

    Finding out whether a copyrighted work is part of the training data is a matter of going through it, and should be the responsibility of the people training the model. I would like to see a case where it has been shown that a copyrighted dataset has been used to train a model, and those violating the copyright by doing so are held responsible.

    • freewheel@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 year ago

      I agree that under the current system of “idea ownership” someone needs to be held responsible, but in my opinion it’s ultimately a futile action. The moment that arbitrary individuals are allowed to download these models and use them independently (HuggingFace, et al), all control of whatever is in the model is lost. Shutting down Open AI or Anthropic doesn’t remove the models from people’s computers, and doesn’t eliminate the knowledge of how to train them.

      I have a gut feeling this is going to change the face of copyright, and it’s going to be painful. We collectively weren’t ready.