• Hawk@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    22
    arrow-down
    1
    ·
    6 months ago

    It’s not what the paper is about at all, seems this is just shit journalism again.

    All the paper says about copyright is that this method is more secure because AI can sometimes spit out training examples.

    • bitfucker@programming.dev
      link
      fedilink
      arrow-up
      5
      arrow-down
      1
      ·
      6 months ago

      Why… why is it more secure? Does it mean AI training is actively abusing copyright law? And this is more secure because they can hide it better?

      • Hawk@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        2
        ·
        6 months ago

        No, you have it the other way around. It means copyright owners can share “corrupted” versions of their works and the AI can still use it. Possible AI leaks won’t return the original work, since it was never used.

        Of course I think this is only one aspect of why artists wouldn’t share their works, but it’s not the point the paper is trying to make. They’re just giving an aspect of how it could be useful.