GenAI tools ‘could not exist’ if firms are made to pay copyright::undefined

  • Valen@lemmy.world
    link
    fedilink
    English
    arrow-up
    151
    arrow-down
    30
    ·
    8 months ago

    So they’re admitting that their entire business model requires them to break the law. Sounds like they shouldn’t exist.

    • Marcbmann@lemmy.world
      link
      fedilink
      English
      arrow-up
      44
      arrow-down
      12
      ·
      8 months ago

      Reproduction of copyrighted material would be breaking the law. Studying it and using it as reference when creating original content is not.

        • hglman@lemmy.ml
          link
          fedilink
          English
          arrow-up
          20
          arrow-down
          11
          ·
          8 months ago

          So if a tool is involved, it’s no longer ok? So, people with glasses cannot consume copyrighted material?

        • hedgehog@ttrpg.network
          link
          fedilink
          English
          arrow-up
          9
          arrow-down
          4
          ·
          8 months ago

          Copyright can only be granted to works created by a human, but I don’t know of any such restriction for fair use. Care to share a source explaining why you think only humans are able to use fair use as a defense for copyright infringement?

          • phdepressed@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            8
            ·
            8 months ago

            Because a human has to use talent+effort to make something that’s fair use. They adapt a product into something that while similar is noticeably different. AI will

            1. make things that are not just similar but not noticeably different.

            2. There’s not an effort in creation. There’s human thought behind a prompt but not on the AI following it.

            3. If allowed to AI companies will basically copyright everything…

            • Harbinger01173430@lemmy.world
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              3
              ·
              8 months ago

              You are aware of the insane amounts of research, human effort and the type of human talent that is required to make a simple piece of software, let alone a complex artificial neural network model whose function is to try and solve whatever stuff…right?

        • Marcbmann@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          5
          ·
          8 months ago

          I don’t agree. The publisher of the material does not get to dictate what it is used for. What are we protecting at the end of the day and why?

          In the case of a textbook, someone worked hard to explain certain materials in a certain way to make the material easily digestible. They produced examples to explain concepts. Reproducing and disseminating that material would be unfair to the author who worked hard to produce it.

          But the author does not have jurisdiction over the knowledge gained. They cannot tell the reader that they are forbidden from using the knowledge gained to tutor another person in calculus. That would be absurd.

          IP law protects the works of the creator. The author of a calculus textbook did not invent calculus. As such, copyright law does not apply.

      • 1Fuji2Taka3Nasubi@lemmy.zip
        link
        fedilink
        English
        arrow-up
        11
        arrow-down
        3
        ·
        8 months ago

        Reproduction of copyrighted material would be breaking the law. Studying it and using it as reference when creating original content is not.

        I’m curious why we think otherwise when it is a student obtaining an unauthorized copy of a textbook to study, or researchers getting papers from sci-hub. Probably because it benefits corporations and they say so?

        • Marcbmann@lemmy.world
          link
          fedilink
          English
          arrow-up
          7
          arrow-down
          2
          ·
          8 months ago

          While I would like to be in a world where knowledge is free, this is apples and oranges.

          OpenAI can purchase a textbook and read it. If their AI uses the knowledge gained to explain maths to an individual, without reproducing the original material, then there’s no issue.

          The difference is the student in your example didn’t buy their textbook. Someone else bought it and reproduced the original for others to study from.

          If OpenAI was pirating textbooks, that would be a wholly separate issue.

          • Blackmist@feddit.uk
            link
            fedilink
            English
            arrow-up
            6
            arrow-down
            3
            ·
            8 months ago

            The fact that the “AI” can spit out whole passages verbatim when given the right prompts, suggests that there is a big problem here and they haven’t a clue how to fix it.

            It’s not “learning” anything other than the probable order of words.

            • FatCrab@lemmy.one
              link
              fedilink
              English
              arrow-up
              3
              ·
              8 months ago

              I really hate this reduction of gpt models. Is the model probabilistic? Absolutely. But it isn’t simply learning a comprehensible probability of words–it is generating a massively complex conditional probability sequence for words. Largely, humans might be said to do the same thing. We make a best guess at the sequence of words we decide to use based on conditional probabilities along a myriad number of conditions (including semantics of the thing we want to say).

            • Marcbmann@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              8 months ago

              Completely agree. And that should be the focal point of the issue.

              Sam Altman is correctly stating that AI is not possible without using copyrighted materials. And I don’t think there’s anything wrong with that.

              His mistake is not redirecting the conversation. He should be talking about the efforts they’re making to stop their machine from reproducing copyrighted works. Not whether or not they should be allowed to use it in the first place.

          • 1Fuji2Taka3Nasubi@lemmy.zip
            link
            fedilink
            English
            arrow-up
            2
            ·
            8 months ago

            I agree that the issues

            • whether AI output are derivative works of its input, and
            • whether input to AI is fair use and requires no compensation

            are separate, but I think they are related, in that AI companies are trying to impose whatever interpretation of copyright that is convenient to them to the rest of the society.

            And indeed Meta pirated books to feed its AI.

            https://www.techspot.com/news/101507-meta-admits-using-pirated-books-train-ai-but.html

      • wewbull@feddit.uk
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        8 months ago

        The model itself is a derivative work. It’s existence is what is under dispute. It’s not about using the model to produce further works

        • Marcbmann@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          2
          ·
          8 months ago

          Then every single student graduating college produces derivative work.

          Everything that required the underlying knowledge gained from the textbooks studied, or research papers read, is derivative work.

          At the core of this, what are we saying? Your machine could only explain calculus because it was provided information from multiple calculus textbooks? Well, that applies to literally everyone.

      • Telodzrum@lemmy.world
        link
        fedilink
        English
        arrow-up
        16
        arrow-down
        2
        ·
        8 months ago

        This ruling only applies to the 2nd Circuit and SCOTUS has yet to take up a case. As soon as there’s a good fact pattern for the Supreme Court of a circuit split, you’ll get nationwide information. You’ll also note that the decision is deliberately written to provide an extremely narrow precedent and is likely restricted to Google Books and near-identical sources of information.

        • hedgehog@ttrpg.network
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          1
          ·
          8 months ago

          Have there been any US ruling stating something along the lines of “The training of general purpose LLMs and/or image generation AIs does not qualify as fair use,” even in a lower court?

        • Eccitaze@yiffit.net
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          8 months ago

          Hell, that article is also all about Google Books, which is an entirely different beast from generative AI. One of the key points from the circuit judge was that Google Books’ use of copyrighted material “…[maintains] respectful consideration for the rights of authors and other creative individuals, and without adversely impacting the rights of copyright holders.” The appeals court, in upholding the ruling that Google Books’ use of copyrighted content is fair use, ruled “the revelations do not provide a significant market substitute for the protected aspects of the originals.”

          If you think that gen AI doesn’t provide a significant market substitute for the artwork created by the artists and authors used to train these models, or that it doesn’t adversely impact their rights, then you’re utterly delusional.