• Jack@slrpnk.net
    link
    fedilink
    arrow-up
    28
    ·
    2 days ago

    I FEEL BAD WHEN BUYING PHYSICAL BOOKS BECAUSE IT FEELS WASTEFUL (usually use e-reader) AND THESE FUCKERS DO THIS…

    • koper@feddit.nl
      link
      fedilink
      arrow-up
      24
      ·
      2 days ago

      Very on brand for AI companies. In a decade they’ll be allowed to give AI agents legal personhood and the right to vote, but only if they first euthanize an equal number of orphans.

    • koper@feddit.nl
      link
      fedilink
      arrow-up
      12
      ·
      2 days ago

      Ultimately, Judge William Alsup ruled that this destructive scanning operation qualified as fair use—but only because Anthropic had legally purchased the books first, destroyed each print copy after scanning, and kept the digital files internally rather than distributing them. The judge compared the process to “conserv[ing] space” through format conversion and found it transformative.

      • mindbleach@sh.itjust.works
        link
        fedilink
        arrow-up
        5
        ·
        2 days ago

        Phrased like it’s a technicality, when it’s just… your rights. You are explicitly allowed to do this.

        This whole article sounds like Jack Valenti shrieking over VCRs. ‘They copied a broadcast! For later!!! That’s skirting copyright law!

        Copyright law suuucks. It needs vicious reform. And yet! These specific things have always been permitted, as a necessary part of protecting consumers, versus an industry that would love to charge rent for the books on your shelf. Those motherfuckers put DRM in cables. And yet: their laws say this is fine.

      • ChicoSuave@lemmy.world
        link
        fedilink
        arrow-up
        6
        ·
        2 days ago

        It’s literally the process that allows digitized media to be safe to possess. Someone read the FBI warnings before movies on VHS. This is some corporate malicious compliance and what the law looks like when taken to an absurd extreme.

    • MartianSands@sh.itjust.works
      link
      fedilink
      arrow-up
      5
      ·
      2 days ago

      That depends on whether you consider an LLM to be reading the text, or reproducing it.

      Outside of the kind of malfunctions caused by overfitting, like when the same text appears again and again in the training data, it’s not difficult to construct an argument that an LLM does the former, not the latter.

      • Arthur Besse@lemmy.mlOP
        link
        fedilink
        English
        arrow-up
        4
        ·
        edit-2
        2 days ago

        models can and do sometimes produce verbatim copies of individual items in their training data, and more frequently produce outputs that are close enough to them that they would clearly constitute copyright infringement if a human produced them.

        the argument that models are not derivative works of their training data is absurd, and the fact that it is being accepted by courts is yet another confirmation that the “justice system” is anything but just and the law simply doesn’t apply when there is enough money at stake.

      • awesomesauce309@midwest.social
        link
        fedilink
        English
        arrow-up
        4
        ·
        2 days ago

        It’s rare a person on social media understands they turn the input into predictive weights, and do not selectively copy and paste out of them.

        • Baggins [he/him]@lemmy.ca
          link
          fedilink
          English
          arrow-up
          4
          ·
          2 days ago

          You’re saying if I encode a copyrighted work into a JPEG it isn’t infringement? It also uses statistics to produce an approximation of the input.

          • awesomesauce309@midwest.social
            link
            fedilink
            English
            arrow-up
            3
            ·
            2 days ago

            You’re saying save one jpeg with the intent to reproduce exactly that image. I’m saying if you have a million images you have turned into weights, it won’t exactly reproduce anything unless there is very limited training data on what you’re having it predict.