• DrCakeEnglish
    arrow-up
    343
    arrow-down
    5
    ·
    1 month ago
    link
    fedilink

    So when’s the ruling against OpenAI and the like using the same copyrighted material to train their models

    • irotsomaEnglish
      arrow-up
      135
      arrow-down
      3
      ·
      1 month ago
      link
      fedilink

      But OpenAI not being allowed to use the content for free means they are being prevented from making a profit, whereas the Internet Archive is giving away the stuff for free and taking away the right of the authors to profit. /s

      Disclaimer: this is the argument that OpenAI is using currently, not my opinion.

    • norimeeEnglish
      arrow-up
      85
      arrow-down
      3
      ·
      1 month ago
      edit-2
      1 month ago
      link
      fedilink

      Ah, I see you got that all wrong.

      Open IA AI uses that content to generate billions in profit on the backs of The People. The Internet Archive just does it for the good of The People.

      We can’t have that. “Good for The People” is not how the economy works, pal. We need profit and exploitation for the world to work

      • v_krishnaEnglish
        arrow-up
        24
        arrow-down
        1
        ·
        1 month ago
        link
        fedilink

        OpenAI is burning billions of dollars not making profit.

        • AgretEnglish
          arrow-up
          4
          arrow-down
          0
          ·
          1 month ago
          link
          fedilink

          Sounds like they are operating the same as all the other big tech companies then

          • ShaggySnacksEnglish
            arrow-up
            6
            arrow-down
            0
            ·
            1 month ago
            link
            fedilink

            Burn a ton a cash to become the only major player in the market and the proceed to enshitify as no one else has anywhere to go.

          • v_krishnaEnglish
            arrow-up
            8
            arrow-down
            0
            ·
            1 month ago
            link
            fedilink

            Eh? That article says nothing about their profit margins. Today they have something like $3.5B in ARR (not really, that’s annualized from their latest peak, in Feb they had like $2B ARR). Meanwhile they have operating costs over $7B. Meaning they are losing money hand over fist and not making a profit.

            I’m not suggesting anything else, just that they are not profitable and personally I don’t see a road to profitability beyond subsidizing themselves with investment.

            • buddascrayonEnglish
              arrow-up
              1
              arrow-down
              4
              ·
              1 month ago
              link
              fedilink

              It’s in the first bloody paragraph. 😮‍💨

              OpenAI is begging the British Parliament to allow it to use copyrighted works because it’s supposedly “impossible” for the company to train its artificial intelligence models — and continue growing its multi-billion-dollar business — without them.

              And if you follow the link the title of the article says it all:

              #OpenAI is set to see its valuation at $80 billion—making it the third most valuable startup in the world

              • danEnglish
                arrow-up
                6
                arrow-down
                1
                ·
                1 month ago
                edit-2
                1 month ago
                link
                fedilink

                Just because the company has a high valuation, doesn’t mean they’re making a profit. They’re indeed losing a lot of money and will go bankrupt if they don’t get new investment and/or increase their ARR soon. Right now, they’ve only got 12 months left before they’re out of money. https://www.windowscentral.com/software-apps/openai-could-be-on-the-brink-of-bankruptcy-in-under-12-months-with-projections-of-dollar5-billion-in-losses

                • buddascrayonEnglish
                  arrow-up
                  2
                  arrow-down
                  3
                  ·
                  1 month ago
                  link
                  fedilink

                  The valuation is based on the expectation of the company to make massive profits. And if you think investor money is not profit for the people running Open AI, you’re crazy. We could only hope that they run out of money and go out of business. But that’ll never happen now with the amount of faith these corporations are putting in AI research.

              • v_krishnaEnglish
                arrow-up
                1
                arrow-down
                0
                ·
                1 month ago
                link
                fedilink

                I take it you don’t understand how startups work?

                OpenAI is not making any profit and is losing money hand over fist today. Valuation and raising investment rounds isn’t profit.

      • RogueBananaEnglish
        arrow-up
        5
        arrow-down
        1
        ·
        1 month ago
        link
        fedilink

        “Good for the people”? You mean COMMUNISM?

      • finitebanjoEnglish
        arrow-up
        3
        arrow-down
        0
        ·
        1 month ago
        link
        fedilink

        I think you accidentally swapped OpenAI and Open IA which happens to initialize Internet Archive, a little confusing.

        • norimeeEnglish
          arrow-up
          4
          arrow-down
          0
          ·
          1 month ago
          link
          fedilink

          I didn’t even realise. Thank you for pointing it out, I fixed it.

    • AnyolduserEnglish
      arrow-up
      23
      arrow-down
      2
      ·
      1 month ago
      link
      fedilink

      Hot on the heels of this one, I’d imagine.

      • shrugsEnglish
        arrow-up
        19
        arrow-down
        1
        ·
        1 month ago
        link
        fedilink

        So, let’s say we create an llm that will be fed will all the copyrighted data and we design it, so that it recalls the originals when asked?! Does that count as piracy or as the kind of legal shananigans openai is doing?

    • PriorityMotifEnglish
      arrow-up
      21
      arrow-down
      15
      ·
      1 month ago
      link
      fedilink

      It’s two different things happening. One is redistribution, which isn’t allowed and the other is fair use, which is allowed. You can’t ban someone from writing a detailed synopsis of your book. That’s all an llm is doing. It’s no different than a human reading the material and then using that to write something similar.

      • xthexderEnglish
        arrow-up
        17
        arrow-down
        1
        ·
        1 month ago
        edit-2
        1 month ago
        link
        fedilink

        the other is fair use

        That’s very much up for debate still.

        (I am personally still undecided)

        • PriorityMotifEnglish
          arrow-up
          9
          arrow-down
          5
          ·
          1 month ago
          link
          fedilink

          The difference is that the llm has the ability to consume and remember all available information whereas a human would have difficulty remembering everything in detail. We still see humans unintentionally remaking things they’ve heard before. Comedians have unintentionally stolen jokes they’ve heard. Every songwriter has unintentionally “discovered” a catchy tune which is actually someone else’s. We have fanfiction and parody. Most people’s personalities are just an amalgamation of everyone and everything they’ve ever seen, not unlike an llm themselves.

          • xthexderEnglish
            arrow-up
            5
            arrow-down
            0
            ·
            1 month ago
            link
            fedilink

            I agree with you for the most part, but when the “person” in charge of the LLM is a big corporation, it just exaggerates many of the issues we have with current copyright law. All the current lawsuits going around signal to me that society as a whole is not so happy with how it’s being used, regardless of how it fits in to current law.

            AI is causing humanity to have to answer a lot of questions most people have been ignoring since the dawn of philosophy. Personally I find it rather concerning how blurry some lines are getting, and I’ve already had to reevaluate how I think about certain things, like what moral responsibilities we’ll have when AIs truely start to become sentient. Is turning them off and deleting them a form of murder? Maybe

            • trafficnabEnglish
              arrow-up
              5
              arrow-down
              2
              ·
              1 month ago
              link
              fedilink

              OpenAI losing their case is how we ensure that the only people who can legally be in charge of an LLM are massive corporations with enough money to license sufficient source material for training, so I’m forced to begrudgingly take their side here

            • greenskyeEnglish
              arrow-up
              2
              arrow-down
              0
              ·
              1 month ago
              link
              fedilink

              Agreed. I keep waffling on my feelings about it. It definitely doesn’t feel like our laws properly handle the scale that LLMs can take advantage of ‘fair use’. It also feels like yet another way to centralize and consolidate wealth, this time not money, but rather art and literary wealth in the hands of a few.

              I already see artists that used to get commissions now replaced by endless AI pictures generated via a Lora specifically aping their style. If it was a human copying you, they’d still be limited by the amount they could produce. But an AI can spit out millions of images all in the style you perfected. Which feels wrong.

          • FerkEnglish
            arrow-up
            2
            arrow-down
            1
            ·
            1 month ago
            edit-2
            1 month ago
            link
            fedilink

            Is “intent” what makes all the difference? I think doing something bad unintentionally does not make it good, right?

            Otherwise, all I need to do something bad is have no bad intentions. I’m sure you can find good intentions for almost any action, but generally, the end does not justify the means.

            I’m not saying that those who act unintentionally should be given the same kind of punishment as those who do it with premeditation what I’m saying is that if something is bad we should try to prevent it in the same level, as opposed to simply allowing it or sometimes even encourage it. And this can be done in the same way regardless of what tools are used. I think we just need to define more clearly what separates “bad” from “good” specifically based on the action taken (as opposed to the tools the actor used).

          • WalnutLumEnglish
            arrow-up
            2
            arrow-down
            4
            ·
            1 month ago
            link
            fedilink

            You’re anthropomorphizing LLMs.

            There’s a philosophical and neuroscuence concept called “Qualia, which helps define the human experience. LLMs have no Qualia.

            • Saik0English
              arrow-up
              2
              arrow-down
              1
              ·
              1 month ago
              link
              fedilink

              You’re anthropomorphizing LLMs.

              No, they’re taking the argument to it’s logical end.

        • FerkEnglish
          arrow-up
          3
          arrow-down
          0
          ·
          1 month ago
          edit-2
          1 month ago
          link
          fedilink

          I think that’s the difference right there.

          One is up for debate, the other one is already heavily regulated currently. Libraries are generally required to have consent if they are making straight copies of copyrighted works. Whether we like it or not.

          What AI does is not really a straight up copy, which is why it’s fuzzy, and much harder to regulate without stepping in our own toes, specially as tech advances and the difference between a human reading something and a machine doing it becomes harder and harder to detect.

      • Gsus4English
        arrow-up
        5
        arrow-down
        2
        ·
        1 month ago
        edit-2
        1 month ago
        link
        fedilink

        The matter is not LLMs reproducing what they have learned, it is that they didn’t pay for the books they read, like people are supposed to do legally.

        This is not about free use, this is about free access, which at the scale of an individual reading books is marketed as “piracy”at the scale of reading all books known to manit’s onmipiracy?

        We need some kind of deal where commercial LLMs have to pay a rent to a fund that distributes that among creators or remain nonprofit, which is never gonnna happen, because it’ll be a bummer for all the grifters rushing into that industry.

        • barsoapEnglish
          arrow-up
          2
          arrow-down
          1
          ·
          1 month ago
          link
          fedilink

          it is that they didn’t pay for the books they read, like people are supposed to do legally.

          If I can read a book from a library, why shouldn’t OpenAI or anybody else?

          but yes from what I’ve heard they (or whoever, don’t remember) actually trained on libgen. OpenAI can be scummy without the general process of feeding AI books you only have read access to being scummy.

          • Gsus4English
            arrow-up
            1
            arrow-down
            1
            ·
            1 month ago
            edit-2
            1 month ago
            link
            fedilink

            This is not like reading a book from a libraryunless you want to force the LLM to only train one book per day and keep no copies after that day.

            • barsoapEnglish
              arrow-up
              2
              arrow-down
              0
              ·
              1 month ago
              link
              fedilink

              They don’t keep copies and learning speed? Why one day? Does it count if I skim through a book?

              • Gsus4English
                arrow-up
                1
                arrow-down
                0
                ·
                1 month ago
                link
                fedilink

                deleted by creator

        • PriorityMotifEnglish
          arrow-up
          2
          arrow-down
          1
          ·
          1 month ago
          link
          fedilink

          I think we need to re-examine what copyright should be. There’s nothing inherently immoral about “piracy” when the original creator gets almost nothing for their work after the initial release.

    • indexEnglish
      arrow-up
      1
      arrow-down
      2
      ·
      1 month ago
      link
      fedilink

      stop asking questions and go back to work