• PrimalAnimist@lemmy.world
    link
    fedilink
    arrow-up
    43
    arrow-down
    1
    ·
    1 year ago

    I feel like this move has nothing to do with investors and everything to do with setting the standard for big corps like Microsoft and Google to be able to scrape their massive amount of data to train next gen AIs. They know they have HUGE amount of data from now and for years and years ago. Content, created by others, then sold for enormous profit.

    • Nina@lemmy.ml
      link
      fedilink
      arrow-up
      24
      arrow-down
      2
      ·
      1 year ago

      I mean AI is already stealing all art and images on the web without paying anything. They could just literally scrape and pay nothing. Web scraping isn’t illegal, they already do it, why would they pay anyone? Unless the law catches up about the rights to manufacture AI content based on ill-gotten data, then why would they pay what they don’t have to?

      • CookieJarObserver@feddit.de
        link
        fedilink
        arrow-up
        4
        arrow-down
        1
        ·
        1 year ago

        Its not even considered stealing to make training data, you can disagree, but that has to be Maschine readable by law, everyone knows that its Algorithms scraping the web for data, if they see a image that says don’t scrape they don’t take it but most images don’t have such a attachment.

        • Nina@lemmy.ml
          link
          fedilink
          arrow-up
          1
          ·
          1 year ago

          Could you please point me to legal definitions, in court or otherwise, that say it is not violating my copyright license to directly use my artwork in any shape or form for a non-fair use product? As in, a service you pay money for to create things based on the training data it has taken from me, is not fair use. Or point me to the legal definitions where I lose my copyright by posting things online? Allowing to scrape is not the same thing as giving derivative copyright license permissions. You aren’t disagreeing with me, you’re disagreeing with my legal rights.

            • hmm. That talks of data mining not of derivative work from the mined data. … I’ll leave the discussion to others, though.
              What effect would a Creative Commons “ND” clause have? Is it reserving the right to make derivatives? And would machine generated stuff even count as a derivative?

              DuckDuckGo translate:

              Copyright and Related Rights Act (Copyright Act) § 44b Text and Data Mining
              (1) Text and data mining is the automated analysis of one or more digital or digitized works in order to obtain information, in particular about patterns, trends and correlations.
              (2) Reproduction of lawfully accessible works for text and data mining is permitted. The reproductions must be deleted when they are no longer required for text and data mining.
              (3) Uses pursuant to subsection (2) sentence 1 shall only be permitted if the rights holder has not reserved them. A reservation of use for works accessible online is only effective if it is made in machine-readable form.

              • CookieJarObserver@feddit.de
                link
                fedilink
                arrow-up
                1
                arrow-down
                1
                ·
                1 year ago

                The stuff the algorithm makes is considered a new thing because the impact of each individual pice of training data is basically unmeasurably small.

            • Nina@lemmy.ml
              link
              fedilink
              arrow-up
              1
              arrow-down
              1
              ·
              edit-2
              1 year ago

              Translation: (1) Text and data mining is the automated analysis of one or more digital or digitized works in order to obtain information, in particular about patterns, trends and correlations.

              (2) Reproduction of lawfully accessible works for text and data mining is permitted. The reproductions must be deleted when they are no longer required for text and data mining.

              (3) Uses pursuant to subsection (2) sentence 1 shall only be permitted if the rights holder has not reserved them. A reservation of use for works accessible online is only effective if it is made in machine-readable form.

              None of that says anything about creating profitable derivative work. In fact is specifies patterns, trends and correlations, which does not lead me to believe it is protecting visual works created from this data, those kind of things are only used to inform things, like information science.

              • CookieJarObserver@feddit.de
                link
                fedilink
                arrow-up
                1
                arrow-down
                1
                ·
                1 year ago

                Algorithms trained with that training data are considered to make new works, so your only change to stop your work getting into a Algorithm is by going against the ones that make the training data. And they will always pull the science card.

                • Nina@lemmy.ml
                  link
                  fedilink
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  edit-2
                  1 year ago

                  Where are algorithms considered as being new and legally allowed derivative works in relation to visual works of art?

                  • CookieJarObserver@feddit.de
                    link
                    fedilink
                    arrow-up
                    1
                    arrow-down
                    1
                    ·
                    1 year ago

                    Germany.

                    And respectively, the algorithm muches up the training data so extremely that less than 0,1% of a training data picture is inside the result, therefore its legally a new work, however that work itself doesn’t fall under copyright protection because that is only a thing for human art. And as already stated, the training data is harvested according to the law i mentioned.

      • Contend6248@feddit.de
        link
        fedilink
        arrow-up
        3
        arrow-down
        2
        ·
        edit-2
        1 year ago

        What do you mean by stealing? The data remains, all they do is learning from something which is public

        What is different to Googles approach, they are just watching and learning Why is it treated so differently when it essentially does nothing new, but uses the data in a different way

        • Nina@lemmy.ml
          link
          fedilink
          arrow-up
          4
          ·
          edit-2
          1 year ago

          they are just watching and learning Why is it treated so differently Because it isn’t human. It isn’t watching and learning, it is being fed my creative content as data that I have not allowed nor have been compensated for, which is then turned around and sold as a service. My work is being consumed for commercial uses by an inhuman who does not have fair use education rights, with the sole intent to create a profitable product, and I’m getting nothing. I have legal rights, no matter where I post my work, to retain my copyrights and I have the right to not consent to improper use of my works that do not align with the licenses I have chosen to give it. Websites ask for a licenses in their ToS to be able to even just display and share my artwork when I upload it. When I create an image, I am given ownership of it’s copyright to control the use, distribution, and right to create derivatives. This isn’t a fuzzy area, it’s very clear. If an artist did not consent to their artwork being used as training data for a non-fair use reason, it is stealing their works.

          And no, it’s not fair use under education. Copyright exists for human protection and uses. It isn’t being used for ‘learning’ it’s used as data to be repackaged and sold. Google’s use of it showing up in search is to link back to posts that contain my work, retain my copyright, and are not derivatives. If you mean by captchas, yeah capchas are pretty bullshit.

          • slacktoid@lemmy.ml
            link
            fedilink
            arrow-up
            1
            ·
            1 year ago

            I feel the bigger problem with these AIs is more how they are solely being used to improve profits and productivity, these only affect the capital owners. None of that is going to improve the laborer (i.e., the artist, the coder, the writer, the people who create value from capital). This is only going to get worse. We are being normalized to automation and AI with the use of self-checkout.

            Also, about Reddit training data, I think they are too late to the party. The weights they were needed for are made. I do not think they are the exclusive source of specialized information, and (I hope) they are going to find out. They are just going to further show how silly the free market and the stock market are. The people who require the data will probably have other ways of getting it. r/datahoarders and people like that come to mind. Reddit is only making new data hard to access which, which they are not (and hopefully never) an exclusive source of.

            • Nina@lemmy.ml
              link
              fedilink
              arrow-up
              3
              ·
              edit-2
              1 year ago

              Yeah, AI can totally exist and be useful, but currently it’s in the hands of tech dudes and admins who have a terrible track record with developing things responsibly and over hyping and masking flaws. It’s used to make a profit at the colossal detriment to humans. It’s used to hurt us currently, not help at all.

              I think the training data from reddit probably only used the API because it was easier and free. And if no longer free, there’s nothing pointing to them actually paying for it. It’s not like reddit is the only data, they very much likely already have web scrapers for other uses that they can just tune for reddit.

    • torturedllama@lemmy.nz
      link
      fedilink
      arrow-up
      10
      ·
      1 year ago

      The thing I worry about whenever someone mentions this angle: What about Lemmy content? As the community moves away from the commercial platforms in favor of Lemmy, Bluesky, Mastodon etc. Then does that lower the legal barrier for AI companies to train on all this content for free? Is that shift in the legal vulnerability of public content something that users consider? Is that desirable to most users? Are people thinking about that?

    • DarkwinDuck@feddit.de
      link
      fedilink
      arrow-up
      9
      ·
      1 year ago

      That’s an interesting but definitely plausible take on the whole thing. 12000$ for 50mio requests is B2B pricing. For a company like Openai/Microsoft that’s not even worth thinking about if you get all of that precious training data for it…

    • argv_minus_one@beehaw.org
      link
      fedilink
      arrow-up
      6
      arrow-down
      1
      ·
      1 year ago

      If he thinks locking down the API is going to stop them, he’s bumped his head. These companies have more than enough manpower to write and maintain an HTML scraper for Reddit.

      • sealneaward@lemmy.ml
        link
        fedilink
        arrow-up
        2
        ·
        1 year ago

        Creating a web scraper vs actually maintaining one that is effective and works is two different things. It’s very easy to fight web scraping if you know what you are doing.

        • argv_minus_one@beehaw.org
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          1 year ago

          Right, but these are big companies with lots of talented programmers on hand. If anyone can overcome such an obstacle, it’s them.

          Also, Google and Microsoft already have a search index full of Reddit content to mine.

          • sealneaward@lemmy.ml
            link
            fedilink
            arrow-up
            1
            ·
            1 year ago

            You are right. You would need a team of skilled scrapers and network engineers though would know how to get around rate limiters with some kind of external load balancer or something along those lines.

            • argv_minus_one@beehaw.org
              link
              fedilink
              arrow-up
              1
              ·
              1 year ago

              That problem is already solved. Google and Microsoft are already fetching every single page on Reddit for search engine indexing.

              • sealneaward@lemmy.ml
                link
                fedilink
                arrow-up
                1
                ·
                1 year ago

                Could they be doing that already because of the still open API of Reddit and that will soon change? I just feel like it’s easier for them currently and it will be tougher once the API changes are implemented.

                • argv_minus_one@beehaw.org
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  1 year ago

                  No. Search engines fetch pages using plain old HTTP GET requests, same as how browsers fetch pages. There is some difficulty in parsing the HTML and extracting meaningful content, but it’s too late: the HTML is already stored on Google/Microsoft servers, ready for extraction, and there’s nothing Reddit can do to stop them.

                  Reddit can make future content harder to extract, but not without also making it invisible to search engines, which would cause Reddit to disappear from Google Search and Bing.

                  That’s why I say trying to charge money for AI training data is a fool’s errand. These facts make it impossible. That doesn’t mean Spez won’t try, but it does mean he won’t succeed.

            • MrPoopyButthole@lemmy.world
              link
              fedilink
              arrow-up
              1
              ·
              1 year ago

              Rate limiters work on IP source. This is easily bypassed with a rotating proxy. There are even SaaS that offer this. The trick is to not use large subnets that can be easily blocked. You have to use a lot of random /32 IPs to be effective.

    • DrQuint@lemmy.ml
      link
      fedilink
      arrow-up
      4
      ·
      1 year ago

      Interesting. I wonder if they already got an offer that matches their new API pricing, and they decided to up everything to match that cost and avoid being sued later.

      Like, there seems to be some urgency between them announcing and upping the price. What was it? Is this the reason? A confirmed, extremely wealthy and extremly naive buyer?

    • CookieJarObserver@feddit.de
      link
      fedilink
      arrow-up
      2
      ·
      1 year ago

      It would be better to just leave the api running as it is now because if you change nothing the data flow will not significantly change either.