cross-posted from: https://lemmy.zip/post/446751

Alternative title: Musk sues data scrapers, blames them for Twitter’s “impaired user experience”

  • spiderman@lemmy.fmhy.ml
    link
    fedilink
    arrow-up
    14
    ·
    1 year ago

    i think what the previous dude said is, people wouldn’t not have resorted to use scrapping if there was a good api. one of the major reason people/organisations choose scrapping is because it’s better than paying insane amounts of money for insane(ly low)amount of api calls.

    • QHC@lemmy.world
      link
      fedilink
      arrow-up
      11
      ·
      1 year ago

      There was a good API before Elon essentially shut it down. But that is irrelevant.

      The reason mass data scraping like OpenAI does would not rely on an API is because they are getting data from the entire Internet (and other sources that aren’t online). They want raw data and they want as much and as varied as possible. It’s much easier, cheaper and practical to build tools that scrape websites generically than to integrate with thousands of completely independent and different APIs.

      It’s the same reason that Reddit complaining about “AI taking all of our data” is bullshit. “AI” is just a convenient excuse and the most recent tech buzzword.

      • pewter@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        They are largely mad because of how effective the AI is. If data was being used just to improve Swype for texting people would care less. I care more about artists’ complaints about getting replaced than big tech companies complaining that content they didn’t create is being used to create things.

        Also I decided to read openAI’s GPT2 paper and they were pretty clear about their created dataset:

        "Instead, we created a new web scrape which emphasizes document quality. To do this we only scraped web pages which have been curated/filtered by humans. Manually filtering a full web scrape would be exceptionally expensive so as a starting point, we scraped all outbound links from Reddit, a social media platform, which received at least 3 karma. This can be thought of as a heuristic indicator for whether other users found the link interesting, educational, or just funny.

        The resulting dataset, WebText, contains the text subset of these 45 million links."

        That’s a nice sized dataset from real people that’s already somewhat filtered by quality. They were totally scraping Reddit very specifically and now that people see it’s effective, anyone else who wants to make their own chatgpt or wants to improve their models will do the same.

        • QHC@lemmy.world
          link
          fedilink
          arrow-up
          2
          ·
          1 year ago

          Very interesting, thanks for sharing!

          It’s worth noting here for context that:

          • scraping the outbound links is very different than all of the user comments, which would put more load on Reddit’s servers and also be easier to claim “belongs” to Reddit (it doesn’t but they argue otherwise)
          • this shows that OpenAI was not using, let alone abusing, Reddit’s API as claimed