• Eager Eagle@lemmy.worldEnglish
    1·
    2 years ago

    Or don’t do anything. There are plenty of crawlers out there and disallowing won’t stop the unethical ones.

    • Zikeji@programming.devEnglish
      1·
      2 years ago

      Just because some people might break into my house doesn’t mean I’ll stop locking my doors.

      • Eager Eagle@lemmy.worldEnglish
        1·
        2 years ago

        that doesn’t lock anything, it’s not a security feature.

        • Arakwar@kbin.social
          1·
          2 years ago

          A house door lock isn’t that much about security either.

          • Zikeji@programming.devEnglish
            1·
            2 years ago

            It’s a deterrent. Which is a pretty apt comparison for robots.txt and user agent blocking.

  • Voyajer@kbin.social
    1·
    2 years ago

    “Please label all of your interesting text so we can flag it with our webcrawler to train on later.”

  • ExpensiveConstant@kbin.social
    1·
    2 years ago

    I mean, you can add their user agent to the robots file but the crawler could just change their user agent or even ignore the robots file if the server isn’t filtering requests by user agent

  • WasPentalive@lemmy.one
    1·
    2 years ago

    Is there some way you could have your web server log who scrapes the site? If you disallow ChatGPT and still find that it has scraped your site would you have cause to sue? @legaleagle (or anyone else too)

    • Cyclohexane@lemmy.ml
      1·
      2 years ago

      It’s gotta be pretty difficult to differentiate human users from bots. If it was easy, you could prevent bots from loading the page altogether.

      • Lmaydev@programming.dev
        1·
        2 years ago

        Exactly what Google are trying to do currently. Just in the worst way possible.