• MonkderVierte@lemmy.ml
    link
    fedilink
    English
    arrow-up
    21
    arrow-down
    1
    ·
    9 hours ago

    Btw, how about limiting clicks per second/minute, against distributed scraping? A user who clicks more than 3 links per second is not a person. Neither, if they do 50 in a minute. And if they are then blocked and switch to the next, it’s still limited in bandwith they can occupy.

    • letsgo@lemm.ee
      link
      fedilink
      English
      arrow-up
      9
      ·
      8 hours ago

      I click links frequently and I’m not a web crawler. Example: get search results, open several likely looking possibilities (only takes a few seconds), then look through each one for a reasonable understanding of the subject that isn’t limited to one person’s bias and/or mistakes. It’s not just search results; I do this on Lemmy too, and when I’m shopping.

      • MonkderVierte@lemmy.ml
        link
        fedilink
        English
        arrow-up
        7
        ·
        8 hours ago

        Ok, same, make it 5 or 10. Since i use Tree Style Tabs and Auto Tab Discard, i do get a temporary block in some webshops, if i load (not just open) too much tabs in too short time. Probably a CDN thing.

      • MonkderVierte@lemmy.ml
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        edit-2
        8 hours ago

        Ah, one request, then the next IP doing one and so on, rotating? I mean, they don’t have unlimited adresses. Is there no way to group them together to a observable group, to set quotas? I mean, in the purpose of defense against AI-DDOS and not just for hurting them.

        • edinbruh@feddit.it
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          1
          ·
          8 hours ago

          There’s always Anubis 🤷

          Anyway, what if they are backed by some big Chinese corporation with some /32 ipv6 and some /16 ipv4? It’s not that unreasonable

            • edinbruh@feddit.it
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              4 hours ago

              my point was that even if they don’t have unlimited ips they might have a lot of them, especially if its ipv6, so you couldn’t just block them. but you can use anubis that doesn’t rely on ip filtering

              • JackbyDev@programming.dev
                link
                fedilink
                English
                arrow-up
                1
                ·
                3 hours ago

                You’re right, and Anubis was the solution they used. I just wanted to mention the IP thing because you did is all.

                I hadn’t heard about Anubis before this thread. It’s cool! The idea of wasting some of my “resources” to get to a webpage sucks, but I guess that’s the reality we’re in. If it means a more human oriented internet then it’s worth it.

                • edinbruh@feddit.it
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  44 minutes ago

                  A lot of FOSS software’s websites are starting to use it lately, starting from the gnome foundation, that’s what popularized it.

                  The idea of proof of work itself came from spam emails, of all places. One proposed but never adopted way of preventing spam was hashcash, which required emails to have a proof of work embedded in the email. Bitcoins came after this borrowing the idea