• Semi-Hemi-Demigod@kbin.social
      link
      fedilink
      arrow-up
      21
      arrow-down
      1
      ·
      1 year ago

      If they’re only able to get the comment counts for a smaller number of threads, those individual numbers would be noisy but lower.

      For example, before they got 100 threads with counts ranging from 0-100. Now they only get 10 threads with the same range of counts.

      • Fades@lemmy.world
        link
        fedilink
        English
        arrow-up
        10
        ·
        edit-2
        1 year ago

        Yes but then what? When you hit the API limit you’re fucking done… There should be big gaps every time this scenario triggers unless what, they have api keys for each sub or something lol

        As of July 1, 2023, we will enforce two different rate limits for those eligible for free access usage of our Data API. The limits are:
        If you are using OAuth for authentication: 100 queries per minute (QPM) per OAuth client id If you are not using OAuth for authentication: 10 QPM

        QPM limits will be an average over a time window (currently 10 minutes) to support bursting requests.

        https://support.reddithelp.com/hc/en-us/articles/16160319875092-Reddit-Data-API-Wiki

        so maybe it’s not that bad, depending on what the scheduler looks like but that would slow the entire tool down considerably if they have to wait many minutes each time after hitting a rate limit, which as you say wouldn’t take long with bigger comment sections

    • bloopernova
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      It could be noisy if the subs are counted in a random order. That way you’d get a different number of comments each day before you hit the API limit.