• Semi-Hemi-Demigod@kbin.social
    link
    fedilink
    arrow-up
    72
    arrow-down
    3
    ·
    1 year ago

    That depends on how it’s written. The fact that it’s not showing zero completely makes me think they’d been iterating over threads counting comments, and they’re hitting API limits. But that’s just a blind guess without having looked at their source code.

      • Semi-Hemi-Demigod@kbin.social
        link
        fedilink
        arrow-up
        19
        arrow-down
        1
        ·
        1 year ago

        If they’re only able to get the comment counts for a smaller number of threads, those individual numbers would be noisy but lower.

        For example, before they got 100 threads with counts ranging from 0-100. Now they only get 10 threads with the same range of counts.

        • Fades@lemmy.world
          link
          fedilink
          English
          arrow-up
          10
          ·
          edit-2
          1 year ago

          Yes but then what? When you hit the API limit you’re fucking done… There should be big gaps every time this scenario triggers unless what, they have api keys for each sub or something lol

          As of July 1, 2023, we will enforce two different rate limits for those eligible for free access usage of our Data API. The limits are:
          If you are using OAuth for authentication: 100 queries per minute (QPM) per OAuth client id If you are not using OAuth for authentication: 10 QPM

          QPM limits will be an average over a time window (currently 10 minutes) to support bursting requests.

          https://support.reddithelp.com/hc/en-us/articles/16160319875092-Reddit-Data-API-Wiki

          so maybe it’s not that bad, depending on what the scheduler looks like but that would slow the entire tool down considerably if they have to wait many minutes each time after hitting a rate limit, which as you say wouldn’t take long with bigger comment sections