I cant seem to get timely post federation to my instance (moist.catsweat.com) from lemmy.world. I might see a new post every few hours instead of every minute or so.

posts from other lemmy instances (lemmy.ca, lemmy.ml, dbzer0, etc) have no problem.

i see a ton of lemmy.world traffic in my nginx logs… it seems very specific to ‘new posts’. voting/comments seem to come through no problem once the post shows up.

ive checked a few other instances similar to mine, and they are not having this issue. so it seems something specific to lemmy.world and my instance.

i do recall a similar issue several months back, and it involved a lemmy.world admin resetting some outbound queue.

any ideas?

  • originalucifer@lemmy.worldOP
    link
    fedilink
    arrow-up
    1
    ·
    6 months ago

    its running on AWS, US. you can see on 4/24 i finally solved for a long running bottleneck (disk writes), and it was smooth sailing until recently.

    im not sure if anyone else has subscribed to any new communities, but ive been subbed to most of the big ones for awhile.

    i don’t see any saturation in the server network activity, connections, etc. my queues are not overloaded, and as i mentioned other instances are very performative.

    maybe i just need to be patient and let it catch up. those graphs are awesome

    • Nothing4You
      link
      fedilink
      English
      arrow-up
      2
      ·
      6 months ago

      lemmy currently only sends one activity per receiving instance at a time, so there is a round trip for every single post, comment, vote, etc., before the next activity will be sent. you wouldn’t see any increased number of connections, as there’s only a single one.

      do you have access logs for /inbox with the lemmy.world’s user agent? you might be able to derive some information from that if requests increased over time or something, maybe also response status codes?