originalucifer

joined 10 months ago
[–] [email protected] 1 points 7 months ago (1 children)

its running on AWS, US. you can see on 4/24 i finally solved for a long running bottleneck (disk writes), and it was smooth sailing until recently.

im not sure if anyone else has subscribed to any new communities, but ive been subbed to most of the big ones for awhile.

i don't see any saturation in the server network activity, connections, etc. my queues are not overloaded, and as i mentioned other instances are very performative.

maybe i just need to be patient and let it catch up. those graphs are awesome

 

I cant seem to get timely post federation to my instance (moist.catsweat.com) from lemmy.world. I might see a new post every few hours instead of every minute or so.

posts from other lemmy instances (lemmy.ca, lemmy.ml, dbzer0, etc) have no problem.

i see a ton of lemmy.world traffic in my nginx logs... it seems very specific to 'new posts'. voting/comments seem to come through no problem once the post shows up.

ive checked a few other instances similar to mine, and they are not having this issue. so it seems something specific to lemmy.world and my instance.

i do recall a similar issue several months back, and it involved a lemmy.world admin resetting some outbound queue.

any ideas?