List of trackers

Thanks, very helpful! In the sense that all seems rather well now. The low score in the table of that one monitor is apparently bogus, because it is reflected neither in the graph, nor in the log data. I noted that since around the time of some recent changes in the pool infrastructure, I am seeing similar artifacts in higher numbers. E.g., there “always” used to be some discrepancies between the overall score values given in various places at the same time, I guess due to different lags of the different ways the score value was obtained from the underlying data for it to be displayed in the different places. But recently, I am seeing much larger discrepancies, for more diverse data points, and for longer, than before those recent changes (not implying a causal link, even though the temporal correlation might indicate there to be one).

Why there was near total outage for most monitors before the morning of 2025-05-05 is strange. Especially given it looks like really a full outage for the majority of monitors, while a small number of other monitors (1 or 2 or so) at least had intermittent connectivity, or recovered noticeably before the majority of the rest.

So I’d suggest you’d consider whether you changed anything on your end on the morning of the fifth, in your approach to manage the influx, or in the configuration of the server (netspeed setting), or anything else, and reflect on how that might have impacted the outcome. That could give a clue as to what was going on. E.g., overzealous blocking, too high a netspeed setting for the capabilities of your system in relation to the zone it is in, …

Not sure you are aware, but there was a major issue with the RU zone not long ago. It wasn’t very stable to begin with, and the roll-out of some buggy software to a huge number of systems (I think it was some kind of wireless speaker system) created a significant surge in traffic that pushed the zone over the edge (see the sharp dip in the number of servers at the end of 2024).

Due to a concerted effort, the surge has been quelled at the source (people from this forum reached out to the vendor, who then fixed the issue and rolled out the new version - one of the rare successful attempts to get an issue fixed at the source), and the number of servers in the zone has increased significantly after a call to action in local communities.

But I don’t have a server in that zone, so cannot judge as to how well that zone is being served right now, i.e., what the effective relation between netspeed, and actual traffic volume is. And how easy it is to knock one-self out by too high a netspeed setting. Though the complete outage of the majority of monitors suggests this might have been caused by more deliberate, indiscriminate blocking, rather than overload (with overload, one would not see a complete outage, but kind of a sawtooth pattern as the server moves into and out of the pool).