I guess I should have taken a screenshot, but I was just in a hurry to fix the mess. About an hour and a half ago, the server stopped responding (yet the crappy monitoring didn’t trigger an alarm…). SSH was still accessible, although quite slow, which allowed me to see a crazy load average of around 170 in top… Yet amazingly, the CPU usage from the process list didn’t seem to add up to more than 150% (the CPU has 2 cores), so no way to kill an obvious culprit. A soft reboot failed, so I ended up doing a hard reboot. Still wondering what could have gone wrong though, I don’t run troublesome stuff here for the moment, apart from Majestic12 but this thing never uses more than one full core per process and is limited to 2 processes (and they seemed to be okay from top). Still, sorry for the downtime. I should find a better monitoring…
See also…
Recent Comments
- asfsafas on Installing Rust in a custom location on Windows
- Doran on Fixing letsencrypt’s “expected xxx.pem to be a symlink”
- patheticcockroach on Removing EasyAntiCheat on Windows
- Anonymous on Removing EasyAntiCheat on Windows
- patheticcockroach on Renewing the Thecus N7510’s TLS certificate
- Giorgos on Renewing the Thecus N7510’s TLS certificate
- patheticcockroach on Various notes / Contact
- patheticcockroach on A brief tutorial to encode in x265 (and Opus) using FFmpeg
Visits since 16 April 2012:
0 Responses
Stay in touch with the conversation, subscribe to the RSS feed for comments on this post.