Hacker Read top | best | new | newcomments | leaders | about | bookmarklet login

My comment was made in jest, the site was initially returning 500 status codes. To their credit they got it back up within half an hour.


sort by: page size:

I mean, no system is perfect. Maybe their status page only queries their site every 5 minutes? Give them a break.

As of 7:20AM PST, it's on there. So it took them 4 whole minutes since this thread was created to get it on there. That's pretty good response.


> You should. If the status page is also down, you know it's going to take a while ;-)

Maybe. I've been shocked by some big companies co-hosting their status pages.


Maybe the server was responding with a 200, but something deeper in the service just wasn't working. I expect these things are complicated and a status page is just an approximation.

The recent outage was very very delayed. I stand by my comment.

Also don't get me wrong, you do a great service. It's just a pet peeve that it seems invariably status pages are a lie.


It's a best effort approximation - this status page is automated via external uptime monitoring.

It won't catch a 1 in 10000 5xx error, but it'll catch the whole site being unavailable for several minutes.


That took over an hour. I always find status pages curious, since they never seem to really reflect the status.

Yes but in that case they took at least 40 minutes before notifying customers and posting on their status blog... I had to contact them on the live chat to be able to know what was happening...

And they could have a script that directly posts to the status page something like "A lot of our servers are down, we are working on it please be patient" when they detect some problem wouldn't take long...


I personally do not believe this page because I've viewed in the past when Reddit was down and it the status showed just that. However, once whatever was wrong got fixed and their system came back up, all those error statuses just disappeared and availability went back to 100% (or whatever) like nothing ever happened.

What's the point of keeping track of these things if downtime is going to be quietly sanitised away after the fact.


If that's the case (I trust your authority in this), then maybe the status page went down because of the spike in traffic.

Yeah, that status page is really frustrating. Didn't anyone consider building it to REALLY check if services were working? not just returning a 200 Status... but actually DOING things?

They're taking the Amazon approach to providing a status page. If you can still load the page, clearly it's not completely down. /s

To be fair my information was not accurate. It was fast but when I said it was a problem with our "backbone" I was wrong (it was a networking problem but not the backbone). I favour speed over accuracy here, but the status page wants to be fast and accurate.

> Do not bother with official status pages.

You should. If the status page is also down, you know it's going to take a while ;-)


What do you mean by status site code?

> Status Page: https://www.githubstatus.com/ -- actually shows red -- it was almost updated as fast as HN.

I think we should applaud this when it actually happens. Far too many services are terrible for this. While I'd rather the service not be down, it makes me feel a bit better if they don't lie about it.


> Next time check: http://www.digitaloceanstatus.com/

Huh... Not the best source really. It took >30 minutes between I got an alert from the monitoring service and an update appearing on the status site.


> Status page was throwing 500's in the first 10 minutes of outage but seems to have recovered.

It's still doing that.

Serving static images (and who knows what else) with "Cache-Control: max-age=1" is probably not a good idea if a million frustrated customers are going to reload your status site every minute.


At least their status page works.

Host my status page on a site that is down about half the time? Hah
next

Legal | privacy