Tell HN: GitHub is down (Update: Back online now)
Getting 500 errors, multiple friends confirming: https://github.com It's like the good old days when the power went down at work and everyone came to the hallways. Good luck to the engineers at GitHub as I know how stressful it can be, but hope everyone else is enjoying a nice break and some socializing According to GH Status (https://www.githubstatus.com/), everything is fine. Gotta love functional status pages. edit: Nevermind, they just reported "degraded performance" for GitHub Actions, Issues, and Pull Requests. Didn't they have actual error rate graphs on that page back in a day? Yes, and some response times too. It was actually useful. They have updated it. i wonder how they even work? Like, do they just display a green css button instead of _actually_ doing a healthcheck? Yes they do, because very rarely would such a healthcheck kind of setup actually work in practice, at a large enough size, for a user-facing dashboard. If you want a healthcheck, look at a Grafana dashboard, not a status page. By the way, I don't know of a single place where this isn't the case, where a human signs off on and updates the status page during large events (at least at the final decision.) Some of it will be automated, sure, like red flags being raised to operators. But at a certain point it is not possible to automate this in some level to achieve second-level accuracy or whatever; the system is rarely (if ever) in a binary state of "working perfectly" or "not working", but somewhere in between. You can't just fire off a big red error bar every time a blip occurs at a place like GitHub. The system is constantly "in motion". The logical conclusion is to just expose your 50+ Grafana dashboards publicly to every user. Isn't that the most honest "overview" of what is happening with your product? Except this often can't tell them useful things either. People on here will also mumble about SLAs but if a customer wants a kickback or is seriously worried about events like these, they're generally talking to account managers, not posting on internet forums. That said, a lot of them get weaselly about that stuff unless you're already negotiating prices with an AM in the first place... When I started work at Amazon in 2001 we had a "gonefishing" page for outages that a human had to flip the site to manually. We actually stopped doing this a year or two later because reporters were setting up monitoring on that page showing up and were reporting on outage statistics based on it. So we just left the site up in whatever degraded state it was in and that made the problem of measuring www.amazon.com uptime externally that little bit more difficult. Probably requires manual updates. It seems like more and more places have moved to this paradigm now that status pages are tied to SLAs which are tied to money. One might call it the politicization of status pages. Politicization, yes; I've never heard of SLAs being tied to status pages. It is like pulling teeth to get most cloud providers to credit the account when they don't meet SLA, and one always has to ask for it; heaven forbid if credits were paid out automatically when service wasn't rendered. Or you get weasel-worded out of it. I had a cloud provider deny a service credit; the SLA stated that the service was only out of SLA if it didn't return 2xx. Well, the API returned "2xx Accepted — your request is being processed", and you could use the API to query the job, and the job … never finished or made any progress at all. But the API returned 2xx the entire time, so that was "within SLA". Correct. SLAs aren't calculated off status pages, there are far better ways of calculating it (running a query over responses, for example). Most modern SLAs are customer initiated anyways, so the customer is writing in to request this rather than automatically calculating them. The status page doesn't need to show anything for a customer to provide logs indicating a QoS less than that promised in the SLA. I don't think it's politics (maybe AWS's is, but GCP wasn't IMO), it's really a function of "in large scale software systems things are constantly failing in all sorts of ways, and it's really hard to output a meaningful automated signal that things are broken. Sure you can set up pingdom type health checks on every endpoint, but even then you're not necessarily guaranteeing that things are working properly. Source: worked at a few cloud providers, paid out a few SLA violations AWS do the same thing. Pretty sure someone updates it manually. Incident is up now - https://www.githubstatus.com/incidents/fz1bdbw24y81 > We are investigating reports of degraded performance for GitHub Actions, Issues, and Pull Requests. Degraded Performance? Their freaky homepage is borked: https://github.com/ yields a 500 error haha. 100% degraded is degraded. Negative growth rate is still growth :) That's a funny way of saying "every page is returning 500" IIRC, that's the default verbiage that goes out when someone pulls the "oh shit" lever. github.com is having issues across the site, the API including git operations (and CLI) still work. Status page is manually updated, and we're working to get it updated. EDIT: it's updated now. EDIT EDIT: github.com is back up and running, apologies for the disruption :( Source: GitHub employee Why manually update vs automatically? Not many places I have worked for allowed for this to be automatic. A lot of it was so they could provide a coherent explanation as to what the current state of internal attention was directed at vs what everyone can plainly see. I guess this makes sense, but I don't understand why you couldn't have it change status automatically, and still allow a person to go in after and manually add an explanation. Mostly because it's more work to manually remove a bunch of spurious robotically added statuses (which look bad, because you're either down more than customers are noticing or your detection is flawed) than it is to manually add and remove real ones. They wait for the HN thread to appear so they don't give false positives. I think it is so they can "hide" small outages that don't rise to the level of making the news sites so they can look better. A lot of sites do this sort of thing these days. Ish? I wrote some stuff up here: https://news.ycombinator.com/item?id=30182591 When I worked at GCP it was all manually updated as well so we could add a sentence about what was actually affected. In any sufficiently large system it's hard to indicate exactly what's broken/how to work around it, so it was just easier to `/status <system> <color> <reason for status>`. Oh dear. Last time a serious incident happened was just 48 hours ago: [0] Now it has gotten critically worse. Is it time to use a self-hosted backup like what GNOME is using? [1] Yup. I've been disrupted by GitHub today, and DockerHub the other day. Crude reminder that the cloud is some company's computer. My own computer wouldn’t be much better. Same. My desktop won't even currently power on after a move... Looks like I get to reseat all of my RAM tonight :) I just refreshed a page from GH that I've had open since last night, and yup, 500. Of course I came to HN first before even visiting their own status page as HN always has an update faster than their official page. Yes, down in Europe too :/ Yep, down since ~10mins here in central europe. I'd like to see a graphic of the traffic spike at times like this from CTRL-R. They even encourage it on the 500 page: "try refreshing". Edit: it's back. Looks like only the website is down. I just used the mobile app and it works fine. Yup, confirmed here as well. Multiple team members also having issues. Too bad federated GitLab still does not exist. I am able to push code but then I went to check the GH Action to make sure it ran and.... error pages. Yup. Not even the unauthenticated home page is showing up. Wew It was down for me too (in Brazil). But it is back now! :) It looks like it's back up now, can anyone confirm? Their 500 page is delightful... More downtime, I say! Right as we were doing a big release, too :) It's currently working in Switzerland. Literally just posted a thread.. Doh! :( graphite.dev (alternative frontend to github code review) still up, it appears to be just their frontend. Works again for me here in Germany. It's up and running now Yes, returning a 500. Looks down to me :( Yep, having issues Same here down for me too down for me can corroborate, very down same