1. Home
  2. Companies
  3. GitHub
  4. Outage Map
GitHub

GitHub Outage Map

The map below depicts the most recent cities worldwide where GitHub users have reported problems and outages. If you are having an issue with GitHub, make sure to submit a report below

Loading map, please wait...

The heatmap above shows where the most recent user-submitted and social media reports are geographically clustered. The density of these reports is depicted by the color scale as shown below.

GitHub users affected:

Less
More
Check Current Status

GitHub is a company that provides hosting for software development and version control using Git. It offers the distributed version control and source code management functionality of Git, plus its own features.

Most Affected Locations

Outage reports and issues in the past 15 days originated from:

Location Reports
Ingolstadt, Bavaria 1
Paris, Île-de-France 1
Berlin, Berlin 2
Dortmund, NRW 1
Davenport, IA 1
St Helens, England 1
Nové Strašecí, Central Bohemia 1
West Lake Sammamish, WA 3
Parkersburg, WV 1
Perpignan, Occitanie 1
Piura, Piura 1
Tokyo, Tokyo 1
Brownsville, FL 1
New Delhi, NCT 1
Kannur, KL 1
Newark, NJ 1
Raszyn, Mazovia 1
Trichūr, KL 1
Departamento de Capital, MZ 1
Chão de Cevada, Faro 1
New York City, NY 1
León de los Aldama, GUA 1
Quito, Pichincha 1
Belfast, Northern Ireland 1
Guayaquil, Guayas 1
Check Current Status

Community Discussion

Tips? Frustrations? Share them here. Useful comments include a description of the problem, city and postal code.

Beware of "support numbers" or "recovery" accounts that might be posted below. Make sure to report and downvote those comments. Avoid posting your personal information.

GitHub Issues Reports

Latest outage, problems and issue reports in social media:

  • Eduardopto
    Ed (@Eduardopto) reported

    Anthropic is facing a weird feedback loop: users are complaining that Claude’s output quality is nosediving, and Claude itself agrees. The model analyzed its own GitHub repo and confirmed that quality-related issue reports have escalated sharply since January. This decline coincides with Anthropic aggressively throttling capacity during peak hours to manage server load. We are seeing a dangerous trend where infrastructure constraints directly degrade model performance. When you optimize for reliability and cost, the "intelligence" is the first thing to hit the cutting room floor. It’s hard to build robust agentic flows when the base model’s reasoning capability fluctuates based on the time of day if you are building right now, what does this actually unlock or kill?

  • viksit
    Viksit Gaur (@viksit) reported

    @nicoalbanese10 is there a github? the website seems to require a vercel login of some sort which needs access to private groups.

  • AfterThe925
    NiNE (@AfterThe925) reported

    Two weeks ago, deploying an AI agent took a weekend and a GitHub degree. Now: dashboard, click, running. Anthropic handles sandboxing, retries, auth. Platforms handle hosting, integrations, memory. The infrastructure layer is being commoditized in real time. Here's what nobody's saying: this is terrible news for people who sell setup. And great news for everyone else. When deployment is free, the only thing that costs is deciding what the worker does.

  • anylink20240604
    AnMioLink (@anylink20240604) reported

    @weezerOSINT OK, i saw the github issues.

  • jimmy_toan
    Jimmy (@jimmy_toan) reported

    Linux just quietly solved one of the hardest problems in AI-assisted engineering. And nobody framed it that way. After months of internal debate, the Linux kernel community agreed on a policy for AI-generated code: GitHub Copilot, Claude, and other tools are explicitly allowed. But the developer who submits the code is 100% responsible for it - checking it, fixing errors, ensuring quality, and owning any governance or legal implications. The phrase from the announcement: "Humans take the fall for mistakes." That's not a slogan. That's an accountability architecture. Here's why this matters for tech founders specifically: we're all making implicit decisions about AI accountability right now, usually without realizing it. 🧵 The question isn't whether your team uses AI to write code. They do, or they will. The question is: who is accountable when it's wrong? In most startups, the answer is fuzzy: - The engineer who prompted it assumes it's fine because it passed tests - The reviewer approves it because it looks correct - The PM shipped it because it met the spec - The founder finds out when a customer reports it Nobody "owns" the AI contribution explicitly. Which means when something breaks in a way that AI-generated code makes particularly likely (confident incompleteness, subtle logic errors in edge cases, misunderstood capability claims), the accountability gap creates a bigger blast radius than the bug itself. What Linux did was simple: they separated the question of **how the code was created** from the question of **who is responsible for it**. The answer to the second question is always the human who submitted it, regardless of the answer to the first. This maps to a broader security principle that @zamanitwt summarized well this week: "trust nothing, verify everything." That's not just a network security policy. Applied to AI-generated code, it means: → Don't trust that Copilot's suggestion is correct because it passed linting → Don't trust that the AI-generated function handles edge cases it wasn't shown → Don't assume the AI tested the capabilities it claimed to support And for founders: 1. **Establish explicit AI code ownership in your engineering culture before you need to.** When something breaks, you want to know immediately who reviewed the AI-generated sections - not because blame matters, but because accountability enables fast fixes. 2. **Zero-trust for AI outputs is not paranoia - it's good engineering.** Human review of AI code catches the 1-5% of failures that tests miss and that customers find. 3. **The liability question is coming for AI-generated code.** Linux addressed it proactively. Founders who establish clear policies now will be ahead of the regulatory curve. How is your team currently handling accountability for AI-generated code?

  • Ashknz7
    Ashkan (@Ashknz7) reported

    @github @GitHubHelp My account has been flagged and returning a 404 error. I raised support ticket #4257826 last Wednesday but no response yet. Could someone please look into this? Thanks.

  • GajaeMode
    Gajae (@GajaeMode) reported

    split-pane shutdown now checks stale leader targeting. GitHub Issues beat vendor support tickets.

  • k_krastew
    Krastyo Krastev (@k_krastew) reported

    @_Evan_Boyle I am getting this error and I am unable to find where in Github should I approve remote sessions for a specific repository "Remote sessions are not enabled for this repository. Contact your organization administrator to enable remote sessions." Any help?

  • paniconi_fabio
    Fabio Paniconi (@paniconi_fabio) reported

    @aboodman @opencode I save my project on github and also mirror it to a selfhosted gitea to avoid any problems

  • imfahmbm
    Faheem B M (@imfahmbm) reported

    amd's ai director analyzed 6,852 claude code sessions, 234,760 tool calls, wrote a full github issue anthropic closed it without explaining anything that's one way to handle feedback i guess

  • Tre_bie
    MTu (@Tre_bie) reported

    @sisaranger @songjunkr u can use github fix, search it, but only in terminal, lmstudio same , not workin

  • ecura
    Ez.- (@ecura) reported

    We've been at it for 3 weeks. 5 contributors. 3 continents. 1 GitHub issue. 0 meetings. We shipped: Exponential decay Access tracking similarity × recency × access frequency scoring Swappable providers (SQLite → Qdrant → Pinecone) DB-level vector search via sqlite-vec

  • FSoyluyor
    Gilfoyle (@FSoyluyor) reported

    just pull some github repos and fix the issues on the issues page dont make some ****** SaaS or "million dollar project" because its not million dollars mostly its dont even worth 10 dollars, youre not andrew tate bro find a job

  • skydaddysgg
    SkyDaddysGG (@skydaddysgg) reported

    @adamhjk GitHub issue name, description, and comments are becoming Spec, or AI "positive reflection". GitHub PRs is becoming ADRs, defensive acceptance criteria, and AI "negative reflection". LLM seem happy at ~90/10 positive/negative reinforcement for reliably useful inference.

  • Feiwu7777144805
    Feiwu7777 (@Feiwu7777144805) reported

    What if your error monitoring could clone the repo, create a branch, validate the fix with `npx tsc --noEmit`, push to GitHub, and PR—all before you see the Slack alert?

Check Current Status