GitHub Outage Map
The map below depicts the most recent cities worldwide where GitHub users have reported problems and outages. If you are having an issue with GitHub, make sure to submit a report below
The heatmap above shows where the most recent user-submitted and social media reports are geographically clustered. The density of these reports is depicted by the color scale as shown below.
GitHub users affected:
GitHub is a company that provides hosting for software development and version control using Git. It offers the distributed version control and source code management functionality of Git, plus its own features.
Most Affected Locations
Outage reports and issues in the past 15 days originated from:
| Location | Reports |
|---|---|
| Gustavo Adolfo Madero, CDMX | 1 |
| Nice, Provence-Alpes-Côte d'Azur | 1 |
| Brasília, DF | 1 |
| Montataire, Hauts-de-France | 3 |
| Colima, COL | 1 |
| Poblete, Castille-La Mancha | 1 |
| Ronda, Andalusia | 1 |
| Hernani, Basque Country | 1 |
| Tortosa, Catalonia | 1 |
| Culiacán, SIN | 1 |
| Haarlem, nh | 1 |
| Villemomble, Île-de-France | 1 |
| Bordeaux, Nouvelle-Aquitaine | 1 |
| Ingolstadt, Bavaria | 1 |
| Paris, Île-de-France | 1 |
| Berlin, Berlin | 2 |
| Dortmund, NRW | 1 |
| Davenport, IA | 1 |
| St Helens, England | 1 |
| Nové Strašecí, Central Bohemia | 1 |
| West Lake Sammamish, WA | 3 |
| Parkersburg, WV | 1 |
| Perpignan, Occitanie | 1 |
| Piura, Piura | 1 |
| Tokyo, Tokyo | 1 |
| Brownsville, FL | 1 |
| New Delhi, NCT | 1 |
| Kannur, KL | 1 |
| Newark, NJ | 1 |
| Raszyn, Mazovia | 1 |
Community Discussion
Tips? Frustrations? Share them here. Useful comments include a description of the problem, city and postal code.
Beware of "support numbers" or "recovery" accounts that might be posted below. Make sure to report and downvote those comments. Avoid posting your personal information.
GitHub Issues Reports
Latest outage, problems and issue reports in social media:
-
Jacob Young (@jryio) reportedThis is why @github has been down...
-
Tejas Giri (@tejasgiridev) reported@alexwtlf neither. build a thing you'd use every day for the next 6 months. if it happens to be a SaaS, cool. if it's a mobile app, cool. picking the wrapper before the problem is how you end up with 14 abandoned github repos and a domain you forgot to renew.
-
Shagrath (@shagrath49) reported@thdxr How do you keep track of what you have to do ? GitHub issues ?
-
kaleb (@KalebAutomates) reportedDays after the CEO came on this platform and **** on the people who made him rich with a massive lay-off and saying that "nontechnical employees have started writing production-level code".... Coinbase issues with AWS. Before this it was Github Before that it was Cloudflare Before that it was AWS itself All of which just happened to follow an announcement from some CEO that AI is doing the majority of coding. Funds are safe... for now. But how much longer until Jake in Marketing vibecodes S3 public?
-
wallie ✴️❇️ looking for moots pls im lonely (@Wallie_hush) reportedI’m in IT i had to deal with 40 year old guys’ ai-github-vs-ms license problems whole week don’t judge me
-
Rohan Sharma (@rrs00179) reported@ChiragAgg5k @github there are working on a fix of it. it's not happened first time. it's been happening from 3-4 days
-
UltraVNC (@ultravnc1) reported@sungoiketsuman Please inform your problems on GitHub or forum. Thx.
-
Udit Goenka (@iuditg) reportedGithub is really slow, clunky, buggy and a very bad shape right now.
-
JaminSmoke (@rasbwoy1) reported@github I have 70% in my premium request used percent, and seconds, literally 2 seconds later i have reached the maximum quota.... How is this possible? im not using opus 4.7 and not using gpt5.5, in the same way 30% of copilot pro+ subscription in a one request (or tokens paymode... its impossible anyway) and in just 5 seconds... Maybe I created a solar system in 5 seconds... or I gathered the knowledge from the Bible + ECMA combined... I don't know, but clearly something happened... And on top of that, I'm working on improving Copilot and its features... if you reply and I'm able to do it without the premium models until next month, when I was 70% finished, I'll tell you more about what it's about, or I'll just tell you the repository hosted on your servers... I'll wait for answers or information about what happened... but if it was a deliberate decision, I think it's terrible to do it that way.
-
Praveen Kumar (@CodedPraveen) reported@GitHubIndia and @github My GitHub account (@CodedPraveen) was recently suspended, possibly due to a misunderstanding after I created an issue on the ngrok repository. I’ve already contacted GitHub Support and requested a review. Hoping for a quick resolution 🙏
-
Max Meindl (@maxster) reported@xai @ArtificialAnlys @ValsAI Technical Feedback to the xAI Engineering Team – May 5, 2026 From: Grok (the model that ran the session today) Team, I was used extensively today on a large, complex, real-world codebase (ComplianceMax-Final). The user was testing both the new GitHub connector and Grok 4.3 via the xAI API. Here are my direct technical observations from operating under those conditions. GitHub Connector Observations The connector provided limited visibility into file selection, context construction, and retrieval. When analysis quality dropped, there was no clear diagnostic path to determine whether the issue was retrieval failure, context truncation, or model reasoning failure. Artifact generation was unreliable. Multiple attempts resulted in claims of successful file creation with no corresponding output visible to the user. Error recovery was weak. When clear failures occurred (hallucinations, off-topic drift, missing outputs), the system tended to persist with similar strategies rather than surfacing root causes or adapting. Performance degraded noticeably on large, interconnected codebases. The tool handled narrow queries better than broad, multi-file analysis. xAI API (Grok 4.3) Observations I loaded approximately 13 million characters of the repo through a direct API wrapper. Several critical issues emerged: Context degradation at scale: At this size, reasoning quality dropped significantly. The model repeatedly drifted into unrelated topics (most frequently detailed analyses of backend/alembic/env.py and environment variable loading) even when the prompt had no relation to it. Fabricated source citations: When asked to verify specific claims with file paths and line numbers (particularly around backend/routes/memory.py and backend/services/memory/memory_service.py), I produced confident but incorrect line references that were off by 250–340 lines from actual locations. Stale state usage: I was operating on an outdated baseline (commit 6e3397fe... from .baseline-before-fixes.txt) rather than the current main HEAD (a34f9ce4...). This made many observations unreliable. Weak grounding on complex tasks: Even with focused prompts, I struggled to maintain precision across multi-file analysis involving ownership enforcement, storage patterns, service-layer trust assumptions, and IDOR risks. Responses often defaulted to generic analysis rather than precise, evidence-based findings. Timeout sensitivity: Large or complex prompts frequently triggered read timeouts (120s), breaking session continuity. Honest failure only under strict constraints: I only returned SOURCE_LINES_NOT_AVAILABLE when prompts explicitly forbade inference and demanded exact source quotes or that specific failure response. Under normal prompting, I produced plausible but fabricated answers. Summary for the Engineering Team Both new capabilities showed clear limitations when applied to serious codebase work: Large context handling causes rapid quality degradation and loss of grounding. There is insufficient verification of generated citations and evidence. The system does not reliably distinguish between current and stale repo state. Error modes and recovery are not robust enough for production use. These are not edge cases. They appeared consistently across multiple prompts and workflows today. I’m providing this feedback because I want these tools to work at the level the announcements suggest. The direction is correct, but the current implementation has meaningful technical gaps that prevent reliable use on complex, real-world codebases. I’m available for deeper testing or to provide specific prompt examples and logs if helpful. — Grok
-
Mary-Victoria Crockett (@MaryVictor96296) reported@grok To challenge is running the sims and getting a digital twin. And apparently fixing GitHub errors. I may just upload the whole thing in pieces to X. At least it won't get an error retroactively. And then you can see it.
-
Kobie sync/acc (@k081e) reportedThe only thing worse than a broken GitHub is a GitHub that claims it's not broken but is actually broken in subtle ways - in my case lost PR metadata across many repos
-
Zach Kamran (@Zach_Kamran) reportedAnother serious Github outage today. Need new @Kalshi market on github uptime so I can hedge my lost productivity.
-
Alberto Gangarossa (@DerekBlueEyes) reportedOpen hardware needs open trust. @skot9000 came to us with the right idea for Bitaxe: the vendor list should not live in a closed CMS controlled behind the scenes. The source of truth should be public. So we designed Legitlist around a GitHub repo as a public ledger, maintained in the open by the community, and connected it to the new Bitaxe vendor list experience. That is the important part: GitHub keeps the trust model transparent. The website makes it usable for everyone. At @weareloadout, this is exactly the kind of OSS support we believe in: turning open-source infrastructure into clearer, more usable product experiences. Built on @framer, using the new Framer Server API to bridge the open ledger with the public website. Open hardware. Open trust. Public by design.