Update - We are continuing to investigate this issue. Oct 31, 2025 - 09:25 UTC
Investigating - We are currently investigating this issue. Oct 31, 2025 - 09:17 UTC
https://console.anthropic.com/api/auth/send_magic_link 500 (Internal Server Error)
I realize this is already a problem for other jobs, which require working with SAAS, but it seems odd to me that now some developers will fall into this "helpless" category as well.
Serious answer: I can write code manually, but it feels like a waste of time. I'll just go for a walk to synthesize my ideas if a service was down, and I don't think not writing actual code for a day is a huge problem. So focus on health and maybe even talk to humans.
We usually try to figure out how to build reliability/redundancy in step with what we require to function as a society under most circumstances without taking outsized losses.
When things go worse than anticipated, we take the hit, try to recover and maybe learn to strengthen the system afterwards. I would rate us roughly okay-ish at that, mostly because I don't know what to compare it to, since we are the only species to do it at this level to my knowledge.
Please let us:
* Sign in with Apple on the website
* Buy subscriptions from iOS In App Purchases
* Remove our payment info from the website
* Give paying subscribers an easy way to get actual support
As a frequent traveller I'm not sure if some of those features are gated by region, because some people said they can do some of those things.
> Elevated errors on claude.ai
> Identified - The issue has been identified and a fix is being implemented. Oct 31, 2025 - 10:18 UTC
> Investigating - We are currently investigating this issue. Oct 31, 2025 - 09:25 UTC
So enjoy your extra capacity :)
TBH I'm not so irritated by this. It keeps me grounded, and saves me from being unconsciously outsourcing all the hard work of thought process to AI.
TBH I'm not so irritated by this. It keeps me grounded, and saves me from being unconsciously outsourcing all the hard work of thought process to AI.
I've also had it suggest solutions e.g. "This C++/OpenGL code doesn't run in wayland can you suggest some solutions".
Other stuff like generating tests is hit and miss.
It randomly fails halfway through a response, sometimes very slow to start, hangs for long periods during a response, and so on.
The Claude chat interface can also slow down with long sessions. I sometimes use Claude code which is better, but I'm not a huge fan of terminal interfaces. I'm aware of third party frontends, but I believe those require api access which I don't like for personal use.
Fix for me: Hopefully will work for you guys too, I logged out of claude, restarted cursor, used the anthropic console login method instead of normal login, when you click the link it gives you an option to signin with chat credentials instead, there it did not work, I pressed the link given in claude a few times and kept trying, finally I was given a pastable code, this took a while to be accepted in terminal but now is logged in.
hope it works for someone else aswell!!
At least their customer service is nice. They forwarded my messages to the dev team and they implemented a lot of things I have suggested.
On a side note, I'm anthropomorphising too much, gonna have to upgrade and get some top rate therapy...
https://docs.z.ai/devpack/tool/claude
Smart move making it a drop-in replacement.
For that purpose? It lets me do things I never would have even tried.
It's not as smart as Opus by the way. Seems to match Sonnet.
Edit: See here, they've got an anthropic-compatible endpoint for this purpose - https://docs.z.ai/devpack/tool/claude
LLMs neither have the mechanical reliability we expect from computers (does it the same way every time) nor the flexible reliability we expect from biological intelligences (solves or works around the unexpected sub-problems as they arise).
I find the debate about LLMs rather exhausting. I find them useful and almost every day someone on social media tells me I'm mistaken, lying or merely deluded .
Plenty on Reddit saying they did. And I did.
Could the outage be a the result of an "unexpected" surge in account activations / use?
Not much of a "welcome back" ;-)
Stack Overflow is dying, it's extremely difficult to get new questions through. Even if they survive moderation then they're unlikely to get answers.
I use Claude Code for programming work, but I choose OpenAI anywhere customer-facing and this cute little outage is making me feel better about that with every passing minute. NOT cool.
"Well, what do you do when the power goes out?", he asked.
"I go home, just like you would.", I said with a smile.
He paused for a moment and nodded, "you know, you're absolutely right".
Even the engineers at these AI companies can't use these LLMs to fix an outage when there is one. Especially SREs.
But if one has to just sit there and "wait" for the outage to subside then perhaps the kitchen timer just went off and declared that these "developers" are cooked.
But what I meant was that the whole response completely disappears. Sometimes the text I wrote previously is pasted back into the text input, but sometimes it's not.
I have this habit of copying my prompt in case it happens.
And no, Claude sucks ass. It's like Anthropic does not want to make money. For a company that's targeting enterprise customers, they are totally unprepared. Like forget customer support, they can't even sell properly. They brag about insane capabilities on the Max plan but good luck trying to buy that on a team plan with company billing.
Even if OpenAI doesn't have the best model, at least they know what to do to make money.
They were broken for a week, I found several people talking about it on Reddit. But no word from Anthropic, no status page info.
I opened a support request, there was no response until 3 or 4 days later when someone messaged to say that it was fixed, and a status page related to it magically appeared.
Other forces are to blame as well, though. In the 80s and 90s there were UI research labs in indistry that did structured testing of user interactions, measuring how well untutored users could accomplish assigned tasks with one UI design versus another, and there were UI-design teams that used the quantitative results of such tests to deign UIs that were demonstrably easier to learn and use.
I don't know whether anyone is doing this anymore, for reasons I'll metion below.
Designing for use is one thing. Designing for sales is another. For sales you want a UI to be visually appealing and approachable. You probably also want it to make the brand memorable.
For actual use you want to hit a different set of marks: you want it to be easy to learn. You want it to be easy to gradually discover and adopt more advanced features, and easy to adapt it to your preferred and developing workflow.
None of these qualities is something that you can notice in the first couple of minutes of interacting with a UI. They require extended use and familiarization before you even know whether they exist, much less how well designed they are.
I think that there has been a general movement away from design for use and toward a design for sales. I think that's perfectly understandable, but tragic. Understandable because if something doesn't sell then it doesn't matter what its features are. Tragic because optimizing for sales doesn't necessarily make a product better for use.
Windows 11, iOS7, iOS26 are just some example of non Web UIs, which focused first on optimizing for sales, i.e. making something look good without thinking about usability implications.
Fortunately usability testing is still pretty much a thing. Good articles here: https://www.nngroup.com/search/?q=usability+testing