frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Launch HN: Freestyle: Sandboxes for AI Coding Agents

https://www.freestyle.sh
36•benswerd•1h ago•9 comments

Germany Doxes "UNKN," Head of RU Ransomware Gangs REvil, GandCrab

https://krebsonsecurity.com/2026/04/germany-doxes-unkn-head-of-ru-ransomware-gangs-revil-gandcrab/
154•Bender•3h ago•64 comments

I won't download your app. The web version is a-ok

https://www.0xsid.com/blog/wont-download-your-app
570•ssiddharth•3h ago•309 comments

sc-im Spreadsheets in Your Terminal

https://github.com/andmarti1424/sc-im
49•m-hodges•1h ago•5 comments

A Cryptography Engineer's Perspective on Quantum Computing Timelines

https://words.filippo.io/crqc-timeline/
66•thadt•2h ago•9 comments

Claude Code is unusable for complex engineering tasks with the Feb updates

https://github.com/anthropics/claude-code/issues/42796
259•StanAngeloff•3h ago•159 comments

81yo Dodgers fan can no longer get tickets because he doesn't have a smartphone

https://twitter.com/Suzierizzo1/status/2040864617467924865
154•josephcsible•1h ago•130 comments

Book Review: There Is No Antimemetics Division

https://www.stephendiehl.com/posts/no_antimimetics/
95•ibobev•3h ago•71 comments

Reducto releases Deep Extract

https://reducto.ai/blog/reducto-deep-extract-agent
16•raunakchowdhuri•1h ago•1 comments

Sky – an Elm-inspired language that compiles to Go

https://github.com/anzellai/sky
17•whalesalad•2h ago•0 comments

What Being Ripped Off Taught Me

https://belief.horse/notes/what-being-ripped-off-taught-me/
202•doctorhandshake•4h ago•126 comments

Show HN: I built a tiny LLM to demystify how language models work

https://github.com/arman-bd/guppylm
766•armanified•17h ago•114 comments

Microsoft hasn't had a coherent GUI strategy since Petzold

https://www.jsnover.com/blog/2026/03/13/microsoft-hasnt-had-a-coherent-gui-strategy-since-petzold/
711•naves•1d ago•490 comments

An open-source 240-antenna array to bounce signals off the Moon

https://moonrf.com/
223•hillcrestenigma•14h ago•45 comments

PostHog (YC W20) Is Hiring

1•james_impliu•4h ago

Gemma 4 on iPhone

https://apps.apple.com/nl/app/google-ai-edge-gallery/id6749645337
793•janandonly•22h ago•222 comments

The Last Quiet Thing

https://www.terrygodier.com/the-last-quiet-thing
21•coinfused•2d ago•4 comments

France pulls last gold held in US for $15B gain

https://www.mining.com/france-pulls-last-gold-held-in-us-for-15b-gain/
472•teleforce•9h ago•261 comments

Show HN: GovAuctions lets you browse government auctions at once

https://www.govauctions.app/
4•player_piano•1h ago•3 comments

The 1987 game “The Last Ninja” was 40 kilobytes

https://twitter.com/exQUIZitely/status/2040777977521398151
240•keepamovin•14h ago•152 comments

Show HN: Real-time AI (audio/video in, voice out) on an M3 Pro with Gemma E2B

https://github.com/fikrikarim/parlor
222•karimf•23h ago•24 comments

One ant for $220: The new frontier of wildlife trafficking

https://www.bbc.com/news/articles/cg4g44zv37qo
94•gmays•4d ago•23 comments

LÖVE: 2D Game Framework for Lua

https://github.com/love2d/love
380•cl3misch•2d ago•193 comments

Signals, the push-pull based algorithm

https://willybrauner.com/journal/signal-the-push-pull-based-algorithm
126•mpweiher•2d ago•32 comments

Drop, formerly Massdrop, ends most collaborations and rebrands under Corsair

https://drop.com/
109•stevebmark•13h ago•52 comments

Running Gemma 4 locally with LM Studio's new headless CLI and Claude Code

https://ai.georgeliu.com/p/running-google-gemma-4-locally-with
365•vbtechguy•1d ago•92 comments

When Virality Is the Message: The New Age of AI Propaganda

https://time.com/article/2026/04/02/when-virality-is-the-message-the-new-age-of-ai-propaganda/
48•virgildotcodes•3h ago•38 comments

Sheets Spreadsheets in Your Terminal

https://github.com/maaslalani/sheets
164•_____k•2d ago•42 comments

Music for Programming

https://musicforprogramming.net
300•merusame•23h ago•152 comments

Show HN: Gemma Gem – AI model embedded in a browser – no API keys, no cloud

https://github.com/kessler/gemma-gem
134•ikessler•17h ago•18 comments
Open in hackernews

Claude Code Down

71•theahura•2h ago
https://www.reddit.com/r/ClaudeCode/comments/1se1p30/broken_again/ https://www.reddit.com/r/ClaudeCode/comments/1se1vn1/oauth_api_key_expiring_daily/ https://www.reddit.com/r/ClaudeCode/comments/1se1ur5/500_error_or_timeout_when_trying_to_reauthorize/

Currently nothing on the status: https://status.claude.com/

Comments

rishabhaiover•2h ago
OAuth is failing, I can't login via claude code.
bronlund•1h ago
Same here.
matheusmoreira•1h ago
Same here. Usage limits are still pretty insane too.
kevinwang•1h ago
Still down for me. (And still nothing on the status page!)
ksajadi•1h ago
Failures all over Code and Chat here too (London UK) and status is showing all green
hgoel•1h ago
Here's hoping they can get it sorted quick. Hopefully these are just growing pains and not indicative of a GitHub style inability to achieve stability.
wojciem•1h ago
What are decent alternatives to ClaudeCode?
Leomuck•1h ago
Probably OpenCode, also works with Claude.
abi•1h ago
Codex is great.
rongenre•1h ago
I've found minimax to be quite good
gedy•1h ago
A keyboard /s
Lucasoato•1h ago
The real issue isn’t that Claude is down, it can happen. The problem is that the status page doesn’t report anything, even if it has been impossible to log in during the past hour. Status pages should be trusted, connected to real metrics and not fake pr stuff :/

EDIT: Now they show the issue, kudos to them! Transparency is the key to build trust. No body expects a perfect service, thanks Claude team for your efforts.

inglor_cz•1h ago
Interesting. I just fixed something using Claude Code. But I am located in Central Europe.
pixl97•1h ago
I've not looked into it, but I'd assume they have more than one data center.
matheusmoreira•1h ago
Me too. Claude Code is running just fine on one of my virtual machines. On another, it logged me out saying the OAuth token expired, and it won't log me back in due to internal server errors. Weird.
esafak•1h ago
You need to use a user-reported status page; the incentives are broken for self reporting.
hn1986•1h ago
it does show issues. https://status.claude.com/
simlevesque•1h ago
It shows issues now. Probably not when the person you're replying to wrote their comment.
rhgraysonii•1h ago
11 minutes elapsed between the comments. There is going to be some actual time before a report and the status page being live in a breaking system.
jakeydus•1h ago
I saw issues starting at least 45 minutes before their status page was updated.
rvz•1h ago
> The problem is that the status page doesn’t report anything, even if it has been impossible to log in during the past hour.

When Claude took an extra day off, he forgot to report his hours to the dashboard when he will be unavailable / unresponsive and this is probably why people here are complaining about no status update.

Wonder where I have seen that before?

chermi•1h ago
This has consistently pissed me off. It seems like we all just accepted that whatever they define as "functioning"/"OK" is suitable. I see the status now shows, but there should be a very loud third party ruthlessly running continuous tests against all of them. Ideally it would also provide proof of the degradation we seem to all agree happens (looking at you Gemini). Like a leaderboard focused on actual live performance. Of course they'd probably quickly game that too. But something showing time to first response, "global capacity reached" etc, effective throttling metric, a intelligence metric. Maybe like crowdsourced stats so they can't focus on improving the metrics for just the IPs associated with this hypothetical third party performance watchdog.

The one that pissed me off the most was Gemini API displaying very clearly 1) user cancelled request in Gemini chat app 2) API showing "user quota reached". Both were blatant lies. In the latter case, you could find the actual global quota cause later in the error message. I don't know why there isn't more outrage. I'm guessing this sort of behavior is not new, but it's never been so visible to me.

CoolGuySteve•1h ago
I'm finding qwen 27b is comparable to sonnet but my self hosting has about 5 more 9s than whatever Anthropic's vibe coding. I also don't have to worry about the quality of the model I'm being served from day to day.

Probably the most damning fact about LLMs is just how poorly written their parent companies' systems are.

tills13•1h ago
What do you run it on? And even then, I'm guessing your tokens per second are not great?
CoolGuySteve•1h ago
I get about 35-40tok/sec on a 3090.

It's actually about the same speed when accounting for how much more responsive my system is to Anthropic's saas infrastructure

jasonjmcghee•1h ago
People keep saying this and idk what I'm doing wrong. Using q8_0 on all the latest and greatest local models and they just don't come close to sonnet.

I've tried different harnesses, building my own etc.

They are reasonably close to haiku? Maybe?

bakugo•1h ago
You're not doing anything wrong, they are not comparable.

Claims to the tune of "this 0.5B local model running on my phone is almost as good as [large expensive model]" are common but greatly exaggerated, it's simply not true beyond the most basic use cases.

Only the much larger models (such as the 744B GLM-5) manage to come close, but nobody's running those locally.

chis•1h ago
Just to make one obvious critique your costs per token are probably about 1000x higher than the ones they provide.

I'm pretty sympathetic to Anthropic/OpenAI just because they are scaling a pretty new technology by 10x every year. It is too bad Google isn't trying to compete on coding models though, I feel like they'd do way better on the infra and stability side.

CoolGuySteve•1h ago
I've owned this GPU for 5 years already, it's fine
kccqzy•1h ago
But do you actually treat LLMs as glorified autocomplete or treat them as puzzle solvers where you give them difficult tasks beyond your own intellect?

Recently I wrote a data transformation pipeline and I added a note that the whole pipeline should be idempotent. I asked Claude to prove it or find a counterexample. It found one after 25 minutes of thinking; I reasonably estimate that it would take me far longer, perhaps one whole day. I couldn’t care less about using Claude to type code I already knew.

CoolGuySteve•1h ago
"give them difficult tasks beyond your own intellect?"

Lol no, I've yet to find a model with those properties. Sounds like a fast track to AI psychosis.

The domain I work in doesn't have enough public documentation for these models to be particularly helpful without a lot of handholding though.

hombre_fatal•57m ago
I've been working on a luks+btrfs+systemd tool (for managing an encrypted raid1 pool). While I have worked with each individually, it's not obvious what kind of cases you have to handle when composing them together. A lot of it is simply emergent, and the status quo has been to do your best and then see what actually happens at runtime.

Documentation is helpful to describe high-level intentions, but the beauty is when you have access to source code. Now a good model can derive behavior from implementation instead of docs which are inherently limited.

I implemented the luks+btrfs part by hand a few years ago, and I resurrected the project a couple months ago. Using source code for local reference, Claude discovered so many major cases I missed, especially in the unhappy-path scenarios. Even in my own hand-written tests. And it helped me set up an amazing NixOS VM test system include reproduction tests on the libraries to see what they do in weird undocumented cases.

So I think "tasks beyond our intellect (and/or time and energy)" can be fitting. Otherwise I'd only be capable of polishing this project if luks+btfs+systemd were specifically my day job. I just can't fit so much in my head and working memory.

zekica•30m ago
And it can fail in great ways. Last example: I asked claude for a non-trivial backup and recovery script using restic. I gave it the whole restic repo and it still made up parameters that don't exist in the code (but exist in a pull request that's been sitting not merged for 10+ months).
hombre_fatal•22m ago
Interesting. I don't think I've seen hallucinations at that level when it's referencing source code.

Though my workflow always starts in plan mode where Claude is clearly more thorough (which is the reason it takes 10x as long as going straight to impl). I rarely skip it.

shimman•1h ago
This says more about you than the "intellect" of these nondeterministic probability programs.

Can you provide actual context to what was beyond your ability and how you're able to determine if the solution was correct?

Finding out that all these comments that reference the "magical incantation" tend to be full of hot air. Maybe yours is different.

kccqzy•6m ago
> how you're able to determine if the solution was correct

I had hundreds of unit tests that did not trigger an assertion I added for idempotency. Claude wrote one that triggered an assertion failure. Simple as that. A counterexample suffices.

wise0wl•1h ago
I've tried a few models and some are decent, including Qwens models. I've tried a few harnesses like Roo Code in VSCode to put things together that in theory emulate the experience I get from VSCode + Claude or Copilot, but I generally find the experience extremely limited and frustrating.

How have you set things up to have a good experience?

CoolGuySteve•1h ago
I'm using the qwen cli tool with a duckduckgo search skill that I made Claude write. It's like bootstrapping I guess

Once it can search for factual information online the smaller model size becomes less noticeable

cyanydeez•1h ago
QWEN3.5-Next-Coder does wonders. It's drawbacks are time to first token is 30 seconds to load the model and OpenCode has an unsolved timeout issue on this load, but otherwise once it's warmed up, it's entirely serviceable.

I've got a AMD395+ with 128GB, so running a ~46GB model gives me about 85k tokens, which gives me easily copy/paste/find/replace behavior; it mocks up new components; it can wire in some functionality, but that's usually at it's limits and requires more debugging.

I've been looking at how to schedule it using systemd to keep a wiki up to date with a long loaded project and breaks the "blank page" issue with extending behaviors in a side project.

I understand some of these larger models can do things faster and smarter, but I don't see how they can implement novel functionality required for the type of app I'm concerned with. If I just wanted to make endless CRUD or TODO apps, I'm betting I could figure out a loop that's mostly hands off.

CharlieDigital•1h ago

    > Probably the most damning fact about LLMs is just how poorly written their parent companies' systems are
I have been working on some work related to MCP and found some gaps in implementation in Claude and Codex. This is a relatively simple, well-defined spec and both Claude Code and Codex CLI have incomplete/incorrect implementations.

During this process of investigation, I checked the CC repo and noticed they had 5000+ issues open. Out of curiosity, I skimmed through them and many point to regressions, real bugs, simple changes, etc. Maybe they have some internal tracker they are using, but you would think that a company with functionally unlimited tokens and access to the best models would be able to use those tokens to get their own house in order.

My sense now is that there is a need for the industry to create a lot of hype right now so we see showmanship like the kernel compiler and the agent swarms building a semi-functional browser, etc....yet their own tooling has not fully implemented their own protocol (MCP) correctly. They need all of us to believe that these agents are more capable than they actually are; the more piles of tangled code you write and the more discipline you cede to their LLMs, the more dependent you are on those LLMs to even know what the code is doing. At some point, teams become incapable of teasing the code apart anymore because no one will understand it.

Peeking at the issues in the repos and seeing big gaps in functionality like Codex's missing support for MCP prompts and resources is like looking behind the curtain at reality.

NitpickLawyer•1h ago
> Probably the most damning fact about LLMs is just how poorly written their parent companies' systems are.

This seems like a popular take, but I think it's the other way around. Them dogfooding cc with cc is proof that it can work, and that "code quality" doesn't really matter in the end.

Before cc claude.ai (equivalent of chatgpt) was meh. They were behind in features, behind in users, behind in mindshare. cc took them from "weirdos who use AI for coding" to "wait, you're NOT using cc? you freak" in ~1 year. And cc is a very big part of them reaching 1-2B$ monthly revenue.

Yes, it's buggy. Yes, the code is a mess (as per the leak, etc). But they're also the most used coding harness. And, on a technical side, having had cc as early as they did, helped them immensely on having users, having real-usage data, real-usage signals and so on. They trained the models on that data, and trained the models in sync with the harness. And it shows, their models are consistently the highest ranked both on benchmarks and on "vibes" from coders. Had they not have that, they would have lacked that real-world data.

And if you look at the competition it's even more clear. Goog is kidna nowhere with their gemini-cli, is all over the place with their antigravity-ex-windsurf, and while having really good generalist models, the general mindshare is just not there for coding. Same for oAI. They have an open-source, rust-based, "solid" cli, they have solid models (esp in code review, planning, architecture, bug fixing, etc) but they are not #1. Claude is with their cc.

So yeah, I think it's really the other way around. Having a vibe-coded, buggy, bad code solution, but being the first to have it, the first to push it, and the first to keep iterating on it is really what sets them apart. Food for thought on the future, and where coding is headed.

sdoering•1h ago
I am currently - and (this post is up 25 minutes as of now) have been using it without noticeable degradation over the last few hours.

Edit: But the status page - at least as of now - is clearly communicating elevated error rates.

buildbot•1h ago
Not that it is the best indicator, but downdector is showing many services with spikes at exactly the same time as Claude Codes issues began.
nightpool•1h ago
downdector always shows spikes when you go to look at it, and then they remove them later retroactively if the spikes are fake.
mistic92•1h ago
I don't have any issues
rvz•1h ago
Claude isn't down. He's on vacation for today and took an extra day off after the weekend.

He'll be back to work by tomorrow.

chermi•1h ago
Yup displays as an "auth" issue to me. Just a nice reminder that my original plan was to be provider agnostic but everything was working so well with cc I lost sight lol.
enraged_camel•1h ago
Claude Code inside the desktop app works for me.
user-•1h ago
I am a believer that everyone should have their main flow be model/provider agnostic at a high level. I often run out of claude tokens and use GLM-5 as backup.

https://gist.github.com/ManveerBhullar/7ed5c01a0850d59188632...

simple script i use to toggle which backend my claude code is using

cyanydeez•1h ago
Interesting; do you find they actually react the same way to the harness?
user-•1h ago
There are differences for sure. Claude models feel the most 'stable' in that I see less tool confusion messages and other mistakes like the one im looking at right now.

"Wait, I'm editing the wrong sections. The edit tool tried to match but replaced with different prop names than what was in the file. Let me re-read the file and understand the current state properly."

And of course models are not 1-to-1 and have different strengths and weaknesses. I know I wont get the same quality plan mode output probably. Its a tradeoff.

fastball•1h ago
Being model and provider agnostic are orthogonal concerns.

e.g. you can run Claude models on AWS Bedrock giving you provider choice for the same model. Whether or not you need model agnosticism at that point seems like a very different question.

user-•1h ago
> . you can run Claude models on AWS Bedrock giving you provider choice for the same model

Is anyone doing this for personal dev that isnt token fed by employers? Coding plans are subzidized for a reason right? If I did API usage from a cloud provider id be out tens of thousands already.

bob1029•1h ago
I tried the agnostic thing for a while, but there are enough quirks between the providers that I gave up trying to normalize it. GPT5.x wipes the floor with other models for my specific tool calling scenarios. I am not going to waste time trying to bridge arbitrary and evolving gaps between providers.

I put my Amex details into OAI, I get tokens, it just works. I really don't understand what the hell is going on with Claude. The $200/m thing is so confusing to me. I'd rather just go buy however many tokens I plan to use. $200 worth of OAI tokens would go a really long way for me (much longer than a month), but perhaps I am holding it wrong.

lovecode007•1h ago
Yup for me too - VSC Claude is def down and not working
lovecode007•1h ago
Def down, keeps saying internal server error
jgrahamc•1h ago
Here in my corner of Europe it seems to be working fine.
kwertyoowiyop•1h ago
For older programmers: this is like when Stack Overflow would go down.

For really old programmers: this is like when Computer Literacy bookstore was closed.

ChrisArchitect•1h ago
Link for up top: https://status.claude.com/incidents/vfjv5x6qkd4j
rishabhaiover•1h ago
The downtime forces me to relook at my utterly dependent relationship with agentic assistance. The inertia to begin engaging with my code is higher than it has ever been.
matheusmoreira•1h ago
Yeah. It's actually starting to make me anxious. I think I got addicted to these agents.
theanonymousone•1h ago
How much is remaining until the last 9 is gone too?
chermi•1h ago
Wtf. Was this just scrubbed/pushed down from frontpage?
hgoel•1h ago
IIRC threads that are just "yup, seeing this too" are not seen as being valuable here. There isn't (or at least wasn't) much discussion happening.
Danielzzzz•1h ago
Seems to be good now. Just logged in successfully. Can't live without Claude nowadays is the life learning I realized in the downtime retro to myself lol.
efficax•55m ago
This is pretty much every monday morning, so it's either scale issues with the busiest window of time (people getting started at work on monday) or it's intentional "outage" that only affects some % of people to take load off the system so that API users (who pay more) can be served during the heaviest usage time of the week.
boleary-gl•53m ago
If you still need access we balance across Claude and AWS via https://kilo.ai/docs/gateway - and you can BYOK for many providers
incomingpain•50m ago
They banned all third party.

Loads of people cancelled their subscriptions.

Should be the least load they have been under in months. Yet unreliable.

Crazy that people are going with their benchmaxxed models.