frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Rust 1.93.0

https://blog.rust-lang.org/2026/01/22/Rust-1.93.0/
1•birdculture•2m ago•0 comments

Show HN: Cryptography, JWT, and ASN1 Debuggers

https://crypto.qkation.com/
1•TheBestTvarynka•3m ago•0 comments

Show HN: Ctx – Context manager for Cloud,K8s VPNs, SSH tunnels, secret managers

https://github.com/vlebo/ctx
2•vlebo•4m ago•1 comments

Show HN: PolyMCP – Simplifying MCP Server Development and Agent Integration

2•justvugg•10m ago•0 comments

China tests hypergravity centrifuge that compresses time and space

https://www.futura-sciences.com/en/china-tests-the-unthinkable-with-a-centrifuge-that-compresses-...
4•walterbell•13m ago•1 comments

Show HN: Hexpiece – a daily chess coverage puzzle

https://hexpiece.com/
2•tothemoon•14m ago•0 comments

Resilience in an Era of Systemic Uncertainty [pdf]

https://static.zegemabeach.com/zegema_beach-2026_resilience_report-20260120.pdf
2•durakot•15m ago•0 comments

Judge rejects DoorDash, Uber bid to block New York City tipping laws

https://www.reuters.com/world/judge-rejects-doordash-uber-bid-block-new-york-city-tipping-laws-20...
3•toomuchtodo•15m ago•0 comments

Arrows to Arrows, Categories to Queries

https://reasonablypolymorphic.com/blog/arrows-to-arrows/
2•surprisetalk•15m ago•0 comments

Maze Algorithms (2017)

http://www.jamisbuck.org/mazes/
4•surprisetalk•15m ago•0 comments

The Math on AI Agents Doesn't Add Up

https://www.wired.com/story/ai-agents-math-doesnt-add-up/
3•brycehalley•17m ago•1 comments

So, why *should* GNOME support server side decorations?

https://blister.zip/posts/gnome-ssd/
3•birdculture•20m ago•0 comments

Umping giants: Fossils show giant prehistoric kangaroos could still hop

https://www.manchester.ac.uk/about/news/jumping-giants-fossils-show-giant-prehistoric-kangaroos-c...
2•gnufx•20m ago•0 comments

Airlines to save big money on fuel as new weight loss pills gain popularity

https://www.cnbc.com/2026/01/14/airlines-to-save-on-fuel-as-weight-loss-pills-grow-popular-wall-s...
3•fortran77•21m ago•0 comments

JSX for AI Video

https://varg.ai/sdk
3•mifydev•22m ago•0 comments

You can now talk to your shell in plain language

https://www.nlsh.dev/
3•JunaidL•23m ago•0 comments

CloudlyKit – a chatkit widget for your site (after struggling with ChatKit)

https://www.cloudlykit.com/
2•jacoblarszon•24m ago•1 comments

Like digging 'your own professional grave': The translators losing work to AI

https://www.cnn.com/2026/01/23/tech/translation-language-jobs-ai-automation-intl
4•methuselah_in•26m ago•0 comments

(Open) Widevine support added to the Chromium port

https://undeadly.org/cgi?action=article;sid=20260118112808
2•speckx•26m ago•0 comments

Show HN: Multi-agent deliberation plugin for Claude Code

https://github.com/BayramAnnakov/agent-tower-plugin
2•Bayram•27m ago•0 comments

Data Structures and Algorithms – Preparing for Interviews

https://tech.stonecharioteer.com/posts/2026/dsa/
2•stonecharioteer•28m ago•0 comments

A Better Practices Guide to Using Claude Code

https://kylestratis.com/posts/a-better-practices-guide-to-using-claude-code/
2•csoham•29m ago•0 comments

"Content Unavailable" – YouTube's latest attempt at breaking ad blockers

https://www.androidpolice.com/youtube-content-unavailable-bug/
5•mikece•31m ago•1 comments

Rust's Standard Library on the GPU

https://www.vectorware.com/blog/rust-std-on-gpu/
2•PaulHoule•32m ago•0 comments

DevCon Fall 2025 – Tracking AI generated code with Git [video]

https://www.youtube.com/watch?v=irK4G2SzhpA
2•addcn•33m ago•0 comments

Gold To 5000 USD, accelerates for more?

https://www.cnbc.com/quotes/@GC.1
4•trilogic•33m ago•4 comments

Show HN: 83 browser-use trajectories, visualized

https://trails-red.vercel.app/viewer
3•wayy•33m ago•0 comments

Steve Yegge Promoting "Gas Town" Crypto Coin

https://steve-yegge.medium.com/bags-and-the-creator-economy-249b924a621a
4•toraway•34m ago•1 comments

Show HN: First autonomous ML and AI engineering Agent

https://marketplace.visualstudio.com/items?itemName=NeoResearchInc.heyneo
2•svij137•34m ago•0 comments

They quit their jobs to bet on current events: A look inside prediction markets

https://www.npr.org/2026/01/21/nx-s1-5672615-e1/they-quit-their-day-jobs-to-bet-on-current-events...
1•paulpauper•38m ago•0 comments
Open in hackernews

Claude.ai silently failing since Jan 14, no official acknowledgment

https://github.com/anthropics/claude-code/issues/18866
119•nurimamedov•1h ago

Comments

measurablefunc•1h ago
Maybe they shouldn't have trusted their LLM to optimize their kernels so much.
OGEnthusiast•1h ago
I feel like Claude Code is starting to fall over from being entirely written by LLMs. How do you even begin to fix precise bugs in a 1M+ LOC codebase all written by AI? It seems like LLMs are great for quickly adding large new features but not great for finding and fixing edge-cases.
throwjjj•1h ago
Wishful thinking. Lol. It’s GPu related
rigel8•49m ago
compaction actually reduces GPU usage?
eunoia•1h ago
This is real. I’ve seen some baffling bugs in prompt based stop hook behavior.

When I investigated I found the docs and implementation are completely out of sync, but the implementation doesn’t work anyway. Then I went poking on GitHub and found a vibed fix diff that changed the behavior in a totally new direction (it did not update the documentation).

Seems like everyone over there is vibing and no one is rationalizing the whole.

heliumtera•52m ago
>Seems like everyone over there is vibing and no one is rationalizing the whole.

Claude Code creator literally brags about running 10 agents in parallel 24/7. It doesn't just seems like it, they confirmed like it is the most positive thing ever.

TrainedMonkey•25m ago
It's software engineering crack. Starting a project feels amazing, features are shipping, a complex feature in the afternoon - ezpz. But AI lacks permanence, for every feature you start over from scratch, except there is more of codebase now, but the context window is still the same. So there is drift, codebase randomizes, edge cases proliferate, and the implementation velocity slows down.

Full disclosure - I am a heavy codex user and I review and understand every line of code. I manually fight spurious tests it tries to add by pointing a similar one already exists and we can get coverage with +1 LOC vs +50. It's exhausting, but personal productivity is still way up.

I think the future is bright because training / fine-tuning taste, dialing down agentic frameworks, introducing adversarial agents, and increasing model context windows all seem attainable and stackable.

MrDarcy•22m ago
I know at least one of the companies behind a coding agent we all have heard of has called in human experts to clean up their vibe coded IAC mess created in the last year.
klodolph•36m ago
I’m happy to throw an LLM at our projects but we also spend time refactoring and reviewing each other’s code. When I look at the AI-generated code I can visualize the direction it’s headed in—lots of copy-pasted code with tedious manual checks for specific error conditions and little thought about how somebody reading it could be confident that the code is correct.

I can’t understand how people would run agents 24/7. The agent is producing mediocre code and is bottlenecked on my review & fixes. I think I’m only marginally faster than I was without LLMs.

gpm•34m ago
> with tedious manual checks for specific error conditions

And specifically: Lots of checks for impossible error conditions - often then supplying an incorrect "default value" in the case of those error conditions which would result in completely wrong behavior that would be really hard to debug if a future change ever makes those branches actually reachable.

klodolph•26m ago
I always thought that the vast majority of your codebase, the right thing to do with an error is to propagate it. Either blindly, or by wrapping it with a bit of context info.

I don’t know where the LLMs are picking up this paranoid tendency to handle every single error case. It’s worth knowing about the error cases, but it requires a lot more knowledge and reasoning about the current state of the program to think about how they should be handled. Not something you can figure out just by looking at a snippet.

stefan_•16m ago
The answer (as usual) is reinforcement learning. They gave ten idiots some code snippets, and all of them went for the "belt and braces" approach. So now thats all we get, ever. It's like the previous versions that spammed emojis everywhere despite that not being a thing whatsoever in their training data. I don't think they ever fixed that, just put a "spare us the emojis" instruction in the system prompt bandaid.
zbentley•5m ago
Training data from junior programmers or introductory programming teaching material. No matter how carefully one labels data, the combination of programming’s subjectivity (damaging human labeling and reinforcement’s effectiveness at filtering around this) and the sheer volume of low-experience code in the input corpus makes this condition basically inevitable.
human_person•23m ago
This is my biggest frustration with the code they generate (but it does make it easy to check if my students have even looked at the generated code). I dont want to fail silently or hard code an error message, it creates a pile of lies to work through for future debugging
colechristensen•18m ago
Writing bad tests and error handling have been the worst performance part of Claude for me.

In particular writing tests that do nothing, writing tests and then skipping them to resolve test failures, and everybody's favorite: writing a test that greps the source code for a string (which is just insane, how did it get this idea?)

skerit•23m ago
I switched to OpenCode, away from Claude-Code, because Claude-Code is _so_ buggy.
einpoklum•14m ago
> When I investigated I found the docs and implementation are completely out of sync, but the implementation doesn’t work anyway.

That is not an uncommon occurrence in human-written code as well :-\

tobyjsullivan•5m ago
Someone said it best after one of those AWS outages from a fat-fingered config change:

Automation is great for scaling work but it also we can break more things faster when things go wrong.

(Paraphrasing because I don’t have the original comment link.)

agumonkey•59m ago
structural (team) recursion and statistical output don't mesh well together ?
egeozcan•53m ago
I think you can keep the vibe coding totally under control with good tests. I'm baffled how such a huge regression would not be caught.

Then again, the google home page was broken on FF on Android for how long?

cyanydeez•52m ago
Whose going to write those good tests, and are they going to write them or disable them cause they dont work.
AstroBen•47m ago
I think you underestimate how impossible of a task it is to write sufficient test coverage to keep AI in line

You can assert that something you want to happen is actually happening

How do you assert all the things it shouldn't be doing? They're endless. And AI WILL mess up

It's enough if you're actively reviewing the code in depth.. but if you're vibe coding? Good luck

gpm•46m ago
Not my experience at all when I occasionally try making something purely coded by AI for fun. It starts off fine but the pile of sub-optimal patterns slowly builds towards an unmaintainable mess with tons of duplication of code, and state that somehow needs to be kept in sync. Tests and linters can't test that the code is actually reasonable code...

Doesn't mean it's not a useful tool - if you read and think about the output you can keep it in check. But the "100% of my contributions to Claude Code were written by Claude Code" claim by the creator makes me doubt this is being done.

gpm•41m ago
PS. In the 5 minutes between starting and finishing writing the parent comment https://claude.ai/settings/usage just stopped displaying my quota usage... fun.

Edit: And 3 minutes later it is back...

AstroBen•39m ago
Everyone has been stressing over losing their job because of AI. I'm genuinely starting to think this will end in 5x more work needing to clean up the mess caused. Who's going to maintain all this generated code?
dawnerd•35m ago
That's what I'm worried about. I hate cleaning up AI code now from contractors. If that's going to be the future of this gig, I'm out.
alephnerd•33m ago
Most of us in the financial side of this space think so as well. This is why AI Ludditism doesn't make sense - CAT Hydraulic Excavators didn't end manual shovelers, it forced them to upskill.

Similarly, Human-in-the-loop utilization of AI/ML tooling in software development is expected and in fact encouraged.

Any IP that is monetizable and requires significant transformation will continue to see humans-in-the-loop.

Weak hiring in the tech industry is for other reasons (macro changes, crappy/overpriced "talent", foreign subsidies, demanding remote work).

AI+Competent Developer paid $300k TC > Competent Developer paid $400k TC >>> AI+Average Developer paid $30k TC >> Average Developer paid $40k TC >>>>> Average Developer paid $200k TC

FeteCommuniste•33m ago
> Who's going to maintain all this generated code?

Other AI agents, I guess. Call Claude in to clean up code written by Gemini, then ChatGPT to clean up the bugs introduced by Claude, then start the cycle over again.

inimino•29m ago
Nobody is going to maintain it, the spec that generated it will be given to better systems and it will be rewritten.
direwolf20•27m ago
It won't be maintained — quality will decrease forever.
ssl-3•11m ago
Or: We throw it all out and call the next iterations "version 2."

If the code is cheap (and it certainly is), then tossing it out and replacing it can also be cheap.

jordanbeiber•3m ago
Using AI doesn’t really change the fact that keeping ones and zeroes in check is like trying to keep quicksand in your hands and shape it.

Shaping of a codebase is the name of the game - this has always been, and still, is difficult. Build something, add to it, refactor, abstraction doesn’t sit right, refactor, semantics change, refactor, etc, etc.

I’m surprised at how so few seem to get this. Working enterprise code, many codebases 10-20 years old could just as well have been produced by LLMs.

We’ve never been good at paying debt and you kind of need a bit of OCD to keep a code base in check. LLM exacerbates the lack of continuous moulding as iterations can be massive and quick.

tyfon•25m ago
I've been trying opencode a bit with gemini pro (and claude via those) with a rust project, and I have a push pre-hook to cargo check the code.

The amount of times I have to "yell" at the llm for adding #[allow] statements to silence the linter instead of fixing the code is crazy and when I point it out they go "Oops, you caught me, let me fix it the proper way".

So the tests don't necessarily make them produce proper code.

egeozcan•16m ago
I added a bunch of lines telling it to never do that in CLAUDE.md and it worked flawlessly.

So I have a different experience with Claude Code, but I'm not trying to say you're holding it wrong, just adding a data point, and then, maybe I got lucky.

OptionOfT•11m ago
You don't. These seems to be this idea that LLMs can do it all, but the reality is that it itself has limited amounts of memory, and thus context.

And this is not tied to the LLMs. It's that to EVERYTHING we do. There are limits everywhere.

And for humans the context window might be smaller, but at least we have developed methods of abstracting different context windows, by making libraries.

Now, as a trade-off of trying to go super-fast, changes need to be made in response to your current prompts, and there is no time validate behavior in cases you haven't considered.

And regardless of whether you have abstractions in libraries, or whether you have inlined code everywhere, you're gonna have issues.

With libraries changes in behavior are going to impact code in places you don't want, but also, you don't necessarily know, as you haven't tested all paths.

With inlined code everywhere you're probably going to miss instances, or code goes on to live its own life and you lose track of it.

They built a skyscraper while shifting out foundational pieces. And now a part of the skyscraper is on the foundation of your backyard shed.

quietsegfault•10m ago
What differences do you see between AI written codebases and a codebase written by engineers? Both parties create buggy code, but I can imagine the types of bugs are different. Is it just that bug fixing doesn't really scale because we don't have the ability to chomp down 1M+ LOC codebases into LLM context?
dataviz1000•8m ago
> I feel like Claude Code is starting to fall over from being entirely written by LLMs.

The degradation is palpable.

I have been using vscode github copilot chat with mostly the claude opus 4.5 model. The underlying code for vscode github copilot chat has turned to shit. It will continuously make mistakes no matter what for 20 minutes. This morning I was researching Claude Code and pricing thinking about switching however this post sounds like it has turned to shit also. I don't mind spending $300-$500 a month for a tool that was a month ago accomplishing in a day what would take me 3-4 days to code. However, the days since the last update have been shit.

Clearly the AI companies can't afford to run these models at profit. Do I buy puts?

heliumtera•1h ago
Vibes too strong, forgot to add "make no mistake" to compact feature, classic
cube00•27m ago
Needs to start with CRITICAL:
esafak•59m ago
I'm just waiting for the formatting to be fixed. Maybe they should have invested in paying off that vibe coding tech debt...
AznHisoka•54m ago
This is precisely why i cancelled my claude max account and switched back to chatgpt. Claude is much better but not when it silently stops workinf
jampa•52m ago
Slightly off topic, but does anyone feel that they nerfed Claude Opus?

It's screwing up even in very simple rebases. I got a bug where a value wasn't being retrieved correctly, and Claude's solution was to create an endpoint and use an HTTP GET from within the same back-end! Now it feels worse than Sonnet.

All the engineers I asked today have said the same thing. Something is not right.

eterm•48m ago
That is a well recognised part of the LLM cycle.

A model or new model version X is released, everyone is really impressed.

3 months later, "Did they nerf X?"

It's been this way since the original chatGPT release.

The answer is typically no, it's just your expectations have risen. What was previously mind-blowing improvement is now expected, and any mis-steps feel amplified.

quentindanjou•38m ago
This is not always true. LLMs do get nerfed, and quite regularly, usually because they discover that users are using them more than expected, because of user abuse or simply because it attract a larger user base. One of the recent nerfs is the Gemini context window, drastically reduced.

What we need is an open and independent way of testing LLMs and stricter regulation on the disclosure of a product change when it is paid under a subscription or prepaid plan.

Analemma_•35m ago
> What we need is an open and independent way of testing LLMs

I mean, that's part of the problem: as far as I know, no claim of "this model has gotten worse since release!" has ever been validated by benchmarks. Obviously benchmarking models is an extremely hard problem, and you can try and make the case that the regressions aren't being captured by the benchmarks somehow, but until we have a repeatable benchmark which shows the regression, none of these companies are going to give you a refund based on your vibes.

landl0rd•28m ago
There's at least one site doing this: https://aistupidlevel.info/

Unfortunately, it's paywalled most of the historical data since I last looked at it, but interesting that opus has dipped below sonnet on overall performance.

jampa•32m ago
I usually agree with this. But I am using the same workflows and skills that were a breeze for Claude, but are causing it to run in cycles and require intervention.

This is not the same thing as a "omg vibes are off", it's reproducible, I am using the same prompts and files, and getting way worse results than any other model.

eterm•18m ago
When I once had that happen in a really bad way, I discovered I had written something wildly incorrect into the readme.

It has a habit of trusting documentation over the actual code itself, causing no end of trouble.

Check your claude.md files (both local and ~user ) too, there could be something lurking there.

Or maybe it has horribly regressed, but that hasn't been my experience, certainly not back to Sonnet levels of needing constant babysitting.

landl0rd•34m ago
I've observed the same random foreign-language characters (I believe chinese or japanese?) interspersed without rhyme or reason that I've come to expect from low-quality, low-parameter-count models, even while using "opus 4.5".

An upcoming IPO increases pressure to make financials look prettier.

epolanski•31m ago
Not really.

In fact as my prompts and documents get better it seems it does increasingly better.

Still, it can't replace a human, I really need to correct it at all, and if I try to one shot a feature I always end up spending more time refactoring it few days later.

Still, it's a huge boost to productivity, but the time it can take over without detailed info and oversight is far away.

paulhebert•51m ago
I tried using Claude Code this week because I have a free account from my work.

However when I try to log in via CLI it takes me to a webpage with an “Authorize” button. Clicking the button does nothing. An error is logged to the console but nothing displays in the UI.

We reached out to support who have not helped.

Not a great first impression

hobofan•28m ago
Sadly their whole frontend seems to be built without QC and mostly blindly assuming a happy path.

For the claude.ai UI, I've never had a single deep research properly transition (and I've done probably 50 or so) to its finished state. I just know to refresh the page after ~10mins to make the report show up.

roywiggins•19m ago
It's had enormous problems in Firefox. For me it would reliably hang the entire tab.

https://github.com/anthropics/claude-code/issues/14222

attheicearcade•20m ago
Do you have API access (platform.claude.com) rather than Claude code (claude.ai)? I had similar issues trying to get Claude CLI working via the second method, not knowing there’s a difference
boringg•50m ago
Oh is this whats been happening? I've been trying to ask question on a fairly long context window and history -- but it fails. No response it kind of acknowledges it received the input but then reprints the last output and then that whole dialogue is essentially dead ... same issue? Happened multiple times - quite frustrating.

Just a pro sub - not max.

Most of the time it gives me a heads up that I'm at 90% but a lot of the times it just failed, no warning, and I assumed it was I hit max.

kingkawn•41m ago
I’ve also been encountering this behavior, coupled with rapidly declining length of use for a pro account now below an hour, and weekly limits getting hit by Wednesday despite achieving very little other than fixing its own mistakes after compressions.
kilroy123•40m ago
Same here. Very bad and frustrating user experience to see nothing.
rvz•49m ago
One day everyone worshipping Claude when it works but once it goes on holiday, vibe-coders don't believe in using their own brains to solve the problem themselves.

Sometimes, poor old Claude wants to go on holiday and that is a problem?!?

daredoes•44m ago
Love [this take](https://github.com/anthropics/claude-code/issues/18866#issue...)

---

> Just my own observation that the same pattern has occurred at least 3 times now:

> release a model; overhype it; provide max compute; sell it as the new baseline

> this attracts a new wave of users to show exponential growth & bring in the next round of VC funding (they only care about MAU going up, couldn’t care less about existing paying users)

> slowly degrade the model and reduce inference

> when users start complaining, initially ignore them entirely then start gaslighting and make official statements denying any degradation

> then frame it as a tiny minority of users experiencing issues then, when pressure grows, blame it on an “accidentally” misconfigured servers that “unintentionally” reduced quality (which coincidentally happened to save the company tonnes of $).

BoredPositron•39m ago
Makes no sense as rationale for the bug at hand.
kace91•43m ago
As many point out, it’s one thing to have a buggy product and another to ignore users.

Businesses like google were already a step in the wrong direction in terms of customer service, but the new wave of AI companies seem to have decided their only relation to clients is collecting their money.

Unclear costs, no support, gaslighting customers when a model is degraded, incoming rug pulls..

cs02rm0•40m ago
So it's not just me.

I cancelled my subscription.

VerifiedReports•39m ago
The VS Code plug-in is broken on Windows. The command-line interface is broken on Windows.

I just signed up as a paying customer, only to find that Claude is totally unusable for my purposes at the moment. There's also no support (shocker), despite their claims that you'll be E-mailed by the support team if you file a report.

copirate•34m ago
There's also this issue[1] with about 300 participants about limits being reached much more quickly since they stopped the 2x limit for the holidays. A few people from Anthropic joined the conversation but didn't say much. Some users say they solved the issue by creating a new account or changing their plan.

[1] https://github.com/anthropics/claude-code/issues/16157

codazoda•28m ago
Something to check is if you’re opted into the test for the 1M context window. A co-worker told me this happened to them. They were burning a lot more tokens in the beta. Seems like creating a new account could track with this (but is obviously the Nuclear option).

I recently put a little money on the API for my personal account. I seem to burn more tokens on my personal account than my day job, in spite of using AI for 4x as long at work, and I’m trying to figure out why.

MicKillah•26m ago
I can definitely vouch for that being the case, for me.
deadbabe•30m ago
I’ve quit Claude Code and this was the final straw. I don’t think it was really even that much better, just different.
elemdos•29m ago
I hope that at some point companies start competing on quality instead of speed. LLMs will never be able to understand a codebase, and the more capable they get the more dangerous it is to just hand them the permission to blindly implement functionality and fix bugs. Bugs should be going down but they seem more prevalent than ever.
whoevercares•26m ago
My guess is SRE culture is a tough sell at Anthropic. When you’re a frontier lab, almost everything else looks more prestigious and more immediately “impactful”.
mccoyb•22m ago
Claude Code gets functionally worse every update. They need to get their shit together, hilarious to see Amodei at Davos talking big game about AGI and the latest update for a TUI application fucking changes observable behavior (like history scrolling with arrow keys), rendering random characters on the "newest" native version rendered in iTerm2, broken status line ... the list goes on and on.

This is the new status quo for software ... changing and breaking beneath your feet like sand.

cheriot•22m ago
I love CC, but there's so many bugs. Even the intended behavior is a mess - CC's VS Code UI bash tool stoped using my .zshrc so now it runs the wrong version of everything.