frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: SmartVPN: A VPN That Hides in Plain Sight over WebSockets

https://typescript.guru/smartvpn-a-vpn-that-hides-in-plain-sight-over-websockets/
1•PhilKunz•29s ago•0 comments

Show HN: LLM Memory Storage that scales, easily integrates, and is smart

https://github.com/colinulin/mind-palace
1•pocketcolin•2m ago•0 comments

Agents over Bubbles

https://stratechery.com/2026/agents-over-bubbles/
1•jonbaer•2m ago•0 comments

Lines of Code as a Productivity Metric in the AI Era

https://keegan.codes/blog/lines-of-code-as-a-productivity-metric-ai-era
1•keegandonley•4m ago•0 comments

Show HN: Stop renting AI – run AI workers on your own dedicated node

https://ainode.sh
1•rzessski•6m ago•1 comments

Northstar CUA Fast, open source 4B CUA model

https://www.tzafon.ai/blog/northstar-cua-fast
1•publius_•7m ago•0 comments

How Tenaciously Palantir Courted Switzerland

https://www.republik.ch/2026/02/18/how-tenaciously-palantir-courted-switzerland
1•sschueller•9m ago•0 comments

Beside

https://aplwiki.com/wiki/Beside
1•tosh•9m ago•0 comments

2 Days to Ship: Codex-Authored Metal Compute Shaders in Draw Things

https://engineering.drawthings.ai/p/2-days-to-ship-codex-authored-metal
1•liuliu•10m ago•0 comments

Open Decision-Making (2021)

https://web.stanford.edu/~ouster/cgi-bin/decisions.php
1•otoolep•10m ago•0 comments

Show HN: Tic-Tac-Word – Can you beat yourself in this tic-tac-toe word game?

https://www.tictacword.com
3•onion92•11m ago•1 comments

Autoresearch for SAT Solvers

https://github.com/iliazintchenko/agent-sat
2•chaisan•11m ago•1 comments

Agents are not thinking: a behavioral study of pwning sonnet

https://technoyoda.github.io/pwning-claude.html
1•sci-genie•11m ago•0 comments

Embedding TeX Hyphenation Patterns for 30 Languages in a 1.1 MB Rust Automaton

https://laurmaedje.github.io/posts/hypher/
2•PaulHoule•12m ago•0 comments

MacBook Neo can be modded to run faster, but you probably shouldn't

https://appleinsider.com/articles/26/03/16/macbook-neo-can-be-modded-to-run-faster-but-you-probab...
3•gslin•13m ago•0 comments

Show HN: Smart glasses that tell me when to stop pouring

https://github.com/RealComputer/GlassKit/tree/main/examples/rokid-overshoot-openai-realtime
3•tash_2s•14m ago•0 comments

All Linus rants from 2012 to 2015

https://github.com/corollari/linusrants
2•stmw•14m ago•0 comments

Pure Functions in APL and J

https://dl.acm.org/doi/10.1145/114054.114065
2•tosh•16m ago•0 comments

Anthropic requires your phone number

https://support.claude.com/en/articles/8287232-verifying-your-phone-number
2•dataspun•16m ago•0 comments

AI campus gender wars ahead

https://hollisrobbinsanecdotal.substack.com/p/the-ai-campus-gender-wars-ahead
1•HR01•17m ago•1 comments

Neal.Fun Constellation Draw

https://neal.fun/constellation-draw/
1•memalign•17m ago•0 comments

Combinatory Logic

https://en.wikipedia.org/wiki/Combinatory_logic
2•tosh•18m ago•0 comments

Chemical safety API – Free for public

https://aletheia.holisticquality.io/
1•Phractal•18m ago•0 comments

Show HN: Mnemosphere – AI chat for the intellectually ambitious

https://www.mnemosphere.ai
1•jntucehyd•20m ago•0 comments

Show HN: GhostVM – macOS containers on Apple Silicon

https://ghostvm.org/
5•CarolineWang•21m ago•0 comments

COBOL Is the Asbestos of Programming Languages

https://www.wired.com/story/cobol-is-the-asbestos-of-programming-languages/
3•chrisaycock•23m ago•0 comments

Show HN: AgentPen – macOS dashboard for managing OpenClaw AI agents

https://agentpen.io
2•sara_builds•25m ago•0 comments

Show HN: Open-source agent first AEO monitoring platform

https://github.com/AINYC/canonry
1•arberx•25m ago•0 comments

Upgrading Hex security – Gleam v1.15.0 released

https://gleam.run/news/upgrading-hex-security/
1•iruoy•26m ago•0 comments

Living Human Brain Cells Play Doom [video]

https://corticallabs.com/doom.html
1•smartmic•28m ago•0 comments
Open in hackernews

Speed at the cost of quality: Study of use of Cursor AI in open source projects

https://arxiv.org/abs/2511.04427
43•wek•2h ago

Comments

rfw300•1h ago
Super interesting study. One curious thing I've noticed is that coding agents tend to increase the code complexity of a project, but simultaneously massively reduce the cost of that code complexity.

If a module becomes unsustainably complex, I can ask Claude questions about it, have it write tests and scripts that empirically demonstrate the code's behavior, and worse comes to worst, rip out that code entirely and replace it with something better in a fraction of the time it used to take.

That's not to say complexity isn't bad anymore—the paper's findings on diminishing returns on velocity seem well-grounded and plausible. But while the newest (post-Nov. 2025) models often make inadvisable design decisions, they rarely do things that are outright wrong or hallucinated anymore. That makes them much more useful for cleaning up old messes.

joshribakoff•1h ago
Bad code has real world consequences. Its not limited to having to rewrite it. The cost might also include sanctions, lost users, attrition, and other negative consequences you don’t just measure in dev hours
SR2Z•1h ago
Right, but that cost is also incurred by human-written code that happens to have bugs.

In theory experienced humans introduce less bugs. That sounds reasonable and believable, but anyone who's ever been paid to write software knows that finding reliable humans is not an easy task unless you're at a large established company.

verdverm•1h ago
There was a recent study posted here that showed AI introduces regressions at an alarming rate, all but one above 50%, which indicates they spend a lot of time fixing their own mistakes. You've probably seen them doing this kind of thing, making one change that breaks another, going and adjusting that thing, not realizing that's making things worse.
MeetingsBrowser•1h ago
The question then becomes, can LLMs generate code close to the same quality as professionals.

In my experience, they are not even close.

MeetingsBrowser•1h ago
This only helps if you notice the code is bad. Especially in overlay complex code, you have to really be paying attention to notice when a subtle invariant is broken, edge case missed, etc.

Its the same reason a junior + senior engineer is about as fast as a senior + 100 junior engineers. The senior's review time becomes the bottleneck and does not scale.

And even with the latest models and tooling, the quality of the code is below what I expect from a junior. But you sure can get it fast.

PeterStuer•1h ago
Interesting from an historical perspective. But data from 4/2025? Might as well have been last century.
happycube•1h ago
I think the gist of it still applies to even Claude Code w/Opus 4.6.

It's basically outsourcing to mediocre programmers - albeit very fast ones with near-infinite patience and little to no ego.

Miraste•53m ago
It doesn't map well to a mediocre human programmer, I think. It operates in a much more jagged world between superhuman, and inhuman stupidity.
matt_heimer•1h ago
Yes, it's not surprising that warnings and complexity increased at a higher rate when paired with increased velocity. Increased velocity == increased lines of code.

Does the study normalize velocity between the groups by adjusting the timeframes so that we could tell if complexity and warnings increased at a greater rate per line of code added in the AI group?

I suspect it would, since I've had to simplify AI generated code on several occasions but right now the study just seems to say that the larger a code base grows the more complex it gets which is obvious.

ex-aws-dude•1h ago
That was my thought as well, because obviously complexity increases when a project grows regardless of AI
bensyverson•1h ago
Yeah, I have a more complex project I'm working on with Claude, but it's not that Claude is making it more complex; it's just that it's so complex I wouldn't attempt it without Claude.
AstroBen•36m ago
"Notably, increases in codebase size are a major determinant of increases in static analysis warnings and code complexity, and absorb most variance in the two outcome variables. However, even with strong controls for codebase size dynamics, the adoption of Cursor still has a significant effect on code complexity, leading to a 9% baseline increase on average compared to projects in similar dynamics but not using Cursor."
AstroBen•1h ago
> On average, Cursor adoption has a modestly significant positive impact on development velocity, particularly in terms of code production volume: Lines added increase by about 28.6% (Table 2). There is no statistically significant effect for the volume of commits.

This doesn't equate to a faster development speed in my eyes? We know that AI code is incredibly verbose.

More lines of code doesn't equate to faster development - even more so when you're comparing apples (human written) to oranges (AI written)

AstroBen•1h ago
They're measuring development speed through lines of code. To show that's true they'd need to first show that AI and humans use the same number of lines to solve the same problem. That hasn't been my experience at all. AI is incredibly verbose.

Then there's the question of if LoC is a reliable proxy for velocity at all? The common belief amongst developers is that it's not.

mellosouls•37m ago
Depends on the nature of the tool I would imagine - eg. Claude Code Terminal (say) would have higher entry requirements in terms of engineering experience (Cursor was sold as newbie-friendly) so I would predict higher quality code than Cursor in a similar survey.

ofc that doesn't take into account the useful high-level and other advantages of IDEs that might mitigate against slop during review, but overall Cursor was a more natural fit for vibe-coders.

This is said without judgement - I was a cheerleader for Cursor early on until it became uncompetitive in value.

mentalgear•16m ago
> We find that the adoption of Cursor leads to a statistically significant, large, but transient increase in project-level development velocity, along with a substantial and persistent increase in static analysis warnings and code complexity. Further panel generalized-method-of-moments estimation reveals that increases in static analysis warnings and code complexity are major factors driving long-term velocity slowdown. Our study identifies quality assurance as a major bottleneck for early Cursor adopters and calls for it to be a first-class citizen in the design of agentic AI coding tools and AI-driven workflows.

So overall seems like the pros and cons of "AI vibe coding" just cancel themselves out.

chris_money202•13m ago
Now someone do a research study where a summary of this research paper is in the AGENTS.md and let’s see if the overall outcomes are better