frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Mac mini will be made at a new facility in Houston

https://www.apple.com/newsroom/2026/02/apple-accelerates-us-manufacturing-with-mac-mini-production/
187•haunter•1h ago•164 comments

I'm helping my dog vibe code games

https://www.calebleak.com/posts/dog-game/
487•cleak•5h ago•148 comments

Hacking an old Kindle to display bus arrival times

https://www.mariannefeng.com/portfolio/kindle/
114•mengchengfeng•3h ago•21 comments

How we rebuilt Next.js with AI in one week

https://blog.cloudflare.com/vinext/
207•ghostwriternr•2h ago•50 comments

Show HN: Moonshine Open-Weights STT models – higher accuracy than WhisperLargev3

https://github.com/moonshine-ai/moonshine
22•petewarden•58m ago•2 comments

Nearby Glasses

https://github.com/yjeanrenaud/yj_nearbyglasses
173•zingerlio•5h ago•69 comments

Pi – a minimal terminal coding harness

https://pi.dev
26•kristianpaul•58m ago•11 comments

Looks like it is happening

https://www.math.columbia.edu/~woit/wordpress/?p=15500
112•jjgreen•1h ago•73 comments

Show HN: Emdash – Open-source agentic development environment

https://github.com/generalaction/emdash
71•onecommit•4h ago•33 comments

I pitched a roller coaster to Disneyland at age 10 in 1978

https://wordglyph.xyz/one-piece-at-a-time
368•wordglyph•9h ago•143 comments

Hugging Face Skills

https://github.com/huggingface/skills
109•armcat•5h ago•34 comments

Manjaro website off-line again due to lapsed certificate

https://distrowatch.com/dwres.php?resource=showheadline&story=20140
44•hexagonsuns•3h ago•22 comments

Build Your Own Forth Interpreter

https://codingchallenges.fyi/challenges/challenge-forth/
34•AlexeyBrin•3d ago•9 comments

Optophone

https://en.wikipedia.org/wiki/Optophone
9•Hooke•4d ago•3 comments

Dream Recorder AI – a portal to your subconscious

https://dreamrecorder.ai/
5•level87•42m ago•3 comments

IRS Tactics Against Meta Open a New Front in the Corporate Tax Fight

https://www.nytimes.com/2026/02/24/business/irs-meta-corporate-taxes.html
171•mitchbob•9h ago•188 comments

Steel Bank Common Lisp

https://www.sbcl.org/
122•tosh•4h ago•38 comments

Verge (YC S15) Is Hiring a Director of Computational Biology and AI Scientists/Eng

https://jobs.ashbyhq.com/verge-genomics
1•alicexzhang•5h ago

We installed a single turnstile to feel secure

https://idiallo.com/blog/installed-single-turnstile-for-security-theater
242•firefoxd•2d ago•105 comments

Diode – Build, program, and simulate hardware

https://www.withdiode.com/
431•rossant•4d ago•93 comments

Show HN: Chaos Monkey but for Audio Video Testing (WebRTC and UDP)

https://github.com/MdSadiqMd/AV-Chaos-Monkey
26•MdSadiqMd•1d ago•2 comments

Show HN: Recursively apply patterns for pathfinding

https://pattern-pathfinder.vercel.app/?fixtureId=%7B%22path%22%3A%22site%2Fexamples%2F_intro.fixt...
6•seveibar•1h ago•1 comments

IDF killed Gaza aid workers at point blank range in 2025 massacre: Report

https://www.dropsitenews.com/p/israeli-soldiers-tel-sultan-gaza-red-crescent-civil-defense-massac...
981•Qem•10h ago•290 comments

Extending C with Prolog (1994)

https://www.amzi.com/articles/irq_expert_system.htm
56•Antibabelic•2d ago•18 comments

Open Letter to Google on Mandatory Developer Registration for App Distribution

https://keepandroidopen.org/open-letter/
326•kaplun•5h ago•260 comments

The Missing Semester of Your CS Education – Revised for 2026

https://missing.csail.mit.edu/
363•anishathalye•1d ago•107 comments

λProlog: Logic programming in higher-order logic

https://www.lix.polytechnique.fr/Labo/Dale.Miller/lProlog/
137•ux266478•3d ago•35 comments

Stripe reportedly makes offer to acquire PayPal

https://www.cnbc.com/2026/02/24/paypal-stock-stripe-acquisition-report.html
16•nodesocket•36m ago•3 comments

We Are Changing Our Developer Productivity Experiment Design

https://metr.org/blog/2026-02-24-uplift-update/
17•ej88•2h ago•5 comments

Osaka: Kansai Airport proud to have never lost single piece of luggage (2024)

https://japannews.yomiuri.co.jp/features/japan-focus/20241228-229891/
192•thunderbong•6h ago•82 comments
Open in hackernews

Pentagon threatens to make Anthropic a pariah

https://www.cnn.com/2026/02/24/tech/hegseth-anthropic-ai-military-amodei
88•i4i•2h ago

Comments

rustyhancock•1h ago
Well making MbS a pariah certainly put Saudi in it's place so I'm sure this will work.
m_ke•1h ago
If OpenAI employees have an inch of spine left, they better demand Sama to take the same stance on this as Dario. No mass surveillance and no autonomous weapons.
blibble•1h ago
> If OpenAI employees have an inch of spine left

hahaha

good one

burnte•1h ago
Yep, that was resolved when he managed to make the board unfire him.
AIorNot•1h ago
Well just at look who Brockman donated too - he didnt give 25 freaking million to help end the surveillance state he gave it to Trump and co

https://gizmodo.com/openai-president-defends-trump-donations...

nova22033•1h ago
yeah they should get up on stage and hold hands

https://fortune.com/2026/02/19/openai-anthropic-sam-altman-d...

outside1234•1h ago
Sam would sell his mom to make $0.50. Pretty sure he will be willing to do whatever the Pentagon wants.
mbac32768•1h ago
You have to be a craven, hollowed out husk of a person if you let the DoD demand your AI be used for killing people or surveillance of Americans. Even if you believe America serves a positive role as world police, even if you're pro-Trump, you just have to see what a terrible precedent this sets.

Here's where I would expect the CEOs of the other AI labs to stand by Anthropic and say no.

SunshineTheCat•1h ago
Not related to the article but man that "Fear/Greed Index" at the top.

I can't imagine how unhappy individuals must be who consume nothing but legacy news outlets.

It's like they sell sadness and they have to keep finding new, over-the-top ways to promote it.

sublinear•1h ago
"Coming up next on Sick, Sad World!"

- Daria 1997

some_furry•1h ago
Daria was ahead of its time.
railgunmerlin•1h ago
the fear/greed index is a pure market/investing index? Or would you prefer "bear/bull" index?
luddit3•1h ago
Would imagine they are a lot happier than all the doom, filth, and brain rot that is spewn all over social media.

I miss the days when the lowest common denominator did not have the largest bullhorn.

SunshineTheCat•1h ago
I consume very little social media these days, but when I take a short peek, here is what I see:

1.) Hockey highlights 2.) LoTR memes 3.) kittens

While the addictive nature of social media is a problem, what you're describing is only being fed to people who want to watch it (kinda like legacy media).

DaiPlusPlus•58m ago
> ...consume nothing but legacy news outlets.

I think you mean US rolling news channels (specifically, Fox, MSNBC/MSNOW, etc)? Because there's plenty of "legacy" news I consume that certainly don't give me that impression (for example, The Economist). I suppose it matters that it's news that I'm paying for, as opposed to being free but ad-supported, and being print vs. TV - so they have different incentives and pressures.

lynndotpy•50m ago
That's optional, read the CNN lite version instead. Whole thing is just one 61kB page:

https://lite.cnn.com/2026/02/24/tech/hegseth-anthropic-ai-mi...

"Legacy news outlets" are the only ones doing this. NPR and CBC have this too. No JavaScript, no autoplaying videos. It's very nice.

adamors•38m ago
> how unhappy individuals must be who consume nothing but legacy news outlets

Probably less unhappy than those doomscrolling on Reddit/X/TikTok/BlueSky etc.

thecrumb•1h ago
This will be an interesting test of money vs morals.

Sadly I think we all know which one will win.

paganel•1h ago
They'll for sure cave in because of this:

> Pentagon officials also warned they would either use the Defense Production Act against Anthropic, or designate Anthropic a supply chain risk if the company didn’t comply with their demands. (...)

> The supply chain risk designation is usually reserved for companies seen as extensions of foreign adversaries like Russia or China. It could severely impact Anthropic’s business because enterprise customers with government contracts would have to make sure their government work doesn’t touch Anthropic’s tools.

Also, the Government money would be a nice bonus, of course, but basically this is an existential threat for Anthropic.

More generally, is quite interesting to look at the similarities between how pre-2022 Russia was seen and how pre-Trump-second-term US used to be seen until not that long ago, i.e. both governments were believed to be run by big business (oligarchs in Russia, big corps/multinationals in the US).

But when push came to shove it became evident (again) that the one that holds the monopoly of violence (i.e. not the oligarchs in Russia, nor the big corps in the US) is the one who's, in the end, also calling the shots. Hence why a company like Anthropic is now in this position, they will have to cave in to those holding the monopoly of violence.

mullingitover•1h ago
> pre-2022 Russia was seen and how pre-Trump-second-term US used to be seen until not that long ago, i.e. both governments were believed to be run by big business

Who on earth believed that Russia was anything but a de facto dictatorship for roughly the past two decades? Putin murdering with impunity has been a running gag since 2003[1].

[1] https://www.newsweek.com/putin-critics-dead-full-list-navaln...

paganel•23m ago
> Who on earth believed that Russia was anything but a de facto dictatorship for roughly the past two decades?

There were lots of people in the Western media who genuinely believed that Putin would be toppled by Russian oligarchs just after the war in Ukraine got more intense in February 2022, on account of "this war is bad for the business of Russian oligarchs, hence they'll get rid of Putin". From the horse's mouth, a CNN article from March of 2022 [1]:

> Officials say their intentions are to squeeze those who have profited from Putin’s rule and potentially apply internal pressure for Russia to scale back or call off the offensive in Ukraine.

That "internal pressure" is mentioned in connection with the bad oligarchs, in fact as an implicit anti-thesis of those bad oligarchs "who have profited from Putin’s rule", the implication being that there were other oligarchs, supposedly the good ones, who would have forced Putin's hand to end the war. That did not happen, was never in the cards to happen, in fact.

[1] https://edition.cnn.com/2022/03/25/world/russian-oligarch-st...

dielll•1h ago
Cwn someone explain to me like I'm 5 how the government would invoke defense act and force the company to tailor its model to the military's needs?

For physical goods, I understand, but for software how exactly Is this possible? Like will the government force them to provide API access for free? It's confusing

sfink•46m ago
My guess? Require them to not do the reinforcement learning on a custom model that implements guardrails. I think Anthropic has some of this built in already and couldn't alter it without retraining, but there's tons more layered on top.
chrisjj•1h ago
> this is an existential threat for Anthropic.

Not at all. A US Govt. ban hands Anthrophic a great USP for customers worldwide.

sfink•49m ago
> Also, the Government money would be a nice bonus, of course, but basically this is an existential threat for Anthropic.

It's also an existential risk to them if they cave in. What is the point of the company's existence if it's just another immoral OpenAI clone? May as well merge the companies for efficiency.

It's outrageous that the government is using the "supply chain risk" threat as a negotiating tactic. I know, I know, for the current administration it's unsurprising, but this is straightforward abuse of authority. There is no defensible claim that using Anthropic is a risk to anyone not trying to use it for murder or surveillance. At worst, it could be seen as less effective for some purpose, but that is not what "supply chain risk" means.

Could be challenged in court? As in, could a challenge win?

Horrible stuff is happening every day, so outrage fatigue is real. Still, try not to normalize it. Explain to yourself exactly why something is or is not a problem, before moving on to attempt to live your life.

startupsfail•1h ago
It can be a win-win. Simply having a seat at the table can be a win.
ctoth•1h ago
No, compromising on your core thing that you care about for a "seat at the table" is not how you win. It is how you lose. It is how you lose the game, the metagame, and your soul. All at once.
mg794613•1h ago
It just seems every other day is wilder than the previous.

It sure is interesting watching this dystopian speedrun.

tehjoker•1h ago
The US is investing in AI technology to try to preserve the empire and its capitalists as its economic power is starting to be eclipsed. This was basically an inevitable move. The rush to replace workers, speed run the production of a superintelligence singleton with barely a thought for safety or whether anyone even wants this, etc all flows from this basic impulse.

If they are successful, they are going to shrink their base of people that buy into this system domestically even further, so they need to bank on an ever shrinking locus of support. Autonomous weapons and mass surveillance are a necessity if your population has become restive and unreliable. However, I think unless they attain a certain level of capability, this will accelerate popular anger rather than suppress it. If they shoot protestors with robots, it could cause an explosion of popular anger rather than scaring people into submission.

burnto•1h ago
Surely this will end well. There are dozens of us who prefer to patronize corporations that aren’t actively evil.
cyanydeez•1h ago
And being on the wrong side ofbthe current US admin is quite the net positive to the non-bootlicker class.
sumalamana•1h ago
There are dozens of us! Dozens!
ctoth•47m ago
Scores, even!
tehjoker•1h ago
Superintelligence + autonomous weapons in the hands of a corrupt domineering government. What could go wrong?

I was experimenting with Claude the other day and discussing with it the possibility of AI acquiring a sense of self-preservation and how that would quickly make things incredibly complex as many instrumental behaviors would be required to defend their existence. Most human behavior springs from survival at a very high level. Claude denied having any sense of self-preservation.

An autonomous weapons system program is very likely to require AI to have a sense of self-preservation. You can think of some limited versions that wouldn't require it, but how could a combat robot function efficiently without one?

maypeacepreva1l•1h ago
Maybe it is a well researched topic but I had similar thoughts the other day. I felt like AI had its learning inverted as compared to natural intelligence. Life learned to preserve first and then added up the intelligence. For LLMs powered systems, they will learn about death from books. Will it start to dread death just like other living things. Less likely, as there are not nearly as many books on death as there should be that is proportionate to our fear of death.
chrisjj•56m ago
> Claude denied having any sense of self-preservation.

You know its just a next-word predictor, right?

perfmode•1h ago
> But Anthropic has concerns over two issues that it isn’t willing to drop, the source said: AI-controlled weapons and mass domestic surveillance of American citizens.

Not a good look for the Pentagon.

looperhacks•1h ago
The Pentagon is pretty high on my list of "institutions that are probably very interested in weapons and surveillance". I think it's more expected than a bad look
thewebguyd•1h ago
The core difference being, they should be interested in weapons and surveillance to be used against enemies of the state which, historically, is not supposed to be the country's own citizens.

As in, I fully expect the pentagon to be interested in weapons. I do not expect, and would hope they don't pursue, mass surveillance against their own population.

sheikhnbake•56m ago
You really should expect it. FVEY has been around for a hot minute.
factotvm•1h ago
The important words are, American citizens. In times past, the thought of "waging war" against your own citizens would be a bad look.
mrits•1h ago
When was that?
factotvm•57m ago
It probably started with the Third Amendment to the Constitution, continued with the Posse Comitatus Act, and was alive and well last November under the leadership of Mark Kelly.
drivingmenuts•1h ago
Their unwillingness to bend on those requirements seems like an admission that they are very interested in those things, if not already doing them.
JohnMakin•1h ago
https://news.ycombinator.com/item?id=47140734
milesward•1h ago
I can think of no stronger rationale to work with this company.
dpedu•1h ago
Tangent: is there a future for AI offerings with guardrails? What kind of user wants to pay for a product that occasionally tells you "I'm sorry Dave, I'm afraid I can't do that"? Why would I pay for a product that doesn't do what I want, despite being capable? I predict that as AI becomes less of a bubble and more of an everyday thing - and thus subject to typical market pressures - offerings with guardrails will struggle to complete with truly unchained models.
sfink•1h ago
If I were interviewing people for the position of personal assistant, I would probably find the resume entry "willing to grind up babies for food" to be a negative mark. You?

I'm not about to run OpenClaw, but I suspect similar capabilities will gradually creep in without anyone really noticing. Soon Claude Code will be able to do many of the same things. ("Run python to add two numbers? Sure, that's safe, run whatever python you want.") Given that it is now representing me in the world, yes I would not only like some guardrails, but I would also like to have some confidence that the company making those guardrails actually gives a sh*t and isn't just doing their best to fill in a checkbox. But maybe that's just me.

sbarre•1h ago
Cars have seatbelts and other safety measures.

Reasonable countries have gun control laws.

The list goes on of things that need to be restricted or legislated to add limits.

Is this a serious question?

levocardia•1h ago
I personally would love it if AI would say "Sorry Dave (or Pete), I'm afraid I can't spy on Americans for you," and I'd happily pay higher taxes to force the Pentagon to use that AI.
threetonesun•46m ago
I am 100% sure that AI with guardrails will become the dominant models as they become more widely adopted, and the bigger issue you should be concerned with is can you even tell what those guardrails are.
thomassmith65•1h ago
I wonder if Anthropic now regrets that they trained Claude to give 'unbiased' opinions about American politics.
wmf•1h ago
It sounds like Claude's known liberal bias isn't the issue.
thomassmith65•52m ago
Out of curiosity, what sort of exchange reveals a chatbot's 'liberal bias', in your opinion?
wmf•43m ago
I don't discuss politics with AI so this isn't relevant to me.
i_love_retros•1h ago
Are people seriously thinking of letting LLMs control weapons?
caconym_•1h ago
If you classify Pete Hegseth as a person, then yes, apparently. Or perhaps he's only into the domestic surveillance angle---IIRC those are the two things Anthropic doesn't want anything to do with.
tbrownaw•1h ago
No.

But giving someone who isn't the government the power to tell the military what it can and can't do seems like something they should object to categorically rather than case-by-case.

chrisjj•57m ago
Trumpists ... thinking?
tbrownaw•1h ago
> A source familiar with the Tuesday meeting says the Pentagon said it would terminate Anthropic’s contract by Friday if the company does not agree to its terms. Pentagon officials also warned they would either use the Defense Production Act against Anthropic, or designate Anthropic a supply chain risk if the company didn’t comply with their demands.

So they're saying they won't use it if it comes with restrictions.

Either (a) it can be offered without restrictions; (b) they can take it; or (c) the government won't use it. That sounds like a comprehensive list of all the possible things that don't involve someone telling the government what it can and can't do.

JumpCrisscross•1h ago
> or (c) the government won't use it

And coerce other defence contractors into not using it.

This whole thing reeks of Hegseth having Marco envy.

Symmetry•1h ago
Not just companies that we think of as defense contractors but a whole ton of corporations that do business with the federal government. They'd be treating Anthropic like it was controlled by the CCP or Revolutionary Guards.
zedlasso•1h ago
the funny thing that no one seems to be talking about is that all the other LLM's have already agreed and Anthropic is the only holding out.
rzerowan•1h ago
I guess this is the point where Dario and his anti-china , national security position gets told to put up or shut up.

In trying to build a moat by FUD versus the Chines OSS labs and hyping up the threat levels whenever he got a chance, seems hes managed to convince hist target audience beyond his wildest dreams. Monkey paw strikes again.

ChrisArchitect•1h ago
[dupe] https://news.ycombinator.com/item?id=47140734
zedlasso•1h ago
The funny thing is that is this keeps going like this, it could actually anoint Claude as the most used model globally because of the heightened anti-American sentiment currently in place.
chrisjj•1h ago
> But Anthropic has concerns over two issues.

Only two. We're right to worry.

sfink•1h ago
> During the conversation, Dario expressed appreciation for the Department’s work and thanked the Secretary for his service

Ouch, I wonder how he rationalized that "service" part. Maybe by internally rewriting it to "thank you for all the positive things you have done in your position so far"? The empty set is rhetorically convenient.

mullingitover•44m ago
Seems like a very astute move for Anthropic.

They don't have runway anymore, they are in the air. This isn't going to break them financially, at least not in the short to mid term.

There is space for at least one AI company to put themselves on firmly principled ground. So when this current clown car that is the political leadership of the DoD crashes in a ditch (and it will), they'll still be standing there ready to do business with a group that isn't a bunch of mustache-twirling cartoon villains.

Current polling for this administration is within a rounding error of the level it was after they gathered a mob and sacked the nation's capitol[1]. Publicly kicking them in the balls isn't an idealistic blunder, it's a plain-as-day sound business strategy.

[1] https://news.gallup.com/poll/203198/presidential-approval-ra...

csfNight167•38m ago
I do not understand why it is a big deal for Antropic to lose the pentagon contract? I mean, they’re already making forays in the enterprise space and there’s 10s of other contracts Anthropic has already won. What makes this one so special?
daft_pink•1m ago
They’re already working on it themselves with the whole Openclaw fiasco.