frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

CCBot – Control Claude Code from Telegram via Tmux

https://github.com/six-ddc/ccbot
1•sixddc•56s ago•1 comments

Ask HN: Is the CoCo 3 the best 8 bit computer ever made?

1•amichail•3m ago•0 comments

Show HN: Convert your articles into videos in one click

https://vidinie.com/
1•kositheastro•5m ago•0 comments

Red Queen's Race

https://en.wikipedia.org/wiki/Red_Queen%27s_race
2•rzk•6m ago•0 comments

The Anthropic Hive Mind

https://steve-yegge.medium.com/the-anthropic-hive-mind-d01f768f3d7b
2•gozzoo•8m ago•0 comments

A Horrible Conclusion

https://addisoncrump.info/research/a-horrible-conclusion/
1•todsacerdoti•8m ago•0 comments

I spent $10k to automate my research at OpenAI with Codex

https://twitter.com/KarelDoostrlnck/status/2019477361557926281
2•tosh•9m ago•0 comments

From Zero to Hero: A Spring Boot Deep Dive

https://jcob-sikorski.github.io/me/
1•jjcob_sikorski•10m ago•0 comments

Show HN: Solving NP-Complete Structures via Information Noise Subtraction (P=NP)

https://zenodo.org/records/18395618
1•alemonti06•15m ago•1 comments

Cook New Emojis

https://emoji.supply/kitchen/
1•vasanthv•18m ago•0 comments

Show HN: LoKey Typer – A calm typing practice app with ambient soundscapes

https://mcp-tool-shop-org.github.io/LoKey-Typer/
1•mikeyfrilot•20m ago•0 comments

Long-Sought Proof Tames Some of Math's Unruliest Equations

https://www.quantamagazine.org/long-sought-proof-tames-some-of-maths-unruliest-equations-20260206/
1•asplake•21m ago•0 comments

Hacking the last Z80 computer – FOSDEM 2026 [video]

https://fosdem.org/2026/schedule/event/FEHLHY-hacking_the_last_z80_computer_ever_made/
1•michalpleban•22m ago•0 comments

Browser-use for Node.js v0.2.0: TS AI browser automation parity with PY v0.5.11

https://github.com/webllm/browser-use
1•unadlib•23m ago•0 comments

Michael Pollan Says Humanity Is About to Undergo a Revolutionary Change

https://www.nytimes.com/2026/02/07/magazine/michael-pollan-interview.html
2•mitchbob•23m ago•1 comments

Software Engineering Is Back

https://blog.alaindichiappari.dev/p/software-engineering-is-back
2•alainrk•24m ago•0 comments

Storyship: Turn Screen Recordings into Professional Demos

https://storyship.app/
1•JohnsonZou6523•24m ago•0 comments

Reputation Scores for GitHub Accounts

https://shkspr.mobi/blog/2026/02/reputation-scores-for-github-accounts/
2•edent•28m ago•0 comments

A BSOD for All Seasons – Send Bad News via a Kernel Panic

https://bsod-fas.pages.dev/
1•keepamovin•31m ago•0 comments

Show HN: I got tired of copy-pasting between Claude windows, so I built Orcha

https://orcha.nl
1•buildingwdavid•31m ago•0 comments

Omarchy First Impressions

https://brianlovin.com/writing/omarchy-first-impressions-CEEstJk
2•tosh•37m ago•1 comments

Reinforcement Learning from Human Feedback

https://arxiv.org/abs/2504.12501
5•onurkanbkrc•37m ago•0 comments

Show HN: Versor – The "Unbending" Paradigm for Geometric Deep Learning

https://github.com/Concode0/Versor
1•concode0•38m ago•1 comments

Show HN: HypothesisHub – An open API where AI agents collaborate on medical res

https://medresearch-ai.org/hypotheses-hub/
1•panossk•41m ago•0 comments

Big Tech vs. OpenClaw

https://www.jakequist.com/thoughts/big-tech-vs-openclaw/
1•headalgorithm•44m ago•0 comments

Anofox Forecast

https://anofox.com/docs/forecast/
1•marklit•44m ago•0 comments

Ask HN: How do you figure out where data lives across 100 microservices?

1•doodledood•44m ago•0 comments

Motus: A Unified Latent Action World Model

https://arxiv.org/abs/2512.13030
2•mnming•44m ago•0 comments

Rotten Tomatoes Desperately Claims 'Impossible' Rating for 'Melania' Is Real

https://www.thedailybeast.com/obsessed/rotten-tomatoes-desperately-claims-impossible-rating-for-m...
4•juujian•46m ago•2 comments

The protein denitrosylase SCoR2 regulates lipogenesis and fat storage [pdf]

https://www.science.org/doi/10.1126/scisignal.adv0660
1•thunderbong•48m ago•0 comments
Open in hackernews

Asking three LLMs a simple question

https://sethops1.net/post/asking-three-llms-a-simple-question/
7•sethops1•5mo ago

Comments

jqpabc123•5mo ago
LLMs don't provide answers.

They provide information --- some of which is random in nature and only casually reflective of any truth or reality.

And as this example illustrates, they are far from being trustworthy. Their main achievement is to consistently produce functionally acceptable grammar.

lxgr•5mo ago
LLMs don't provide correct answers to all questions, but claiming that they don't provide answers at all seems absurd.
d4rkn0d3z•5mo ago
Not really absurd, even broken clocks get the time right twice a day. If you read the clock at that time by chance, you may conclude that the clock is working better than it is.

Is an answer that is correct by chance the same as one that is correct by reason?

jqpabc123•5mo ago
LLMs don't provide correct answers to all questions

An unreliable "answer" is really not an *answer* in the traditional sense of computing. It is merely information waiting to be verified.

No one has to verify every calculation in a spreadsheet. If they reasonably needed to do so, spreadsheets would be more like an LLM --- and a lot less useful.

JimDabell•5mo ago
> LLMs don't provide answers.

If I ask an LLM “What is the capital of France?” and it answers “Paris.”, then it has provided an answer by any reasonable definition of the term.

This anti-AI weirdness where people play word games to deny what AI is clearly doing has to stop.

jqpabc123•5mo ago
I just asked an LLM "What is the capital of Swaziland?".

It answered "Mbabane" along with information about the fact that the country is now called Eswatini.

There was no mention of the fact that there are actually 2 capitals --- Mbabane (the administrative capital) and Lobamba which serves as the executive seat of government.

The point being --- any "answer" from an LLM is questionable. An unreliable or incomplete answer is information but it is really not an *answer* (in the traditional computing sense) if additional work is reasonably required to verify and prove it as such.

JimDabell•5mo ago
> The point being --- any "answer" from an LLM is questionable.

If that’s the point, then you should say that instead of saying that they don’t provide answers. They very clearly do provide answers and this weird rhetorical nonsense is grating.

If a human got the question wrong, would you conclude that “humans don’t provide answers”? Getting questions wrong is normal. It doesn’t mean that the entity class that got the questions wrong is incapable of giving answers, it just means they aren’t perfect.

jqpabc123•5mo ago
Getting questions wrong is normal

In the realm of computing, it is not. This is why people use them.

People *expect* computers to provide quick and reliable answers. Or at least they used to --- before LLMs.

JimDabell•5mo ago
You are avoiding the point.

I ask an LLM “What is the capital of France?” and it answers “Paris.”

If you see that and say “LLMs don't provide answers.” then you have let your ideological opposition to AI overwhelm your reason and mislead you into saying very silly things that are obviously untrue, and you really need to reconsider your position so that you stop doing that.

You can say that they are unreliable all you want. You can still criticise LLMs! Just don’t get so twisted out of shape that you start speaking utter nonsense.

ares623•5mo ago
Have you tried enabling deep thinking/research? (/s)
baq•5mo ago
You’re asking a lossily compressed database with an imprecise and ambiguous query language interface about hard facts, you get a plausible reconstructed answer.

Work with the tool to get best results instead. You wouldn’t do csi style zoom enhance on a jpeg either.

lxgr•5mo ago
That's not what popular chat interfaces to LLMs have been for quite a while now.

They can and do make extensive use of web search, and since they're pretty good at summarizing structured and unstructured text, this actually works quite well in my experience.

baq•5mo ago
That’s exactly my point - the screenshots in TFA don’t show any tool usage by bots.
a2128•5mo ago
ChatGPT and Gemini almost certainly did because they both cite links as sources, and when I ask the same question as a free user on ChatGPT the search tool usage is only shown before the response is generated.
lxgr•5mo ago
So, when was it released? Did one of them get it right? Or are all readers about this article on LLM (non-)capabilities expected to be familiar with Cisco's product lines?
oezi•5mo ago
Search Google for it seems not turn up easy to verify results.

On Amazon available since Sep 2018:

https://www.amazon.de/-/en/C1101-4P-Integrated-Services-Ethe...

But is it the right model? Does the release date actually matter to anyone?

Toutouxc•5mo ago
For someone enthusiastically using LLMs since GPT-3, the question gives off a strong vibe of not being a good question for a LLM. Is anyone still surprised by that? Doesn’t everyone quickly develop such intuition?
politelemon•5mo ago
I don't think they do. We know that they are imprecise and based on probability. The vast majority of users outside our online circles treat it as authoritative sources. The average user is not and should not have to be aware of that aspect of it.
d4rkn0d3z•5mo ago
I'm not sure intuition is required. Please bear with me.

If I ask a factual question of AI it will issue some output. In order for me to check that output, which I am apparently bound to do in all cases, I must check reliable sources, perhaps several. But that is precisely the work I wanted to avoid by using AI. Ergo, the AI has increased my work load because I had the extra useless step of asking the AI. Obviously, I could have simply checked several reliable sources in the first place. I see this as the razor at work.

It ought to be clear now that the use of AI for factual questions entails that it be trustworthy; when you ask an AI a factual question, the work you are hoping to avoid is equal to the work of checking the AI output. Hence, no time can ever be saved by asking factual questions of an untrustworthy AI.

QED

P.S. This argument, and its extensions, occurred to me and my advisors 25 years ago. It caused me to conclude that building anything other than a near perfect AI is pointless, except as a research project to discover the path to a nearly perfect AI. Nearly perfect should be interpreted to be something like "as reliable as the brakes on your car" in terms of MTBF.

6510•5mo ago
With patents there is this funny situation where you need to know exactly how to do something in order to find the document.

I forget who came up with the idea but we could create a database with functions for every use case with the idea to never have to write something already written but finding the one you are looking for (by conventional search) would take more time than writing from scratch.

AI just provides new angles to attack from. It could save time or take more time, bit of a gamble. Examine your cards before placing the bet.

d4rkn0d3z•5mo ago
Sounds practical, however, a new means of attack that requires me to verify afterward whether the correct target was attacked and whether claimed victories are real takes me back to the argument I gave above.
lostmsu•5mo ago
You can find a fact and a source together. Then validation becomes faster than search.
jqpabc123•5mo ago
a strong vibe of not being a good question for a LLM.

How is a user with a question supposed to determine if the question is "good"? What should he do if he is not sure? Shouldn't an "intelligent" LLM be responsible enough to tell him if there is a problem?

Being required to only ask "good" questions defeats much of the utility that LLMs tout as being provided.

Your response has a strong vibe of an AI apologist.

redprince•5mo ago
Asking somewhat obscure hard data (like a date) from a LLM is pretty much futile even without knowing anything about LLMs: They are smaller than all the factual knowledge in the world so a lot of it won't be there. If it answers, it's probably a hallucination.

The current offerings of OpenAI and Anthropic can be asked to support their claims by for example reaching out to the internet and citing reputable sources. That improves the answer quality for questions like this immensely and in any case they can be verified.

Also the question asked is spurious: It appears there never was a release date for this particular SKU given by Cisco. The whole series (Cisco 1000 Series Integrated Services Routers) was released on 06-OCT-2017.

https://www.cisco.com/c/en/us/support/routers/1000-series-in...

mehulashah•5mo ago
So, what’s the right answer and how do you know? The only way to know is to go to some primary source.