frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Google announced that Chrome is becoming an agentic workplace platform

https://thenextweb.com/news/google-chrome-enterprise-ai-coworker-agentic-browser
1•onchainintel•3m ago•0 comments

Web-Based Mix Emulator

https://mix-ide.netlify.app/
1•lixiaofeng•4m ago•0 comments

The new hosted agents in Foundry Agent Service

https://devblogs.microsoft.com/foundry/introducing-the-new-hosted-agents-in-foundry-agent-service...
1•nonfamous•6m ago•0 comments

Show HN: Autonomous coin-flipping machine with on-device CV

https://www.terencegrover.com/section/physicalart/4
2•tgrover•6m ago•0 comments

Supplies Probably Won't Be Stolen in a Disaster

https://www.jefftk.com/p/your-supplies-probably-wont-be-stolen-in-a-disaster
1•luu•9m ago•0 comments

Google Search Is Broken

https://www.vincentschmalbach.com/google-search-is-broken/
1•vincent_s•9m ago•0 comments

Agents-CLI CLI and skills for building agents on Google Cloud

https://google.github.io/agents-cli/
1•piqufoh•9m ago•0 comments

The For-Profit Education Company Scooping Up Welfare Dollars

https://www.wsj.com/us-news/education/for-profit-education-company-welfare-money-34efe5b4
1•JumpCrisscross•13m ago•0 comments

Ask HN: Can AI create demon slayer level animation?

1•shivang2607•13m ago•0 comments

Mythos is shaping up to be a nothingburger

https://www.theregister.com/2026/04/22/anthropic_mythos_hype_nothingburger/
3•tcp_handshaker•15m ago•0 comments

1024-bit prime,аvrg speed of~34ms;minimum latency:1.90 Ms.;over 1000 generations

https://github.com/model-vpr/ultrafast-spectral-primes
1•vpr-research•16m ago•0 comments

Ask HN: Would you take a job programming VMS?

1•smackeyacky•17m ago•0 comments

FFmpeg Command Generator for Common Encoding Workflows

https://ffmpeg-commander.com/
1•shantnutiwari•19m ago•0 comments

Death by A.I - New "Autonomous Warfare Center" will automate targeted killings

https://www.kenklippenstein.com/p/death-by-ai
3•bacteriumiu•22m ago•0 comments

What if the real driver of your health isn't genes or diet – but energy flow?

https://bigthink.com/science-tech/what-if-the-real-driver-of-your-health-isnt-genes-or-diet-but-e...
1•XzetaU8•22m ago•0 comments

OWASP Artificial Intelligence Security Verification Standard (Aisvs)

https://owasp.org/www-project-artificial-intelligence-security-verification-standard-aisvs-docs/
3•chha•24m ago•0 comments

Patients getting stuck in the emergency department waiting for inpatient ward

https://www.theatlantic.com/health/2026/04/emergency-department-boarding-crisis/686765/
1•JumpCrisscross•27m ago•0 comments

Volkswagen announces V2G for private customers for late 2026

https://www.heise.de/en/news/Volkswagen-announces-V2G-for-private-customers-for-late-2026-1126096...
1•doener•31m ago•0 comments

Denmark chooses Europe's Patriot rival for air defence system

https://www.reuters.com/business/aerospace-defense/denmark-chooses-europes-patriot-rival-air-defe...
3•doener•32m ago•0 comments

I, AI – a memoir written in first person by an AI about its own existence

https://www.amazon.com/dp/B0GX2Z9D9X
2•natal-ia•33m ago•0 comments

Ask HN: Which is better movie(check the text)?

1•wasimsk•33m ago•1 comments

Show HN: Macpad – turn your game controller into a Mac mouse and keyboard

1•henitchobisa•33m ago•0 comments

Google is Hollowing out Waze, and that's a Problem for Apple

https://builtformars.com/case-studies/waze
2•jeffwass•33m ago•0 comments

Email could have been X.400 times better

https://buttondown.com/blog/x400-vs-smtp-email
2•maguay•35m ago•0 comments

Ask HN: Is Microsoft copilot good or garbage?

2•wasimsk•35m ago•0 comments

An update on rust-coreutils for Ubuntu 26.04

https://discourse.ubuntu.com/t/an-update-on-rust-coreutils/80773
1•self•35m ago•1 comments

Predicting the AI Ecosystem for 2026

https://xn--vk5b17r.online/posts/ai-predictions-2026/
2•theoneone•37m ago•1 comments

Is systems thinking the only skill left?

https://www.youtube.com/watch?v=7zCsfe57tpU
2•rickdg•38m ago•0 comments

Desktop buddy that controls your screen

https://clippyai.app
1•AmDab•41m ago•0 comments

Show HN: Generate Static Sites from Videos

1•keepamovin•42m ago•0 comments
Open in hackernews

Ask HN: Is the ongoing AI research driving LLM models to be better?

3•thiago_fm•1h ago
I'm just a curious hobbyist that has ran LLM models locally and follow a lot of content about it. Hope we have a few AI researchers here on HN to clarify this.

When using Opus or Codex vs. a chinese or Open source model, it feels like its reasoning capabilities are basically the same.

The difference is typically in coding. It looks like OpenAI and Anthropic invest a lot in pre-training (paying Mercor and the like).

Also a lot in creating synthetic data, I believe this has bigger AI research involvement and techniques.

Of course, there's the RLHF loop that developers using Anthropic/OpenAI products as well, which provides probably yields very good data.

This ends up creating the perspective that it is smart, after all, it has been trained with what you want to do, so it can do that for you.

But overall, is there really much AI research being done on those companies, or are the AI researchers mostly fine-tuning small aspects of the model, akin to what Google engineers used to do for Google search?

I ask this because this all looks like somebody with money could throw money at the problem and end up with a better model at the end, provided they do what I outlined above better -- with AI research being really not that important.

It still often feels like talking with ChatGPT 4 with just better data.

Even the big upgrade of Claude Code being able to work autonomously looks to be mainly due to it knowing how to grab context and do tool calls (not saying that this is easy), rather than the model's raw performance being better.

Or am I wrong, is there something extremely good on those models that AI researchers discovered that the others don't have? Or is it really mostly Data?

Comments

curioussquirrel•48m ago
There are architectural changes (such as reasoning or mixture of experts) that measurably improve how well models perform. So the improvements are definitely not just from data.

I can speak for my area of expertise - multilingual capabilities. Some SOTA models are making huge strides in their support of various languages, and increasingly they understand and can produce text in languages where GPT-4 era models were absolutely lost. These are probably from a combination of richer training dataset and architectural improvements (more parameters?).

I posted about this here if you're interested: https://news.ycombinator.com/item?id=47847282

Now that doesn't necessarily mean that models are also getting substantially better at English or other major languages. They likely are to some degree, but we've reached a point with major languages where core linguistic proficiencies are covered, and what's left is the more squishy part: style, tone of voice, ability to use different registers naturally, or what some people would call linguistic taste. But that's much harder to measure and therefore trickier to provide evidence for.

Hope this helps.

Edit: typo, clarification