frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

Start your own Internet Resiliency Club

https://bowshock.nl/irc/
68•todsacerdoti•1h ago•14 comments

Jokes and Humour in the Public Android API

https://voxelmanip.se/2025/06/14/jokes-and-humour-in-the-public-android-api/
98•todsacerdoti•8h ago•40 comments

NesDev.org – A community of homebrew game devs for NES and other retro consoles

https://www.nesdev.org/
50•ibobev•3h ago•2 comments

Lisp-stat: Lisp environment for statistical computing

https://lisp-stat.dev/about/
58•oumua_don17•1d ago•15 comments

Is Gravity Just Entropy Rising? Long-Shot Idea Gets Another Look

https://www.quantamagazine.org/is-gravity-just-entropy-rising-long-shot-idea-gets-another-look-20250613/
48•pseudolus•8h ago•16 comments

Why SSL was renamed to TLS in late 90s (2014)

https://tim.dierks.org/2014/05/security-standards-and-name-changes-in.html
296•Bogdanp•18h ago•135 comments

Modifying an HDMI dummy plug's EDID using a Raspberry Pi

https://www.downtowndougbrown.com/2025/06/modifying-an-hdmi-dummy-plugs-edid-using-a-raspberry-pi/
237•zdw•16h ago•65 comments

Twin – A Textmode WINdow Environment

https://github.com/cosmos72/twin
93•kim_rutherford•12h ago•12 comments

Canyon.mid

https://canyonmid.com/
285•LorenDB•19h ago•165 comments

Childhood leukemia: how a deadly cancer became treatable

https://ourworldindata.org/childhood-leukemia-treatment-history
197•surprisetalk•19h ago•52 comments

DARPA program sets distance record for power beaming

https://www.darpa.mil/news/2025/darpa-program-distance-record-power-beaming
54•gnabgib•9h ago•25 comments

Telephone Exchanges in the UK

https://telephone-exchanges.org.uk/
119•petecooper•13h ago•45 comments

Datalog in Rust

https://github.com/frankmcsherry/blog/blob/master/posts/2025-06-03.md
283•brson•21h ago•29 comments

Real-time CO2 monitoring without batteries or external power

https://news.kaist.ac.kr/newsen/html/news/?mode=V&mng_no=47450
34•gnabgib•10h ago•4 comments

First 2D, non-silicon computer developed

https://www.psu.edu/news/research/story/worlds-first-2d-non-silicon-computer-developed
100•giuliomagnifico•3d ago•19 comments

Chemical knowledge and reasoning of large language models vs. chemist expertise

https://www.nature.com/articles/s41557-025-01815-x
57•bookofjoe•1d ago•17 comments

Simplest C++ Callback, from SumatraPDF

https://blog.kowalczyk.info/a-stsj/simplest-c-callback-from-sumatrapdf.html
121•jandeboevrie•15h ago•97 comments

How to modify Starlink Mini to run without the built-in WiFi router

https://olegkutkov.me/2025/06/15/how-to-modify-starlink-mini-to-run-without-the-built-in-wifi-router/
290•LorenDB•19h ago•84 comments

Datalog in miniKanren

https://deosjr.github.io/dynamicland/datalog.html
95•deosjr•16h ago•9 comments

Accumulation of cognitive debt when using an AI assistant for essay writing task

https://arxiv.org/abs/2506.08872
160•stephen_g•5h ago•94 comments

Nanonets-OCR-s – OCR model transforms documents into structured markdown

https://huggingface.co/nanonets/Nanonets-OCR-s
51•PixelPanda•2h ago•10 comments

Reinventing circuit breakers with supercritical CO2

https://spectrum.ieee.org/sf6-gas-replacement
75•rbanffy•9h ago•28 comments

Fields where Native Americans farmed a thousand years ago discovered in Michigan

https://www.smithsonianmag.com/smart-news/massive-field-where-native-american-farmers-grew-corn-beans-and-squash-1000-years-ago-discovered-in-michigan-180986758/
185•CoopaTroopa•3d ago•83 comments

Unprecedented optical clock network lays groundwork for redefining the second

https://phys.org/news/2025-06-unprecedented-optical-clock-network-lays.html
11•wglb•3d ago•4 comments

Solving LinkedIn Queens with APL

https://pitr.ca/2025-06-14-queens
6•pitr•1d ago•0 comments

Random Walk: A Modern Introduction (2010) [pdf]

https://www.math.uchicago.edu/~lawler/srwbook.pdf
36•Anon84•3d ago•4 comments

Foundations of Computer Vision (2024)

https://visionbook.mit.edu
195•tzury•22h ago•10 comments

The Hewlett-Packard Archive

https://hparchive.com
18•joebig•6h ago•2 comments

Cyborg Embryos Offer New Insights into Brain Growth

https://spectrum.ieee.org/embryo-electrode-array
23•rbanffy•3d ago•2 comments

The Art of Lisp and Writing (2003)

https://www.dreamsongs.com/ArtOfLisp.html
179•Bogdanp•1d ago•70 comments
Open in hackernews

Chemical knowledge and reasoning of large language models vs. chemist expertise

https://www.nature.com/articles/s41557-025-01815-x
57•bookofjoe•1d ago

Comments

fuzzfactor•1d ago
Nothing to see here unless you have some kind of unsatisfied interest in the future of AI :\

This is all highly academic, and I'm highly industrial so take this with a grain of salt. Sodium salt or otherwise, your choice ;)

If you want things to be accomplished at the bench, you want any simulation to be made by those who have not been away from the bench for that many decades :)

Same thing with the industrial environment, some people have just been away from it for too long regardless of how much familiarity they once had. You need to brush up, sometimes the same plant is like a whole different world if you haven't been back in a while.

mistrial9•7h ago
BASF Group - will they speak in public? probably not, given what is at stake IMHO
calibas•5h ago
I'm sure an LLM knows more about computer science than a human programmer.

Not to say the LLM is more intelligent or better at coding, but that computer science is an incredibly broad field (like chemistry). There's simply so much to know that the LLM has an inherent advantage. It can be trained with huge amounts of generalized knowledge far faster than a human can learn.

Do you know every common programing language? The LLM does, plus it can code in FRACTRAN, Brainfuck, Binary lambda calculus, and a dozen other obscure languages.

It's very impressive, until you realize the LLM's knowledge is a mile wide and an inch deep. It has vast quantities of knowledge, but lacks depth. A human that specializes in a field is almost always going to outperform an LLM in that field, at least for the moment.

esafak•4h ago
But the LLM can already connect things that you can not, by virtue of its breadth. Some may disagree, but I think it will soon go deeper too.
timschmidt•4h ago
> Do you know every common programing language? The LLM does, plus it can code in FRACTRAN, Brainfuck, Binary lambda calculus, and a dozen other obscure languages.

Not only this, but they're surprisingly talented at reading compiled binaries in a dozen different machine and bytecodes. I have seen one one-shot an applet rewrite from compiled java bytecode to modern javascript.

anthk•2h ago
Binwalk, Unicorn... as if it that was advanced wizardry. Unix systems have file(1) since forever and binutils from and to every arch.
mumbisChungo•4h ago
It's impressive until you realize its limitations.

Then it becomes impressive again once you understand how to productively use it as a tool, given its limitations.

anthk•2h ago
So impressive that every complex SUBLEQ code I've tried with an LLM failed really fast.
logifail•2h ago
> Do you know every common programing language?

A long time ago my OH was introduced to someone who claimed "to speak seven languages fluently".

Her response at the time was was "Do they have anything interesting to say in any of them?"

6LLvveMx2koXfwn•3h ago
Received 01 April 2024

Accepted 26 March 2025

Published 20 May 2025

Probably normal but shows the built in obsolescence of the peer review journal article model in such a fast moving field.

Jimmc414•3h ago
shows the value of preprint servers like arxiv.org and chemrxiv.org
eesmith•2h ago
How so?

To me it looks like the paper was submitted last year but the peer reviewers identified issues with the paper which required revision before the final acceptance in March.

We can see the paper was updated since the 1 April 2024 version as it includes o1-preview (released September 2024, I believe), and GPT‑3.5 Turbo from August. I think a couple of other tested versions also post-date 1 April.

Thus, one possible criticism might have been (and I stress that I am making this up) that the original paper evaluated only 3 systems, and didn't reflect the fully diversity of available tools.

In any case, the main point of the paper was not the specific results of AI models available by the end of last year, but the development of a benchmark which can be used to evaluated models in general.

How has that work been made obsolete?

bufferoverflow•1h ago
How so? All the models they've tested are obsolete, multiple generations behind high-end versions.

(Though even these obsolete models did better than the best humans and domain experts).

eesmith•1h ago
As I wrote, the main point of the paper was not the specific model evaluation, but the development of a benchmark which can be used to test new models.

Good benchmark development is hard work. The paper goes into the details of how it was carried out.

Now that the benchmark is available, you or anyone else could use it to evaluate the current high-end versions, and measure how the performance has changed over time.

You could also use their paper to help understand how to develop a new benchmark, perhaps to overcome some limitations in the benchmark.

That benchmark and the contents of that paper are not obsolete until there is a better benchmark and description of how to build benchmarks.

pu_pe•56m ago
Nice benchmark but the human comparison is a little lacking. They claim to have surveyed 19 experts, though the vast majority of them have only a master's degree. This would be akin to comparing LLM programming expertise to a sample of programmers with less than 5 years of experience.

I'm also not sure it's a fair comparison to average human results like that. If you quiz physicians on a broad variety of topics, you shouldn't expect cardiologists to know that much about neurology and vice-versa. This is what they did here, it seems.

KSteffensen•31m ago
I'll get some downvotes for this but PhD vs master's degree difference is mostly work experience, an element of workload hazing and snobbery.

Somebody with a masters degree and 5 years of work experience will likely know more than a freshly graduated PhD

eesmith•20m ago
Sure, but all we know is that these "13 have a master’s degree (and are currently enroled in Ph.D. studies)". We only know they have at least "2 years of experience in chemistry after their first university-level course in chemistry."

How does that qualify them as "domain experts"? What domain is their expertise? All of chemistry?