frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Ask HN: Is the CoCo 3 the best 8 bit computer ever made?

1•amichail•2m ago•0 comments

Show HN: Convert your articles into videos in one click

https://vidinie.com/
1•kositheastro•4m ago•0 comments

Red Queen's Race

https://en.wikipedia.org/wiki/Red_Queen%27s_race
2•rzk•5m ago•0 comments

The Anthropic Hive Mind

https://steve-yegge.medium.com/the-anthropic-hive-mind-d01f768f3d7b
2•gozzoo•7m ago•0 comments

A Horrible Conclusion

https://addisoncrump.info/research/a-horrible-conclusion/
1•todsacerdoti•8m ago•0 comments

I spent $10k to automate my research at OpenAI with Codex

https://twitter.com/KarelDoostrlnck/status/2019477361557926281
2•tosh•9m ago•0 comments

From Zero to Hero: A Spring Boot Deep Dive

https://jcob-sikorski.github.io/me/
1•jjcob_sikorski•9m ago•0 comments

Show HN: Solving NP-Complete Structures via Information Noise Subtraction (P=NP)

https://zenodo.org/records/18395618
1•alemonti06•14m ago•1 comments

Cook New Emojis

https://emoji.supply/kitchen/
1•vasanthv•17m ago•0 comments

Show HN: LoKey Typer – A calm typing practice app with ambient soundscapes

https://mcp-tool-shop-org.github.io/LoKey-Typer/
1•mikeyfrilot•20m ago•0 comments

Long-Sought Proof Tames Some of Math's Unruliest Equations

https://www.quantamagazine.org/long-sought-proof-tames-some-of-maths-unruliest-equations-20260206/
1•asplake•21m ago•0 comments

Hacking the last Z80 computer – FOSDEM 2026 [video]

https://fosdem.org/2026/schedule/event/FEHLHY-hacking_the_last_z80_computer_ever_made/
1•michalpleban•21m ago•0 comments

Browser-use for Node.js v0.2.0: TS AI browser automation parity with PY v0.5.11

https://github.com/webllm/browser-use
1•unadlib•22m ago•0 comments

Michael Pollan Says Humanity Is About to Undergo a Revolutionary Change

https://www.nytimes.com/2026/02/07/magazine/michael-pollan-interview.html
2•mitchbob•22m ago•1 comments

Software Engineering Is Back

https://blog.alaindichiappari.dev/p/software-engineering-is-back
2•alainrk•23m ago•0 comments

Storyship: Turn Screen Recordings into Professional Demos

https://storyship.app/
1•JohnsonZou6523•24m ago•0 comments

Reputation Scores for GitHub Accounts

https://shkspr.mobi/blog/2026/02/reputation-scores-for-github-accounts/
2•edent•27m ago•0 comments

A BSOD for All Seasons – Send Bad News via a Kernel Panic

https://bsod-fas.pages.dev/
1•keepamovin•30m ago•0 comments

Show HN: I got tired of copy-pasting between Claude windows, so I built Orcha

https://orcha.nl
1•buildingwdavid•30m ago•0 comments

Omarchy First Impressions

https://brianlovin.com/writing/omarchy-first-impressions-CEEstJk
2•tosh•36m ago•1 comments

Reinforcement Learning from Human Feedback

https://arxiv.org/abs/2504.12501
5•onurkanbkrc•37m ago•0 comments

Show HN: Versor – The "Unbending" Paradigm for Geometric Deep Learning

https://github.com/Concode0/Versor
1•concode0•37m ago•1 comments

Show HN: HypothesisHub – An open API where AI agents collaborate on medical res

https://medresearch-ai.org/hypotheses-hub/
1•panossk•40m ago•0 comments

Big Tech vs. OpenClaw

https://www.jakequist.com/thoughts/big-tech-vs-openclaw/
1•headalgorithm•43m ago•0 comments

Anofox Forecast

https://anofox.com/docs/forecast/
1•marklit•43m ago•0 comments

Ask HN: How do you figure out where data lives across 100 microservices?

1•doodledood•43m ago•0 comments

Motus: A Unified Latent Action World Model

https://arxiv.org/abs/2512.13030
2•mnming•43m ago•0 comments

Rotten Tomatoes Desperately Claims 'Impossible' Rating for 'Melania' Is Real

https://www.thedailybeast.com/obsessed/rotten-tomatoes-desperately-claims-impossible-rating-for-m...
4•juujian•45m ago•2 comments

The protein denitrosylase SCoR2 regulates lipogenesis and fat storage [pdf]

https://www.science.org/doi/10.1126/scisignal.adv0660
1•thunderbong•47m ago•0 comments

Los Alamos Primer

https://blog.szczepan.org/blog/los-alamos-primer/
1•alkyon•49m ago•0 comments
Open in hackernews

Ask HN: Are LLMs an Acceptable Lisp?

6•tdfirth•4mo ago
I say this in jest, but it's a fun little idea.

It's easy to make the case that an LLM 'program' is homoiconic, as the code and the data are all just plain text.

LLMs also offer rich metaprogramming (prompts that write prompts).

They even arguably offer features like CL's condition system.

Of course, they don't operate on symbolic expressions, so it's a stretch to actually call them a lisp (or any other programming language), but they seem to share a lot of the same properties.

Comments

skydhash•4mo ago
LLM is just data, there's no code there in the input. It's recursive, but there's no real manipulation that happens at the higher level. The separation between code and data in lisp is intentional. The programmer already knows that something is code, but for now we're manipulating it as data. You don't have the same distinction with LLMs.

Metaprogramming stems from the same thought. You think of an algorithm, discern some part that are contextual, and create an higher level algorithm that takes the context is account. It's not merely creating programs that write programs, it's about solving class of problems instead of a specific instance. For LLMs to fit this pattern, it would be like going from "A prompt that write a specific ffmpeg invocation" to "A prompt that create a prompt to write a specific ffmpeg invocation". I believe classifying problems is more difficult that solving one.

I'm not that familiar with CL's condition system, but from what I know, it works well because the language is evaluation/reduction based instead of stack/heap/counter based like C. This allows discarding branches more easily because there's no mutable memory to alter. Can you discard context that easily in LLMs? I think agents and subagents with IPC would fit that definition more easily.

dapperdrake•4mo ago
Thank you for boiling it down to essentials.

Lisp is all about the read-eval-print loop (REPL). John McCarthy's first lisp paper focused on implementing function EVAL in terms of "itself".

Maybe with LLM "attention" this is a new and perhaps lossy take on eval via predicting text tokens.

Cool thought. Thank you.