frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

BookTalk: A Reading Companion That Captures Your Voice

https://github.com/bramses/BookTalk
1•_bramses•54s ago•0 comments

Is AI "good" yet? – tracking HN's sentiment on AI coding

https://www.is-ai-good-yet.com/#home
1•ilyaizen•1m ago•1 comments

Show HN: Amdb – Tree-sitter based memory for AI agents (Rust)

https://github.com/BETAER-08/amdb
1•try_betaer•2m ago•0 comments

OpenClaw Partners with VirusTotal for Skill Security

https://openclaw.ai/blog/virustotal-partnership
1•anhxuan•2m ago•0 comments

Show HN: Seedance 2.0 Release

https://seedancy2.com/
1•funnycoding•3m ago•0 comments

Leisure Suit Larry's Al Lowe on model trains, funny deaths and Disney

https://spillhistorie.no/2026/02/06/interview-with-sierra-veteran-al-lowe/
1•thelok•3m ago•0 comments

Towards Self-Driving Codebases

https://cursor.com/blog/self-driving-codebases
1•edwinarbus•3m ago•0 comments

VCF West: Whirlwind Software Restoration – Guy Fedorkow [video]

https://www.youtube.com/watch?v=YLoXodz1N9A
1•stmw•4m ago•1 comments

Show HN: COGext – A minimalist, open-source system monitor for Chrome (<550KB)

https://github.com/tchoa91/cog-ext
1•tchoa91•5m ago•1 comments

FOSDEM 26 – My Hallway Track Takeaways

https://sluongng.substack.com/p/fosdem-26-my-hallway-track-takeaways
1•birdculture•5m ago•0 comments

Show HN: Env-shelf – Open-source desktop app to manage .env files

https://env-shelf.vercel.app/
1•ivanglpz•9m ago•0 comments

Show HN: Almostnode – Run Node.js, Next.js, and Express in the Browser

https://almostnode.dev/
1•PetrBrzyBrzek•9m ago•0 comments

Dell support (and hardware) is so bad, I almost sued them

https://blog.joshattic.us/posts/2026-02-07-dell-support-lawsuit
1•radeeyate•10m ago•0 comments

Project Pterodactyl: Incremental Architecture

https://www.jonmsterling.com/01K7/
1•matt_d•10m ago•0 comments

Styling: Search-Text and Other Highlight-Y Pseudo-Elements

https://css-tricks.com/how-to-style-the-new-search-text-and-other-highlight-pseudo-elements/
1•blenderob•12m ago•0 comments

Crypto firm accidentally sends $40B in Bitcoin to users

https://finance.yahoo.com/news/crypto-firm-accidentally-sends-40-055054321.html
1•CommonGuy•13m ago•0 comments

Magnetic fields can change carbon diffusion in steel

https://www.sciencedaily.com/releases/2026/01/260125083427.htm
1•fanf2•13m ago•0 comments

Fantasy football that celebrates great games

https://www.silvestar.codes/articles/ultigamemate/
1•blenderob•13m ago•0 comments

Show HN: Animalese

https://animalese.barcoloudly.com/
1•noreplica•14m ago•0 comments

StrongDM's AI team build serious software without even looking at the code

https://simonwillison.net/2026/Feb/7/software-factory/
3•simonw•14m ago•0 comments

John Haugeland on the failure of micro-worlds

https://blog.plover.com/tech/gpt/micro-worlds.html
1•blenderob•15m ago•0 comments

Show HN: Velocity - Free/Cheaper Linear Clone but with MCP for agents

https://velocity.quest
2•kevinelliott•15m ago•2 comments

Corning Invented a New Fiber-Optic Cable for AI and Landed a $6B Meta Deal [video]

https://www.youtube.com/watch?v=Y3KLbc5DlRs
1•ksec•17m ago•0 comments

Show HN: XAPIs.dev – Twitter API Alternative at 90% Lower Cost

https://xapis.dev
2•nmfccodes•17m ago•1 comments

Near-Instantly Aborting the Worst Pain Imaginable with Psychedelics

https://psychotechnology.substack.com/p/near-instantly-aborting-the-worst
2•eatitraw•23m ago•0 comments

Show HN: Nginx-defender – realtime abuse blocking for Nginx

https://github.com/Anipaleja/nginx-defender
2•anipaleja•24m ago•0 comments

The Super Sharp Blade

https://netzhansa.com/the-super-sharp-blade/
1•robin_reala•25m ago•0 comments

Smart Homes Are Terrible

https://www.theatlantic.com/ideas/2026/02/smart-homes-technology/685867/
2•tusslewake•26m ago•0 comments

What I haven't figured out

https://macwright.com/2026/01/29/what-i-havent-figured-out
1•stevekrouse•27m ago•0 comments

KPMG pressed its auditor to pass on AI cost savings

https://www.irishtimes.com/business/2026/02/06/kpmg-pressed-its-auditor-to-pass-on-ai-cost-savings/
1•cainxinth•27m ago•0 comments
Open in hackernews

From tokens to thoughts: How LLMs and humans trade compression for meaning

https://arxiv.org/abs/2505.17117
124•ggirelli•8mo ago

Comments

valine•8mo ago
>> For each LLM, we extract static, token-level embeddings from its input embedding layer (the ‘E‘matrix). This choice aligns our analysis with the context-free nature of stimuli typical in human categorization experiments, ensuring a comparable representational basis.

They're analyzing input embedding models, not LLMs. I'm not sure how the authors justify making claims about the inner workings of LLMs when they haven't actually computed a forward pass. The EMatrix is not an LLM, its a lookup table.

Just to highlight the ridiculousness of this research, no attention was computed! Not a single dot product between keys and queries. All of their conclusions are drawn from the output of an embedding lookup table.

The figure showing their alignment score correlated with model size is particularly egregious. Model size is meaningless when you never activate any model parameters. If Bert is outperforming Qwen and Gemma something is wrong with your methodology.

blackbear_•8mo ago
Note that the token embeddings are also trained, therefore their values do give some hints on how a model is organizing information.

They used token embeddings directly and not intermediate representations because the latter depend on the specific sentence that the model is processing. Data on human judgment was however collected without any context surrounding each word, thus using the token embeddings seem to be the most fair comparison.

Otherwise, what sentence(s) would you have used to compute the intermediate representations? And how would you make sure that the results aren't biased by these sentences?

navar•8mo ago
You can process a single word through a transformer and get the corresponding intermediate representations.

Though it sounds odd there is no problem with it and it would indeed return the model's representation of that single word as seen by the model without any additional context.

valine•8mo ago
Embedding models are not always trained with the rest of the model. That’s the whole idea behind VLLMs. First layer embeddings are so interchangeable you can literally feed in the output of other models using linear projection layers.

And like the other commenter said, you can absolutely feed single tokens through the model. Your point doesn’t make any sense though regardless. How about priming the model with “You’re a helpful assistant” just like everyone else does.

boroboro4•8mo ago
It’s mind blowing LeCun is listed as one of the authors.

I would expect model size to correlate with alignment score because usually model sizes correlate with hidden dimension. But also opposite can be true - bigger models might shift more basic token classification logic into layers and hence embedding alignment can go down. Regardless feels like pretty useless research…

danielbln•8mo ago
Leaves a bit of a taste considering LeCun's famously critical stance on auto-regressive transformer LLMs.
throwawaymaths•8mo ago
the llm is also a lookup table! but your point is correct. they should have looked at subsequent layers that aggregate information over distance.
andoando•8mo ago
Am I the only one that is lost on how the calculations are made?

From what I can tell this is limited in scope to categorizing nouns (robin is a bird).

fusionadvocate•8mo ago
Open a bank account. Open your heart. Open a can. Open to new experiences.

Words are a tricky thing to handle.

an0malous•8mo ago
OpenAI agrees
esafak•8mo ago
And models since BERT and ELMo capture polysemy!

https://aclanthology.org/2020.blackboxnlp-1.15/

bluefirebrand•8mo ago
And that is just in English

Other languages have similar but fundamentally different oddities which do not translate cleanly

suddenlybananas•8mo ago
Not sure how they're fundamentally different. What do you mean?
bluefirebrand•8mo ago
Think about the work of localizing a joke that relies on wordplay or similar sounding words to be funny. Or simply how words rhyme

Try explaining why tough and rough rhyme but bough doesn't

You know? Language has a ton of idiosyncrasies.

Qworg•8mo ago
To make it more concrete - here's an example in Chinese: https://en.wikipedia.org/wiki/Grass_Mud_Horse
mock-possum•8mo ago
> 2009, renowned artist Ai Weiwei published an image of himself nude with only a 'Caonima' hiding his genitals, with a caption "草泥马挡中央" (cǎonímǎ dǎng zhōngyāng; 'a Grass Mud Horse covering the center'. One interpretation of the caption is: "fuck your mother, Communist Party Central Committee"). Political observers speculated that the photo may have contributed to Ai's arrest in 2011 by angering Chinese Communist Party hardliners.

How did I never hear about this detail??

Scarblac•8mo ago
ChatGPT is horrible at producing Dutch rhymes (for Sinterklaas poems) until you realize that the words it comes up with do rhyme when translated to English.
suddenlybananas•8mo ago
Right but I wouldn't call those things fundamentally different. That's just having different words; the categories of idiosyncrasies are still the same.
thesz•8mo ago
As most languages allow expressions of algorithms, they are all Turing complete and, thus, are not fundamentally different. The complexity of expressions of some concepts is different, though.

My favorite thing is a "square." I put that name to an enumeration that allows me to compare and contrast things with two different qualities expressed by two extremes.

One such square is "One can (not) do (not do) something." Both "not"'s can be present and absent, just like a truth table.

"One can do something", "one can not do something", "one can do not do something" and, finally, "one can not help but do something."

Why should we use "help but" instead of "do not"?

While this does not preclude one from enumerating possibilities thinking in English, it makes that enumeration harder than it can be in other languages. For example, in Russian the "square" is expressible directly.

Also, "help but" is not shorter than "do not," it is longer. Useful idioms usually expressed in shorter forms, thus, apparently, "one can not help but do something" is considered by Englishmen as not useful.

falcor84•8mo ago
I agree in general, but I think that "open" is actually a pretty straightforward word.

As I see it, "Open your heart", "Open a can" and "Open to new experiences" have very similar meanings for "Open", being essentially "make a container available for external I/O", similar to the definition of an "open system" in thermodynamics. "Open a bank account" is a bit different, as it creates an entity that didn't exist before, but even then the focus is on having something that allows for external I/O - in this case deposits and withdrawals.

johnnyApplePRNG•8mo ago
This paper is interesting, but ultimately it's just restating that LLMs are statistical tools and not cognitive systems. The information-theoretic framing doesn’t really change that.
Nevermark•8mo ago
> LLMs are statistical tools and not cognitive systems

I have never understood broad statements that models are just (or mostly) statistical tools.

Certainly statistics apply, minimizing mismatches results in mean (or similar measure) target predictions.

But the architecture of a model is the difference between compressed statistics vs. forcing a model to translate information in a highly organized way reflecting the actual shape of the problem to get any accuracy at all.

In both cases, statistics are relevant, but in the latter it's not a particularly insightful way to talk about what a model has learned.

Statistical accuracy, prediction, etc. are basic problems to solve. The training criteria being optimized. But they don't limit the nature of solutions. They both leave problem difficulty, and solution sophistication unbounded.

larodi•8mo ago
my take from the paper is that the authors overstate the following:

LLMs cannot zoom in and out of certain details, while we humans subconsciously zoom in and out of context and even go through conceptual spaces in the very hermetic sense of it. LLMs do not. they work on the present context, which kickstarts them into expounding likeliness in the same contextual space.

catchnear4321•8mo ago
incomplete inaccurate off misleading meandering not quite generation prediction removal of superfluous fast but spiky

this isn’t talking about that.

xwat•8mo ago
Stochastic parrots