frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Anthropic Economic Index economic primitives

https://www.anthropic.com/research/anthropic-economic-index-january-2026-report
30•malshe•1h ago

Comments

mips_avatar•1h ago
Every single AI economic analysis talks about travel planning but none of the AI labs have the primitives (transit routing, geocoding, etc.) in a semantic interface for the models to use.
malshe•44m ago
Coincidentally, YouTube demos on vibe coding commonly make travel planning apps!
bix6•1h ago
> These “primitives”—simple, foundational measures of how Claude is used, which we generate by asking Claude specific questions about anonymized Claude.ai and first-party (1P) API transcripts

I just skimmed but is there any manual verification / human statistical analysis done on this or we just taking Claude’s word for it?

sdwr•39m ago
Looks like they are relying on Claude for it, which is interesting. I bet social scientists are going to love this approach
adverbly•1h ago
This is very cool but it's not quite what I expected out of economic primitives.

I expected to see measures of the economic productivity generated as a result of artificial intelligence use.

Instead, what I'm seeing is measures of artificial intelligence use.

I don't really see how this is measuring the most important economic primitives. Nothing related to productivity at all actually. Everything about how and where and who... This is just demographics and usage statistics...

ossa-ma•46m ago
There is a chapter titled 'Chapter 4: Tasks and productivity'. They discuss speedup, success rates, productivity growth across different occupations. Also theres more in the full report.
kurttheviking•46m ago
agree, was similarly hoping for something akin to a total factor productivity argument
brap•1h ago
All of this performative bullshit coming out of Anthropic is slowly but surely making them my least favorite AI company.

We get it guys the very scary future is here any minute now and you’re the only ones taking it super seriously and responsibly and benevolently. That’s great. Now please just build the damn thing

ossa-ma•42m ago
These are economic studies on AI's impact on productivity, jobs, wages, global inequality. It's important to UNDERSTAND who benefits from technology and who gets left behind. Even putting the positive impacts of a study like this aside - this kinda due diligence is critical for them to understand developing markets and how to reach them.
brap•27m ago
Ok Dario
mlsu•57m ago
> This also highlights the importance of model design and training. While Claude is able to respond in a highly sophisticated manner, it tends to do so only when users input sophisticated prompts.

If the output of the model depends on the intelligence of the person picking outputs out of its training corpus, is the model intelligent?

This is kind of what I don't quite understand when people talk about the models being intelligent. There's a huge blindspot, which is that the prompt entirely determines the output.

thousand_nights•49m ago
i don't know, are we intelligent?

you could argue that our input (senses) entirely define the output (thoughts, muscle movements, etc)

wat10000•48m ago
A smart person will tailor their answers to the perceived level of knowledge of the person asking, and the sophistication of the question is a big indicator of this.
TrainedMonkey•44m ago
Humans also respond differently when prompted in different ways. For example, politeness often begets politeness. I would expect that to be reflected in training data.
mlsu•7m ago
If I, a moron, hire a PhD to crack a tough problem for me, I don't need to go back and forth prompting him at a PhD level. I can set him loose on my problem and he'll come back to me with a solution.
zozbot234•35m ago
What is a "sophisticated prompt"? What if I just tack on "please think about this a lot and respond in a highly sophisticated manner" to my question/prompt? Anyone can do this once they're made aware of this potential issue. Sometimes the UX layer even adds this for you in the system prompt, you just have to tick the checkbox for "I want a long, highly sophisticated answer".
mlsu•17m ago
They have a chart that shows it. The education level of the input determines the education level of the output.

These things are supposed to have intelligence on tap. I'll imagine this in a very simple way. Let's say "intellignce" is like a fluid. It's a finite thing. Intelligence is very valuable, it's the substrate for real-world problem solving that makes these things ostensibly worth trillions of dollars. Intelligence comes from interaction with the world; someone's education and experience. You spend some effort and energy feeding someone, clothing them, sending them to college. And then you get something out, which is intelligence that can create value for society.

When you are having a conversation with the AI, is the intelligence flowing out of the AI? Or is it flowing out of the human operator?

The answer to this question is extremely important. If the AI can be intelligent "on its own" without a human operator, then it will be very valuable -- feed electricity into a datacenter and out comes business value. But if a model is only intelligent as someone using it, well, the utility seems to be very harshly capped. At best it saves a bit of time, but it will never do anything novel, it will never create value on its own, independently, it will never scale beyond a 1:1 "human picking outputs".

If you must encode intelligence into the prompt to get intelligence out of the model, well, this doesn't quite look like AGI does it?

mlsu•12m ago
ofc what I'm getting at is, you can't get something from nothing. There is no free lunch.

You spend energy distilling the intelligence of the entire internet into a set of weights, but you still had to expend the energy to have humans create the internet first. And on top of this, in order to pick out what you want from the corpus, you have to put some energy in: first, the energy of inference, but second and far more importantly, the energy of prompting. The model is valuable because the dataset is valuable; the model output is valuable because the prompt is valuable.

So wait then, where does this exponential increase in value come from again?

dingdingdang•45m ago
The title actually cringes me out a bit, it reads like early report titles in academia where young students (myself no doubt incl back when) try their hardest at making a title sound clever but in actuality only achieve obscuration of their own material.
Herring•45m ago
I got halfway through then asked the LLM to extract any surprising or particularly interesting results, below. Verified each one manually.

1. Accelerated US Diffusion vs. Global Stagnation

The report estimates that usage parity across US states could be achieved in 2–5 years. This rate of diffusion is approximately 10 times faster than that of major 20th-century technologies (e.g., electricity or the automobile). However, this rapid convergence is not observed globally. International adoption remains strictly correlated with GDP per capita, with no evidence of lower-income countries "catching up." This suggests AI may currently exacerbate rather than narrow the digital divide between nations.

2. The Education Mirror Effect

A high correlation (r > 0.92) exists between the education level required to write a prompt and the education level of Claude's response. This implies that the model's sophisticated capabilities are only unlocked by users who already possess high formal education. Rather than acting as a cognitive equalizer, AI appears to function as a capital multiplier for existing high-skill workers.

3. Productivity "Haircut" via Reliability

While raw "speedup" data suggest significant time savings (e.g., a 12x speedup for college-level tasks), these gains are substantially offset by task failure. Adjusting aggregate productivity growth estimates for task reliability reduces the projected impact from 1.8 percentage points to approximately 1.0 percentage point of annual labor productivity growth. This 44% "haircut" highlights that current model unreliability is a primary bottleneck for macroeconomic impact.

4. Selective Deskilling of White-Collar Work

The report finds that AI disproportionately covers high-education tasks. In a task-displacement model, this leads to a net "deskilling" effect for several professions. For example, Technical Writers may lose their most complex analytical tasks to AI, leaving behind only routine illustrative or observational work. Conversely, occupations like Property Management may experience "upskilling" as AI handles bookkeeping, leaving managers to focus on high-stakes negotiations.

5. Multi-Turn "Task Horizon" Extension

The "task horizon"—the maximum task duration at which AI achieves a 50% success rate—varies wildly by interface. For single-turn API interactions, the horizon is 3.5 hours. For multi-turn Claude.ai conversations, it extends to 19 hours. This suggests that the iterative, human-in-the-loop chat interface is significantly more effective at managing complex, long-duration tasks than programmatic automation.

Critical Uncertainties:

Temporal limitations: The data was collected over a single week in November 2025 and predates the release of Opus 4.5.

Self-selection bias: Success rates reflect only the tasks users choose to bring to AI. If users avoid tasks they expect the AI to fail at, success rates are artificially inflated.

Measurement of "Education Years": The report uses a Ridge regression model to predict education requirements from task embeddings. If the training data (BLS occupation levels) is a lagging indicator of actual skill requirements, the deskilling/upskilling analysis may be skewed.

blibble•43m ago
> How is AI reshaping the economy?

oh I know this one!

it's created mountains of systemic risk for absolutely no payoff whatsoever!

andy_xor_andrew•38m ago
no payoff whatsoever? I just asked Claude to do a task that would have previously taken me four days. Then I got up and got lunch, and when I was back, it was done.

I would never make the argument that there are no risks. But there's also no way you can make the argument there are no payoffs!

siliconc0w•35m ago
Skimmed, some notes for a more 'bear' case:

* value seems highly concentrated in a sliver of tasks - the top ten accounting for 32%, suggesting a fat long-tail where it may be less useful/relevant.

* productivity drops to a more modest 1-1.2% productivity gain once you account for humans correcting AI failure. 1% is still plenty good, especially given the historical malaise here of only like 2% growth but it's not like industrial revolution good.

* reliability wall - 70% success rate is still problematic and we're getting down to 50% with just 2+ hours of task duration or about "15 years" of schooling in terms of complexity for API. For web-based multi-turn it's a bit better but I'd imagine that would at least partly due to task-selection bias.

xiphias2•29m ago
,,1% is still plenty good, especially given the historical malaise here of only like 2% growth but it's not like industrial revolution good.''

You can't compare the speed of AI improvements to the speed of technical improvements during the industrial revolution. ChatGPT is 3 years old.

ossa-ma•30m ago
I'm not an economist so can someone explain whether this stat is significant:

> a sustained increase of 1.0 percentage point per year for the next ten years would return US productivity growth to rates that prevailed in the late 1990s and early 2000s

What can it be compared to? Is it on the same level of productivity growth as computers? The internet? Sliced bread?

GPTZero finds 100 new hallucinations in NeurIPS 2025 accepted papers

https://gptzero.me/news/neurips/
650•segmenta•8h ago•353 comments

Show HN: isometric.nyc – giant isometric pixel art map of NYC

https://cannoneyed.com/isometric-nyc/
551•cannoneyed•7h ago•143 comments

Viking Ship Museum in Denmark announces the discovery of the largest cog

https://www.medievalists.net/2025/12/medieval-ship-discovered-copenhagen/
30•PaulHoule•2h ago•13 comments

Qwen3-TTS family is now open sourced: Voice design, clone, and generation

https://qwen.ai/blog?id=qwen3tts-0115
437•Palmik•10h ago•132 comments

Anthropic Economic Index economic primitives

https://www.anthropic.com/research/anthropic-economic-index-january-2026-report
30•malshe•1h ago•26 comments

Compiling Scheme to WebAssembly

https://eli.thegreenplace.net/2026/compiling-scheme-to-webassembly/
59•chmaynard•5d ago•8 comments

Composing APIs and CLIs in the LLM era

https://walters.app/blog/composing-apis-clis
14•zerf•6h ago•0 comments

My first year in sales as technical founder

https://www.fabiandietrich.com/blog/first-year-in-sales.html
51•f3b5•5d ago•16 comments

Extracting a UART Password via SPI Flash Instruction Tracing

https://zuernerd.github.io/blog/2026/01/07/switch-password.html
26•Eduard•2h ago•3 comments

CSS Optical Illusions

https://alvaromontoro.com/blog/68091/css-optical-illusions
122•ulrischa•6h ago•11 comments

'Active' sitting is better for brain health: review of studies

https://www.sciencealert.com/not-all-sitting-is-equal-one-type-was-just-linked-to-better-brain-he...
56•mikhael•4h ago•25 comments

Recent discoveries on the acquisition of the highest levels of human performance

https://www.science.org/doi/abs/10.1126/science.adt7790
90•colincooke•5h ago•45 comments

Why does SSH send 100 packets per keystroke?

https://eieio.games/blog/ssh-sends-100-packets-per-keystroke/
224•eieio•4h ago•157 comments

Your app subscription is now my weekend project

https://rselbach.com/your-sub-is-now-my-weekend-project
168•robteix•4d ago•141 comments

Tree-sitter vs. Language Servers

https://lambdaland.org/posts/2026-01-21_tree-sitter_vs_lsp/
199•ashton314•9h ago•55 comments

Show HN: First Claude Code client for Ollama local models

https://github.com/21st-dev/1code
29•SerafimKorablev•6h ago•14 comments

Capital One to acquire Brex for $5.15B

https://www.reuters.com/legal/transactional/capital-one-buy-fintech-firm-brex-515-billion-deal-20...
42•personjerry•2h ago•43 comments

Preserved Fish, Boss of New York City

https://signoregalilei.com/2025/12/21/preserved-fish-boss-of-new-york-city/
22•surprisetalk•3d ago•5 comments

A Year of 3D Printing

https://brookehatton.com/blog/making/a-year-of-3d-printing/
76•nindalf•5d ago•75 comments

AnswerThis (YC F25) Is Hiring

https://www.ycombinator.com/companies/answerthis/jobs/r5VHmSC-ai-agent-orchestration
1•ayush4921•6h ago

Show HN: Text-to-video model from scratch (2 brothers, 2 years, 2B params)

https://huggingface.co/collections/Linum-AI/linum-v2-2b-text-to-video
36•schopra909•7h ago•9 comments

Show HN: CLI for working with Apple Core ML models

https://github.com/schappim/coreml-cli
30•schappim•3h ago•1 comments

Launch HN: Constellation Space (YC W26) – AI for satellite mission assurance

28•kmajid•6h ago•11 comments

Design Thinking Books (2024)

https://www.designorate.com/design-thinking-books/
263•rrm1977•12h ago•120 comments

Keeping 20k GPUs healthy

https://modal.com/blog/gpu-health
73•jxmorris12•4d ago•30 comments

Show HN: I've been using AI to analyze every supplement on the market

https://pillser.com/
47•lilouartz•9h ago•17 comments

Mote: An Interactive Ecosystem Simulation [video]

https://www.youtube.com/watch?v=Hju0H3NHxVI
55•evakhoury•1d ago•8 comments

Reverse engineering Lyft Bikes for fun (and profit?)

https://ilanbigio.com/blog/lyft-bikes.html
46•ibigio•7h ago•13 comments

FIPS dependencies and prebuilt binaries

https://www.docker.com/blog/fips-dependencies-and-prebuilt-binaries/
7•LaurentGoderre•48m ago•2 comments

It looks like the status/need-triage label was removed

https://github.com/google-gemini/gemini-cli/issues/16728
269•nickswalker•7h ago•68 comments