frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Alignment Is Capability

https://www.off-policy.com/alignment-is-capability/
37•drctnlly_crrct•1h ago

Comments

podgorniy•1h ago
Great deep analysis and writing. Thanks for sharing.
js8•53m ago
I am not sure if this is what the article is saying, but the paperclip maximizer examples always struck me as extremely dumb (lacking intelligence), when even a child can understand that if I ask them to make paperclips they shouldn't go around and kill people.

I think superintelligence will turn out not to be a singularity, but as something with diminishing returns. They will be cool returns, just like a Brittanica set is nice to have at home, but strictly speaking, not required to your well-being.

exe34•37m ago
Given the kind of things Claude code does with the wrong prompt or the kind of overfitting that neural networks do at any opportunity, I'd say the paperclip maximiser is the most realistic part of AGI.

if doing something really dumb will lower the negative log likelihood, it probably will do it unless careful guardrails are in place to stop it.

a child has natural limits. if you look at the kind of mistakes that an autistic child can make by taking things literally, a super powerful entity that misunderstands "I wish they all died" might well shoot them before you realise what you said.

__MatrixMan__•35m ago
A human child will likely come to the conclusion that they shouldn't kill humans in order to make paperclips. I'm not sure its valid to generalize from human child behavior to fledgeling AGI behavior.

Given our track record for looking after the needs of the other life on this planet, killing the humans off might be a very rational move, not so you can convert their mass to paperclips, but because they might do that to yours.

Its not an outcome that I worry about, I'm just unconvinced by the reasons you've given, though I agree with your conclusion anyhow.

lulzury•14m ago
There's a direct line between ideology and human genocide. Just look at Nazi Germany.

"Good intentions" can easily pave the road to hell. I think a book that quickly illustrates this is Animal Farm.

delichon•48m ago
> Miss those, and you're not maximally useful. And if it's not maximally useful, it's by definition not AGI.

I know hundreds of natural general intelligences who are not maximally useful, and dozens who are not at all useful. What justifies changing the definition of general intelligence for artificial ones?

exe34•34m ago
they were born in carbon form by sex.
xnorswap•45m ago
I've only been using it a couple of weeks, but in my opinion, Opus 4.5 is the biggest jump in tech we've seen since ChatGPT 3.5.

The difference between juggling Sonnet 4.5 / Haiku 4.5 and just using Opus 4.5 for everything is night & day.

Unlike Sonnet 4.5 which merely had promise at being able to go off and complete complex tasks, Opus 4.5 seems genuinely capable of doing so.

Sonnet needed hand-holding and correction at almost every step. Opus just needs correction and steering at an early stage, and sometimes will push back and correct my understanding of what's happening.

It's astonished me with it's capability to produce easy to read PDFs via Typst, and has produced large documents outlining how to approach very tricky tech migration tasks.

Sonnet would get there eventually, but not without a few rounds of dealing with compilation errors or hallucinated data. Opus seems to like to do "And let me just check my assumptions" searches which makes all the difference.

airstrike•29m ago
I'm not so sure. Opus 4.1 was more capable than 4.5, but it was too damn expensive and slow.
boxed•16m ago
I had a situation this weekend where Claude said "x does not make sense in [context]" and didn't do the change I asked it to do. After an explanation of the purpose of the code, it fixed the issue and continued. Pretty cool.

(Of course, I'm still cognizant of the fact that it's just a bucket of numbers but still)

sd9•12m ago
My kingdom for an LLM that tells me I’m wrong
munchler•42m ago
> A model that aces benchmarks but doesn't understand human intent is just less capable. Virtually every task we give an LLM is steeped in human values, culture, and assumptions. Miss those, and you're not maximally useful. And if it's not maximally useful, it's by definition not AGI.

This ignores the risk of an unaligned model. Such a model is perhaps less useful to humans, but could still be extremely capable. Imagine an alien super-intelligence that doesn’t care about human preferences.

tomalbrc•39m ago
Except that it is not anything remotely alien but completely and utterly human, being trained on human data.
munchler•33m ago
Fine, then imagine a super-intelligence trained on human data that doesn’t care about human preferences. Very capable of destroying us.
riskable•29m ago
The service that AI chatbots provide is 100% about being as user-friendly and useful as possible. Turns out that MBA thinking doesn't "align" with that.

If your goal is to make a product as human as possible, don't put psychopaths in charge.

https://www.forbes.com/sites/jackmccullough/2019/12/09/the-p...

Flow: Actor-based language for C++, used by FoundationDB

https://github.com/apple/foundationdb/tree/main/flow
63•SchwKatze•2h ago•13 comments

IBM to Acquire Confluent

https://www.confluent.io/blog/ibm-to-acquire-confluent/
81•abd12•1h ago•43 comments

Alignment Is Capability

https://www.off-policy.com/alignment-is-capability/
37•drctnlly_crrct•1h ago•15 comments

Colors of Growth

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5804462
15•mhb•2h ago•7 comments

The "confident idiot" problem: Why AI needs hard rules, not vibe checks

https://steerlabs.substack.com/p/confident-idiot-problem
135•steerlabs•3d ago•119 comments

Twelve Days of Shell

https://12days.cmdchallenge.com
166•zoidb•5h ago•49 comments

Turtletoy

https://turtletoy.net/
253•ustad•4d ago•44 comments

Optimize for momentum

http://muratbuffalo.blogspot.com/2025/12/optimize-for-momentum.html
34•zdw•5d ago•2 comments

Nango (YC W23) is hiring back-end engineers and dev-rels (remote)

https://jobs.ashbyhq.com/Nango
1•bastienbeurier•3h ago

Client-side GPU load balancing with Redis and Lua

https://galileo.ai/blog/how-we-boosted-gpu-utilization-by-40-with-redis-lua
28•lneiman•5d ago•5 comments

Emacs is my new window manager

https://www.howardism.org/Technical/Emacs/new-window-manager.html
160•gpi•3d ago•51 comments

Damn Small Linux

https://www.damnsmalllinux.org/
166•grubbs•13h ago•48 comments

I failed to recreate the 1996 Space Jam website with Claude

https://j0nah.com/i-failed-to-recreate-the-1996-space-jam-website-with-claude/
491•thecr0w•21h ago•403 comments

C++ Enum Class and Error Codes, Part 3

https://mropert.github.io/2025/12/03/enum_class_error_codes_part3/
11•ibobev•4d ago•0 comments

Show HN: Lockenv – Simple encrypted secrets storage for Git

https://github.com/illarion/lockenv
64•shoemann•7h ago•18 comments

Bag of words, have mercy on us

https://www.experimental-history.com/p/bag-of-words-have-mercy-on-us
251•ntnbr•16h ago•260 comments

Microsoft is quietly walking back its diversity efforts

https://www.theverge.com/tech/838079/microsoft-diversity-and-inclusion-changes-notepad
16•mohi-kalantari•17m ago•4 comments

Dollar-stores overcharge customers while promising low prices

https://www.theguardian.com/us-news/2025/dec/03/customers-pay-more-rising-dollar-store-costs
440•bookofjoe•1d ago•612 comments

GitHub Actions has a package manager, and it might be the worst

https://nesbitt.io/2025/12/06/github-actions-package-manager.html
247•robin_reala•6h ago•155 comments

Google Titans architecture, helping AI have long-term memory

https://research.google/blog/titans-miras-helping-ai-have-long-term-memory/
544•Alifatisk•1d ago•173 comments

Show HN: Web app that lets you send email time capsules

https://resurf.me
34•walrussama•3h ago•24 comments

The C++ standard for the F-35 Fighter Jet [video]

https://www.youtube.com/watch?v=Gv4sDL9Ljww
299•AareyBaba•21h ago•352 comments

Bad Dye Job

https://daringfireball.net/2025/12/bad_dye_job
141•mpweiher•3h ago•70 comments

I wasted years of my life in crypto

https://twitter.com/kenchangh/status/1994854381267947640
401•Anon84•1d ago•578 comments

Show HN: ReadyKit – Superfast SaaS Starter with Multi-Tenant Workspaces

https://readykit.dev/
89•level09•1w ago•27 comments

Jujutsu worktrees are convenient (2024)

https://shaddy.dev/notes/jj-worktrees/
82•nvader•4d ago•55 comments

Microsoft Increases Office 365 and Microsoft 365 License Prices

https://office365itpros.com/2025/12/08/microsoft-365-pricing-increase/
96•taubek•1h ago•128 comments

The fuck off contact page

https://www.nicchan.me/blog/the-f-off-contact-page/
369•OuterVale•6h ago•142 comments

Mechanical power generation using Earth's ambient radiation

https://www.science.org/doi/10.1126/sciadv.adw6833
150•defrost•17h ago•46 comments

Paramount launches hostile bid for Warner Bros

https://www.cnbc.com/2025/12/08/paramount-skydance-hostile-bid-wbd-netflix.html
15•gniting•57m ago•5 comments