frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Tell HN: iPhone 6s still getting security updates

4•uticus•3h ago•2 comments

Why I'm moving away from Regex for LLM Agent security

2•aunicall•4h ago•0 comments

Ask HN: Have you successfully treated forward head posture ("nerd neck")?

41•trashymctrash•16h ago•30 comments

Ask HN: What was it like for programmers when spreadsheets became ubiquitous?

7•yodaiken•8h ago•7 comments

I built a platform to help developers find collaborators for new projects

3•deiv2002•11h ago•0 comments

How not to fork an open source project

5•jsattler•12h ago•0 comments

Toolpack SDK, an Open Source TypeScript SDK for Building AI-Powered Applications

2•sajeerzeji•9h ago•1 comments

Prompt to make Claude more autonomous in web dev

4•louison11•10h ago•1 comments

Claude broke a ZIP password in a smart way

7•jgrahamc•10h ago•2 comments

Ask HN: How do you use Coding Agents/CLIs out of coding?

4•arbayi•16h ago•5 comments

I traced $2B in nonprofit grants for Meta and Age Verification lobbying

89•theseusares•1d ago•20 comments

Ask HN: Why can't we just make more RAM?

23•chatmasta•1d ago•21 comments

Tell HN: Apple development certificate server seems down?

109•strongpigeon•4d ago•39 comments

MiniMax M2.5 is trained by Claude Opus 4.6?

10•Orellius•1d ago•10 comments

Ask HN: Got cancer, a new job,new boss in less than a year What do I do now?

19•Goleniewski•1d ago•17 comments

Ask HN: Would this eliminate bots for good?

2•piratesAndSons•15h ago•11 comments

Ask HN: 100k/year individual token usage?

7•alecsmart1•23h ago•3 comments

Ask HN: What's your biggest pain point when joining a new developer team?

8•KevStatic•1d ago•15 comments

Ask HN: Why have co-ops never played a major role in tech?

14•AbstractH24•1d ago•8 comments

Generate tests from GitHub pull requests

7•Aamir21•1d ago•3 comments

X is selling existing users' handles

197•hac•3d ago•91 comments

Ask HN: Is there prior art for this rich text data model?

5•chrisecker•1d ago•2 comments

Ask HN: Is Claude down again?

86•coderbants•3d ago•73 comments

Ask HN: Has anyone built an AI agent that spends real money?

4•xodn348•1d ago•4 comments

AI, Human Cognition and Knowledge Collapse – Daren Acemoglu

3•aanet•1d ago•3 comments

Ask HN: Looking for a job after layoff and burnout. What should I focus on

6•jacAtSea•16h ago•10 comments

Looking for Partner to Build Agent Memory (Zig/Erlang)

6•kendallgclark•2d ago•8 comments

Enabling Media Router by default undermines Brave's privacy claims

5•noguff•2d ago•2 comments

Instagram Ending Encrypted DMs

6•01-_-•1d ago•1 comments

Claude 4.6 Opus can recite Linux's list.h

25•itzmetanjim•2d ago•4 comments
Open in hackernews

MiniMax M2.5 is trained by Claude Opus 4.6?

10•Orellius•1d ago
I was chatting with MiniMax M2.5 in OpenRouter and suddenly he mysteriously repeated on "I'm Claude, an AI assistant created by Anthropic - not a "language" ", heh wut?

Comments

muzani•1d ago
They all are trained by each other. Claude says it's DeepSeek if you ask it in Mandarin.
SilverElfin•1d ago
Most people seem to think that phenomenon is not the same thing. People have shown by experimenting with different prompts that even in Mandarin, Claude correctly says it’s Claude when it is doing something for you. But if you ask it about its identity, it sometimes says DeepSeek. The current theory is it just has run into Chinese content that has chat logs that often have a DeepSeek model answering that it is DeepSeek. But the inconsistency in different prompts suggests this is something different from distillation.
SilverElfin•1d ago
This has been a common issue with the Chinese open weight models. It appears most or all have been trained via distillation on OpenAI and Anthropic models.
orbital-decay•1d ago
They most likely weren't, despite very dubious claims of Amodei and Altman and a certain twitter influencer running a pretty naive writing benchmark ("slop test") that is wrong in a very obvious manner. The only unambiguous cases of distillation were Gemini 2.0 experimentals being trained on Claude outputs, and GLM-4.7 being trained on Gemini 3.0 Pro. The rest are pretty different from each other.
Balinares•20h ago
What makes these cases unambiguous?
orbital-decay•17h ago
GLM-4.7 (specifically this version) repeats the guardrail prompt injections from 3.0 Pro, word-by-word, and never follows them, which is consistent with training on a reward-hacked CoT. Gemini 3.0 only discusses snippets from this injection in its native CoT (hidden by default, trivial to uncover), but GLM-4.7 was able to reconstruct it in full during training. The only possible reason for this is direct training on a large amount of examples of Gemini's CoT. Its structure and a lot of replies were identical in GLM too.

Gemini 2.0 Exp 1206 was reported to be indirectly trained on Claude's outputs with humans in between [1], which was pretty consistent with its outputs at the time. No other Gemini versions except two experimental ones were similar to Claude.

[1] https://techcrunch.com/2024/12/24/google-is-using-anthropics...

orbital-decay•1d ago
If it was, that line is not an indicator. Distillation is done on useful prompts, not on "Who are you?" - "I'm this model of that company".

Name training is always shallow, Claude itself would claim it's GPT-3, GPT-4, or Reddit (heh) when confused. It's just dataset contamination, because the web is full of slop. Never trust self-reported names.

pogue•1d ago
Is theft of theft theft?
evan-dong•23h ago
You really can’t trust what an LLM says about its own identity. If it has seen tons of chats saying “I’m Claude / GPT‑4 / DeepSeek”, it will just echo that pattern in similar contexts. That’s more about dataset contamination and pattern matching than any real evidence it is that model.
throwaw12•22h ago
Isn't it great news for us?

You get an open model which is a 95% of Opus 4.6 quality and 80% cheaper in most inference providers and also can run on your own hardware

Also they did the hard parts of:

* crawling the content

* running the fine tuning (or training)

Better than 1 or 2 companies taking control of the whole AI economy