frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Ask HN: Anyone Using a Mac Studio for Local AI/LLM?

49•UmYeahNo•2d ago•31 comments

Ask HN: Opus 4.6 ignoring instructions, how to use 4.5 in Claude Code instead?

3•Chance-Device•9h ago•0 comments

Ask HN: Ideas for small ways to make the world a better place

21•jlmcgraw•1d ago•22 comments

Ask HN: Non AI-obsessed tech forums

35•nanocat•1d ago•28 comments

Ask HN: Who wants to be hired? (February 2026)

139•whoishiring•5d ago•527 comments

Ask HN: 10 months since the Llama-4 release: what happened to Meta AI?

45•Invictus0•1d ago•11 comments

Ask HN: Who is hiring? (February 2026)

313•whoishiring•5d ago•515 comments

LLMs are powerful, but enterprises are deterministic by nature

5•prateekdalal•18h ago•7 comments

Tell HN: Another round of Zendesk email spam

105•Philpax•3d ago•54 comments

AI Regex Scientist: A self-improving regex solver

7•PranoyP•1d ago•1 comments

Ask HN: Is Connecting via SSH Risky?

19•atrevbot•2d ago•37 comments

Ask HN: Has your whole engineering team gone big into AI coding? How's it going?

18•jchung•2d ago•14 comments

Ask HN: Non-profit, volunteers run org needs CRM. Is Odoo Community a good sol.?

2•netfortius•1d ago•1 comments

Ask HN: Is there anyone here who still uses slide rules?

123•blenderob•4d ago•122 comments

Kernighan on Programming

171•chrisjj•5d ago•62 comments

Ask HN: Mem0 stores memories, but doesn't learn user patterns

9•fliellerjulian•3d ago•6 comments

Ask HN: How does ChatGPT decide which websites to recommend?

5•nworley•2d ago•11 comments

Ask HN: Why LLM providers sell access instead of consulting services?

5•pera•1d ago•13 comments

Ask HN: Is it just me or are most businesses insane?

8•justenough•2d ago•7 comments

Ask HN: What is the most complicated Algorithm you came up with yourself?

3•meffmadd•1d ago•7 comments

Ask HN: Anyone Seeing YT ads related to chats on ChatGPT?

2•guhsnamih•2d ago•4 comments

We built a serverless GPU inference platform with predictable latency

5•QubridAI•2d ago•1 comments

Ask HN: Does a good "read it later" app exist?

8•buchanae•4d ago•18 comments

Ask HN: Does global decoupling from the USA signal comeback of the desktop app?

5•wewewedxfgdf•2d ago•3 comments

Ask HN: Have you been fired because of AI?

17•s-stude•4d ago•15 comments

Ask HN: Anyone have a "sovereign" solution for phone calls?

12•kldg•4d ago•1 comments

Ask HN: Cheap laptop for Linux without GUI (for writing)

15•locusofself•4d ago•16 comments

GitHub Actions Have "Major Outage"

53•graton•5d ago•17 comments

Ask HN: Has anybody moved their local community off of Facebook groups?

23•madsohm•5d ago•20 comments

Ask HN: OpenClaw users, what is your token spend?

14•8cvor6j844qw_d6•5d ago•6 comments
Open in hackernews

Qwen3-Max-Thinking Drops: 36T Tokens

4•SilasYee•1w ago
Alibaba has officially launched Qwen3-Max-Thinking, a trillion-parameter MoE flagship LLM pretrained on 36T tokens—double the corpus of Qwen 2.5—and it’s already matching or outperforming top-tier models like GPT-5.2-Thinking, Claude-Opus-4.5, and Gemini 3 Pro across 19 authoritative benchmarks. Its two core technical breakthroughs are what truly set it apart.

First, Adaptive Tool Calling: No manual prompts are needed—it autonomously invokes search engines, memory tools, and code interpreters based on task demands. This cuts down on hallucinations and boosts real-time problem-solving; for instance, coding tasks trigger automatic error correction loops, while research tasks combine search with context synthesis. Second, Test-Time Scaling (TTS): It outperforms standard parallel sampling by refining reasoning through iterative insights, with measurable jumps in key benchmarks—GPQA rose from 90.3 to 92.8, LiveCodeBench v6 hit 91.4 from 88.0, and IMO-AnswerBench climbed to 91.5 from 89.5.

Notably, its preview version even achieved 100% accuracy in tough math contests like AIME 25 and HMMT 25. The model runs smoothly on web/desktop demos, and its API is production-ready with adjustable thinking budgets (up to 80K tokens by default) to balance depth and speed. This isn’t just an incremental update—it’s a leap that closes the gap in reasoning and tool integration for real-world academic and engineering tasks.

Check it out: https://chat.qwen.ai/

Comments

imovie4•1w ago
if it was the llm u used to generate this i can't say i'm impressed
ChrisArchitect•1w ago
[dupe] Discussion: https://news.ycombinator.com/item?id=46766741