frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

CEO pay and stock buybacks have soared at the 100 largest low-wage corporations

https://ips-dc.org/report-executive-excess-2025/
25•hhs•20m ago•0 comments

Crimes with Python's Pattern Matching (2022)

https://www.hillelwayne.com/post/python-abc/
66•agluszak•2h ago•9 comments

How does the US use water?

https://www.construction-physics.com/p/how-does-the-us-use-water
79•juliangamble•10h ago•50 comments

AI tooling must be disclosed for contributions

https://github.com/ghostty-org/ghostty/pull/8289
400•freetonik•3h ago•189 comments

DeepSeek-v3.1 Release

https://api-docs.deepseek.com/news/news250821
183•wertyk•3h ago•36 comments

An interactive guide to SVG paths

https://www.joshwcomeau.com/svg/interactive-guide-to-paths/
143•joshwcomeau•3d ago•14 comments

A Decoder Ring for AI Job Titles

https://www.dbreunig.com/2025/08/21/a-guide-to-ai-titles.html
30•dbreunig•3h ago•26 comments

Beyond sensor data: Foundation models of behavioral data from wearables

https://arxiv.org/abs/2507.00191
182•brandonb•7h ago•37 comments

Text.ai (YC X25) Is Hiring Founding Full-Stack Engineer

https://www.ycombinator.com/companies/text-ai/jobs/OJBr0v2-founding-full-stack-engineer
1•RushiSushi•1h ago

Miles from the ocean, there's diving beneath the streets of Budapest

https://www.cnn.com/2025/08/18/travel/budapest-diving-molnar-janos-cave
75•thm•3d ago•9 comments

My other email client is a daemon

https://feyor.sh/blog/my-other-email-client-is-a-mail-daemon/
60•aebtebeten•13h ago•14 comments

Weaponizing image scaling against production AI systems

https://blog.trailofbits.com/2025/08/21/weaponizing-image-scaling-against-production-ai-systems/
295•tatersolid•10h ago•75 comments

Beyond the Logo: How We're Weaving Full Images Inside QR Codes

https://blog.nitroqr.com/beyond-the-logo-how-were-weaving-full-images-inside-qr-codes
11•bhasinanant•3d ago•5 comments

How well does the money laundering control system work?

https://www.journals.uchicago.edu/doi/10.1086/735665
150•PaulHoule•9h ago•146 comments

D4D4

https://www.nmichaels.org/musings/d4d4/d4d4/
426•csense•4d ago•49 comments

Using Podman, Compose and BuildKit

https://emersion.fr/blog/2025/using-podman-compose-and-buildkit/
229•LaSombra•11h ago•65 comments

Building AI products in the probabilistic era

https://giansegato.com/essays/probabilistic-era
74•sdan•3h ago•43 comments

The power of two random choices (2012)

https://brooker.co.za/blog/2012/01/17/two-random.html
33•signa11•3d ago•3 comments

Show HN: OS X Mavericks Forever

https://mavericksforever.com/
262•Wowfunhappy•3d ago•110 comments

The contrarian physics podcast subculture

https://timothynguyen.org/2025/08/21/physics-grifters-eric-weinstein-sabine-hossenfelder-and-a-crisis-of-credibility/
124•Emerson1•5h ago•151 comments

The Core of Rust

https://jyn.dev/the-core-of-rust/
123•zdw•5h ago•102 comments

Launch HN: Skope (YC S25) – Outcome-based pricing for software products

32•benjsm•7h ago•27 comments

Mark Zuckerberg freezes AI hiring amid bubble fears

https://www.telegraph.co.uk/business/2025/08/21/zuckerberg-freezes-ai-hiring-amid-bubble-fears/
630•pera•11h ago•629 comments

Privately-Owned Rail Cars

https://www.amtrak.com/privately-owned-rail-cars
60•jasoncartwright•9h ago•91 comments

Adding my home electricity uptime to status.href.cat

https://aggressivelyparaphrasing.me/2025/08/21/adding-my-home-electricity-uptime-to-status-href-cat/
34•todsacerdoti•6h ago•29 comments

Uv format: Code Formatting Comes to uv (experimentally)

https://pydevtools.com/blog/uv-format-code-formatting-comes-to-uv-experimentally/
65•tanelpoder•1h ago•65 comments

In the long run, LLMs make us dumber

https://desunit.com/blog/in-the-long-run-llms-make-us-dumber/
59•speckx•3h ago•43 comments

Show HN: ChartDB Cloud – Visualize and Share Database Diagrams

https://app.chartdb.io
77•Jonathanfishner•9h ago•11 comments

Libre-Chip Awarded NLnet Grant to Prototype a CPU Isn't Vulnerable to Spectre

https://www.phoronix.com/news/Libre-Chip-NLnet-Grant
11•Bender•1h ago•0 comments

I forced every engineer to take sales calls and they rewrote our platform

https://old.reddit.com/r/Entrepreneur/comments/1mw5yfg/forced_every_engineer_to_take_sales_calls_they/
200•bilsbie•6h ago•144 comments
Open in hackernews

DeepSeek-v3.1 Release

https://api-docs.deepseek.com/news/news250821
182•wertyk•3h ago

Comments

hodgehog11•2h ago
For reference, here is the terminal-bench leaderboard:

https://www.tbench.ai/leaderboard

Looks like it doesn't get close to GPT-5, Claude 4, or GLM-4.5, but still does reasonably well compared to other open weight models. Benchmarks are rarely the full story though, so time will tell how good it is in practice.

seunosewa•2h ago
The DeepSeek R1 in that list is the old model that's been replaced. Update: Understood.
yorwba•1h ago
Yes, and 31.3% is given in the announcement as the performance of the new v3.1, which would put it in sixteenth place.
coliveira•1h ago
My personal experience is that it produces high quality results.
amrrs•1h ago
Any example or prompt you use to make this statment?
imachine1980_•1h ago
I remember asking for quotes about the Spanish conquest of South America because I couldn't remember who said a specific thing. The GPT model started hallucinating quotes on the topic, while DeepSeek responded with, "I don't know a quote about that specific topic, but you might mean this other thing." or something like that then cited a real quote in the same topic, after acknowledging that it wasn't able to find the one I had read in an old book. i don't use it for coding, but for things that are more unique i feel is more precise.
mycall•10m ago
I wonder if Conway's law is at all responsible for that, in the similarity it is based on; regional trained data which has concept biases which it sends back in response.
YetAnotherNick•1h ago
Depends on the agent. Rank 5 and 15 are claude 4 sonnet, and this stands close to 15th.
guluarte•1h ago
tbh companies like anthopic, openai, create custom agents for specific benchmarks
bazmattaz•1h ago
Do you have a source for this? I’m intrigued
guluarte•1h ago
https://www-cdn.anthropic.com/07b2a3f9902ee19fe39a36ca638e5a... "we iteratively refine prompting by analyzing failure cases and developing prompts to address them."
segmondy•47m ago
garbage benchmark, inconsistent mix of "agent tools" and models. if you wanted to present a meaningful benchmark, the agent tools will stay the same and then we can really compare the models.

there are plenty of other benchmarks that disagree with these, with that said. from my experience most of these benchmarks are trash. use the model yourself, apply your own set of problems and see how well it fairs.

tonyhart7•29m ago
Yeah but the pricing is insane, I don't care about SOTA if its not break my bank
seunosewa•2h ago
It's a hybrid reasoning model. It's good with tool calls and doesn't think too much about everything, but it regularly uses outdated tool formats randomly instead of the standard JSON format. I guess the V3 training set has a lot of those.
ivape•1h ago
What formats? I thought the very schema of json is what allows these LLMs to enforce structured outputs at the decoder level? I guess you can do it with any format, but why stray from json?
seunosewa•1h ago
Sometimes it will randomly generate something like this in the body of the text: ``` <tool_call>executeshell <arg_key>command</arg_key> <arg_value>echo "" >> novels/AI_Voodoo_Romance/chapter-1-a-new-dawn.txt</arg_value> </tool_call> ```

or this: ``` <|toolcallsbegin|><|toolcallbegin|>executeshell<|toolsep|>{"command": "pwd && ls -la"}<|toolcallend|><|toolcallsend|> ```

Prompting it to use the right format doesn't seem to work. Claude, Gemini, GPT5, and GLM 4.5, don't do that. To accomodate DeepSeek, the tiny agent that I'm building will have to support all the weird formats.

darrinm•10m ago
Did you try the strict (beta) function calling? https://api-docs.deepseek.com/guides/function_calling
esafak•2h ago
It seems behind Qwen3 235B 2507 Reasoning (which I like) and gpt-oss-120B: https://artificialanalysis.ai/models/deepseek-v3-1-reasoning

Pricing: https://openrouter.ai/deepseek/deepseek-chat-v3.1

bigyabai•2h ago
Those Qwen3 2507 models are the local creme-de-la-creme right now. If you've got any sort of GPU and ~32gb of RAM to play with, the A3B one is great for pair-programming tasks.
pdimitar•1h ago
Do you happen to know if it can be run via an eGPU enclosure with f.ex. RTX 5090 inside, under Linux?

I'm considering buying a Linux workstation lately and I want it full AMD. But if I can just plug an NVIDIA card via an eGPU card for self-hosting LLMs then that would be amazing.

gunalx•1h ago
You would still need drivers and all the stuff difficult with nvidia in linux with a egpu. (Its not nessecarily terrible just suboptimal) Rather just add the second GPU in the Workstation, or just run the llm in your AMD GPU.
pdimitar•1h ago
Oh, we can run LLMs efficiently with AMD GPUs now? Pretty cool, I haven't been following, thank you.
DarkFuture•42m ago
I've been running LLM models on my Radeon 7600 XT 16GB for past 2-3 months without issues (Windows 11). I've been using llama.cpp only. The only thing from AMD I installed (apart from latest Radeon drivers) is the "AMD HIP SDK" (very straight forward installer). After unzipping (the zip from GitHub releases page must contain hip-radeon in the name) all I do is this:

llama-server.exe -ngl 99 -m Qwen3-14B-Q6_K.gguf

And then connect to llamacpp via browser to localhost:8080 for the WebUI (its basic but does the job, screenshots can be found on Google). You can connect more advanced interfaces to it because llama.cpp actually has OpenAI-compatible API.

bigyabai•1h ago
Sure, though you'll be bottlenecked by the interconnect speed if you're tiling between system memory and the dGPU memory. That shouldn't be an issue for the 30B model, but would definitely be an issue for the 480B-sized models.
oktoberpaard•1h ago
I’m running Ollama on 2 eGPUs over Thunderbolt. Works well for me. You’re still dealing with an NVDIA device, of course. The connection type is not going to change that hassle.
pdimitar•1h ago
Thank you for the validation. As much as I don't like NVIDIA's shenanigans on Linux, having a local LLM is very tempting and I might put my ideological problems to rest over it.

Though I have to ask: why two eGPUs? Is the LLM software smart enough to be able to use any combination of GPUs you point it at?

arcanemachiner•3m ago
Yes, Ollama is very plug-and-play when it comes to multi GPU.

llama.cpp probably is too, but I haven't tried it with a bigger model yet.

tomr75•1h ago
With qwen code?
decide1000•1h ago
I use it on a 24gb gpu Tesla P40. Very happy with the result.
hkt•56m ago
Out of interest, roughly how many tokens per second do you get on that?
edude03•49m ago
Like 4. Definitely single digit. The P40s are slow af
abtinf•1h ago
Unrelated, but it would really be nice to have a chart breaking down Price Per Token Per Second for various model, prompt, and hardware combinations.
theuurviv467456•48m ago
Sweet. I wish there guys weren't bound by the idiotic "nationalist" () bans so that they could do their work unrestricted.

Only idiots who are completely drowned in US's dark propaganda would think this is about anything but keeping China down.

simianparrot•46m ago
As if the CCP needs help keeping its own people down. Please.
tonyhart7•18m ago
Every country acting in its own best interest, US is not unique in this regard

wait until you find out that China also acting the same way toward the rest of the world (surprise pikachu face)

danielhanchen•3m ago
For local runs, I made some GGUFs! You need around RAM + VRAM >= 250GB for good perf for dynamic 2bit (2bit MoE, 6-8bit rest) - can also do SSD offloading but it'll be slow.

./llama.cpp/llama-cli -hf unsloth/DeepSeek-V3.1-GGUF:UD-Q2_K_XL -ngl 99 --jinja -ot ".ffn_.*_exps.=CPU"

More details on running + optimal params here: https://docs.unsloth.ai/basics/deepseek-v3.1