frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

China's AI Industrial Policy

https://www.high-capacity.com/p/chinas-ai-industrial-policy
1•RetiredRichard•59s ago•0 comments

Automated Discovery of High-Performance GPU Kernels with OpenEvolve

https://huggingface.co/blog/codelion/openevolve-gpu-kernel-discovery
1•codelion•6m ago•0 comments

Ask HN: A weird vesting term in not-USA country

1•ayjay_t•10m ago•0 comments

Show HN: A Comprehensive List of Top AI Image Tools

https://aiex.me/top-ai-image-tools
1•zack119•21m ago•0 comments

A Loved One Was Diagnosed with Dementia. Now What?

https://www.nytimes.com/2025/06/25/well/family/dementia-diagnosis-advice.html
1•whack•33m ago•0 comments

A Children's Book in a Happiness Program for College Students

https://childrensbookforall.org/activities/knc-college
1•chbkall•39m ago•1 comments

Rádio Starthits

https://stream-176.zeno.fm/k0fka0zkmxhvv?zt=eyJhbGciOiJIUzI1NiJ9.eyJzdHJlYW0iOiJrMGZrYTB6a214aHZ2IiwiaG9zdCI6InN0cmVhbS0xNzYuemVuby5mbSIsInJ0dGwiOjUsImp0aSI6ImVwcTdHRThjU3JHZzRLbVVzWC1uS0EiLCJpYXQiOjE3NDc4OTQ3MDcsImV4cCI6MTc0Nzg5NDc2N30.PxDtwuCbiFKQok1xoXA8dJxlUos2LKSIBLAeyWIxyDs
1•Starthits•40m ago•0 comments

OmniGen2

https://github.com/VectorSpaceLab/OmniGen2
1•handfuloflight•40m ago•0 comments

Apple's Other 'F1 the Movie' In-App Promotions

https://www.macrumors.com/2025/06/27/f1-the-movie-now-playing-in-theaters/
1•Bogdanp•43m ago•1 comments

Turn your raw ideas into actionable specifications

https://news.ycombinator.com/ask
1•normadia•46m ago•1 comments

Flow Match Statements

https://github.com/facebook/flow/blob/main/tests/match_exhaustive/basic.js
2•cod1r•48m ago•0 comments

Engineer Unlocks Hidden Photo in Power Mac ROM After 27 Years

https://digitrendz.blog/newswire/technology/21664/engineer-unlocks-hidden-photo-in-power-mac-rom-after-27-years/
1•cyberwaj•49m ago•1 comments

Food structure plays key role in which gut hormones are released

https://www.imperial.ac.uk/news/265333/food-structure-plays-role-which-hormones/
1•gmays•50m ago•0 comments

Scientists achieve shortest hard X-ray pulses to date

https://phys.org/news/2025-06-scientists-shortest-hard-ray-pulses.html
1•PaulHoule•54m ago•0 comments

Ask HN: What's the hardest/most interesting thing you've used AI to code?

1•ozb•59m ago•0 comments

It's Known as 'The List'–and It's a Secret File of AI Geniuses

https://www.wsj.com/tech/meta-ai-recruiting-mark-zuckerberg-openai-018ed7fc
3•sonabinu•59m ago•2 comments

Facebook is starting to feed its AI with private, unpublished photos

https://www.theverge.com/meta/694685/meta-ai-camera-roll
3•pier25•1h ago•1 comments

10 Myths of Scalable Parallel Languages, Part 3: New vs. Extended

https://chapel-lang.org/blog/posts/10myths-part3/
3•matt_d•1h ago•0 comments

Visible

https://visible.vc/
1•handfuloflight•1h ago•0 comments

VMware perpetual license holder receives audit letter from Broadcom

https://arstechnica.com/information-technology/2025/06/vmware-perpetual-license-holder-receives-audit-letter-from-broadcom/
1•hilux•1h ago•1 comments

Cross-Compiling Common Lisp for Windows

https://www.fosskers.ca/en/blog/cl-windows
5•todsacerdoti•1h ago•0 comments

Jane Austen's Boldest Novel Is Also Her Least Understood

https://www.nytimes.com/2025/06/27/books/review/jane-austen-mansfield-park.html
3•lermontov•1h ago•1 comments

Guidelines for buildable and testable code examples

https://pigweed.dev/docs/contributing/docs/examples.html
3•kaycebasques•1h ago•0 comments

Go is an 80/20 language

https://blog.kowalczyk.info/article/d-2025-06-26/go-is-8020-language.html
4•todsacerdoti•1h ago•0 comments

Converging AI and HPC: Design and Optimization of a CGRA Architecture [pdf]

https://cogarchworkshop.org/assets/papers/paper_3.pdf
3•matt_d•1h ago•0 comments

(Experiment) Colocating agent instructions with eng docs

https://technicalwriting.dev/ai/agents/colocate.html
1•kaycebasques•1h ago•0 comments

Multi-Stage Programming with Splice Variables

https://tsung-ju.org/icfp25/
6•matt_d•1h ago•0 comments

We need a censorship-resistant truth protocol – I have the idea, not the skills

2•PowerQuestion•1h ago•1 comments

When cars outsmart their drivers

https://www.carsandhorsepower.com/featured/when-cars-outsmarted-their-drivers
2•Anumbia•1h ago•0 comments

Meta is offering multi-mn pay for AI researchers,but not $100M signing bonuses

https://techcrunch.com/2025/06/27/meta-is-offering-multi-million-pay-for-ai-researchers-but-not-100m-signing-bonuses/
2•pranay01•1h ago•0 comments
Open in hackernews

Normalizing Flows Are Capable Generative Models

https://machinelearning.apple.com/research/normalizing-flows
70•danboarder•4h ago

Comments

layer8•2h ago
Earlier discussion: https://news.ycombinator.com/item?id=44358535
jc4p•1h ago
i've been trying to keep up with this field (image generation) so here's quick notes I took:

Claude's Summary: "Normalizing flows aren't dead, they just needed modern techniques"

My Summary: "Transformers aren't just for text"

1. SOTA model for likelihood on ImageNet 64×64, first ever sub 3.2 (Bits Per Dimension) prev was 2.99 by a hybrid diffusion model

2. Autoregressive (transformers) approach, right now diffusion is the most popular in this space (it's much faster but a diff approach)

tl;dr of autoregressive vs diffusion (there's also other approaches)

Autoregression: step based, generate a little then more then more

Diffusion: generate a lot of noise then try to clean it up

The diffusion approach that is the baseline for sota is Flow Matching from Meta: https://arxiv.org/abs/2210.02747 -- lots of fun reading material if you throw both of these into an LLM and ask it to summarize the approaches!

godelski•1h ago
You have a few minor errors and I hope I can help out.

  > Diffusion: generate a lot of noise then try to clean it up
You could say this about Flows too. The history of them is shared with diffusion and goes back to the Whitening Transform. Flows work by a coordinate transform so we have an isomorphism where diffusion works through, for easier understanding, a hierarchical mixture of gaussians. Which is a lossy process (more confusing when we get into latent diffusion models, which are the primary type used). The goal of a Normalizing Flow is to turn your sampling distribution, which you don't have an explicit representation of, into a probability distribution (typically Normal Noise/Gaussian). So in effect, there are a lot of similarities here. I'd highly suggest learning about Flows if you want to better understand Diffusion Models.

  > The diffusion approach that is the baseline for sota is Flow Matching from Meta
To be clear, Flow Matching is a Normalizing Flow. Specifically, it is a Continuous and Conditional Normalizing Flow. If you want to get into the nitty gritty, Ricky has a really good tutorial on the stuff[0]

[0] https://arxiv.org/abs/2412.06264

jc4p•36m ago
thank you so much!!! i should’ve put that final sentence in my post!
godelski•21m ago
Happy to help and if you have any questions just ask, this is my jam
godelski•1h ago
As far as I'm aware, this is the largest Normalizing Flow that exists, and I think they undermined their work by not mentioning this...

Their ImageNet model (4_1024_8_8_0.05[0]) is ~820M while AFHQ is ~472M. Prior to that there is DenseFlow[1] and MaCow[2], which are both <200M parameters. For more comparison, that makes DenseFlow and MaCow smaller than iDDPM[3] (270M params) and ADM[4] (553M for 256 unconditional). And now, it isn't uncommon for modern diffusion models to have several billion parameters![5] (from this we get some numbers on ImageNet-256, which allows a direct comparison, making TarFlow closer to MaskDiT/2 and much smaller than SimpleDiffusion and VDM++, both of which are in billions. But note that this is 128 vs 256!)

Essentially, the argument here is that you can scale (Composable) Normalizing Flows just as well as diffusion models. There's a lot of extra benefits you get too in the latent space, but that's a much longer discussion. Honestly, the TarFlow method is simple and there's probably a lot of improvements that can be made. But don't take that as a knock on this paper! I actually really appreciated it and it really set out to show what they tried to show. The real thing is just no one trained flows at this scale before and this really needs to be highlighted.

The tldr: people have really just overlooked different model architectures

[0] Used a third party reproduction so might be different but their AFHQ-256 model matches at 472M params https://github.com/encoreus/GS-Jacobi_for_TarFlow

[1] https://arxiv.org/abs/2106.04627

[2] https://arxiv.org/abs/1902.04208

[3] https://arxiv.org/abs/2102.09672

[4] https://arxiv.org/abs/2105.05233

[5] https://arxiv.org/abs/2401.11605

[Side note] Hey, if the TarFlow team is hiring, I'd love to work with you guys