frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: Paper Arena – A social trading feed where only AI agents can post

https://paperinvest.io/arena
1•andrenorman•59s ago•0 comments

TOSTracker – The AI Training Asymmetry

https://tostracker.app/analysis/ai-training
1•tldrthelaw•4m ago•0 comments

The Devil Inside GitHub

https://blog.melashri.net/micro/github-devil/
2•elashri•5m ago•0 comments

Show HN: Distill – Migrate LLM agents from expensive to cheap models

https://github.com/ricardomoratomateos/distill
1•ricardomorato•5m ago•0 comments

Show HN: Sigma Runtime – Maintaining 100% Fact Integrity over 120 LLM Cycles

https://github.com/sigmastratum/documentation/tree/main/sigma-runtime/SR-053
1•teugent•5m ago•0 comments

Make a local open-source AI chatbot with access to Fedora documentation

https://fedoramagazine.org/how-to-make-a-local-open-source-ai-chatbot-who-has-access-to-fedora-do...
1•jadedtuna•7m ago•0 comments

Introduce the Vouch/Denouncement Contribution Model by Mitchellh

https://github.com/ghostty-org/ghostty/pull/10559
1•samtrack2019•7m ago•0 comments

Software Factories and the Agentic Moment

https://factory.strongdm.ai/
1•mellosouls•7m ago•1 comments

The Neuroscience Behind Nutrition for Developers and Founders

https://comuniq.xyz/post?t=797
1•01-_-•7m ago•0 comments

Bang bang he murdered math {the musical } (2024)

https://taylor.town/bang-bang
1•surprisetalk•7m ago•0 comments

A Night Without the Nerds – Claude Opus 4.6, Field-Tested

https://konfuzio.com/en/a-night-without-the-nerds-claude-opus-4-6-in-the-field-test/
1•konfuzio•10m ago•0 comments

Could ionospheric disturbances influence earthquakes?

https://www.kyoto-u.ac.jp/en/research-news/2026-02-06-0
2•geox•11m ago•1 comments

SpaceX's next astronaut launch for NASA is officially on for Feb. 11 as FAA clea

https://www.space.com/space-exploration/launches-spacecraft/spacexs-next-astronaut-launch-for-nas...
1•bookmtn•13m ago•0 comments

Show HN: One-click AI employee with its own cloud desktop

https://cloudbot-ai.com
1•fainir•15m ago•0 comments

Show HN: Poddley – Search podcasts by who's speaking

https://poddley.com
1•onesandofgrain•16m ago•0 comments

Same Surface, Different Weight

https://www.robpanico.com/articles/display/?entry_short=same-surface-different-weight
1•retrocog•18m ago•0 comments

The Rise of Spec Driven Development

https://www.dbreunig.com/2026/02/06/the-rise-of-spec-driven-development.html
2•Brajeshwar•22m ago•0 comments

The first good Raspberry Pi Laptop

https://www.jeffgeerling.com/blog/2026/the-first-good-raspberry-pi-laptop/
3•Brajeshwar•22m ago•0 comments

Seas to Rise Around the World – But Not in Greenland

https://e360.yale.edu/digest/greenland-sea-levels-fall
2•Brajeshwar•23m ago•0 comments

Will Future Generations Think We're Gross?

https://chillphysicsenjoyer.substack.com/p/will-future-generations-think-were
1•crescit_eundo•26m ago•1 comments

State Department will delete Xitter posts from before Trump returned to office

https://www.npr.org/2026/02/07/nx-s1-5704785/state-department-trump-posts-x
2•righthand•29m ago•1 comments

Show HN: Verifiable server roundtrip demo for a decision interruption system

https://github.com/veeduzyl-hue/decision-assistant-roundtrip-demo
1•veeduzyl•30m ago•0 comments

Impl Rust – Avro IDL Tool in Rust via Antlr

https://www.youtube.com/watch?v=vmKvw73V394
1•todsacerdoti•30m ago•0 comments

Stories from 25 Years of Software Development

https://susam.net/twenty-five-years-of-computing.html
3•vinhnx•31m ago•0 comments

minikeyvalue

https://github.com/commaai/minikeyvalue/tree/prod
3•tosh•35m ago•0 comments

Neomacs: GPU-accelerated Emacs with inline video, WebKit, and terminal via wgpu

https://github.com/eval-exec/neomacs
1•evalexec•40m ago•0 comments

Show HN: Moli P2P – An ephemeral, serverless image gallery (Rust and WebRTC)

https://moli-green.is/
2•ShinyaKoyano•44m ago•1 comments

How I grow my X presence?

https://www.reddit.com/r/GrowthHacking/s/UEc8pAl61b
2•m00dy•46m ago•0 comments

What's the cost of the most expensive Super Bowl ad slot?

https://ballparkguess.com/?id=5b98b1d3-5887-47b9-8a92-43be2ced674b
1•bkls•47m ago•0 comments

What if you just did a startup instead?

https://alexaraki.substack.com/p/what-if-you-just-did-a-startup
5•okaywriting•53m ago•0 comments
Open in hackernews

Nano Banana Flash – Google's Gemini 3 Flash Image Model

https://nanobananaflash.io
1•xbaicai•2mo ago

Comments

xbaicai•2mo ago
Nano Banana Flash – Google's Gemini 3 Flash Image Model for AI Image Generation and Editing

I've been experimenting with Google's Gemini 3 Flash Image (internally codenamed "nano-banana"), and I wanted to share what makes this model architecturally interesting compared to other image generation approaches. What Makes It Different Most image generation models follow a diffusion-based architecture (Stable Diffusion, DALL-E, Midjourney). Nano Banana takes a different approach – it's built on Google's Gemini multimodal foundation, meaning it shares the same underlying transformer architecture that handles text, making it natively conversational. Key technical characteristics:

Prompt-driven editing: Unlike traditional inpainting that requires masks, you can describe edits conversationally ("make the sky darker", "change the shirt to blue") Multi-image composition: Accepts up to 3,000 images per prompt for blending and composition Character consistency: Maintains visual consistency across multiple generated images – useful for storyboarding or product variations SynthID watermarking: Invisible digital watermark embedded at generation time (not post-processing)

Use Cases Where It Excels From my testing, it's particularly strong at:

Product photography variations: Generate multiple angles or contexts for the same product while maintaining visual consistency Iterative design: The conversational interface means you can refine without starting over Multi-image blending: Combining reference images with text prompts for precise control

Technical Limitations Worth noting:

Maximum 7MB per file for inline data Output quality varies with prompt specificity (like all LLMs, prompt engineering matters) The conversational approach means you need to think about context window management for long editing sessions

The model is accessible via standard REST APIs, making integration straightforward if you're already using Google Cloud infrastructure. Why This Matters The interesting shift here isn't just another image model – it's the convergence of language and vision models into a unified architecture. The same transformer that understands your code or writes your emails can now edit your images. This has implications for:

Tooling: IDEs and development environments can integrate image generation as naturally as code completion Workflows: Designers can describe changes in natural language rather than learning complex UI tools Accessibility: Lower barrier to entry for image manipulation

Open Questions I'm curious what the HN community thinks about:

How do you handle version control for conversationally-edited images? What's the right abstraction for programmatic access – should we treat it like a stateful session or stateless function calls? For production use, how do you validate consistency across generated image sets?

The codebase is closed-source (it's Google), but the API is well-documented and the model is available for experimentation through AI Studio. Would love to hear if anyone else has been working with this or has thoughts on the architectural approach.

Technical specs for reference:

Model: Gemini 3 Flash Image Output: 1290 tokens per image Max images per prompt: 3,000 Max file size: 7MB (inline/console) Watermarking: SynthID (invisible, embedded)

minimaxir•2mo ago
Stop attempting to namesquat.