frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Cloudflare Global Network experiencing issues

https://www.cloudflarestatus.com/?t=1
1924•imdsm•5h ago•1288 comments

Gemini 3 Pro Preview Live in AI Studio

https://aistudio.google.com/prompts/new_chat?model=gemini-3-pro-preview
257•preek•1h ago•96 comments

Gemini 3 for developers: New reasoning, agentic capabilities

https://blog.google/technology/developers/gemini-3-developers/
130•janpio•36m ago•16 comments

Google Antigravity, a New Era in AI-Assisted Software Development

https://antigravity.google/blog/introducing-google-antigravity
54•meetpateltech•34m ago•9 comments

Gemini 3

https://blog.google/products/gemini/gemini-3/
140•meetpateltech•40m ago•24 comments

Nearly all UK drivers say headlights are too bright

https://www.bbc.com/news/articles/c1j8ewy1p86o
342•YeGoblynQueenne•2h ago•323 comments

Short Little Difficult Books

https://countercraft.substack.com/p/short-little-difficult-books
64•crescit_eundo•2h ago•24 comments

How Quake.exe got its TCP/IP stack

https://fabiensanglard.net/quake_chunnel/index.html
324•billiob•8h ago•60 comments

Do Not Put Your Site Behind Cloudflare If You Don't Need To

https://huijzer.xyz/posts/123/do-not-put-your-site-behind-cloudflare-if-you-dont
262•huijzer•3h ago•197 comments

Google Antigravity

https://antigravity.google/
94•Fysi•53m ago•60 comments

A Day at Hetzner Online in the Falkenstein Data Center

https://www.igorslab.de/en/a-day-at-hetzner-online-in-the-falkenstein-data-center-insights-into-s...
5•speckx•22m ago•0 comments

Experiment: Making TypeScript Immutable-by-Default

https://evanhahn.com/typescript-immutability-experiment/
45•ingve•2h ago•32 comments

Gemini 3 Pro Model Card

https://pixeldrain.com/u/hwgaNKeH
357•Topfi•5h ago•232 comments

Show HN: Optimizing LiteLLM with Rust – When Expectations Meet Reality

https://github.com/neul-labs/fast-litellm
3•ticktockten•8m ago•0 comments

The Miracle of Wörgl

https://scf.green/story-of-worgl-and-others/
87•simonebrunozzi•5h ago•45 comments

Ruby 4.0.0 Preview2 Released

https://www.ruby-lang.org/en/news/2025/11/17/ruby-4-0-0-preview2-released/
129•pansa2•3h ago•37 comments

Mathematics and Computation (2019) [pdf]

https://www.math.ias.edu/files/Book-online-Aug0619.pdf
35•nill0•4h ago•6 comments

GoSign Desktop RCE flaws affecting users in Italy

https://www.ush.it/2025/11/14/multiple-vulnerabilities-gosign-desktop-remote-code-execution/
41•ascii•4h ago•17 comments

5 Things to Try with Gemini 3 Pro in Gemini CLI

https://developers.googleblog.com/en/5-things-to-try-with-gemini-3-pro-in-gemini-cli/
3•keithba•31m ago•0 comments

Beauty in/of mathematics: tessellations and their formulas

https://www.tandfonline.com/doi/full/10.1080/00036811.2025.2510472
9•QueensGambit•5d ago•0 comments

Google Brings Gemini 3 AI Model to Search and AI Mode

https://blog.google/products/search/gemini-3-search-ai-mode/
3•CrypticShift•32m ago•0 comments

How many video games include a marriage proposal? At least one

https://32bits.substack.com/p/under-the-microscope-ncaa-basketball
299•bbayles•5d ago•70 comments

Show HN: I built a synth for my daughter

https://bitsnpieces.dev/posts/a-synth-for-my-daughter/
1229•random_moonwalk•6d ago•205 comments

Langfuse (YC W23) Hiring OSS Support Engineers in Berlin and SF

https://jobs.ashbyhq.com/langfuse/5ff18d4d-9066-4c67-8ecc-ffc0e295fee6
1•clemo_ra•9h ago

The Uselessness of "Fast" and "Slow" in Programming

https://jerf.org/iri/post/2025/the_uselessness_of_fast/
89•zdw•6d ago•47 comments

I've Wanted to Play That 'Killer Shark' Arcade Game Briefly Seen in 'Jaws'

https://www.remindmagazine.com/article/15694/jaws-arcade-video-game-killer-shark-atari-sega-elect...
12•speckx•3d ago•4 comments

Multiple Digital Ocean services down

https://status.digitalocean.com/incidents/lgt5xs2843rx
95•inanothertime•3h ago•33 comments

Azure hit by 15 Tbps DDoS attack using 500k IP addresses

https://www.bleepingcomputer.com/news/microsoft/microsoft-aisuru-botnet-used-500-000-ips-in-15-tb...
440•speckx•23h ago•283 comments

Ditch your (mut)ex, you deserve better

https://chrispenner.ca/posts/mutexes
111•commandersaki•6d ago•127 comments

The surprising benefits of giving up

https://nautil.us/the-surprising-benefits-of-giving-up-1248362/
156•jnord•11h ago•122 comments
Open in hackernews

Gemini 3 Pro Preview Live in AI Studio

https://aistudio.google.com/prompts/new_chat?model=gemini-3-pro-preview
250•preek•1h ago

Comments

nilsingwersen•1h ago
Feeling great to see something confidential
RobinL•1h ago
- Anyone have any idea why it says 'confidential'?

- Anyone actually able to use it? I get 'You've reached your rate limit. Please try again later'. (That said, I don't have a paid plan, but I've always had pretty much unlimited access to 2.5 pro)

[Edit: working for me now in ai studio]

sd9•1h ago
How long does it typically take after this to become available on https://gemini.google.com/app ?

I would like to try the model, wondering if it's worth setting up billing or waiting. At the moment trying to use it in AI Studio (on the Free tier) just gives me "Failed to generate content, quota exceeded: you have reached the limit of requests today for this model. Please try again tomorrow."

Squarex•1h ago
Today I guess. They were not releasing the preview models this time and it seems the want to synchronize the release.
mpeg•1h ago
Allegedly it's already available in stealth mode if you choose the "canvas" tool and 2.5. I don't know how true that is, but it is indeed pumping out some really impressive one shot code

Edit: Now that I have access to Gemini 3 preview, I've compared the results of the same one shot prompts on the gemini app's 2.5 canvas vs 3 AI studio and they're very similar. I think the rumor of a stealth launch might be true.

sd9•59m ago
Thanks for the hint about Canvas/2.5. I have access to 3.0 in AI Studio now, and I agree the results are very similar.
csomar•1h ago
It's already available. I asked it "how smart are you really?" and it gave me the same ai garbage template that's now very common on blog posts: https://gist.githubusercontent.com/omarabid/a7e564f09401a64e...
magicalhippo•43m ago
> https://gemini.google.com/app

How come I can't even see prices without logging in... they doing regional pricing?

mil22•1h ago
It's available to be selected, but the quota does not seem to have been enabled just yet.

"Failed to generate content, quota exceeded: you have reached the limit of requests today for this model. Please try again tomorrow."

"You've reached your rate limit. Please try again later."

Update: as of 3:33 PM UTC, Tuesday, November 18, 2025, it seems to be enabled.

misiti3780•1h ago
seeing the same issue.
sottol•1h ago
you can bring your google api key to try it out, and google used to give $300 free when signing up for billing and creating a key.

when i signed up for billing via cloud console and entered my credit card, i got $300 "free credits".

i haven't thrown a difficult problem at gemini 3 pro it yet, but i'm sure i got to see it in some of the A/B tests in aistudio for a while. i could not tell which model was clearly better, one was always more succinct and i liked its "style" but they usually offered about the same solution.

lousken•1h ago
I hope some users will switch from cerebras to free up those resources
sarreph•1h ago
Looks to be available in Vertex.

I reckon it's an API key thing... you can more explicitly select a "paid API key" in AI Studio now.

CjHuber•1h ago
For me it’s up and running. I was doing some work with AI Studio when it was released and reran a few prompts already. Interesting also that you can now set thinking level low or high. I hope it does something, in 2.5 increasing maximum thought tokens never made it think more
r0fl•50m ago
Works for me.
informal007•1h ago
It seem that Google doesn't prepare well to release Gemini 3 but leak many contents, include the model card early today and gemini 3 on aistudio.google.com
guluarte•1h ago
it is live in the api

> gemini-3-pro-preview-ais-applets

> gemini-3-pro-preview

spudlyo•5m ago
Can confirm. I was able to access it using GPTel in Emacs using 'gemini-3-pro-preview' as the model name.
__jl__•1h ago
API pricing is up to $2/M for input and $12/M for output

For comparison: Gemini 2.5 Pro was $1.25/M for input and $10/M for output Gemini 1.5 Pro was $1.25/M for input and $5/M for output

hirako2000•1h ago
Google went from the phase of loss leader, to bait-to-switch.

They have started lock in with studio, I would say they are still in market penetration but stakeholders want to see path to profit so they are starting the price skimming, it's just the beginning.

mupuff1234•59m ago
I assume the model is just more expensive to run.
hirako2000•35m ago
Likely. The point is we would never know.
jhack•47m ago
With this kind of pricing I wonder if it'll be available in Gemini CLI for free or if it'll stay at 2.5.
raincole•46m ago
Still cheaper than Sonnet 4.5: $3/M for input and $15/M for output.
brianjking•43m ago
It is so impressive that Anthropic has been able to maintain this pricing still.
Aeolun•8m ago
Because every time I try to move away I realize there’s nothing equivalent to move to.
fosterfriends•4m ago
Thrilled to see the cost is competitive with Anthropic.
DeathArrow•1h ago
It generated a quite cool pelican on a bike: https://imgur.com/a/yzXpEEh
rixed•9m ago
2025: solve the biking pelican problem

2026: cure cancer

GodelNumbering•1h ago
And of course they hiked the API prices

Standard Context(≤ 200K tokens)

Input $2.00 vs $1.25 (Gemini 3 pro input is 60% more expensive vs 2.5)

Output $12.00 vs $10.00 (Gemini 3 pro output is 20% more expensive vs 2.5)

Long Context(> 200K tokens)

Input $4.00 vs $2.50 (same +60%)

Output $18.00 vs $15.00 (same +20%)

CjHuber•55m ago
Is it the first time long context has separate pricing? I hadn’t encountered that yet
Topfi•53m ago
Google has been doing that for a while.
brianjking•47m ago
Google has always done this.
CjHuber•46m ago
Ok wow then I‘ve always overlooked that.
1ucky•38m ago
Anthropic is also doing this for long context >= 200k Tokens on Sonnet 4.5
panarky•51m ago
Claude Opus is $15 input, $75 output.
aliljet•1h ago
When will this be available in the cli?
_ryanjsalva•40m ago
Gemini CLI team member here. We'll start rolling out today.
aliljet•38m ago
This is the heroic move everyone is waiting for. Do you know how this will be priced?
Sammi•17m ago
I'm already seeing it in https://aistudio.google.com/
skerit•1h ago
Not the preview crap again. Haven't they tested it enough? When will it be available in Gemini-CLI?
CjHuber•49m ago
Honestly I liked 2.5 Pro preview much more than the final version
prodigycorp•1h ago
I'm sure this is a very impressive model, but gemini-3-pro-preview is failing spectacularly at my fairly basic python benchmark, which involves type analysis. In fact, gemini-2.5-pro gets a lot closer (but is still wrong).

For reference: gpt-5.1-thinking passes, gpt-5.1-instant fails, gpt-5-thinking fails, gpt-5-instant fails, sonnet-4.5 passes, opus-4.1 passes (lesser claude models fail).

This is a reminder that benchmarks are meaningless – you should always curate your own out-of-sample benchmarks. A lot of people are going to say "wow, look how much they jumped in x, y, and z benchmark" and start to make some extrapolation about society, and what this means for others. Meanwhile.. I'm still wondering how they're still getting this problem wrong.

m00dy•49m ago
that's why everyone using AI for code should code in rust only.
Filligree•47m ago
What's the benchmark?
petters•42m ago
Good personal benchmarks should be kept secret :)
ahmedfromtunis•41m ago
I don't think it would be a good idea to publish it on a prime source of training data.
Hammershaft•34m ago
He could post an encrypted version and post the key with it to avoid it being trained on?
benterix•26m ago
What makes you think it wouldn't end up in the training set anyway?
prodigycorp•39m ago
nice try!
mupuff1234•43m ago
Could also just be rollout issues.
prodigycorp•42m ago
Could be. I'll reply to my comment later with pass/fail results of a re-run.
ddalex•40m ago
I moved to using the model from python coding to golang coding and got incredible speedups in writing the correct version of the code
mring33621•30m ago
I agree that benchmarks are noise. I guess, if you're selling an LLM wrapper, you'd care, but as a happy chat end-user, I just like to ask a new model about random stuff that I'm working on. That helps me decide if I like it or not.

I just chatted with gemini-3-pro-preview about an idea I had and I'm glad that I did. I will definitely come back to it.

IMHO, the current batch of free, free-ish models are all perfectly adequate for my uses, which are mostly coding, troubleshooting and learning/research.

This is an amazing time to be alive and the AI bubble doomers that are costing me some gains RN can F-Off!

testartr•28m ago
and models are still pretty bad at playing tic-tac-toe, they can do it, but think way too much

it's easy to focus on what they can't do

prodigycorp•25m ago
Except I'm not nit picking at some limitations with tokenization, like "how many as are there in strawberry" . If you "understand" the code, you shouldn't be getting it wrong.
benterix•24m ago
> This is a reminder that benchmarks are meaningless – you should always curate your own out-of-sample benchmarks.

Yeah I have my own set of tests and the results are a bit unsettling in the sense that sometimes older models outperform newer ones. Moreover, they change even if officially the model doesn't change. This is especially true of Gemini 2.5 pro that was performing much better on the same tests several months ago vs. now.

Iulioh•18m ago
A lot of newer models are geared towards efficency and if you add the fact that more efficent models are trained on the output of less efficent (but more accurate) models....

GPT4/3o might be the best we will ever have

WhitneyLand•19m ago
>>benchmarks are meaningless

No they’re not. Maybe you mean to say they don’t tell the whole story or have their limitations, which has always been the case.

>>my fairly basic python benchmark

I suspect your definition of “basic” may not be consensus. Gpt-5 thinking is a strong model for basic coding and it’d be interesting to see a simple python task it reliably fails at.

Rover222•18m ago
curious if you tried grok 4.1 too
sosodev•14m ago
How can you be sure that your benchmark is meaningful and well designed?

Is the only thing that prevents a benchmark from being meaningful publicity?

prodigycorp•8m ago
I didn't tell you what you should think about the model. All I said is that you should have your own benchmark.

I think my benchmark is well designed. It's well designed because it's a generalization of a problem I've consistently had with LLMs on my code. Insofar that it encapsulates my coding preferences and communication style, that's the proper benchmark for me.

My benchmark says that I will stick with Codex CLI for the foreseeable future.

thefourthchime•12m ago
I like to ask "Make a pacman game in a single html page". No model has ever gotten a decent game in one shot. My attempt with Gemini3 was no better than 2.5.
ofa0e•4m ago
Your benchmarks should not involve IP.
nickandbro•1h ago
What we have all been waiting for:

"Create me a SVG of a pelican riding on a bicycle"

https://www.svgviewer.dev/s/FfhmhTK1

Thev00d00•58m ago
That is pretty impressive.

So impressive it makes you wonder if someone has noticed it being used a benchmark prompt.

burkaman•52m ago
Simon says if he gets a suspiciously good result he'll just try a bunch of other absurd animal/vehicle combinations to see if they trained a special case: https://simonwillison.net/2025/Nov/13/training-for-pelicans-...
jmmcd•36m ago
"Pelican on bicycle" is one special case, but the problem (and the interesting point) is that with LLMs, they are always generalising. If a lab focussed specially on pelicans on bicycles, they would as a by-product improve performance on, say, tigers on rollercoasters. This is new and counter-intuitive to most ML/AI people.
ddalex•34m ago
https://www.svgviewer.dev/s/TVk9pqGE giraffe in a ferrari
rixed•12m ago
I have tried combinations of hard to draw vehicle and animals (crocodile, frog, pterodactly, riding a hand glider, tricycle, skydiving), and it did a rather good job in every cases (compared to previous tests). Whatever they have done to improve on that point, they did it in a way that generalise.
bitshiftfaced•51m ago
It hadn't occurred to me until now that the pelican could overcome the short legs issue by not sitting on the seat and instead put its legs inside the frame of the bike. That's probably closer to how a real pelican would ride a bike, even if it wasn't deliberate.
xnx•47m ago
Very aero
CjHuber•1h ago
Interesting that they added an option to select your own API key right in AI studio‘s input field. I sincerely hope the times of generous free AIstudio usage are not over
golfer•56m ago
Supposedly this is the model card. Very impressive results.

https://pbs.twimg.com/media/G6CFG6jXAAA1p0I?format=jpg&name=...

Also, the full document:

https://archive.org/details/gemini-3-pro-model-card/page/n3/...

tweakimp•52m ago
Every time I see a table like this numbers go up. Can someone explain what this actually means? Is there just an improvement that some tests are solved in a better way or is this a breakthrough and this model can do something that all others can not?
rvnx•45m ago
This is a list of questions and answers that was created by different people.

The questions AND the answers are public.

If the LLM manages through reasoning OR memory to repeat back the answer then they win.

The scores represent the % of correct answers they recalled.

stavros•28m ago
I estimate another 7 months before models start getting 115% on Humanity's Last Exam.
HardCodedBias•32m ago
If you believe another thread the benchmarks are comparing Gemini-3 (probably thinking) to GPT-5.1 without thinking.

The person also claims that with thinking on the gap narrows considerably.

We'll probably have 3rd party benchmarks in a couple of days.

samuelknight•56m ago
"Gemini 3 Pro Preview" is in Vertex
ponyous•43m ago
Can’t wait to test it out. Been running a tons of benchmarks (1000+ generations) for my AI to CAD model project and noticed:

- GPT-5 medium is the best

- GPT-5.1 falls right between Gemini 2.5 Pro and GPT-5 but it’s quite a bit faster

Really wonder how well Gemini 3 will perform

santhoshr•36m ago
Pelican riding a bicycle: https://pasteboard.co/CjJ7Xxftljzp.png
mohsen1•29m ago
Some time I think I should spend $50 on Upwork to get a real human artist to do it first to know what is that we're going for. What a good pelican riding a bicycle SVG is actually looking like?
AstroBen•14m ago
IMO it's not about art, but a completely different path than all these images are going down. The pelican needs tools to ride the bike, or a modified bike. Maybe a recumbent?
robterrell•26m ago
At this point I'm surprised they haven't been training on thousands of professionally-created SVGs of pelicans on bicycles.
notatoad•14m ago
i think anything that makes it clear they've done that would be a lot worse PR than failing the pelican test would ever be.
bn-l•11m ago
It’s a good pelican. Not great but good.
Der_Einzige•35m ago
When will they allow us to use modern LLM samplers like min_p, or even better samplers like top N sigma, or P-less decoding? They are provably SOTA and in some cases enable infinite temperature.

Temperature continues to be gated to maximum of 0.2, and there's still the hidden top_k of 64 that you can't turn off.

I love the google AI studio, but I hate it too for not enabling a whole host of advanced features. So many mixed feelings, so many unanswered questions, so many frustrating UI decisions on a tool that is ostensibly aimed at prosumers...

ttul•34m ago
My favorite benchmark is to analyze a very long audio file recording of a management meeting and produce very good notes along with a transcript labeling all the speakers. 2.5 was decently good at generating the summary, but it was terrible at labeling speakers. 3.0 has so far absolutely nailed speaker labeling.
iagooar•26m ago
What prompt do you use for that?
zurfer•17m ago
It also tops LMSYS leaderboard across all categories. However knowledge cutoff is Jan 2025. I do wonder how long they have been pre-training this thing :D.
mihau•15m ago
@simonw wen pelican
Workaccount2•13m ago
It still failed my image identification test ([a photoshopped picture of a dog with 5 legs]...please count the legs) that so far every other model has failed agonizingly, even failing when I tell them they are failing, and they tend to fight back at me.

Gemini 3 however, while still failing, at least recognized the 5th leg, but thought the dog was...well endowed. The 5th leg however is clearly a leg, despite being where you would expect the dogs member to be. I'll give it half credit for at least recognizing that there was something there.

Still though, there is a lot of work that needs to be done on getting these models to properly "see" images.

bespokedevelopr•6m ago
Wow so the polymarket insider bet was true then..

https://old.reddit.com/r/wallstreetbets/comments/1oz6gjp/new...

mikeortman•5m ago
Its available for me now in gemini.google.com.... but its failing so bad at accurate audio transcription.

Its transcribing the meeting but hallucinates badly... both in fast and thinking mode. Fast mode only transcribed about a fifth of the meeting before saying its done. Thinking mode completely changed the topic and made up ENTIRE conversations. Gemini 2.5 actually transcribed it decently, just occasional missteps when people talked over each other.

I'm concerned.