frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

FunctionGemma 270M Model

https://blog.google/technology/developers/functiongemma/
83•mariobm•3h ago

Comments

canyon289•2h ago
Hi all, I'm a research lead on this model. Same as every model release post, I enjoy working at Google for a multitude of reasons, and opinions here are my own.

Happy to answer whatever technical questions I can!

xnx•2h ago
Cool game! Amazing it can run in the browser. My mind was blown when I saw you could give goal based commands vs prescriptive ones. https://huggingface.co/spaces/webml-community/FunctionGemma-...
canyon289•1h ago
So I didn't even know this was going to be made until recently, and when I saw it, it also blew my mind. I didn't realize how far along web ml community had pushed things, and was impressed by the creativity of the HF folks with visuals and "game flow".

Personally speaking its really neat to see other people who take these models and run with them, creating things I could haven't have imagined. I'm hoping many others in the open community do the same in the coming weeks and the new year

carlcortright•2h ago
Very cool model! Congrats on the work!
canyon289•1h ago
Thank you much for the kind words
NitpickLawyer•1h ago
Wen gemma4? :)

But on a serious note, I'm happy to see more research going into vSLMs (very small...) My "dream" scenario is to have the "agentic" stuff run locally, and call into the "big guns" as needed. Being able to finetune these small models on consumer cards is awesome, and can open up a lot of niche stuff for local / private use.

canyon289•1h ago
Trust me as a daily at home Gemma user myself, I'm just excited for what's upcoming as you are, maybe even more because I have some hints for what's to come.

>My "dream" scenario is to have the "agentic" stuff run locally, and call into the "big guns" as needed.

FunctionGemma 270m is your starter pack for this, train your own functions to call out to whatever larger models you choose. It's been quite effective my testing, and the finetuning guides should show you how to add in your own capabilities.

Speaking from the research side its incredible how so many small models, not just Gemma, are achieving performance levels of must larger models from just a year or two ago. It's personally why I stay in this space.

xnx•1h ago
Not FunctionGemma related, but would love to see an open weights model from Google for speech to text transcription (diarization, timestamps, etc.).

Whisper is old and resource intensive for the accuracy it provides.

canyon289•13m ago
I'm not specifically promising anything but I do want to say 2026 is going to be a great year! Many of my colleagues are shipping models too, such as t5gemma which is on the front page, and I'm personally excited to see what we're all collectively going to release in the coming year.
lukeinator42•1h ago
Very cool! I was wondering, is a separate model performing speech recognition for the voice demos such as the game? The FunctionGemma model card only seems to show text input/output.
canyon289•47m ago
Yes a separate model is performing ASR in this case. Gemma270m (base, function, and others) are not multimodal out of the box.

That being said if someone in the community wanted to use other encoders like siglip and plug them into Gemma270m to make it multimodal that'd be a great way to have fun over break and build up an AI Eegineer resume :)

zikani_03•59m ago
Thanks for all the great work. How good is the model at composing actions and is there a way to say, give the model ability to scope actions, for example if actions are related to permissions or some other context? Would one need to pass the role or permission as context or finetune separately?

I hope those questions make sense

canyon289•34m ago
> How good is the model at composing actions?

I think you mean taking the results of one function call and putting it into another? We saw some promise but didn't heavily train for this use case in the base model. The thing we noticed with the 270m sized models, and the performance expectations of AI models in 2025, is that these size models perform best for _specific users_ when finetuned to that specific use case.

What I suggest is mocking some data either by hand or using some automated tool and finetuning in this kind of use case and using the finetuning colab setup.

> is there a way to give the model ability to scope action for example if actions are related to permissions

Permissions depend on your system architecture more than the model. The model itself just takes in tokens and outputs tokens. Permissions are defined by your security/system setup in which the model itself is running.

vessenes•47m ago
Hey! Love the Gemma series. Question that came to mind reading the announcement post - the proposal there is that you can use this as a local backbone and have it treat a larger model as a 'tool call' when more reasoning is needed.

In my mind we want a very smart layer frontier model orchestrating, but not slowing everything down by doing every little thing; this seems like the opposite - a very fast layer that can be like "wait a minute, I'm too dumb for this, need some help".

My question is - does the Gemma team use any evaluation around this particular 'call a (wiser) friend' strategy? How are you thinking about this? Is this architecture flow more an accommodation to the product goal - fast local inference - or do you guys think it could be optimal?

canyon289•29m ago
We evaluate many things that you alluded to, such as speed on device, output correctness, and also "is this something that would be useful" the last one being a bit abstract.

The way we think about it is what do we think developers and users need, and is there a way we can fill that gap in a useful way. With this model we had the hypothesis you had, there are fantastic larger models out there pushing the frontier of AI capabilities, but there's also a nice for smaller customizable model that's quick to run and quick to tune.

What is optimal then ultimately falls to you and your use cases (which I'm guessing at here), you have options now between Gemini and Gemma.

exacube•34m ago
Some fine tuning data questions:

i see the the dataset Google published in this notebook https://github.com/google-gemini/gemma-cookbook/blob/main/Fu... -- from looking at the dataset on huggingface, it looks synthetically generated.

1. do you recommend any particular mix or focus in the dataset for finetuning this model, without losing too much generality?

2. do you have any recommendations for how many examples per-tool?

thank you for your (and your teams) work!

canyon289•15m ago
> Do you recommend any particular mix or focus in the dataset for finetuning this model, without losing too much generality?

Astute questions, there's sort of two ways to think about finetuning, 1. Obliterate any general functionality and train the model on your general commands 2. As you asked maintain generality trying to preserve initial model ability

For 2 typically low learning rate or LORA is a good strategy. We show an example in our the finetuning tutorial in the blog.

> 2. do you have any recommendations for how many examples per-tool? This depends on the tool complexity and the variety of user inputs. So a simple tool like turn_flashlight_on(), with no args, will get taught quickly, especially if say you're only prompting in English.

But if you have a more complex function like get_weather(lat, lon, day, region, date) and have prompts coming in in English, Chinese, Gujarati and spanish, the model needs to do a lot more "heavy lifting" to both translate a request and fill out a complex query. We know as programmers date by themselves are insanely complex in natural language (12/18/2025 vs 18/12/2025).

To get this right it'll help the model if it was trained on data that shows it the versions of variations of inputs possible.

Long answer but I hope this makes sense.

mrinterweb•6m ago
I have often wondered how much a specialized local LLM could benefit an agentic tool like Gemini CLI. I would think there could be a good win for speed and minimizing token use if coding agents used a local model. A local model could handle a lot of the low level system interaction type tasks and then send the prompts that require deeper reasoning to frontier models. It seems wasteful and slow to use frontier models to figure out how to grep a codebase, run tests, git diff, etc.

Might Gemini CLI offload some of its prompts to FunctionGemma?

nateb2022•2h ago
Ollama link too: https://ollama.com/library/functiongemma
homarp•1h ago
llama.cpp link https://huggingface.co/ggml-org/functiongemma-270m-it-GGUF
SpaceManNabs•1h ago
can you run this from n8n?
canyon289•39m ago
I just looked through their webpage and github and I'm not sure. But maybe someone should make a feature request!

https://github.com/n8n-io/n8n

orliesaurus•1h ago
edit: Im so dumb...
canyon289•44m ago
Its already on the phone! Check out the demo videos and colab that show you how to load this model onto a device relatively easily.

On this project I was lucky enough to work with the Google AI Edge team who have deep expertise in edge deployments on device. Check out this app they built which loads in the Gemma 270m models and runs them on your phone.

https://play.google.com/store/apps/details?id=com.google.ai....

You also can finetune your own models and load them onto device with the sameworkflow. Scroll to the bottom to see the instructions and a screenshot example https://ai.google.dev/gemma/docs/mobile-actions

xnx•51m ago
Unbelievable shipping velocity from Google in December, and it sounds like they're not done for the week: https://x.com/osanseviero/status/2001723652635541566
eachro•45m ago
Do you think this would be appropriate for a command line tool that hits various apis as the function calls? Ex: "what's the weather in SF tomorrow?" Or "daily price change of apple, Tesla stock for past week"? (Let's assume I have documented the apis thoroughly somewhere that the model has access to or fine tuned it on this data)
milenf•25m ago
Hi, also on the FunctionGemma team! Something like this would be a good use case for the model. Based on how complicated the API is you might need to finetune it (we released a colab that guides you through the experience + how to export/run it locally). Generally better tool descriptions help although if it is something very complicated finetuning would be better.

Beginning January 2026, all ACM publications will be made open access

https://dl.acm.org/openaccess
989•Kerrick•5h ago•107 comments

We pwned X, Vercel, Cursor, and Discord through a supply-chain attack

https://gist.github.com/hackermondev/5e2cdc32849405fff6b46957747a2d28
282•hackermondev•2h ago•97 comments

GPT-5.2-Codex

https://openai.com/index/introducing-gpt-5-2-codex/
227•meetpateltech•3h ago•141 comments

Skills for organizations, partners, the ecosystem

https://claude.com/blog/organization-skills-and-directory
196•adocomplete•4h ago•120 comments

Texas is suing all of the big TV makers for spying on what you watch

https://www.theverge.com/news/845400/texas-tv-makers-lawsuit-samsung-sony-lg-hisense-tcl-spying
153•tortilla•2d ago•94 comments

T5Gemma 2: The next generation of encoder-decoder models

https://blog.google/technology/developers/t5gemma-2/
44•milomg•1h ago•7 comments

Delty (YC X25) Is Hiring an ML Engineer

https://www.ycombinator.com/companies/delty/jobs/MDeC49o-machine-learning-engineer
1•lalitkundu•26m ago

How China built its ‘Manhattan Project’ to rival the West in AI chips

https://www.japantimes.co.jp/business/2025/12/18/tech/china-west-ai-chips/
68•artninja1988•2h ago•65 comments

Classical statues were not painted horribly

https://worksinprogress.co/issue/were-classical-statues-painted-horribly/
476•bensouthwood•9h ago•236 comments

How did IRC ping timeouts end up in a lawsuit?

https://mjg59.dreamwidth.org/73777.html
66•dvaun•1d ago•8 comments

FunctionGemma 270M Model

https://blog.google/technology/developers/functiongemma/
84•mariobm•3h ago•27 comments

Show HN: Stop AI scrapers from hammering your self-hosted blog (using porn)

https://github.com/vivienhenz24/fuzzy-canary
58•misterchocolat•2d ago•12 comments

Show HN: Picknplace.js, an alternative to drag-and-drop

https://jgthms.com/picknplace.js/
36•bbx•2d ago•26 comments

Meta Segment Anything Model Audio

https://ai.meta.com/samaudio/
90•megaman821•2d ago•9 comments

Your job is to deliver code you have proven to work

https://simonwillison.net/2025/Dec/18/code-proven-to-work/
508•simonw•6h ago•434 comments

TRELLIS.2: state-of-the-art large 3D generative model (4B)

https://github.com/microsoft/TRELLIS.2
33•dvrp•1d ago•4 comments

Firefox will have an option to disable all AI features

https://mastodon.social/@firefoxwebdevs/115740500373677782
116•twapi•3h ago•122 comments

Using TypeScript to obtain one of the rarest license plates

https://www.jack.bio/blog/licenseplate
116•lafond•6h ago•117 comments

I've been writing ring buffers wrong all these years (2016)

https://www.snellman.net/blog/archive/2016-12-13-ring-buffers/
19•flaghacker•2d ago•2 comments

How I wrote JustHTML, a Python-based HTML5 parser, using coding agents

https://friendlybit.com/python/writing-justhtml-with-coding-agents/
31•simonw•4d ago•17 comments

Interactive Fluid Typography

https://electricmagicfactory.com/articles/interactive-fluid-typography/
19•list•1h ago•0 comments

Launch HN: Pulse (YC S24) – Production-grade unstructured document extraction

30•sidmanchkanti21•5h ago•31 comments

The Scottish Highlands, the Appalachians, Atlas are the same mountain range

https://vividmaps.com/central-pangean-mountains/
30•lifeisstillgood•2h ago•10 comments

Ringspace: A proposal for the human web

https://taggart-tech.com/ringspace/
17•todsacerdoti•17h ago•2 comments

Please just try HTMX

http://pleasejusttryhtmx.com/
351•iNic•7h ago•312 comments

How to hack Discord, Vercel and more with one easy trick

https://kibty.town/blog/mintlify/
47•todsacerdoti•1h ago•10 comments

Oliver Sacks put himself into his case studies – what was the cost?

https://www.newyorker.com/magazine/2025/12/15/oliver-sacks-put-himself-into-his-case-studies-what...
7•barry-cotter•36m ago•51 comments

The <time> element should do something

https://nolanlawson.com/2025/12/14/the-time-element-should-actually-do-something/
38•birdculture•2d ago•7 comments

The immortality of Microsoft Word

https://theredline.versionstory.com/p/on-the-immortality-of-microsoft-word
22•jpbryan•6h ago•33 comments

Military standard on software control levels

https://entropicthoughts.com/mil-std-882e-software-control
48•ibobev•4h ago•21 comments