frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

Gemma 3 270M: Compact model for hyper-efficient AI

https://developers.googleblog.com/en/introducing-gemma-3-270m/
437•meetpateltech•5h ago•181 comments

We Rewrote the Ghostty GTK Application

https://mitchellh.com/writing/ghostty-gtk-rewrite
40•tosh•40m ago•1 comments

Streaming services are driving viewers back to piracy

https://www.theguardian.com/film/2025/aug/14/cant-pay-wont-pay-impoverished-streaming-services-are-driving-viewers-back-to-piracy
199•nemoniac•5h ago•178 comments

Steve Wozniak: Life to me was never about accomplishment, but about happiness

https://yro.slashdot.org/comments.pl?sid=23765914&cid=65583466
335•MilnerRoute•3h ago•227 comments

Org-social is a decentralized social network that runs on Org Mode

https://github.com/tanrax/org-social
79•tanrax•1d ago•36 comments

I made a real-time C/C++/Rust build visualizer

https://danielchasehooper.com/posts/syscall-build-snooping/
135•dhooper•5h ago•43 comments

New protein therapy shows promise as antidote for carbon monoxide poisoning

https://www.medschool.umaryland.edu/news/2025/new-protein-therapy-shows-promise-as-first-ever-antidote-for-carbon-monoxide-poisoning.html
197•breve•10h ago•47 comments

OneSignal (YC S11) Is Hiring Engineers

https://onesignal.com/careers
1•gdeglin•58m ago

What's the strongest AI model you can train on a laptop in five minutes?

https://www.seangoedecke.com/model-on-a-mbp/
460•ingve•2d ago•167 comments

Show HN: OWhisper – Ollama for realtime speech-to-text

https://docs.hyprnote.com/owhisper/what-is-this
63•yujonglee•6h ago•25 comments

Airbrush art of the 80s was Chrome-tastic (2015)

https://www.coolandcollected.com/airbrush-art-of-the-80s-was-chrome-tastic/
22•Michelangelo11•2h ago•3 comments

Architecting large software projects [video]

https://www.youtube.com/watch?v=sSpULGNHyoI
59•jackdoe•2d ago•27 comments

Show HN: I built a free alternative to Adobe Acrobat PDF viewer

https://github.com/embedpdf/embed-pdf-viewer
122•bobsingor•6h ago•31 comments

All Souls exam questions and the limits of machine reasoning

https://resobscura.substack.com/p/all-souls-exam-questions-and-the
33•benbreen•1d ago•13 comments

Blood oxygen monitoring returning to Apple Watch in the US

https://www.apple.com/newsroom/2025/08/an-update-on-blood-oxygen-for-apple-watch-in-the-us/
292•thm•8h ago•216 comments

Homekit-steam-user-switcher: A way to remotely switch Steam users using HomeKit

https://github.com/rcarmo/homekit-steam-user-switcher
12•rcarmo•3d ago•0 comments

Lambdas, Nested Functions, and Blocks

https://thephd.dev/lambdas-nested-functions-block-expressions-oh-my
5•zaikunzhang•2d ago•0 comments

Launch HN: Cyberdesk (YC S25) – Automate Windows legacy desktop apps

46•mahmoud-almadi•6h ago•31 comments

1976 Soviet edition of 'The Hobbit' (2015)

https://mashable.com/archive/soviet-hobbit
229•us-merul•3d ago•75 comments

Reverse Proxy Deep Dive: Why Load Balancing at Scale Is Hard

https://startwithawhy.com/reverseproxy/2025/08/08/ReverseProxy-Deep-Dive-Part4.html
27•miggy•3d ago•2 comments

Bluesky: Updated Terms and Policies

https://bsky.social/about/blog/08-14-2025-updated-terms-and-policies
67•mschuster91•5h ago•85 comments

Show HN: MCP Security Suite

https://github.com/NineSunsInc/mighty-security
11•jodoking•1h ago•8 comments

"Privacy preserving age verification" is bullshit

https://pluralistic.net/2025/08/14/bellovin/
170•Refreeze5224•4h ago•111 comments

What does Palantir actually do?

https://www.wired.com/story/palantir-what-the-company-does/
142•mudil•22h ago•112 comments

DINOv3

https://github.com/facebookresearch/dinov3
16•reqo•1h ago•6 comments

How to rig elections [video]

https://media.ccc.de/v/why2025-218-how-to-rig-elections
114•todsacerdoti•9h ago•94 comments

Nyxt: The Emacs-like web browser

https://lwn.net/Articles/1001773/
112•signa11•3d ago•25 comments

500 days of math

https://gmays.com/500-days-of-math/
139•gmays•2d ago•80 comments

Big Tech's A.I. Data Centers Are Driving Up Electricity Bills for Everyone

https://www.nytimes.com/2025/08/14/business/energy-environment/ai-data-centers-electricity-costs.html
15•moneycantbuy•1h ago•1 comments

Show HN: Modelence – Supabase for MongoDB

https://github.com/modelence/modelence
25•artahian•5h ago•8 comments
Open in hackernews

Show HN: OWhisper – Ollama for realtime speech-to-text

https://docs.hyprnote.com/owhisper/what-is-this
63•yujonglee•6h ago
Hello everyone. This is Yujong from the Hyprnote team (https://github.com/fastrepl/hyprnote).

We built OWhisper for 2 reasons: (Also outlined in https://docs.hyprnote.com/owhisper/what-is-this)

(1). While working with on-device, realtime speech-to-text, we found there isn't tooling that exists to download / run the model in a practical way.

(2). Also, we got frequent requests to provide a way to plug in custom STT endpoints to the Hyprnote desktop app, just like doing it with OpenAI-compatible LLM endpoints.

The (2) part is still kind of WIP, but we spent some time writing docs so you'll get a good idea of what it will look like if you skim through them.

For (1) - You can try it now. (https://docs.hyprnote.com/owhisper/cli/get-started)

  bash
  brew tap fastrepl/hyprnote && brew install owhisper
  owhisper pull whisper-cpp-base-q8-en
  owhisper run whisper-cpp-base-q8-en

If you're tired of Whisper, we also support Moonshine :) Give it a shot (owhisper pull moonshine-onnx-base-q8)

We're here and looking forward to your comments!

Comments

yujonglee•3h ago
Happy to answer any questions!

These are list of local models it supports:

- whisper-cpp-base-q8

- whisper-cpp-base-q8-en

- whisper-cpp-tiny-q8

- whisper-cpp-tiny-q8-en

- whisper-cpp-small-q8

- whisper-cpp-small-q8-en

- whisper-cpp-large-turbo-q8

- moonshine-onnx-tiny

- moonshine-onnx-tiny-q4

- moonshine-onnx-tiny-q8

- moonshine-onnx-base

- moonshine-onnx-base-q4

- moonshine-onnx-base-q8

phkahler•1h ago
I thought whisper and others took large chunks (20-30 seconds) of speech, or a complete wave file as input. How do you get real-time transcription? What size chunks do you feed it?

To me, STT should take a continuous audio stream and output a continuous text stream.

yujonglee•1h ago
I use VAD to chunk audio.

Whisper and Moonshine both works in a chunk, but for moonshine:

> Moonshine's compute requirements scale with the length of input audio. This means that shorter input audio is processed faster, unlike existing Whisper models that process everything as 30-second chunks. To give you an idea of the benefits: Moonshine processes 10-second audio segments 5x faster than Whisper while maintaining the same (or better!) WER.

Also for kyutai, we can input continuous audio in and get continuous text out.

- https://github.com/moonshine-ai/moonshine - https://docs.hyprnote.com/owhisper/configuration/providers/k...

mijoharas•1h ago
Something like that, in a cli tool, that just gives text to stdout would be perfect for a lot of use cases for me!

(maybe with an `owhisper serve` somewhere else to start the model running or whatever.)

yujonglee•1h ago
Are you thinking about the realtime use-case or batch use-case?

For just transcribing file/audio,

`owhisper run <MODEL> --file a.wav` or

`curl httpsL//something.com/audio.wav | owhisper run <MODEL>`

might makes sense.

mijoharas•1h ago
agreed, both of those make sense, but I was thinking realtime. (pipes can stream data, I'd like and find useful something that can stream tts to stdout in realtime.)
yujonglee•1h ago
It's open-source. Happy to review & merge if you can send us PR!

https://github.com/fastrepl/hyprnote/blob/8bc7a5eeae0fe58625...

alkh•42m ago
Sorry, maybe I missed it but I didn't see this list on your website. I think it is a good idea to add this info there. Besides that, thank you for the effort and your work! I will definetely give it a try
yujonglee•39m ago
got it. fyi if you run `owhisper pull --help`, this info is printed
JP_Watts•2h ago
I’d like to use this to transcribe meeting minutes with multiple people. How could this program work for that use case?
yujonglee•2h ago
If your use-case is meeting, https://github.com/fastrepl/hyprnote is for you. OWhisper is more like a headless version of it.
JP_Watts•2h ago
Can you describe how it pick different voices? Does it need separate audio channels, or does it recognize different voices on the same audio input?
yujonglee•2h ago
It separate mic/speaker as 2 channel. So you can reliably get "what you said" vs "what you heard".

For splitting speaker within channel, we need AI model to do that. It is not implemented yet, but I think we'll be in good shape somewhere in September.

Also we have transcript editor that you can easily split segment, assign speakers.

sxp•2h ago
If you want to transcribe meeting notes, whisper isn't the best tool because it doesn't separate the transcribe by speakers. There are some other tools that do that, but I'm not sure what the best local option is. I've used Google's cloud STT with the diarization option and manually renamed "Speaker N" after the fact.
solarkraft•2h ago
Wait, this is cool.

I just spent last week researching the options (especially for my M1!) and was left wishing for a standard, full-service (live) transcription server for Whisper like OLlama has been for LLMs.

I’m excited to try this out and see your API (there seems to be a standard vaccuum here due to openai not having a real time transcription service, which I find to be a bummer)!

Edit: They seem to emulate the Deepgram API (https://developers.deepgram.com/reference/speech-to-text-api...), which seems like a solid choice. I’d definitely like to see a standard emerging here.

yujonglee•1h ago
Correct. About the deepgram-compatibility: https://docs.hyprnote.com/owhisper/deepgram-compatibility

Let me know how it goes!

clickety_clack•1h ago
Please find a way to add speaker diarization, with a way to remember the speakers. You can do it with pyannote, and get a vector embedding of each speaker that can be compared between audio samples, but that’s a year old now so I’m sure there’s better options now!
yujonglee•1h ago
yeah that is on the roadmap!
mijoharas•1h ago
Ok, cool! I was actually one of the people on the hyprnote HN thread asking for a headless mode!

I was actually integrating some whisper tools yesterday. I was wondering if there was a way to get a streaming response, and was thinking it'd be nice if you can.

I'm on linux, so don't think I can test out owhisper right now, but is that a thing that's possible?

Also, it looks like the `owhisper run` command gives it's output as a tui. Is there an option for a plain text response so that we can just pipe it to other programs? (maybe just `kill`/`CTRL+C` to stop the recording and finalize the words).

Same question for streaming, is there a way to get a streaming text output from owhisper? (it looks like you said you create a deepgram compatible api, I had a quick look at the api docs, but I don't know how easy it is to hook into it and get some nice streaming text while speaking).

Oh yeah, and diarisation (available with a flag?) would be awesome, one of the things that's missing from most of the easiest to run things I can find.

mijoharas•1h ago
Oh wait, maybe you do support linux for owhisper: https://github.com/fastrepl/homebrew-hyprnote/blob/main/Form...

Can you help me out to find where the code you've built is? I can see the folder in github[0], but I can't see the code for the cli for instance? unless I'm blind.

[0] https://github.com/fastrepl/hyprnote/tree/main/owhisper

yujonglee•1h ago
This is CLI entry point:

https://github.com/fastrepl/hyprnote/blob/8bc7a5eeae0fe58625...

yujonglee•1h ago
> I'm on linux

I didn't tested on Linux yet, but we have linux build: http://owhisper.hyprnote.com/download/latest/linux-x86_64

> also, it looks like the `owhisper run` command gives it's output as a tui. Is there an option for a plain tex

`owhisper run` is more like way to quickly trying it out. But I think piping is definitely something that should work.

> Same question for streaming, is there a way to get a streaming text output from owhisper?

You can use Deepgram client to talk to `owhisper serve`. (https://docs.hyprnote.com/owhisper/deepgram-compatibility) So best resource might be Deepgram client SDK docs.

> diarisation

yeah on the roadmap

mijoharas•38m ago
Nice stuff, had a quick test on linux and it works (built directly, I didn't check out the brew). I ran into a small issue with moonshine and opened an issue on github.

Great work on this! excited to keep an eye on things.

DiabloD3•40m ago
I suggest you don't brand this "Ollama for X". They've become a commercial operation that is trying to FOSS-wash their actions through using llama.cpp's code and then throwing their users under the bus when they can't support them.

I see that you are also using llama.cpp's code? That's cool, but make sure you become a member of that community, not an abuser.

yujonglee•34m ago
yeah we use whisper.cpp for whisper inference. this is more like a community-focused project, not a commercial product!