frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

The Scallenge

https://herbertlui.net/the-scallenge/
1•mooreds•26s ago•0 comments

Implementing Logic Programming

https://btmc.substack.com/p/implementing-logic-programming
1•sirwhinesalot•3m ago•1 comments

Openvino-Plugins-AI-Audacity

https://github.com/intel/openvino-plugins-ai-audacity
1•petethomas•3m ago•0 comments

(LLM self fine-tuning) Unsupervised Elicitation of Language Models

https://arxiv.org/abs/2506.10139
1•xianshou•6m ago•0 comments

Ask HN: Creatives – want a portfolio site that feels like you?

1•pomdevv•7m ago•0 comments

Flohmarkt v0.10.0 Released

https://codeberg.org/flohmarkt/flohmarkt/releases/tag/0.10.0
1•midzer•8m ago•0 comments

Fiber-Optic Drones the New Must-Have in Ukraine War

https://www.rferl.org/a/ukraine-fiber-optic-drones-russia/33344310.html
2•austinallegro•10m ago•1 comments

JavaFlow – highly concurrent, async programming with deterministic execution

https://github.com/panghy/javaflow
1•simonpure•11m ago•0 comments

Show HN: Yupp – Every AI for Everyone

https://yupp.ai/
1•wennyyustalim•13m ago•0 comments

A collection of sample agents built with Agent Development (ADK)

https://github.com/google/adk-samples
1•tanelpoder•13m ago•0 comments

Show HN: WaveGen – turn blog articles into text overlay videos, not slops

https://wavegen.ai
1•lululi1989•13m ago•0 comments

Humpback Whales Are Way Cooler Than You

https://nautil.us/humpback-whales-are-way-cooler-than-you-1216796/
2•dnetesn•15m ago•0 comments

When Monsters Came for Mathematics

https://nautil.us/when-monsters-came-for-mathematics-1217312/
1•dnetesn•16m ago•0 comments

China AI Companies Dodge US Chip Curbs by Flying Suitcases of Hard Drives Abroad

https://www.wsj.com/tech/china-ai-chip-curb-suitcases-7c47dab1
4•walterbell•19m ago•1 comments

Recognizing and Communicating with diverse intelligences [video]

https://www.youtube.com/watch?v=OD5TOsPZIQY
1•downboots•25m ago•0 comments

Introducing Sulka, the Hardened Yocto Distro

https://ejaaskel.dev/introducing-sulka-the-hardened-yocto-distro/
5•FrankSansC•27m ago•1 comments

How we built our multi-agent research system

https://www.anthropic.com/engineering/built-multi-agent-research-system
1•mfiguiere•27m ago•0 comments

Show HN: Start building and monetizing your Bluesky audience

https://bluesky-bot.com/
1•cranberryturkey•27m ago•0 comments

Scientists discover 230 new giant viruses that shape ocean life and health

https://news.miami.edu/rosenstiel/stories/2025/06/scientists-discover-230-new-giant-viruses-that-shape-ocean-life-and-health.html
2•gmays•29m ago•0 comments

Google using ZK for age assurance

https://blog.google/around-the-globe/google-europe/age-assurance-europe/
3•abhv•29m ago•0 comments

StillOS 10 Preview – Brand New Distro Aimed to Be as Consumer Ready as Possible

https://www.stillhq.io
3•pizzalovingnerd•30m ago•1 comments

Step-by-Step Guide: Posts by Month in Bear Blog

https://yordi.me/step-by-step-guide-posts-by-month-in-bear-blog/
1•Froodooo•31m ago•0 comments

The State of React and the Community in 2025

https://blog.isquaredsoftware.com/2025/06/react-community-2025/
1•switz•32m ago•0 comments

We found a germ that 'feeds' on hospital plastic – new study

https://theconversation.com/we-found-a-germ-that-feeds-on-hospital-plastic-new-study-256945
2•PaulHoule•34m ago•0 comments

GPU-accelerated Llama3.java inference in pure Java using TornadoVM

https://github.com/beehive-lab/GPULlama3.java
1•pjmlp•35m ago•0 comments

Show HN: Omi Desktop – open-source meeting summarizer

https://github.com/BasedHardware/omi/tree/main/app
1•kodjima33•36m ago•1 comments

3Blue1Brown Follow-Up: From Hypothetical Examples to LLM Circuit Visualization

https://peterlai.github.io/gpt-circuits/
1•peterlai•36m ago•1 comments

Building Foundation Models into Banks

https://building.nubank.com/foundation-models-ai-nubank-transformation/
1•rafaepta•37m ago•0 comments

The Stress of Wall Street Is Sending Men to Pelvic Floor Therapy

https://www.bloomberg.com/news/features/2025-06-13/wall-street-stress-sends-nyc-men-to-pelvic-floor-therapy
1•petethomas•38m ago•0 comments

Indian scientists search for the perfect Apple

https://www.bbc.com/news/articles/c0l05762elpo
1•1659447091•40m ago•0 comments
Open in hackernews

Ask HN: Can anybody clarify why OpenAI reasoning now shows non-English thoughts?

21•johnnyApplePRNG•22h ago
People have noticed for a while now that Google's Bard/Gemini has inserted random hindi/bengali words often. [0]

I just caught this in an o3-pro thought process: "and customizing for low difficulty. কাজ করছে!"

That last set of chars is apparently Bengali for "working!".

I just find it curious that similar "errors" are appearing from multiple different models... what is the training method or reasoning that these alternate languages can creep in, does anyone know?

[0] https://www.reddit.com/r/Bard/comments/18zk2tb/bard_speaking_random_languages/

Comments

yen223•21h ago
I have no idea what's going on with ChatGPT, but I can say it's pretty common for multilingual people to be thinking about things in a different language from what they are currently speaking.
johnnyApplePRNG•21h ago
Interesting thanks yea I forgot that even I used to be able to think in another language long ago!
latentsea•19h ago
Language itself structures how to think about things too. There are some thoughts that are easier to have in one language vs another because the language naturally expresses an idea a particular way that is possible, but less natural to express in another.
puttycat•21h ago
Multilingual LLMs dont have a clear boundary between languages. They will appear to have one since they maximize likelihoods, so asking something in English will most likely produce an English continuation, etc.

In other circumstances they might take a different path (in terms of output probability decoding) through other character sets, if the probabilities justify this.

johnnyApplePRNG•21h ago
I understand that but how often/common could it possibly be to mix in a single bengali word/phrase like that into a larger english one?

Perhaps it's more common in the parts of the world where bengali and english are more commonly spoken in general?

Why so much bengali/hindi then and why not other languages?

epa•21h ago
There are many users in India training these models. There is also a lot more content out there the models are consuming.
groby_b•21h ago
And not to forget, many (most?) Indians are bilingual. Multilingual speakers tend to skip languages within conversation if both parties are fluent -> training material includes those switches.
daeken•20h ago
This has been really interesting to me. I've been learning Spanish for a while and will mix un poco español en my sentences with ChatGPT all the time, and it's cool to see the same thing reflected back to me. It's not uncommon for a response to be 75% in English with 25% Spanish at the beginnings and ends especially. All of my conversation titles are in Spanish because I always start them with "Hola", so whatever model sets the title just assumes Spanish for it, regardless of what the rest of the message is.
outside1234•19h ago
I’m so glad I am not the only one that does this!
pixl97•19h ago
I was on a vacation in the last week and was waiting at a restaurant for a while. The lady behind me was seemingly switching between english and spanish every few sentences which caught my attention. I can only assume the person on the other side was a bilingual also. Someone in their family had a medical emergency and was in the hospital. What's interesting is it seemed like the sentences talking about medical stuff were in very good english, and the spanish sentences were about other things (I can't interpret very fast at all). With the speed and fluency of their conversation it seemed like there was not any cost on their part using either language.
tstrimple•18h ago
Language is incredibly interesting to me. Especially when it’s blended or becomes its own pidgin dialect. Multilingual societies are fascinating.
tehlike•17h ago
I am bilingual.

My phrases switch to the language I learned them on very easily.

Computer terms are almost always English.

A lot of idioms I learned in my adult life are going to stay English, even if a Turkish equivalent exists and I later learned about them.

BrandoElFollito•10h ago
I am bilingual as well, my children are trilingual.

I find out that it is way easier for me to translate to or from English (not a native speaker) to any of the languages I am bilingual in, than between these languages. It is very hard for me to listen to one, and speak the other.

BrandoElFollito•10h ago
I spent a good time in the middle east and loved to listen to my friends arguing in Arabic.

To my French ear or sounded like they were sentencing me to terrible things (and were always surprised they sounded like this :)), up until the random "router" or "framework" which was the core of the fight.

I love to listen to languages I do not understand (a great source is Radio Green) and try to get from the words what they are talking about.

Another one is one of my closest friend, a German, who speaks a very soft English. This until he described me how to drive somewhere (pre-GPS era) and the names he was using were like lashes.

Speaking various languages is a blessing

ASalazarMX•19h ago
I usually interact with LLMs in English. A few weeks ago I made a Gemini gem that tries to consider two opposite sides, moderator included. Somehow it started including bits of Spanish in some of its answers, which I actually don't mind because that's my primary language.

I assumed it knew I speak Spanish from other conversations, my Google profile, geolocation, etc. Maybe my English has enough hints that it was learned by a native Spanish speaker?

hiAndrewQuinn•16h ago
There have been a nonzero number of times that asking Gemini about something in Finnish about the demoscene or early 1990s tech has returned much more... colorful answers than what I saw with equivalent questioning in English.
Vilian•16h ago
Colorful answers?
ipsum2•20h ago
Models like O3 are rewarded for the final output, not the intermediary thinking steps. So whatever it generates as "thoughts" that gives a better answer gets a higher score.

The DeepSeek-R1 paper has a section on this, where they 'punish' the model if it thinks in a different language to make the thinking tokens more readable. Probably Anthropic does this too.

jmward01•20h ago
It would be interesting to study when this type of behavior emerges to see what the patterns are. It could give insights into language or culture specific reasoning patterns and subjects that are easier to convey in one language or another. Is it easier to understand math word problems in XXX or YYY? What about relationships?
atlex2•20h ago
Definitely curious what circuits light-up from a Neuralese perspective. We want reasoning traces that are both faithful to the thought process and also interpretable. If the other language segments are lighting up meanings much different than their translations, that would raise questions for me.
tough•20h ago
I've seen also russian and chinese which i certainly have never speaked to it nor understand
janalsncm•20h ago
Others have mentioned that DeepSeek R1 also noticed this “problem”. I believe there are two things going on here.

One, the model is no longer being trained to output likely tokens or tokens likely to satisfy pairwise preferences. So the model doesn’t care. You have to explicitly punish the model for language switching, which dilutes the reasoning reward.

Two, I believe there has been some research on models representing similar ideas in multiple languages in similar areas. Sparse autoencoders have shown this. So if the translated text makes sense, I think this is why. If not, I have no idea.

NooneAtAll3•20h ago
I remember watching video mentioning it (https://www.youtube.com/shorts/Vv5Ia6C5vYk)

The main suspicion is that it's more compact?

neilv•19h ago
If the reasoning didn't need to be exposed to a user, are there any ways in which you get better performance or effect by using the same LLM methods, but using a language better suited to that? (Existing language or bespoke.)

(Inspired by movies and TV shows, when characters switch from English to a different language, such as French or Mandarin, to better express something. Maybe there's a compound word in German for that.)

Bjorkbat•19h ago
I don't actually think this is the case, but nonetheless I think it would be kind of funny if LLMs somehow "discovered" linguistic relativity (https://en.wikipedia.org/wiki/Linguistic_relativity).
mindcrime•19h ago
LLM's aren't humans and there's no reason to expect their "thinking"[1] to behave exactly - or even much - like human thinking. In particular, they don't need to "think" in one language. More concretely, in the DeepSeek R1 paper[2] they observed this "thought language mixing" and did some experiments on suppressing it... and the model results got worse. So I wouldn't personally think of it as an "error", but rather as just an artifact of how these things work.

[1]: By this I mean "whatever it is they do that can be thought of as sorta kind roughly analogous to what we generally call thinking." I'm not interested in getting into a debate (here) about the exact nature of thinking and whether or not it's "correct" to refer to LLM's as "thinking". It's a colloquialism that I find useful in this context, nothing more.

[2]: https://arxiv.org/pdf/2501.12948

diwank•18h ago
This isn’t entirely surprising. Language-model “reasoning” is basically the model internally exploring possibilities in token-space. These models are trained on enormous multilingual datasets and optimized purely for next-token prediction, not language purity. When reasoning traces or scratchpads are revealed directly (as OpenAI occasionally does with o-series models or DeepSeek-R1-zero), it’s common to see models slip into code-switching or even random language fragments, simply because it’s more token-efficient in their latent space.

For example, the DeepSeek team explicitly reported this behavior in their R1-zero paper, noting that purely unsupervised reasoning emerges naturally but brings some “language mixing” along. Interestingly, they found a small supervised fine-tuning (SFT) step with language-consistency rewards slightly improved readability, though it came with trade-offs (DeepSeek blog post).

My guess is OpenAI has typically used a smaller summarizer model to sanitize reasoning outputs before display (they mentioned summarization/filtering briefly at Dev Day), but perhaps lately they’ve started relaxing that step, causing more multilingual slips to leak through. It’d be great to get clarity from them directly on whether this is intentional experimentation or just a side-effect.

[1] DeepSeek-R1 paper that talks about poor readability and language mixing in R1-zero’s raw reasoning https://arxiv.org/abs/2501.12948

[2] OpenAI “Detecting misbehavior in frontier reasoning models” — explains use of a separate CoT “summarizer or sanitizer” before showing traces to end-users https://openai.com/index/chain-of-thought-monitoring/

rerdavies•18h ago
Reminds me of the son of a friend of mine, who was raised bilingually (English and French). When he was 3, he would sometimes ask "is this English, or the other language?"
CMCDragonkai•17h ago
Multilingual humans do this too, so not surprising that AI does this.
CMCDragonkai•17h ago
In fact monolingual humans have quite a limited understanding of the world.
nsonha•17h ago
No such thing as a monolingual human. Any language can be broken down to subsets that are associated with different ways of thinking. Another thing is globalization and culture export.
Incipient•15h ago
I know plenty of bilingual people that have a very limited understanding of the world, and conversely monolinguists that have a very broad view.

One could even say assuming someone's level of worldly understanding based on how many languages they speak shows a fairly limited world view.

ta20240528•15h ago
As I speaker of five languages, all but one fluently: why does my understanding of the world magically increase when I learn a new noun so say "sparrow" in the fifth that I'm learning?

Is it linear (25% more understanding for the fifth) or asymptotically? Does it increase across all domains equally (geology, poetry, ethics) or asymmetrically?

Seriously, explain it to me?

dpiers•16h ago
Languages are thought encodings.

Most people can only encode/decode a single language but an LLM can move between them fluidly.

muzani•15h ago
I do some AI training as a side gig and there has been a few recent updates on code-switching (i.e. speaking two languages at the same time) in the last few months. It's possible that these changes may have caused such behavior recently.
drivingmenuts•9h ago
I see this as a problem. You can't make an LLM "unlearn" something; once it's in there, it's in there. If I have a huge database, I can easily delete swathes of useless data, but I cannot do the same with an LLM. It's not a living, thinking being - it's a program running on a computer; a device that we, in other circumstances, can add information to or remove it from. We can suppress certain things, but that information is still in there, taking up space and can still possibly be accessed.

We are intentionally undoing one of the things that makes computers useful.

throwpoaster•8h ago
Multilingual humans do this too. Sometimes a concept is easier to shorthand in one language versus another. It’s somehow “closer”.