frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Teaching LLMs to compose math symbolically, not execute it

2•CheerfulDreamer•2mo ago
Right now LLMs cannot be counted on to perfectly perform math. The solution I propose is to teach LLMs to instead of executing the math, just compose the mathematical equations correctly and leave the execution to a post-processing step.

My core method would be: Use a single special token ᶜ (U+1D9C) before each element that needs computation, and then compute the result afterwards. For known math that doesn't need to be computed the ᶜ is not added or present.

Thus we would see in the output:

Normal (already computed): 847 * 293 = 248171

Requesting computation: ᶜ847 ᶜ* ᶜ293 ᶜ= ᶜx

The Core Mechanic: Post-Process Computation

This is what makes everything work: Model generates output with ᶜ-marked expressions (fast, no blocking) Generation completes Parse all ᶜ-marked expressions Execute computations with perfect precision Substitute results back into the output Show user the final result with normal mathematical notation

The model never waits for computation results. It reasons symbolically with variables, and values are computed after generation is complete.

Multi-step example:

Model generates: "First, ᶜ847 ᶜ* ᶜ293 ᶜ= ᶜa, then ᶜa ᶜ+ ᶜ150 ᶜ= ᶜb. The answer is ᶜb."

Post-processing: - Execute: 847 * 293 = 248171 (bind to 'ᶜa') - Execute: 248171 + 150 = 248321 (bind to 'ᶜb') - Substitute: ...

User sees: "First, 847 * 293 = 248171, then 248171 + 150 = 248321. The answer is 248321." This is how the model can compose complex calculations without blocking - it's just manipulating symbols, and we handle execution separately.

Training Path 1: During Base Model Training

We augment the training set such that:

Instructional ("Calculate 847 × 293") → add ᶜ tokens Expository ("The result 847 × 293 = 248,171 shows...") → leave as-is

The model learns both patterns during pretraining. When it generates ᶜ-marked expressions during training, they get post-processed (executed and substituted) before computing the loss. The model learns that ᶜ notation leads to computed results.

Training Path 2: Fine-Tuning Existing Models

*If you already have a trained base model:*

1. *Add ᶜ token to vocabulary* 2. *Generate synthetic training data:* ``` Q: "What is 847 multiplied by 293?" A: "Let me calculate: ᶜ847 ᶜ* ᶜ293 ᶜ= ᶜx. The result is x."

Post-process: → "Let me calculate: 847 * 293 = 248171. The result is 248171."

Train on the post-processed version.

Loss and rewards:

High penalty: arithmetic errors without using ᶜ Small penalty: unnecessary ᶜ use (like for 2+2) Reward: correct ᶜ usage and accurate composition

The model learns: "I already know math notation from base training. Now I'm learning to mark computations with ᶜ and let the execution engine handle them."

Fine-tuning is faster since the model already understands mathematical notation - you're just teaching when to use the ᶜ pattern.

Why This Works Separation of concerns:

Model: mathematical composition, when to calculate, symbolic reasoning Execution engine: precise arithmetic, guaranteed correctness Post-processing is the key: The model never waits for results during generation. It composes symbolically, we compute separately. The model doesn't waste parameters learning that 847 × 293 = 248,171. It learns "multiplication is needed here" and delegates execution.

Extensions Same pattern for any deterministic operation:

Dates: ᶜdate_2023 ᶜ- ᶜdate_2022 ᶜ= ᶜdays

Counting: ᶜcount ᶜ( ᶜlist ᶜ) ᶜ= ᶜn

Memory: ᶜstore ᶜ( ᶜslot ᶜ, ᶜvalue ᶜ)

Public Domain Anyone may use, implement, modify, or build upon this approach for any purpose, commercial or non-commercial, without restriction. I specifically disclaim any patent rights and intend this publication to serve as prior art preventing future patent restrictions.

My goal is to help advance AI capabilities in a way that benefits everyone. All praise to Jesus and God who created this amazing universe for us to enjoy.

AI Overviews are killing the web search, and there's nothing we can do about it

https://www.neowin.net/editorials/ai-overviews-are-killing-the-web-search-and-theres-nothing-we-c...
2•bundie•3m ago•0 comments

City skylines need an upgrade in the face of climate stress

https://theconversation.com/city-skylines-need-an-upgrade-in-the-face-of-climate-stress-267763
2•gnabgib•4m ago•0 comments

1979: The Model World of Robert Symes [video]

https://www.youtube.com/watch?v=HmDxmxhrGDc
1•xqcgrek2•9m ago•0 comments

Satellites Have a Lot of Room

https://www.johndcook.com/blog/2026/02/02/satellites-have-a-lot-of-room/
1•y1n0•9m ago•0 comments

1980s Farm Crisis

https://en.wikipedia.org/wiki/1980s_farm_crisis
3•calebhwin•10m ago•1 comments

Show HN: FSID - Identifier for files and directories (like ISBN for Books)

https://github.com/skorotkiewicz/fsid
1•modinfo•15m ago•0 comments

Show HN: Holy Grail: Open-Source Autonomous Development Agent

https://github.com/dakotalock/holygrailopensource
1•Moriarty2026•22m ago•1 comments

Show HN: Minecraft Creeper meets 90s Tamagotchi

https://github.com/danielbrendel/krepagotchi-game
1•foxiel•29m ago•1 comments

Show HN: Termiteam – Control center for multiple AI agent terminals

https://github.com/NetanelBaruch/termiteam
1•Netanelbaruch•30m ago•0 comments

The only U.S. particle collider shuts down

https://www.sciencenews.org/article/particle-collider-shuts-down-brookhaven
2•rolph•32m ago•1 comments

Ask HN: Why do purchased B2B email lists still have such poor deliverability?

1•solarisos•33m ago•2 comments

Show HN: Remotion directory (videos and prompts)

https://www.remotion.directory/
1•rokbenko•35m ago•0 comments

Portable C Compiler

https://en.wikipedia.org/wiki/Portable_C_Compiler
2•guerrilla•37m ago•0 comments

Show HN: Kokki – A "Dual-Core" System Prompt to Reduce LLM Hallucinations

1•Ginsabo•37m ago•0 comments

Software Engineering Transformation 2026

https://mfranc.com/blog/ai-2026/
1•michal-franc•39m ago•0 comments

Microsoft purges Win11 printer drivers, devices on borrowed time

https://www.tomshardware.com/peripherals/printers/microsoft-stops-distrubitng-legacy-v3-and-v4-pr...
3•rolph•39m ago•1 comments

Lunch with the FT: Tarek Mansour

https://www.ft.com/content/a4cebf4c-c26c-48bb-82c8-5701d8256282
2•hhs•42m ago•0 comments

Old Mexico and her lost provinces (1883)

https://www.gutenberg.org/cache/epub/77881/pg77881-images.html
1•petethomas•45m ago•0 comments

'AI' is a dick move, redux

https://www.baldurbjarnason.com/notes/2026/note-on-debating-llm-fans/
5•cratermoon•47m ago•0 comments

The source code was the moat. But not anymore

https://philipotoole.com/the-source-code-was-the-moat-no-longer/
1•otoolep•47m ago•0 comments

Does anyone else feel like their inbox has become their job?

1•cfata•47m ago•1 comments

An AI model that can read and diagnose a brain MRI in seconds

https://www.michiganmedicine.org/health-lab/ai-model-can-read-and-diagnose-brain-mri-seconds
2•hhs•50m ago•0 comments

Dev with 5 of experience switched to Rails, what should I be careful about?

2•vampiregrey•53m ago•0 comments

AlphaFace: High Fidelity and Real-Time Face Swapper Robust to Facial Pose

https://arxiv.org/abs/2601.16429
1•PaulHoule•54m ago•0 comments

Scientists discover “levitating” time crystals that you can hold in your hand

https://www.nyu.edu/about/news-publications/news/2026/february/scientists-discover--levitating--t...
3•hhs•56m ago•0 comments

Rammstein – Deutschland (C64 Cover, Real SID, 8-bit – 2019) [video]

https://www.youtube.com/watch?v=3VReIuv1GFo
1•erickhill•56m ago•0 comments

Tell HN: Yet Another Round of Zendesk Spam

5•Philpax•56m ago•1 comments

Postgres Message Queue (PGMQ)

https://github.com/pgmq/pgmq
1•Lwrless•1h ago•0 comments

Show HN: Django-rclone: Database and media backups for Django, powered by rclone

https://github.com/kjnez/django-rclone
2•cui•1h ago•1 comments

NY lawmakers proposed statewide data center moratorium

https://www.niagara-gazette.com/news/local_news/ny-lawmakers-proposed-statewide-data-center-morat...
2•geox•1h ago•0 comments