frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Show HN: Continual Learning with .md

https://github.com/SunAndClouds/ReadMe
23•wenhan_zhou•9h ago
I have a proposal that addresses long-term memory problems for LLMs when new data arrives continuously (cheaply!). The program involves no code, but two Markdown files.

For retrieval, there is a semantic filesystem that makes it easy for LLMs to search using shell commands.

It is currently a scrappy v1, but it works better than anything I have tried.

Curious for any feedback!

Comments

sudb•8h ago
I really like the simplicity of this! What's retrieval performance and speed like?
wenhan_zhou•35m ago
Minimalism is my design philosophy :-)

Good question. Since it is just an LLM reading files, it depends entirely on how fast it can call tools, so it depends on the token/s of the model.

Haven't done a formal benchmark, but from the vibes, it feels like a few seconds for GPT-5.4-high per query.

There is an implicit "caching" mechanism, so the more you use it, the smoother it will feel.

namanyayg•8h ago
I've seen a lot of such systems come and go. One of my friends is working on probably the best (VC-funded) memory system right now.

The problem always is that when there are too many memories, the context gets overloaded and the AI starts ignoring the system prompt.

Definitely not a solved problem, and there need to be benchmarks to evaluate these solutions. Benchmarks themselves can be easily gamed and not universally applicable.

xwowsersx•5h ago
What is the memory system you are referring to? I've been trying Memori with OpenClaw. Haven't had a ton of time to really kick the tires on it, so the jury's still out.
natpalmer1776•5h ago
The armchair ML engineer in me says our current context management approach is the issue. With a proper memory management system wired up to it’s own LLM-driven orchestrator, memories should be pulled in and pushed out between prompts, and ideally, in the middle of a “thinking” cycle. You can enhance this to be performant using vector databases and such but the core principle remains the same and is oft repeated by parents across the world: “Clean up your toys before you pull a new one out!”

Also since I thought for another 30 seconds, the “too many memories!” Problem imo is the same problem as context management and compaction and requires the same approach: more AI telling AI what AI should be thinking about. De-rank “memories” in the context manager as irrelevant and don’t pass them to the outer context. If a memory is de-ranked often and not used enough it gets purged.

dummydummy1234•4h ago
Mid thinking cycle seems dangerous as it will probably kill caching.
natpalmer1776•4h ago
The mid thinking cycle would require significant architecture change to current state of art and imo is a key blocker to AGI
wenhan_zhou•45m ago
Fair concern.

ReadMe does support loading memories mid-reasoning! It is simply an agent reading files.

Although GPT-5.4 currently likes to explore a lot upfront, and only then responds. But that is more of a model behaviour (adjustable through prompting) rather than an architectural limitation.

natpalmer1776•43m ago
Ah, I mean bi-directional management of context. Add and remove. Basically just the remove bit since we have adding down.
wenhan_zhou•51m ago
Context bloat is real, but the architecture has the potential to solve it.

You need clever naming for the filesystem and exploration policy in AGENTS.md. (not trivial!)

The benchmark is definitely the core bottleneck. I don't know any good benchmark for this, probably an open research question in itself.

alexbike•7h ago
The markdown approach has a real advantage people underestimate: you can read and edit the memory yourself. With vector DBs and embeddings the memory becomes opaque — you can't inspect or correct what the model "knows". Plain files keep the human in the loop.

The hard part is usually knowing what +not+ to write down. Every system I've seen eventually drowns in low-signal entries.

in-silico•7h ago
This assumes that the model's behavior and memories are faithful to their english/human language representation, and don't stray into (even subtle) "neuralese".
verdverm•6h ago
Is there anything (besides plumbing) that prevents both? i.e. when the file is edited, all the representations are updated
wenhan_zhou•59m ago
The editability is surely an underrated advantage, both for the program itself and the memories it generated.

I think in terms of noise, it is less problematic here because not everything is being retrieved. The agent can selectively explore subsets of the tree (plus you can edit the exploration policy by yourself).

Since there is no context bloat, it is quite forgivable to just write things down.

dhruv3006•4h ago
I love how you approached this with markdown !

I guess the markdown approach really has a advantage over others.

PS : Something I built on markdown : https://voiden.md/

wenhan_zhou•42m ago
Yep. Markdown is the future :-)

Show HN: Ithihāsas – a character explorer for Hindu epics, built in a few hours

https://www.ithihasas.in
137•cvrajeesh•10h ago•33 comments

Show HN: Deflect One – command line dashboard for managing Linux servers via SSH

https://github.com/Frytskyy/deflect-one
2•whitemanv•1h ago•1 comments

Show HN: Continual Learning with .md

https://github.com/SunAndClouds/ReadMe
23•wenhan_zhou•9h ago•16 comments

Show HN: Prmana – OIDC SSH Login for Linux with DPoP (Rust, Apache 2.0)

https://github.com/prodnull/prmana
3•cbchhaya•3h ago•0 comments

Show HN: boringBar – a taskbar-style dock replacement for macOS

https://boringbar.app/
503•a-ve•1d ago•287 comments

Show HN: Excalicharts – Charting Library for Excalidraw

https://github.com/tombedor/excalicharts
4•jjfoooo4•5h ago•0 comments

Show HN: Mcptube – Karpathy's LLM Wiki idea applied to YouTube videos

https://github.com/0xchamin/mcptube
7•0xchamin•13h ago•0 comments

Show HN: Lythonic – Compose Python functions into data-flow pipelines

https://github.com/walnutgeek/lythonic
4•walnutgeek•9h ago•0 comments

Show HN: I built a social media management tool in 3 weeks with Claude and Codex

https://github.com/brightbeanxyz/brightbean-studio
177•JanSchu•20h ago•120 comments

Show HN: OQP – A verification protocol for AI agents

https://github.com/OranproAi/open-qa-protocol
5•Aamir21•7h ago•1 comments

Show HN: Oberon System 3 runs natively on Raspberry Pi 3 (with ready SD card)

https://github.com/rochus-keller/OberonSystem3Native/releases
237•Rochus•1d ago•93 comments

Show HN: Claudraband – Claude Code for the Power User

https://github.com/halfwhey/claudraband
117•halfwhey•1d ago•42 comments

Show HN: Equirect – a Rust VR video player

https://github.com/greggman/equirect
11•greggman65•21h ago•1 comments

Show HN: Farchive – SQLite-backed history-preserving compressed archive

https://github.com/eliask/farchive
4•ekns•13h ago•0 comments

Show HN: Deconflict – Open-source WiFi planner with physics-based walls

https://deconflict.pages.dev
3•s_e__a___n•9h ago•1 comments

Show HN: Pardonned.com – A searchable database of US Pardons

496•vidluther•2d ago•272 comments

Show HN: I benchmarked Gemma 4 E2B – the 2B model beat the 12B on multi-turn

https://aiexplr.com/post/gemma-4-e2b-benchmark
6•mailharishin•10h ago•0 comments

Show HN: GDL – I built an AI-powered invention engine

https://groundeddiscoverylabs.com/
2•Whyachi•10h ago•0 comments

Show HN: Lint-AI by RooAGI, a Rust CLI for AI Doc Retrieval

https://github.com/RooAGI/Lint-AI
3•rooagi•10h ago•0 comments

Show HN: Encrypted, nothing stored, nothing repeated face-gated asset sharing

https://veylt.net/
4•bekkerza•12h ago•0 comments

Show HN: pg_grpc – Call gRPC services directly from PostgreSQL

https://github.com/CSenshi/pg_grpc
7•csenshi•12h ago•1 comments

Show HN: Hitoku Draft – context aware local macOS assistant

https://github.com/Saladino93/hitokudraft
3•lostathome•13h ago•0 comments

Show HN: Asthi – Damn good asset tracker

https://www.asthi.app/
3•suvamsh•13h ago•0 comments

Show HN: Bloomberg Terminal for LLM ops – free and open source

7•amans9712•13h ago•1 comments

Show HN: Aeolus – a library for unified access to air quality sensor networks

https://github.com/southlondonscientific/aeolus
3•ruaraidh•13h ago•0 comments

Show HN: Dbg – One CLI debugger for every language (AI-agent ready)

https://redknightlois.github.io/dbg/
4•redknight666•13h ago•0 comments

Show HN: FluidCAD – Parametric CAD with JavaScript

https://fluidcad.io/
155•maouida•3d ago•37 comments

Show HN: 15 yrs of Django in prod: patterns I keep using (agent skills)

https://github.com/dvf/opinionated-django
6•vanflymen•14h ago•1 comments

Show HN: Rekal – Long-term memory for LLMs in a single SQLite file

https://github.com/janbjorge/rekal
7•jeeybee•1d ago•8 comments

Show HN: Crafto – AI carousel post generator for LinkedIn and Instagram

https://crafto.ai
4•xz_tie•15h ago•1 comments