We built Anchor Engine because LLMs have no persistent memory. Every conversation is a fresh start—yesterday's discussion, last week's project notes, even context from another tab—all gone. Context windows help, but they're ephemeral and expensive.
The STAR algorithm (Semantic Traversal And Retrieval) takes a different approach. Instead of embedding everything into vector space, STAR uses deterministic graph traversal. But before traversal comes atomization—our lightweight process for extracting just enough conceptual structure from text to build a traversable semantic graph.
Atomization, not exhaustive extraction.
Projects like Kanon 2 are doing incredible work extracting every entity, citation, and clause from documents with remarkable precision. That's valuable for document intelligence. Anchor Engine takes a different path: we extract only the core concepts and relationships needed to support semantic memory. The result is a graph that's just rich enough for an LLM to retrieve relevant context, but lightweight enough to run offline in <3GB RAM—even on a Raspberry Pi or in a browser via WASM.
Why graph traversal instead of vector search?
Embeddings drift over time and across models
Similarity scores are opaque and nondeterministic
Vector search often requires GPUs or cloud APIs
You can't inspect why something was retrieved
STAR gives you deterministic, inspectable results. Same graph, same query, same output—every time. And because the graph is built through atomization, it stays small and portable.
Key technical details:
Runs entirely offline in <3GB RAM. No API calls, no GPUs.
Compiled to WASM – embed it anywhere, including browsers.
Recursive architecture – we used Anchor Engine to help write its own code. The dogfooding is real.
AGPL-3.0 – open source, always.
What it's not: It's not a replacement for LLMs or vector databases. It's a memory layer—a deterministic, inspectable substrate that gives LLMs persistent context without cloud dependencies. And it's not a competitor to deep extraction models like Kanon 2; they could even complement each other (Kanon 2 builds the graph, Anchor Engine traverses it for memory).
We're especially interested in feedback from people who've built RAG systems, experimented with symbolic memory, or worked on graph-based AI. If you've ever wanted to give an LLM long-term memory that fits on a Raspberry Pi, this might be interesting to you.
BERTmackliin•1h ago
The STAR algorithm (Semantic Traversal And Retrieval) takes a different approach. Instead of embedding everything into vector space, STAR uses deterministic graph traversal. But before traversal comes atomization—our lightweight process for extracting just enough conceptual structure from text to build a traversable semantic graph.
Atomization, not exhaustive extraction. Projects like Kanon 2 are doing incredible work extracting every entity, citation, and clause from documents with remarkable precision. That's valuable for document intelligence. Anchor Engine takes a different path: we extract only the core concepts and relationships needed to support semantic memory. The result is a graph that's just rich enough for an LLM to retrieve relevant context, but lightweight enough to run offline in <3GB RAM—even on a Raspberry Pi or in a browser via WASM.
Why graph traversal instead of vector search?
Embeddings drift over time and across models
Similarity scores are opaque and nondeterministic
Vector search often requires GPUs or cloud APIs
You can't inspect why something was retrieved
STAR gives you deterministic, inspectable results. Same graph, same query, same output—every time. And because the graph is built through atomization, it stays small and portable.
Key technical details:
Runs entirely offline in <3GB RAM. No API calls, no GPUs.
Compiled to WASM – embed it anywhere, including browsers.
Recursive architecture – we used Anchor Engine to help write its own code. The dogfooding is real.
AGPL-3.0 – open source, always.
What it's not: It's not a replacement for LLMs or vector databases. It's a memory layer—a deterministic, inspectable substrate that gives LLMs persistent context without cloud dependencies. And it's not a competitor to deep extraction models like Kanon 2; they could even complement each other (Kanon 2 builds the graph, Anchor Engine traverses it for memory).
The whitepaper goes deep on the graph traversal math and includes benchmarks vs. vector search: https://github.com/RSBalchII/anchor-engine-node/blob/d9809ee...
We're especially interested in feedback from people who've built RAG systems, experimented with symbolic memory, or worked on graph-based AI. If you've ever wanted to give an LLM long-term memory that fits on a Raspberry Pi, this might be interesting to you.