How Cilow works

Cilow is built around a tri-store architecture: a vector index for semantic similarity, a temporal knowledge graph for relationships and timeline, and a structured key-value store for fast lookup. A unified query planner across all three assembles the minimal context a model needs for any inference call.

Write path: Data sources → Ingest → Extract → Structure → Index ↓ Query path: Intent → Multi-lane retrieval → Fusion → Assembly → Model ↑ Memory: Reinforcement / Decay / Supersession

Three layers of context

Write path

01
Ingest

Accepts documents, chats, code, APIs, and structured data — any source, any format.

02
Extract

Entity recognition, relationship extraction, and timeline anchoring from raw input.

03
Structure

Builds the knowledge graph, indexes vectors, and updates the key-value store.

04
Index

All three stores updated atomically — vector, graph, and structured in one pass.

Query path

01
Intent classification

Routes the query to the right retrieval lanes based on what kind of answer is needed.

02
Multi-lane retrieval

Vector, graph, and structured retrieval run in parallel — each contributing different signal.

03
Fusion

N-way result fusion surfaces the best candidates from across all three lanes.

04
Assembly

Builds the minimal working set for the current inference call — nothing more.

Memory lifecycle

01
Reinforcement

Context that proves useful gets stronger signal — the system learns what helps.

02
Decay

Context that is never retrieved or acted on fades over time rather than accumulating as noise.

03
Supersession

Newer, more specific facts retire older ones — no contradictions accumulate silently.

The storage layer

Vector index

Semantic similarity search using high-dimensional embeddings. Retrieves what is conceptually close to the query.

Temporal knowledge graph

Entities, relationships, and how they change over time — the structural layer that RAG does not have. Tracks when facts held true and when they were superseded.

Structured key-value store

Fast exact lookup, scalar filters, and secondary indexes. The right tool when you need precision over similarity.

Why three stores and not one: each store handles a different kind of retrieval need. Combining them with a query planner gives better results than any single store alone — because semantic similarity, structural relationships, and exact lookup are three distinct problems.

How Cilow scores context

Cilow ranks candidates across multiple signals before assembly. No single signal dominates — the system combines them to surface what is genuinely useful for the current call.

Recency

More recent information scores higher by default. Information that was true last week ranks above information that was true two years ago.

Relevance

Semantic match to the current query — the signal vector search specializes in.

Frequency

Context used often by the model gets reinforced. Repeated utility is a strong signal of lasting value.

Causal role

Information that led to a correct outcome is weighted up. The system tracks which facts contributed to good answers.

Past utility

The system learns over time what kinds of context actually helped, and adjusts scoring accordingly.

Context assembly

The query planner builds the working set from candidates across all three stores — vector, graph, and structured retrieval contribute in parallel.
Assembly is constrained: the goal is the minimal set that answers the question, not the maximal set. Everything that does not contribute is excluded.
Smaller working sets mean lower latency, lower cost, and less noise for the model to reason through.

Benchmarks

In LongMemEval testing, Cilow achieved 94.17% accuracy (113/120) — see the full results.

→ Benchmarks

Ready to build on top of Cilow's tri-store architecture?

Start building → Docs
Cilow