System Architecture Overview
m1nd is a local MCP runtime for coding agents. Internally it is still built on a graph-and-analysis system, but the public runtime is organized around an operational contract:
- graph-grounded retrieval and ranking
- proof-aware structural flows
- next-step handoff on high-value tools
- continuity through trails and perspectives
- observable multi-file writes through
apply_batch - repair-friendly errors when a tool call goes wrong
The workspace exposes the live MCP tool surface over JSON-RPC stdio, with an HTTP/UI surface in the default build. Use tools/list for the exact count in your current build.
Four-Crate Workspace
The system is organized as a Cargo workspace with three core crates plus one auxiliary bridge crate:
[workspace]
members = ["m1nd-core", "m1nd-ingest", "m1nd-mcp", "m1nd-openclaw"]
resolver = "2"
[workspace.dependencies]
thiserror = "2"
serde = { version = "1", features = ["derive"] }
serde_json = "1"
smallvec = { version = "1.13", features = ["serde"] }
parking_lot = "0.12"
rayon = "1.10"
static_assertions = "1"
| Crate | Role | Key Dependency |
|---|---|---|
| m1nd-core | Graph engine, activation, plasticity, XLR, resonance, temporal, semantic | parking_lot, smallvec, static_assertions |
| m1nd-ingest | File walking, language-specific extraction, reference resolution, diff | rayon, walkdir, regex |
| m1nd-mcp | JSON-RPC transport, tool dispatch, session management, persistence | tokio, serde_json |
| m1nd-openclaw | Auxiliary OpenClaw bridge and transport integration | m1nd-core, m1nd-mcp |
Dependencies flow strictly downward: m1nd-mcp depends on both m1nd-core and m1nd-ingest; m1nd-ingest depends on m1nd-core; m1nd-core has no internal crate dependencies.
Data Flow
graph TD
subgraph "m1nd-mcp (Transport)"
STDIO["JSON-RPC stdio<br/>dual: framed + line"]
DISPATCH["Tool Dispatch<br/>live tool surface"]
SESSION["SessionState<br/>SharedGraph + Engines"]
PERSIST["Auto-Persist<br/>every 50 queries"]
end
subgraph "m1nd-ingest (Extraction)"
WALK["DirectoryWalker<br/>walkdir + binary detect"]
EXTRACT["Parallel Extraction<br/>rayon threadpool"]
RESOLVE["ReferenceResolver<br/>proximity disambiguation"]
DIFF["GraphDiff<br/>incremental updates"]
MEMORY["MemoryAdapter<br/>markdown → graph"]
UNIVERSAL["Universal + Document Adapters<br/>canonical docs → graph"]
end
subgraph "m1nd-core (Engine)"
GRAPH["CSR Graph<br/>forward + reverse"]
NODES["NodeStorage (SoA)<br/>hot|warm|cold paths"]
ACT["Activation Engine<br/>hybrid heap/wavefront"]
SEM["Semantic Engine<br/>trigram TF-IDF + PPMI"]
TEMP["Temporal Engine<br/>co-change + decay"]
PLAST["Plasticity Engine<br/>Hebbian LTP/LTD"]
XLR["XLR Engine<br/>spectral noise cancel"]
RES["Resonance Engine<br/>standing waves"]
SNAP["Snapshot<br/>atomic JSON persist"]
end
STDIO --> DISPATCH
DISPATCH --> SESSION
SESSION --> PERSIST
PERSIST --> SNAP
SESSION -->|ingest| WALK
WALK --> EXTRACT
EXTRACT --> RESOLVE
RESOLVE --> GRAPH
UNIVERSAL --> GRAPH
DIFF -->|incremental| GRAPH
SESSION -->|activate| ACT
SESSION -->|impact/predict| TEMP
SESSION -->|learn| PLAST
SESSION -->|activate+xlr| XLR
SESSION -->|resonate| RES
ACT --> GRAPH
ACT --> SEM
ACT --> NODES
SEM --> GRAPH
TEMP --> GRAPH
PLAST --> GRAPH
XLR --> GRAPH
RES --> GRAPH
GRAPH --> NODES
Request Lifecycle
A typical structural workflow now has two layers: graph execution and agent handoff.
Example: activate
- Transport: JSON-RPC message arrives on stdin (either Content-Length framed or raw line JSON).
- Dispatch:
McpServer.serve()parses the JSON-RPC request, matches the tool name, extracts parameters. - Session: The tool handler acquires a read lock on
SharedGraph(Arc<parking_lot::RwLock<Graph>>). - Seed Finding:
SeedFinderlocates matching nodes via a 5-level cascade: exact label, prefix, substring, tag, fuzzy trigram. - Activation:
HybridEngineauto-selects heap or wavefront strategy based on seed ratio and average degree. - Dimensions: Four dimensions run: Structural (BFS/heap propagation), Semantic (trigram TF-IDF + co-occurrence PPMI), Temporal (decay + velocity), Causal (forward/backward with discount).
- Merge:
merge_dimensions()combines results with adaptive weights[0.35, 0.25, 0.15, 0.25]and resonance bonus (4-dim: 1.5x, 3-dim: 1.3x). - XLR: If enabled,
AdaptiveXlrEngineruns spectral noise cancellation with dual hot/cold pulses and sigmoid gating. - Plasticity:
PlasticityEngine.update()runs the 5-step Hebbian cycle on co-activated edges. - Response: Results serialized to JSON-RPC response, written to stdout in the same transport mode as the request.
Example: guided structural flow
A more agent-native flow such as trace -> view -> surgical_context_v2 -> validate_plan -> apply_batch adds two extra kinds of behavior on top of the graph work:
- Cognitive state: tools can expose
proof_statesuch astriaging,proving, orready_to_edit - Next-step guidance: tools can return
next_suggested_tool,next_suggested_target, andnext_step_hint
This is one of the biggest architectural changes in v0.6.0: the MCP layer is no longer just a transport for graph answers. It is a transport for graph answers plus workflow guidance.
Example: code plus document workflow
The current architecture also supports a second operating loop:
- ingest code
- ingest or auto-ingest document roots through the universal lane
- resolve canonical document artifacts
- bind documents to likely code
- inspect document/code drift after implementation moves
This is what turns m1nd from a code-only graph into a local code + knowledge runtime.
Key Design Decisions
Compressed Sparse Row (CSR) Graph
The graph uses CSR format rather than adjacency lists or adjacency matrices. CSR provides O(1) neighbor iteration start, cache-friendly sequential access, and compact memory layout. Edge weights are stored as AtomicU32 (bit-reinterpreted f32) for lock-free plasticity updates via CAS with a 64-retry limit.
Forward and reverse CSR arrays are maintained in parallel, enabling both outgoing and incoming edge traversal without full graph scans.
Struct-of-Arrays Node Storage
NodeStorage uses SoA layout with explicit hot/warm/cold path separation:
- Hot path (every query):
activation: Vec<[FiniteF32; 4]>,pagerank: Vec<FiniteF32> - Warm path (plasticity):
plasticity: Vec<PlasticityNode> - Cold path (seed finding, export):
label,node_type,tags,last_modified,change_frequency,provenance
This layout ensures that activation queries touch only hot-path arrays, maximizing L1/L2 cache utilization.
FiniteF32 Type Safety
All floating-point values in the graph are wrapped in FiniteF32, a #[repr(transparent)] newtype that makes NaN and Infinity impossible by construction:
#![allow(unused)]
fn main() {
#[derive(Clone, Copy, Default, PartialEq)]
#[repr(transparent)]
pub struct FiniteF32(f32);
impl FiniteF32 {
#[inline]
pub fn new(v: f32) -> Self {
debug_assert!(v.is_finite(), "FiniteF32::new received non-finite: {v}");
Self(if v.is_finite() { v } else { 0.0 })
}
}
}
Because NaN is excluded, FiniteF32 can safely implement Ord, Eq, and Hash – operations that are unsound on raw f32. Related newtypes PosF32 (strictly positive), LearningRate (0, 1]), and DecayFactor (0, 1]) provide compile-time invariant enforcement for their respective domains.
String Interning
All node labels, tags, and relation names are interned via StringInterner. Once interned, string comparisons become u32 == u32 (zero-allocation, single CPU cycle). The interner maps strings to InternedStr(u32) handles and provides O(1) bidirectional lookup.
Parallel Extraction, Sequential Building
Ingestion uses rayon for parallel file reading and language-specific extraction across all CPU cores, but graph construction is single-threaded. This avoids the complexity of concurrent graph mutation while still saturating I/O bandwidth during the most expensive phase (parsing hundreds of files).
Atomic Persistence
Graph and plasticity state are saved via atomic write: serialize to a temporary file, then rename() over the target. This prevents corruption on crash (FM-PL-008). The NaN firewall at the export boundary rejects any non-finite values that might have leaked through.
Performance Characteristics
There are two truths worth separating:
- engine/runtime timings
- workflow-level benchmark truth
Engine timings still matter, especially for ingest and structural queries. Workflow benchmarks matter when the goal is to reduce context churn, false starts, and repair loops.
Representative measured engine timings:
| Operation | Time | Notes |
|---|---|---|
| Full ingest | ~910ms | Walk + parallel extract + resolve + finalize (CSR + PageRank) |
| Activate query | ~31ms | 4-dimension with XLR, top-20 results |
| Impact analysis | ~5ms | BFS blast radius, 3-hop default |
| Trace analysis | ~3.5ms | Stacktrace to ranked suspects |
| Trail resume | ~0.2ms | Restore continuity and next-step hints |
| Apply batch | ~165ms | Atomic multi-file write before deeper verification |
Workflow benchmark truth lives on the Benchmarks page. The current recorded warm-graph corpus shows:
10518 -> 5182aggregate token proxy50.73%aggregate reduction14 -> 0false starts39guided follow-throughs12successful recovery loops
Memory footprint scales linearly with graph size. A 10K-node graph with 25K edges uses approximately 15MB of heap (graph + all engine indexes). The CSR representation is 3-5x more compact than equivalent adjacency list representations.
Scaling Bounds
- Node limit:
NodeId(u32)supports up to ~4 billion nodes. Practical limit is themax_nodesconfig (default: 500K). - Edge weights:
AtomicU32CAS with 64-retry limit. Under high contention (>32 concurrent plasticity updates on the same edge), CAS may exhaust retries and returnCasRetryExhausted. - Co-occurrence index: Disabled above 50K nodes (
COOCCURRENCE_MAX_NODES) to avoid O(N * walks * length) random walk cost. - Co-change matrix: Hard budget of 500K entries with per-row cap of 100 entries.
- Resonance pulse budget: 50K pulses per standing wave analysis to prevent combinatorial explosion in dense subgraphs.
Multi-Agent Model
m1nd runs as a single process serving multiple agents through the same JSON-RPC stdio channel. All agents share one graph and one set of engines. Writes (plasticity updates, ingestion) are immediately visible to all readers through SharedGraph = Arc<parking_lot::RwLock<Graph>>.
Each agent gets its own AgentSession tracking first/last seen timestamps and query count. The perspective system, trail system, and shared runtime state provide isolation where needed without duplicating the underlying graph.
The important modern detail is that agent-facing state is not limited to ownership and transport. Several flows now expose cognitive runtime state directly, including:
proof_state- next-step guidance fields
- trail continuation hints
- batch progress and handoff metadata
parking_lot::RwLock is used instead of std::sync::RwLock to prevent writer starvation – a critical property when plasticity updates (writes) must interleave with activation queries (reads).