IDEs
MCP tools for recall before a task and store after, without changing the flow teams already use.
Reasoning traces that carry forward: shorter runs, sharper answers on repeat work, and less token spend on solved paths through recall, inject, and store.
Overview
TraceBase captures every solved problem as a reasoning trace and feeds it back into future runs. Your agents don't just execute — they accumulate expertise. Every run is built on every run before it.
MCP tools for recall before a task and store after, without changing the flow teams already use.
Wrap OpenAI or Anthropic clients once; middleware runs recall, then your completion, then store.
Similar incidents surface prior traces instead of forcing the model to reason from zero each time.
Local SQLite by default, with optional embeddings when you want broader semantic retrieval.
TraceBase is included in the Daytona Startup Grid — Daytona's program backing early teams that ship AI-native developer infrastructure.
Before / After
TraceBase does not change the model choice. It changes what the model starts with and what the next run gets back after a fix has already shipped.
Same model, no layer
Cold path steps
No recall path means every repeat case still pays for fresh exploration. The model can solve it, but it does not start from what already worked.
With TraceBase
TraceBase flow steps
Recall surfaces strong matches before tokens hit the model, then successful runs are written back so the next repeat case starts grounded instead of blank.
How it works
Recall before the call.
The middleware sits between your code and the LLM. Before each call it checks memory. After each call it stores the result. No manual work.
01
Check memory for similar problems solved before.
02
Add prior solution to system prompt as a hint.
03
LLM solves faster with context. Fewer tokens.
04
New trace captured. Memory grows automatically.
Integrations
Works with any agent.
MCP for IDE-native agents, middleware for wrapped SDKs, and a clean path for custom runtimes when retrieval needs to stay under your control.


Under the hood
Fingerprint, BM25, Jaccard, structural, cosine. Two-stage retrieval.
Thompson Sampling learns optimal signal weights from your feedback.
Middleware recalls and injects prior solutions automatically.
Full stream:true support. Traces captured after stream completes.
SQLite with WAL. Sub-millisecond recall. Data stays on your machine.
Optional cosine similarity via OpenAI text-embedding-3-small.
Setup
Three ways.
Pick the layer that fits your stack: wrapped SDKs, one-command IDE rollout, or direct control inside custom agents.
OpenAI / Anthropic
Wrap the client once and let recall run before each call while store captures the resolved path after completion.
const layer = new ReasoningLayer() const openai = wrapOpenAI(new OpenAI(), layer) await openai.chat.completions.create(...)
Custom integrations
When your models already sit inside a larger workflow, the right move is usually not another generic widget. It's a reasoning layer shaped around the way your pipeline actually behaves.
We can plug retrieval, injection, and memory capture into existing agent loops, internal tools, or product surfaces without forcing a rewrite.
01
We identify where your agents repeat work, where prompts drift, and where cost spikes come from avoidable re-exploration.
02
SDK middleware, MCP tools, or a custom orchestration layer — whatever matches the way your product already routes model calls.
03
Resolved paths get stored, recall quality improves, and repeat workflows start grounded instead of blank.
Pricing
Open source now.
Self-hosted is available today. The startup and enterprise tiers below are planned managed rollout packaging, shown here as draft launch pricing rather than live checkout.
Choose billing cadence
Annual keeps pricing a little lower for teams committing to a production rollout.
Self-hosted memory, MCP, SDK middleware, and local SQLite storage.
Managed traces, analytics, and team access for small product teams shipping weekly.
Private deployment, SSO, and support for higher-volume or regulated environments.
One install. Agents that get better with every run.