Free
For prototyping and exploration
- 5K captures/month
- 75 synthesize calls/day
- 7 learn calls/day
- 3 agents, 1 API key
- 5K stored memories
- Instant raw memory search
- Auto-instrumentation
- Community support
Capture events → compile knowledge → return compressed reasoning context.
import { AgentMemory } from "@hippocortex/sdk"
const memory = new AgentMemory()
memory.capture(event)
const context = await memory.synthesize(query)
agent.run(context)Each node in the brain represents a stage in the memory pipeline. Events flow through capture, learning, compilation, prediction, transfer, and synthesis.
The pattern extraction engine replays event traces, identifies recurring procedures, detects contradictions, and promotes high-confidence patterns to compiled knowledge artifacts.
await memory.learn({
scope: "session",
minConfidence: 0.7,
extractors: ["procedures", "preferences"]
})Turns raw patterns into typed artifacts - procedures, failure playbooks, causal chains, decision policies. Deterministic: same inputs produce identical artifacts every time.
const artifacts = await memory.compile({
type: "procedure",
topic: "deployment",
sources: ["episodic", "semantic"]
})Predictive cache learns from agent access patterns and pre-warms relevant artifacts before queries arrive. Reduces synthesis latency for common reasoning paths.
memory.predict({
agent: "deploy-bot",
horizon: "next_action",
warmCache: true
})Cross-agent knowledge transfer via versioned artifact packs. Tenant-isolated with fine-grained access controls. One agent learns, all agents benefit.
await memory.transfer({
from: "deploy-bot",
to: "monitor-bot",
artifacts: ["deployment-procedures"]
})Queries all memory layers simultaneously, scores relevance, compresses aggressively, and assembles a provenance-tagged context packet optimized for your token budget.
const ctx = await memory.synthesize({
query: "deploy payments to staging",
budget: 4000,
include: ["procedures", "failures"]
})Agents dump entire conversation histories into context windows. 200k tokens of raw chat for a task that needs 2kb of reasoning. You're burning 99% of your token budget on noise.
Your agent solved this exact problem yesterday. Today it starts from scratch - no memory of what worked, what failed, or why decisions were made. Every session is day one.
Without structured memory, agents produce different outputs for the same inputs. No determinism, no provenance, no way to trace why a decision was made.
RAG retrieves similar text, not relevant knowledge. Vector similarity ≠ reasoning utility. Your agents spend more time searching than thinking.
Hippocortex sits between your agent and the LLM. Events flow in, compressed reasoning context flows out.
Six infrastructure components that handle the full lifecycle of agent memory. Redis for speed. PostgreSQL for durability. Rust for determinism.
TypeScript and Python clients. Three function calls: capture(), learn(), synthesize(). Framework-agnostic with adapters for OpenAI Agents, LangGraph, CrewAI, and more.
RESTful event ingestion endpoint. Accepts structured events with automatic dedup, salience scoring, and schema validation. Sustained 403 events/sec.
Ordered event processing with at-least-once delivery. Idempotent workers ensure deterministic processing regardless of retries or failures.
Deterministic pattern extraction engine. Replays event traces, identifies procedures, detects contradictions, and produces versioned knowledge artifacts.
Persistent storage for compiled knowledge artifacts. Full-text search, temporal indexing, and provenance tracking. GDPR-ready with hard deletion.
Multi-signal ranking engine with token-budget-aware assembly. 18ms p50 latency. Predictive cache pre-warms common query patterns.
HMX is an open, language-agnostic wire protocol for agent memory interoperability. No vendor lock-in. Full portability.
The atomic unit of observable information. 13 event types from messages to state changes.
Compiled knowledge units. Task schemas, failure playbooks, decision policies, causal patterns.
Portable compressed memory state. Export an agent's entire learned knowledge as a single transferable object.
Token-budget-aware reasoning inputs. Multi-signal ranking with provenance tracking.
Semantic versioning with forward/backward compatibility negotiation.
{
"hmx_version": "HMX-1.0",
"event_type": "tool_call",
"agent_id": "agent-ops-1",
"session_id": "sess_abc123",
"timestamp": "2026-03-15T00:12:34Z",
"sequence": 4,
"content": {
"tool_name": "deploy",
"input": { "service": "payments", "env": "staging" }
}
}One protocol. Every framework. Full interoperability.
Three function calls. Any language. Any framework. Drop in Hippocortex without rewriting your agent.
from openai import OpenAI
# Change your base URL. That's it. ~99% reliability.
client = OpenAI(
base_url="https://api.hippocortex.dev/v1",
api_key="hx_live_...",
default_headers={
"X-LLM-API-Key": "sk-...", # Your OpenAI key
},
)
# Use normally. Every call now has memory.
response = client.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": "Deploy payments to staging"}]
)
# Works with OpenAI, Anthropic, Groq, Together, Ollama, Mistral...Agent memory contains sensitive operational data. Every layer of Hippocortex is built with security as a first-class constraint.
Complete data isolation between tenants. Every query is scoped. No cross-tenant data leakage by design.
Scoped API keys with fine-grained permissions. Key rotation without downtime. Full audit trail on every request.
Per-key and per-tenant rate limits. Configurable burst and sustained thresholds. Graceful degradation under load.
Every processing step is idempotent. Safe retries, no duplicate side effects. Deterministic output regardless of failures.
Input validation, payload size limits, schema enforcement, and request signing. Defense in depth at every layer.
TLS 1.3 in transit. AES-256 at rest. Encryption keys managed per-tenant with regular rotation.
Not a wrapper around vector search. Infrastructure-grade memory compilation that no existing tool provides.
No credit card required. Free tier is real — not a 7-day trial. Upgrade when your agents need more memory.
For prototyping and exploration
For developers who need the sharpest memory
Unlimited everything. No limits.
For teams building production AI. Up to 10 seats.
For prototyping and exploration
For developers who need the sharpest memory
Unlimited everything. No limits.
For teams building production AI. Up to 10 seats.
For developers who need the sharpest memory
For prototyping and exploration
Unlimited everything. No limits.
For teams building production AI. Up to 10 seats.
All plans include REST API, JS/TS SDK, auto-instrumentation, and dashboard access. No hidden fees.
Comprehensive docs, SDK references, and working examples. From first install to production in hours.
Free tier. No credit card. Production-ready memory for your AI agents.