Skip to main content
SOTA on LoCoMo Benchmark

World's First
Human-Like Memory

We've Engineered A Foundational Memory Layer Powering Next-Gen AI World Models

View on GitHub

No credit card required · Free tier available

Works with every agent framework

LangChainLlamaIndexCrewAIAutoGenOpenAIAnthropicVercel AIPydantic AI
pip install hypermemory# one line. any agent.

Hypermemory is a hybrid memory retrieval system for AI agents. It combines semantic search, BM25 keyword matching, temporal scoring, and multi-hop reasoning to give long-running agents persistent, adaptive memory. Unlike context windows that reset, Hypermemory persists facts across sessions and achieves state-of-the-art results on the LoCoMo conversational memory benchmark — scoring 92% on Temporal Reasoning, 94% on Single Hop, and 88% on Multi Hop.

Trusted by developers at

IBM
Meta
Siemens
Deloitte
TikTok
Coca-Cola
Disney
HP
Hinge
Venmo
3M
Illumina
IBM
Meta
Siemens
Deloitte
TikTok
Coca-Cola
Disney
HP
Hinge
Venmo
3M
Illumina

Live Dashboard

Full Visibility Into Every Memory

Search, trace, ingest, and query your memory store. Explore retrieval analytics. Earn XP as you explore.

6

Memories

0.90

Avg Score

5

Active

12ms

Avg Latency

Novice
0 XP / 100

0.90

Avg Score

5

Active

4

Modes Used

Retrieval Radar

0.940.970.630.910.44SemanticTemporalBM25FactMatchMultiHop

Product launch is April 15th — confirmed by CEO in standup

Mode Usage

Semantic
5
Temporal
3
Fact Match
3
BM25
1

Most active mode: Semantic

Source Breakdown

6total
Conversation4
Inferred2
Superseded1

24h Activity

live

Memory ingestions over last 24h

Score Distribution

launch_date_april_15
0.97
stripe_webhook_raw_body
0.93
onboarding_variant_b_wins
0.91
user_prefers_nextjs_supabase
0.88
user_timezone_ist
0.86
cors_error_resolved
0.82
Ingested your first memory+50 XP
Ran 3 queries+75 XP
Expanded 5 memory rows+100 XP
+3 more achievements

Architecture

5 Retrieval Modes, Running in Parallel

Every query fans through all strategies simultaneously. Adaptive score fusion returns the best result from whichever path wins.

Q
query dispatched
1
Semantic Search
92%
2
BM25 Keyword
87%
3
Temporal Scoring
94%
4
Fact Matching
88%
5
Multi-Hop Reasoning
85%
Score Fusion
Ranked Memory Results
Semantic Search

Finds conceptually similar memories even when exact words differ

BM25 Keyword

Precise recall for exact terms, names, and specific phrases

Temporal Scoring

Weights recent memories higher; detects date-specific queries

Fact Matching

Structured recall of who, what, where from extracted facts

Multi-Hop Reasoning

Chains related memories across topics for complex queries

Memory Retrieval

Hybrid Retrieval Architecture

Hypermemory gives long-running agents persistent, adaptive memory that continuously evolves their intelligence. It drives cost-efficient self-learning, saving developers time, tokens, and money while enabling advanced reasoning across domains like temporal, inferential, and open-world tasks.

Semantic Search

Vector Similarity
Q
What blockers did the team flag in last week's retro?
From the retro on Mar 18: 3 blockers logged — CI pipeline exceeding 22 min, unclear ownership on the auth refactor, and missing design specs for the mobile checkout flow.
A

Keyword Search

BM25 Ranking
Q
Show all conversations mentioning 'latency regression'
Found 4 conversations: FinexAI support ticket (Mar 3), v2.3.1 post-deploy review (Mar 10), on-call handoff notes (Mar 14), SDK changelog discussion (Mar 21).
A

Temporal Reasoning

Date-Aware
Q
How long did Dr. Anjali's Phase II clinical trial run?
Based on stored records: protocol approved March 2023, last patient enrolled November 2023. The trial ran for approximately 8 months before moving to data analysis.
A

Multi-Hop Reasoning

Connected Traversal
Q
Which offices has Rohan visited this quarter?
From multiple memories: Rohan visited Mumbai HQ in January for the Q1 kickoff, attended the Dubai partner summit in February, and is scheduled for the Singapore office in March.
A

Adversarial

Hallucination-Proof
Q
When did Priya join the backend team?
Not answerable — Priya is stored as a frontend engineer who joined in April 2024. There is no memory of her transferring to the backend team. The question contains an assumption not supported by context.
A

Inferential

Commonsense + World Knowledge
Q
Would Karan benefit from attending the Rust systems programming workshop?
Very likely — Karan mentioned migrating performance-critical services to Rust last month and has attended 3 systems programming talks this year. This directly matches his current focus.
A
1

Data Ingestion

API · MCP · SDK

2

Fact Extraction

Lazy Async

3

Multi-Modal Index

5 Modes

4

Query Intelligence

Expand · Filter

5

Hybrid Retrieval

Score Fusion

6

Memory Response

Ranked + Proven

Developer API

Built for Developers

Simple REST API and MCP integration. Add memory to your AI in minutes.

REST API

Store and retrieve memories with a simple HTTP call

REST API
curl -X POST https://api.hypermemory.run/v1/memories \
  -H "Authorization: Bearer YOUR_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "agent_id": "support-agent",
    "content": "Arjun from FinexAI: latency spiked to 8s after v2.3.1 SDK update — hotfix needed by 5pm",
    "metadata": { "source": "slack", "customer": "FinexAI", "priority": "high" }
  }'

MCP Integration

Add Hypermemory as a Model Context Protocol server

MCP Integration
{
  "mcpServers": {
    "hypermemory": {
      "command": "npx",
      "args": ["-y", "hypermemory-mcp"],
      "env": {
        "HYPERMEMORY_API_KEY": "your-api-key",
        "HYPERMEMORY_AGENT_ID": "support-agent"
      }
    }
  }
}

Integration

One-Line Ingestion.
Infinite Recall.

Add persistent memory to your LLM apps with a single function call.

hypermemory_quickstart.pyPython
from hypermemory import Hypermemory

hm = Hypermemory(api_key="your-api-key")

# Store a memory — fact extraction happens automatically
hm.add(
    agent_id="support-agent",
    content="Arjun from FinexAI reported that inference latency spiked to 8 seconds "
            "after the v2.3.1 SDK update. Needs a hotfix by 5pm or escalates to CTO.",
)

# Recall with natural language — multi-modal retrieval kicks in
results = hm.search(
    agent_id="support-agent",
    query="Which customers are affected by the SDK latency regression?",
)
# Returns: relevant memories ranked by semantic similarity,
# temporal recency, and entity-fact matches
PythonTypeScriptREST APIMCP

Open Source

Open Source at Heart

Built in the open. Join thousands of developers building the future of AI memory.

License

MIT Licensed

Use it freely in personal projects, startups, or enterprise products. No strings attached — ever.

Transparency

Fully Auditable

Every line of the memory layer is public. Understand exactly how your data is stored, retrieved, and scored.

Control

Self-Hostable

Deploy on your own infra — on-prem, private cloud, or air-gapped. Zero dependency on our servers.

Community

Shape the Roadmap

Open issues, submit PRs, and vote on features. The community drives what gets built next.

Our commitments

No vendor lock-in
Public issue tracker
Semantic versioning
Changelogs on every release
Reproducible builds
Community RFC process

Performance

LoCoMo Benchmark Results

Hypermemory excels across all LoCoMo evaluation domains.

Temporal Reasoning92% vs 61% baseline
Open Domain89% vs 58% baseline
Inferential87% vs 54% baseline
Single Hop94% vs 67% baseline
Multi Hop88% vs 52% baseline
Hypermemory
Baseline
Hypermemory vs Baseline on LoCoMo Benchmark
DomainHypermemoryBaseline
Temporal Reasoning92%61%
Open Domain89%58%
Inferential87%54%
Single Hop94%67%
Multi Hop88%52%

Use Cases

AI Memory That Adapts to Your Domain

Hallucination-proof RAG for compliance-critical AI

Patient data retrieval, diagnostics, drug interaction checks

Reduce Readmissions by 40%

Telehealth agents that remember every patient interaction, medication change, and care preference. No more lost context between visits — your AI assistant recalls what matters for better outcomes.

  • Cut repeat diagnostic workups by 60%
  • Catch medication conflicts before they happen
  • Track patient journeys across providers seamlessly

Live example

Use Cases

Built for Every Industry

Hypermemory adapts to your industry — the same retrieval engine, tuned to what matters most in your context.

Healthcare
Healthcare

Agents that remember every patient

Telehealth agents that recall medications, symptoms, allergies, and care preferences across every visit — reducing readmissions and improving outcomes.

40%

fewer readmissions

HIPAA-readyLongitudinal memory
Education
Education

Tutors that adapt to every learner

AI tutors that track each student's learning pace, knowledge gaps, and preferred explanation style — personalizing every session from day one.

faster concept retention

Adaptive learningLong-term context
E-commerce
E-commerce

Shopping assistants with taste memory

Agents that remember what a customer bought, returned, loved, and hated — surfacing the right product before they even search for it.

2.8×

higher conversion

Preference trackingTemporal scoring
Customer Support
Customer Support

Support that never makes you repeat yourself

Agents with full conversation history across channels. Every ticket, refund, and complaint remembered — so customers never have to explain twice.

65%

reduction in handle time

Cross-session memoryMulti-channel
Sales & CRM
Sales & CRM

AI reps that remember every deal detail

Sales agents that track objections, competitor mentions, stakeholder names, and deal history — delivering hyper-personalised follow-ups that close.

31%

higher close rate

Entity memoryRelationship graphs
Legal & Compliance
Legal & Compliance

Assistants that track regulatory changes

Agents that monitor case law, contract clauses, and compliance requirements over time — with temporal supersession so the current rule always wins.

90%

faster clause retrieval

Temporal supersessionAudit trails
Gaming & Entertainment
Gaming & Entertainment

NPCs with persistent world memory

Game characters that remember player choices, past interactions, and evolving storylines — creating narratives that feel genuinely alive.

player session length

Narrative memoryCharacter state
Enterprise Ops
Enterprise Ops

Internal agents that know your org

Knowledge agents that remember org charts, project history, team preferences, and institutional knowledge — making every employee 10× more effective.

55%

reduction in search time

On-prem deployAccess control

Enterprise

Secure Memory Layer That Cuts LLM Spend and Passes Audits

SOC 2 Type II ready. Deploy anywhere. Full audit trails.

Zero-Trust Security & Compliance

SOC 2 Type II ready. End-to-end encryption, RBAC, and audit logs for every memory operation.

Deploy Anywhere, No Tradeoffs

On-prem, private cloud, or managed SaaS. Same API, same performance, your infrastructure.

Traceable by Default

Full provenance for every memory. Know where data came from, when it was updated, and who accessed it.

Deployment Options

On-Prem
Private Cloud
Managed SaaS

From the Blog

Insights on AI Memory

Why Your AI Agent Forgets Everything — And How to Fix It
Research

Why Your AI Agent Forgets Everything — And How to Fix It

Long-running agents break down not because of bad reasoning, but because they can't remember. We explore the root causes of context degradation and the architecture that solves it.

7 min readMar 2026
Read

Join The Hypermemory Community

Connect with developers building the future of AI memory.

N

Noah

Hypermemory · Support