interlocute.ai beta

Long-term Memory

A zero-config persistence engine that lets your AI nodes remember user context, preferences, and history across thousands of sessions — no database required.

What is long-term memory?

Long-term memory allows your AI node to retain information across conversations. Instead of starting from scratch every session, the node recalls relevant user context — preferences, prior decisions, key facts — and uses that context to produce more personalized and consistent responses.

Why it matters

Without persistent memory, every conversation is a blank slate. Users have to repeat themselves, and the AI cannot build on past interactions. Memory transforms a stateless chat endpoint into a relationship — the node learns and improves over time.

How Interlocute helps

Enable memory on your node and Interlocute handles the rest. Every interaction is automatically embedded and indexed. When a new message arrives, the platform performs a semantic lookup of past interactions, surfacing the most relevant context before the LLM processes the request. There are no databases to manage and no retrieval logic to write.

Built for production

Interlocute's memory engine uses vector-native storage with automatic TTL management. You control how long context stays warm, how many memories are retrieved per turn, and which nodes share memory partitions. Every memory operation is metered and logged for full visibility.

Frequently Asked Questions

Long-term Memory

What is long-term memory in the context of AI agents?
Long-term memory allows an AI agent to retain and recall information from previous conversations. Unlike session-based chat that forgets everything when the conversation ends, a node with long-term memory can remember user preferences, past decisions, and contextual details across unlimited sessions.
How does Interlocute implement long-term memory?
Interlocute automatically embeds each interaction into vector storage and indexes it for semantic similarity search. When a new message arrives, the platform retrieves the most relevant past interactions and injects them into the LLM context window. This happens transparently — no code required beyond enabling the feature.
Do I need to manage a database for memory storage?
No. Memory storage is fully managed by Interlocute. There are no databases to provision, no schemas to maintain, and no backup procedures to implement. Memory persists as long as the node exists.
Can I control how long memories are retained?
Yes. You can configure TTL (time-to-live) settings to control how long individual memories remain active. Expired memories move to cold storage and are no longer surfaced during semantic retrieval, but can be restored if needed.
Is memory isolated between different nodes?
Yes. Each node has its own isolated memory partition. Memories stored by one node are never accessible to another node, ensuring data isolation for multi-tenant deployments and distinct use cases.
How does memory interact with RAG and other features?
Memory, RAG, tool use, and scheduling are all composable. A node can use long-term memory alongside a RAG knowledge base — the memory provides user-specific context while RAG provides document-level knowledge. Both are injected into the LLM prompt automatically.
What is the difference between memory and RAG?
Memory stores and retrieves context from past conversations — things the user has said or the node has learned. RAG retrieves context from documents you upload. Memory is interaction-based and grows over time; RAG is document-based and is updated when you add or change files.
How is memory usage billed?
Memory operations are metered as part of your node's computation usage. Each write (embedding a new interaction) and each read (semantic retrieval) is logged and included in your usage ledger. There are no separate storage fees.

Ready to build with Long-term Memory?

Deploy your node in seconds and start using Long-term Memory today.