Skip to main content

Vector store-backed memory

VectorStoreRetrieverMemory stores memories in a VectorDB and queries the top-K most "salient" docs every time it is called.

This differs from most of the other Memory classes in that it doesn't explicitly track the order of interactions.

In this case, the "docs" are previous conversation snippets. This can be useful to refer to relevant pieces of information that the AI was told earlier in the conversation.

npm install @langchain/openai
import { OpenAI, OpenAIEmbeddings } from "@langchain/openai";
import { VectorStoreRetrieverMemory } from "langchain/memory";
import { LLMChain } from "langchain/chains";
import { MemoryVectorStore } from "langchain/vectorstores/memory";
import { PromptTemplate } from "@langchain/core/prompts";

const vectorStore = new MemoryVectorStore(new OpenAIEmbeddings());
const memory = new VectorStoreRetrieverMemory({
// 1 is how many documents to return, you might want to return more, eg. 4
vectorStoreRetriever: vectorStore.asRetriever(1),
memoryKey: "history",

// First let's save some information to memory, as it would happen when
// used inside a chain.
await memory.saveContext(
{ input: "My favorite food is pizza" },
{ output: "thats good to know" }
await memory.saveContext(
{ input: "My favorite sport is soccer" },
{ output: "..." }
await memory.saveContext({ input: "I don't the Celtics" }, { output: "ok" });

// Now let's use the memory to retrieve the information we saved.
await memory.loadMemoryVariables({ prompt: "what sport should i watch?" })
{ history: 'input: My favorite sport is soccer\noutput: ...' }

// Now let's use it in a chain.
const model = new OpenAI({ temperature: 0.9 });
const prompt =
PromptTemplate.fromTemplate(`The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context. If the AI does not know the answer to a question, it truthfully says it does not know.

Relevant pieces of previous conversation:

(You do not need to use these pieces of information if not relevant)

Current conversation:
Human: {input}
const chain = new LLMChain({ llm: model, prompt, memory });

const res1 = await{ input: "Hi, my name is Perry, what's up?" });
console.log({ res1 });
res1: {
text: " Hi Perry, I'm doing great! I'm currently exploring different topics related to artificial intelligence like natural language processing and machine learning. What about you? What have you been up to lately?"

const res2 = await{ input: "what's my favorite sport?" });
console.log({ res2 });
{ res2: { text: ' You said your favorite sport is soccer.' } }

const res3 = await{ input: "what's my name?" });
console.log({ res3 });
{ res3: { text: ' Your name is Perry.' } }

API Reference:

Help us out by providing feedback on this documentation page: