What is Memory in LangChain: Explained with Examples
memory is a way to store and recall information during a conversation or task, helping the AI remember past inputs or outputs. It acts like a short-term notebook that keeps track of what happened before to make interactions more natural and context-aware.How It Works
Memory in LangChain works like a friendly assistant who takes notes during a chat. Imagine talking to someone who remembers what you said earlier, so they don't ask the same questions again and can build on previous answers. This memory stores key details from the conversation or task and feeds them back to the AI model when needed.
Technically, memory modules keep track of past inputs, outputs, or other relevant data in a structured way. When the AI needs to respond, it looks at this stored information to provide answers that fit the ongoing context. This makes conversations feel smoother and more human-like.
Example
This example shows how to use a simple memory in LangChain to remember user input and include it in the next AI response.
from langchain.memory import ConversationBufferMemory from langchain.chat_models import ChatOpenAI from langchain.chains import ConversationChain # Create a memory object to store conversation memory = ConversationBufferMemory() # Initialize the chat model chat = ChatOpenAI(temperature=0) # Create a conversation chain with memory conversation = ConversationChain(llm=chat, memory=memory) # Simulate a conversation print(conversation.predict(input="Hi, my name is Alex.")) print(conversation.predict(input="What is my name?"))
When to Use
Use memory in LangChain when you want your AI to remember details during a conversation or multi-step task. This is helpful for chatbots, virtual assistants, or any app where context matters.
For example, a customer support bot can remember a user's problem across messages, or a personal assistant can recall your preferences without asking repeatedly. Memory helps create a more natural and efficient user experience.
Key Points
- Memory stores conversation or task context for better AI responses.
- It acts like a short-term notebook for the AI.
- Helps avoid repeating questions and keeps interactions natural.
- Commonly used in chatbots, assistants, and multi-step workflows.