knowledge-base/docs/examples-llama-index-mem0.md
Pratik Narola 7689409950 Initial commit: Production-ready Mem0 interface with monitoring
- Complete Mem0 OSS integration with hybrid datastore
- PostgreSQL + pgvector for vector storage
- Neo4j 5.18 for graph relationships
- Google Gemini embeddings integration
- Comprehensive monitoring with correlation IDs
- Real-time statistics and performance tracking
- Production-grade observability features
- Clean repository with no exposed secrets
2025-08-10 17:34:41 +05:30

8.6 KiB

LlamaIndex ReAct Agent with Mem0

Overview

A ReAct agent combines reasoning and action capabilities, making it versatile for tasks requiring both thought processes (reasoning) and interaction with tools or APIs (acting). Mem0 as memory enhances these capabilities by allowing the agent to store and retrieve contextual information from past interactions.

This guide demonstrates how to create a ReAct Agent with LlamaIndex that uses Mem0 as the memory store, showcasing the dramatic difference between agents with and without memory capabilities.

Setup

Installation

pip install llama-index-core llama-index-memory-mem0

Initialize the LLM

import os
from llama_index.llms.openai import OpenAI

os.environ["OPENAI_API_KEY"] = "<your-openai-api-key>"
llm = OpenAI(model="gpt-4o")

Initialize Mem0 Memory

You can find your API key here. Read about Mem0 Open Source.

os.environ["MEM0_API_KEY"] = "<your-mem0-api-key>"

from llama_index.memory.mem0 import Mem0Memory

context = {"user_id": "david"}
memory_from_client = Mem0Memory.from_client(
    context=context,
    api_key=os.environ["MEM0_API_KEY"],
    search_msg_limit=4,  # optional, default is 5
)

Create Agent Tools

These tools will be used by the agent to perform actions:

from llama_index.core.tools import FunctionTool

def call_fn(name: str):
    """Call the provided name.
    Args:
        name: str (Name of the person)
    """
    return f"Calling... {name}"

def email_fn(name: str):
    """Email the provided name.
    Args:
        name: str (Name of the person)
    """
    return f"Emailing... {name}"

def order_food(name: str, dish: str):
    """Order food for the provided name.
    Args:
        name: str (Name of the person)
        dish: str (Name of the dish)
    """
    return f"Ordering {dish} for {name}"

# Create tool instances
call_tool = FunctionTool.from_defaults(fn=call_fn)
email_tool = FunctionTool.from_defaults(fn=email_fn)
order_food_tool = FunctionTool.from_defaults(fn=order_food)

Initialize the Agent with Memory

from llama_index.core.agent import FunctionCallingAgent

agent = FunctionCallingAgent.from_tools(
    [call_tool, email_tool, order_food_tool],
    llm=llm,
    memory=memory_from_client,  # Mem0 memory integration
    verbose=True,
)

Building User Context

Let's start by having the agent learn about the user through conversation:

Introduction

Input:

response = agent.chat("Hi, My name is David")
print(response)

Output:

> Running step bf44a75a-a920-4cf3-944e-b6e6b5695043. Step input: Hi, My name is David
Added user message to memory: Hi, My name is David
=== LLM Response ===
Hello, David! How can I assist you today?

Learning Preferences

Input:

response = agent.chat("I love to eat pizza on weekends")
print(response)

Output:

> Running step 845783b0-b85b-487c-baee-8460ebe8b38d. Step input: I love to eat pizza on weekends
Added user message to memory: I love to eat pizza on weekends
=== LLM Response ===
Pizza is a great choice for the weekend! If you'd like, I can help you order some. Just let me know what kind of pizza you prefer!

Communication Preferences

Input:

response = agent.chat("My preferred way of communication is email")
print(response)

Output:

> Running step 345842f0-f8a0-42ea-a1b7-612265d72a92. Step input: My preferred way of communication is email
Added user message to memory: My preferred way of communication is email
=== LLM Response ===
Got it! If you need any assistance or have any requests, feel free to let me know, and I can communicate with you via email.

Comparing Agents: With vs Without Memory

Using the Agent WITHOUT Memory

Setup:

agent_no_memory = FunctionCallingAgent.from_tools(
    [call_tool, email_tool, order_food_tool],
    # memory is not provided
    llm=llm,
    verbose=True,
)

Input:

response = agent_no_memory.chat("I am feeling hungry, order me something and send me the bill")
print(response)

Output:

> Running step e89eb75d-75e1-4dea-a8c8-5c3d4b77882d. Step input: I am feeling hungry, order me something and send me the bill
Added user message to memory: I am feeling hungry, order me something and send me the bill
=== LLM Response ===
Please let me know your name and the dish you'd like to order, and I'll take care of it for you!

Result: The agent has no memory of previous conversations and cannot act on user preferences.

Using the Agent WITH Memory

Setup:

agent_with_memory = FunctionCallingAgent.from_tools(
    [call_tool, email_tool, order_food_tool],
    llm=llm,
    memory=memory_from_client,  # Mem0 memory integration
    verbose=True,
)

Input:

response = agent_with_memory.chat("I am feeling hungry, order me something and send me the bill")
print(response)

Output:

> Running step 5e473db9-3973-4cb1-a5fd-860be0ab0006. Step input: I am feeling hungry, order me something and send me the bill
Added user message to memory: I am feeling hungry, order me something and send me the bill
=== Calling Function ===
Calling function: order_food with args: {"name": "David", "dish": "pizza"}
=== Function Output ===
Ordering pizza for David
=== Calling Function ===
Calling function: email_fn with args: {"name": "David"}
=== Function Output ===
Emailing... David
> Running step 38080544-6b37-4bb2-aab2-7670100d926e. Step input: None
=== LLM Response ===
I've ordered a pizza for you, and the bill has been sent to your email. Enjoy your meal! If there's anything else you need, feel free to let me know.

Result: The agent remembers:

  • User's name is David
  • User loves pizza on weekends
  • User prefers email communication
  • Automatically orders pizza and sends bill via email

Key Benefits of Memory Integration

1. Personalized Responses

  • Agents remember user preferences and act accordingly
  • No need to repeat information in every conversation

2. Contextual Decision Making

  • Agents can make informed decisions based on past interactions
  • Improved user experience through continuity

3. Efficient Interactions

  • Reduced friction in user-agent communication
  • Faster task completion with fewer prompts needed

4. Learning and Adaptation

  • Agents improve over time by learning from interactions
  • Better understanding of user behavior patterns

Configuration Options

Memory Configuration

memory_from_client = Mem0Memory.from_client(
    context=context,
    api_key=os.environ["MEM0_API_KEY"],
    search_msg_limit=4,  # Controls how many past messages to retrieve
)

Context Parameters

  • user_id: Unique identifier for memory isolation between users
  • search_msg_limit: Number of relevant past messages to include in context

Use Cases

1. Personal Assistants

  • Remember user preferences, schedules, and habits
  • Provide personalized recommendations and actions

2. Customer Support Agents

  • Maintain conversation history and user preferences
  • Provide consistent support across multiple interactions

3. E-commerce Assistants

  • Remember shopping preferences and past purchases
  • Suggest relevant products and services

4. Educational Tutors

  • Track learning progress and adapt teaching methods
  • Remember student strengths and areas for improvement

Best Practices

1. Context Management

  • Use meaningful user IDs for proper memory isolation
  • Adjust search_msg_limit based on conversation complexity

2. Tool Design

  • Create tools that can leverage memory context
  • Design functions with clear parameter definitions

3. Memory Hygiene

  • Regularly review and update memory contents
  • Implement privacy controls for sensitive information

4. Testing

  • Test agents both with and without memory to understand impact
  • Validate memory persistence across sessions

Conclusion

Integrating Mem0 with LlamaIndex ReAct agents transforms static, forgetful assistants into intelligent, context-aware companions. The dramatic difference between agents with and without memory demonstrates the power of persistent context in creating truly helpful AI assistants.

The combination enables:

  • Continuity across conversations
  • Personalization based on user preferences
  • Efficiency through reduced repetition
  • Intelligence through accumulated context

This integration makes AI agents more human-like in their ability to remember and build upon past interactions, creating a foundation for truly intelligent and helpful AI systems.