Zep AI (YC W24)

Zep AI (YC W24)

Software Development

The Foundational Memory Layer for AI

About us

Long-Term Memory for AI Assistants. Recall, understand, and extract data from chat histories. Power personalized AI experiences.

Industry
Software Development
Company size
2-10 employees
Type
Privately Held
Founded
2023

Employees at Zep AI (YC W24)

Updates

  • Zep AI (YC W24) reposted this

    View profile for Lior Sinclair, graphic
    Lior Sinclair Lior Sinclair is an Influencer

    Covering the latest in AI R&D • ML-Engineer • MIT Lecturer • Building AlphaSignal, a newsletter read by 200,000+ AI engineers.

    Zep just introduced a game-changing way for AI agents to remember and learn. Unlike other systems that only retrieve static documents, Zep uses a temporal knowledge graph to combine conversations and structured business data, keeping track of how things change over time. Here’s why it matters: > It’s more accurate: 94.8% on DMR (better than MemGPT’s 93.4%). > It’s faster: Cuts response time by 90%. > It handles complex tasks like remembering across sessions and reasoning over time. > Lower token costs, making it scalable and enterprise-ready. Read it here: https://fnf.dev/4ars9sP

    • No alternative text description for this image
  • Zep AI (YC W24) reposted this

    Zep AI (YC W24) is the new state of the art in agent memory. It's a memory layer for AI agents that continuously learns from user interactions and changing business data. By providing agents with a complete, holistic view of each user, Zep enables developers to build applications that tackle complex, personalized tasks. In research published today, Zep demonstrated that it delivers up to 18.5% higher accuracy with 90% lower latency when compared to tools like MemGPT, excelling in both the Deep Memory Retrieval (DMR) and LongMemEval benchmarks. https://lnkd.in/gBhxW-_a

  • Big News Day: Zep is officially the state-of-the-art in AI agent memory. In research we're publishing today, we demonstrate that Zep outperforms the current state-of-the-art memory system, MemGPT (Letta AI), in the Deep Memory Retrieval (DMR) benchmark—the primary evaluation metric used by the Letta/MemGPT team. More significantly, Zep excels in the LongMemEval benchmark, a comprehensive and challenging chat history memory evaluation that better reflects real-world enterprise use cases. In this benchmark, Zep delivers aggregate accuracy improvements of up to 18.5%, with individual evaluations showing gains exceeding 100% compared to using the full chat transcript in the context window, all while reducing response latency by 90%. In short, these findings demonstrate that Zep is capable of offering rich context retrieval, even in the complex data scenarios faced by large enterprises. Read more here: https://buff.ly/4jlis35

    • No alternative text description for this image
  • Amazing article on building agents with memory. Also, check out that diagram! Woo 🚀

    View profile for Nikki Siapno, graphic

    Engineering Manager at Canva | Co-Founder of Level Up Coding

    How to Build Your Own AI Agent: AI agents are one of the biggest growth areas in 2025. This means understanding and working with them will be one of the most valuable skills of the year. 𝗦𝗼 𝗵𝗼𝘄 𝗱𝗼 𝘄𝗲 𝗰𝗿𝗲𝗮𝘁𝗲 𝗼𝗻𝗲? First, AI agents typically fall into two categories: Those without memory — 𝘀𝘁𝗮𝘁𝗲𝗹𝗲𝘀𝘀 𝗮𝗴𝗲𝗻𝘁𝘀. They react only to immediate input, like a blank slate every time. Those with memory — 𝘀𝘁𝗮𝘁𝗲𝗳𝘂𝗹 𝗮𝗴𝗲𝗻𝘁𝘀. They leverage past interactions to deliver context-aware, smarter, and highly personalized responses—making them exponentially more powerful and capable of highly complex workflows. 𝗠𝗲𝗺𝗼𝗿𝘆 𝗶𝘀 𝘄𝗵𝗲𝗿𝗲 𝘁𝗵𝗲 𝗺𝗮𝗴𝗶𝗰 𝗵𝗮𝗽𝗽𝗲𝗻𝘀. Here's a great read on memory → https://lnkd.in/gejyFrF5 The article above by Ken Collins and Zep AI (YC W24) provides a fantastic explanation on: 🔹 How memory works 🔹 How to integrate memory into an AI agent Now let’s walk through the key steps to build an AI agent with memory: 𝟭) 𝗗𝗲𝗳𝗶𝗻𝗲 𝘁𝗵𝗲 𝗔𝗜 𝗮𝗴𝗲𝗻𝘁'𝘀 𝗽𝘂𝗿𝗽𝗼𝘀𝗲 Start by clarifying its role, such as a personal assistant or customer service bot, and determine the data it needs to process (eg; user profiles or task history). 𝟮) 𝗖𝗵𝗼𝗼𝘀𝗲 𝗟𝗟𝗠 𝗺𝗼𝗱𝗲𝗹𝘀 Select an LLM or LLMs (e.g; OpenAI, Hugging Face) that match your requirements. 𝟯) 𝗣𝗹𝗮𝗻 𝗺𝗲𝗺𝗼𝗿𝘆 𝗿𝗲𝗾𝘂𝗶𝗿𝗲𝗺𝗲𝗻𝘁𝘀 Design for both short-term memory (conversation context) and long-term memory (persistent user knowledge). 𝟰) 𝗜𝗻𝘁𝗲𝗴𝗿𝗮𝘁𝗲 𝗭𝗲𝗽 𝗳𝗼𝗿 𝗺𝗲𝗺𝗼𝗿𝘆 𝗺𝗮𝗻𝗮𝗴𝗲𝗺𝗲𝗻𝘁 Use Zep’s framework to manage memory, sessions, and knowledge graphs. 𝟱) 𝗕𝘂𝗶𝗹𝗱 𝘂𝘀𝗲𝗿 𝗽𝗿𝗼𝗳𝗶𝗹𝗲𝘀 𝗮𝗻𝗱 𝘀𝗲𝘀𝘀𝗶𝗼𝗻𝘀 Create user IDs and maintain session continuity for seamless interactions. 𝟲) 𝗗𝗲𝘀𝗶𝗴𝗻 𝗮 𝘄𝗼𝗿𝗸𝗳𝗹𝗼𝘄 Establish processes for memory management. Define workflows to retrieve, update, and use memory for informed responses. 𝟳) 𝗜𝗻𝗰𝗼𝗿𝗽𝗼𝗿𝗮𝘁𝗲 𝗮 𝗸𝗻𝗼𝘄𝗹𝗲𝗱𝗴𝗲 𝗴𝗿𝗮𝗽𝗵 Use Zep's knowledge graph to store and query relationships, and enrich responses. 𝟴) 𝗗𝗲𝘃𝗲𝗹𝗼𝗽 𝗮𝗻𝗱 𝘁𝗲𝘀𝘁 𝗰𝗼𝗻𝘁𝗲𝘅𝘁𝘂𝗮𝗹 𝗽𝗿𝗼𝗺𝗽𝘁𝘀 Create prompts that dynamically use memory and knowledge graphs for relevant, secure responses. 𝟵) 𝗦𝗲𝗰𝘂𝗿𝗲 𝗱𝗮𝘁𝗮 Implement robust security measures and restrict data access. 𝟭𝟬) 𝗠𝗼𝗻𝗶𝘁𝗼𝗿, 𝗶𝗺𝗽𝗿𝗼𝘃𝗲, 𝗮𝗻𝗱 𝘀𝗰𝗮𝗹𝗲 Use analytics and feedback to refine memory systems and scale capabilities. These steps provide a high-level guide for building stateful AI agents. If you want to dive deeper into how memory works or integrating it, here's a great article: https://lnkd.in/gejyFrF5 💭 Over to you. Have you worked on projects with AI agents? 💬

    • No alternative text description for this image
  • Zep AI (YC W24) reposted this

    View profile for Santiago Valdarrama, graphic

    Computer scientist and writer. I teach hard-core Machine Learning at ml.school.

    Knowledge graphs are huge for AI Agents! A knowledge graph is the difference between a dumb AI agent and one that blows everyone's mind. Agents need memory and must know how to keep it updated over time (This is difficult, and it's the main reason most agents you've seen get dumber overnight!) This is where a knowledge graph helps. A knowledge graph is a network of connected points, each representing a piece of information. It's a very efficient structure for capturing complex relationships between data. Google uses a (huge) knowledge graph as part of Search. (Probably the largest knowledge graph in the world.) It was arguably one of the best improvements to Search since it was created. For building AI agents, knowledge graphs have two advantages: 1. They make it easier to extract facts from memory 2. They make it easier to update facts as they change The second point is crucial: You want agents to keep up with the world and update old facts as they discover new information. Here is a recommendation that will teach you how to use a knowledge graph as the memory layer of an AI agent: Ken Collins wrote an excellent article in which he builds a chat history for Llama 3 using Zep AI (YC W24)'s AI Memory (backed by a knowledge graph.) Here is a link to the article: https://fnf.dev/4fPAXtx This article is a great example of how to build agents that keep up with change. Zep is an open-source library that will serve as your agent's memory. You can connect it to any agent framework, model, or platform. The article's source code is in TypeScript, but you can use Zep with Python or Go as well. In a few bullet points: 1. You send messages to your AI agent 2. Zep synthesizes the information into a knowledge graph 3. You can retrieve any relevant facts from memory extremely fast Thanks to the Zep team for sponsoring this post.

  • Thanks for the shout out, Kesha Williams!

    View profile for Kesha Williams, graphic

    AI Advisor • Head of Enterprise Architecture • AWS Hero (Machine Learning) • Award-Winning Engineer • International Keynote Speaker • Host LAItency Unplugged Podcast

    Have you heard of Zep AI (YC W24)? In the fast-evolving world of AI, Zep is the memory layer for AI agents we've been waiting for. The ability of AI agents to retain and utilize contextual information over time is paramount. Zep addresses this need by providing a robust memory layer that enhances the capabilities of AI agents and assistants. ➡️ Temporal Knowledge Graph: At the heart of Zep lies a temporal knowledge graph, a dynamic structure that models the evolving relationships between complex entities such as users and products. This graph enables AI agents to understand and reason about changes over time, ensuring their responses remain accurate and contextually relevant. ➡️ Low-Latency Memory Retrieval: Zep distinguishes itself with its low-latency memory retrieval system. By avoiding reliance on large language model (LLM)-based agentic behavior for memory access, Zep ensures that AI agents can swiftly recall pertinent information, leading to more responsive and efficient interactions. ➡️ Platform Independence: Designed with flexibility, Zep is platform, model, and framework-independent. Developers can integrate Zep into their AI systems regardless of the underlying technologies, making it a versatile solution for many applications. ➡️ Getting Started with Zep: Embarking on the journey with Zep is straightforward. The platform offers comprehensive documentation and SDKs in multiple programming languages, including Python, TypeScript, and Go, facilitating seamless integration into existing AI projects. And the best part? Zep is available now as a community edition, with a hosted version launching soon. Whether you're an AI developer or a product owner, this is your chance to explore how Zep can redefine what's possible in AI memory. 🔗 Ready to dive in? Start your Zep journey today: https://fnf.dev/4gyzIiY 

  • Zep AI (YC W24) reposted this

    If you give a RAG agent chunks of space articles from 1994, it’ll tell you Pluto’s a planet. But what happens if you give it chunks of articles from 2004, and chunks from an article from 2024 that says Pluto is no longer a planet? The LLM might get confused, or bias towards the more frequent mention of Pluto being a planet. Vector databases have enabled AI agents to perform incredibly well at information retrieval. But RAG doesn’t do a great job of encoding time. Vector databases match for keyword and semantic relevance…not “How recent is this information?” or “Is this fact still valid?” Unless you design your system to account for this, you’ll get inaccurate results. That’s fine if time isn’t an issue for your AI agent. But if you’re designing an enterprise product that talks to users, it probably is: - Your customer support agent should know past issues your customers ran into - Your product recommendation agent should know the customer bought Nikes, then returned them because they “didn’t like them” - Your company docs agent should know you used to support Salesforce in v1, but deprecated support in v2 LLMs, surprisingly enough, do a great job understanding this. But they can only reason with the information they have, and if the vector database doesn't retrieve temporally relevant chunks, it doesn’t matter. Luckily, we solved this with Graphiti, Zep’s open-source graph library. DM if you’re interested in learning more.

  • Zep AI (YC W24) reposted this

    I think all the hype around LLMs has caused us to overlook Small Language Models. Let me explain - you might actually realize you need them. We’re building the foundational memory later for AI agents. That means our customers expect low latency - no one wants to wait more than a few seconds for a response from an AI agent. Just imagine trying to build a sales bot that takes 15 seconds to answer your question. You’d just buy from somebody else. So the ability to do high-leverage tasks (like sales) quickly is extremely powerful. This is where smaller-language models can give you leverage. Fine-tuning for specific tasks and using accurate memory to ensure right data is provided can result in faster results and with the same accuracy as frontier models. Example models that come to mind: Microsoft Phi-3, Llama3.2 11B Just don’t feel like you have to throw the largest and most expensive model at a problem.

Similar pages

Browse jobs

Funding

Zep AI (YC W24) 2 total rounds

Last Round

Pre seed

US$ 500.0K

See more info on crunchbase