
Create an AI app on your own data in a minute
Embedchain is an open-source RAG (Retrieval-Augmented Generation) framework that enables developers to build AI applications powered by their own data in minutes. It follows a "Conventional but Configurable" design principle, abstracting away the complexity of data ingestion, chunking, embedding, and vector storage so developers can focus on building. The project was later rebranded as Mem0, shifting focus toward a persistent memory layer for AI agents, while the original Embedchain repository remains a widely used RAG framework.
Ingest data from PDFs, web pages, YouTube videos, CSV, JSON, Markdown, Word documents, Notion, GitHub, Slack, Discord, Gmail, PostgreSQL, MySQL, Sitemaps, images, audio, and more.
Works with OpenAI, Anthropic Claude, Cohere, Hugging Face models, Mistral, Llama, and Ollama for local deployment.
Supports multiple vector databases including ChromaDB (default), Zilliz/Milvus, and others for embedding storage and retrieval.
Handles the full pipeline automatically — segmenting documents into optimally sized chunks, generating embeddings, and storing them for fast semantic retrieval.
Provides distinct APIs for question answering, contextual information extraction, and interactive chat conversations, all grounded in the user's own data.
Supports multiple embedding providers including OpenAI, Cohere, Hugging Face, and Ollama, letting developers optimize for cost, latency, or privacy.
Teams ingest internal documentation, wikis, Notion pages, and Slack history to create a chatbot that answers employee questions using company-specific knowledge.
Developers build applications that allow users to upload PDFs, Word documents, or entire websites and receive accurate, source-grounded answers through semantic search.
Companies index their product documentation, FAQs, and support articles to power AI support agents that respond contextually to customer queries.
Researchers and content teams ingest YouTube channels, news sites, academic PDFs, and RSS feeds to query and summarize large bodies of content quickly.
Sensible defaults make it usable out of the box for rapid prototyping, while a comprehensive configuration system allows deep customization for production use.
Integrates with popular AI orchestration frameworks including LangChain compatibility layers, making it easy to incorporate into existing pipelines.
Install via pip and build a working RAG application with just a few lines of Python code, dramatically reducing the barrier to entry.
AI developers and product teams use Embedchain to prototype and validate RAG-based product ideas in hours rather than days.

The world's fastest AI inference � 20x faster than GPU clouds