An opinionated RAG framework for integrating generative AI into applications, supporting any LLM, vector store, and file type.
Quivr is an opinionated Retrieval-Augmented Generation (RAG) framework that simplifies adding generative AI capabilities to applications. It handles document ingestion, retrieval, and generation workflows, allowing developers to integrate AI assistants or knowledge bases without building RAG systems from scratch. The framework supports multiple LLMs, vector stores, and file formats for flexibility.
Developers and product teams building AI-powered applications like chatbots, knowledge assistants, or document analysis tools who want a pre-built, customizable RAG solution.
Quivr reduces development time by providing a production-ready RAG core with extensive customization options, avoiding vendor lock-in through support for any LLM or vector store. Its opinionated design ensures best practices while allowing integration into existing products.
Opiniated RAG for integrating GenAI in your apps 🧠 Focus on your product rather than the RAG. Easy integration in existing products with customisation! Any LLM: GPT4, Groq, Llama. Any Vectorstore: PGVector, Faiss. Any Files. Anyway you want.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.
Provides a pre-built, fast RAG implementation that reduces development overhead, letting developers focus on product features rather than RAG complexities.
Supports multiple providers like OpenAI, Anthropic, and local models via Ollama, avoiding vendor lock-in and offering flexibility.
Handles various file types including PDF, TXT, and Markdown, with options for custom parsers to extend data ingestion.
Allows extensions like internet search and tools through YAML configuration, enabling tailored RAG applications beyond basic retrieval.
Requires YAML files and Python code for customization, which can be steep for developers unfamiliar with RAG concepts or workflow orchestration.
The project is actively evolving, as noted in the README, which might lead to breaking changes and incomplete features during updates.
Relies on third-party services for LLMs and rerankers (e.g., Cohere), introducing potential latency, cost, and reliability issues in production.