Showing 9 of 9 projects
An open platform for training, serving, and evaluating large language model based chatbots.
A CLI and library for evaluating, red-teaming, and comparing LLM prompts, agents, and RAGs with simple declarative configs.
An open-source platform for debugging, evaluating, and monitoring LLM applications, RAG systems, and agentic workflows with tracing and automated evaluations.
An open-source platform for debugging, evaluating, and monitoring LLM applications, RAG systems, and agentic workflows.
A framework and open-source registry for evaluating large language models (LLMs) and LLM systems.
An open-source LLMOps platform unifying gateway, observability, evaluation, optimization, and experimentation for industrial-grade LLM applications.
A command-line tool for red-teaming and vulnerability scanning of large language models (LLMs).
An open-source Python framework to evaluate, test, and monitor ML and LLM systems with 100+ built-in metrics.
An open-source LLMOps platform for prompt management, evaluation, and observability to build reliable LLM applications faster.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.