Showing 12 of 12 projects
A platform to run, manage, and serve open-source large language models (LLMs) locally or on your own infrastructure.
A platform to run, manage, and serve open-source large language models locally with a simple CLI and REST API.
A high-throughput, memory-efficient inference and serving engine for large language models (LLMs).
A web UI and optimization library for running and fine-tuning open-source AI models locally with 2x faster training and 70% less VRAM.
An open-source AI engine that runs LLMs, vision, voice, and image/video models on any hardware with drop-in OpenAI API compatibility.
A high-performance serving framework for large language models and multimodal models, delivering low-latency and high-throughput inference.
An open-source Java library that simplifies integrating LLMs into Java applications through a unified API and comprehensive toolbox.
An open-source pipeline for training medical domain GPT models using PT, SFT, RLHF, DPO, ORPO, and GRPO methods.
A lightweight, single-binary Rust inference server providing 100% OpenAI-API compatible endpoints for local GGUF models.
A C#/.NET library for efficient local inference of LLaMA and other large language models, based on llama.cpp.
A JAX/Flax-based framework for easy and scalable pre-training, fine-tuning, evaluation, and serving of large language models.
A React library for building smooth, customizable user interfaces for LLM-powered applications with streaming output.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.