A curated list of open-source large language models licensed for commercial use, including models for general text and code generation.
Open LLMs is a curated GitHub repository listing open-source large language models that are licensed for commercial use. It solves the problem of fragmented information by providing a centralized, structured directory where developers can compare models based on license, size, capabilities, and release details to find suitable models for their projects.
AI researchers, machine learning engineers, and product developers who need to identify and integrate commercially usable open-source LLMs into their applications, tools, or research workflows.
Developers choose Open LLMs because it saves significant research time by aggregating critical model information in one place, with a strict filter for commercial licensing, enabling confident selection and deployment without legal uncertainty.
📋 A list of open LLMs available for commercial use.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.
Exclusively lists models with permissive licenses like Apache 2.0, MIT, or OpenRAIL-M, saving developers from legal research when deploying commercially. The README includes a dedicated table explaining license implications.
Presents models in a detailed table with columns for parameters, context length, and release dates, enabling quick side-by-side evaluation. For example, it highlights ChatGLM3's 128k context vs. Llama 3's 8k.
Separates code-generation models like Code Llama and StarCoder into a dedicated table, helping developers find tools tailored for programming tasks without sifting through general-purpose models.
Includes curated datasets for pre-training and instruction-tuning, plus links to external leaderboards like LMSYS and Hugging Face's Open LLM Leaderboard, providing a holistic view for model development.
As a community-maintained list, it may not instantly reflect new model releases or updates, risking outdated information in a fast-moving field. The README admits improvements like adding training code are pending.
Performance data is only linked externally, requiring users to navigate away for evaluations. This lacks direct, curated comparisons or synthesized metrics, which could streamline decision-making.
Provides model checkpoints and paper links but no tutorials or code snippets for deployment, fine-tuning, or serving. Developers must rely on external resources for practical integration steps.