Showing 3 of 3 projects
A collection of libraries to optimize AI model performance through inference acceleration, infrastructure efficiency, and fine-tuning optimization.
A collection of libraries to optimize AI model performance through inference, infrastructure, and fine-tuning techniques.
A semantic cache library for LLM queries that reduces API costs by 10x and boosts response speed by 100x.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.