A modular PyTorch library for state-of-the-art diffusion models to generate images, audio, and 3D molecular structures.
🤗 Diffusers is a PyTorch library that provides state-of-the-art pretrained diffusion models for generating images, audio, and 3D molecular structures. It solves the problem of accessing and using advanced generative AI models by offering a modular toolbox that supports both inference and custom training. The library simplifies working with diffusion models through high-level pipelines and interchangeable components.
Machine learning researchers, AI developers, and data scientists working on generative AI projects, particularly those focused on image, audio, or molecular structure generation using diffusion models.
Developers choose 🤗 Diffusers for its comprehensive collection of pretrained models, modular design that balances usability with customizability, and strong community support from Hugging Face. It stands out by offering both simple inference pipelines and low-level components for building custom systems.
🤗 Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.
Integrates with Hugging Face Hub for access to 30,000+ pretrained checkpoints, including popular models like Stable Diffusion and Kandinsky, enabling quick experimentation.
Separates models, schedulers, and pipelines, allowing users to build custom diffusion systems as shown in the low-level quickstart example with UNet2DModel and DDPMScheduler.
High-level pipelines like DiffusionPipeline enable image generation in just a few lines of code, making state-of-the-art AI accessible without deep expertise.
Offers detailed guides for loading, inference, optimization, and training, supporting a wide range of tasks from basic usage to advanced customization.
Admits in its philosophy that performance is secondary to usability, so default configurations may require optimization for speed or memory efficiency, as highlighted in separate optimization guides.
Relies exclusively on PyTorch, making it unsuitable for projects standardized on other frameworks like TensorFlow, despite its modular design.
Training custom models involves significant configuration and understanding of diffusion theory, as indicated by the separate, in-depth training guides that assume prior ML knowledge.