A Python library for offline deep reinforcement learning with support for state-of-the-art algorithms and user-friendly APIs.
d3rlpy is an offline deep reinforcement learning library that enables training RL agents using pre-existing datasets without requiring online environment interaction. It solves the problem of applying RL in domains where real-time interaction is costly or impossible, such as robotics and healthcare. The library supports both offline and online RL algorithms through a unified API.
Machine learning practitioners and researchers working on reinforcement learning projects, particularly those focused on offline RL, robotics, or data-driven control systems.
Developers choose d3rlpy for its comprehensive support of state-of-the-art offline RL algorithms, user-friendly API that abstracts deep learning complexities, and unique features like distributional Q functions and distributed training capabilities.
An offline deep reinforcement learning library
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.
Supports a wide range of state-of-the-art offline and online RL algorithms, including distributional Q functions across all, as highlighted in the README's algorithm table.
Abstracts deep learning complexities with user-friendly APIs, allowing implementation without deep knowledge of libraries like PyTorch, evidenced by simple code examples for training.
Enables data-parallel distributed offline RL training with multiple GPUs or nodes, facilitating large-scale experiments, as shown in the distributed training example.
Offers full documentation with tutorials and reproduction scripts, including Google Colab notebooks for easy onboarding, ensuring accessibility for practitioners.
Version 2 introduces breaking changes, requiring careful version management and potential code adjustments, as warned in the README's important note.
Several algorithms like BEAR, AWAC, and IQL lack support for discrete control tasks, limiting their use in domains like Atari games without workarounds.
Pins specific versions such as gymnasium==1.0.0, which can cause compatibility issues in projects integrating other RL libraries or newer dependencies.
Relies solely on GitHub Issues for technical support with no direct developer contact, potentially slowing resolution for urgent or complex problems.