A cross-platform, high-performance accelerator for machine learning inference and training with ONNX models.
ONNX Runtime is a cross-platform inference and training accelerator for machine learning models in the ONNX format. It enables faster model execution and lower deployment costs by optimizing performance across different hardware platforms while supporting models from popular frameworks like PyTorch, TensorFlow, and scikit-learn.
Machine learning engineers and developers who need to deploy trained models in production environments with high performance requirements across diverse hardware configurations.
Developers choose ONNX Runtime for its exceptional performance optimizations, broad framework compatibility, and ability to leverage hardware accelerators without rewriting model code, making it ideal for production ML deployments.
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.
Runs on Windows, Linux, macOS and leverages various hardware accelerators like NVIDIA GPUs, enabling optimal performance across diverse environments as stated in the README.
Supports models from PyTorch, TensorFlow, scikit-learn, LightGBM, and XGBoost, allowing deployment without rewriting code for different frameworks, highlighting its versatility.
Reduces inference latency and costs through graph optimizations and hardware acceleration, with documentation emphasizing faster customer experiences.
Speeds up transformer training on multi-node NVIDIA GPUs with a one-line addition in PyTorch scripts, as mentioned in the README for minimal code changes.
Models must be converted to ONNX format, which can be complex and may not support all operations or latest framework features, adding a step to deployment pipelines.
While broad, support for non-standard or custom models might require additional work, and not all ML libraries export seamlessly to ONNX, potentially causing compatibility issues.
Integrating ONNX Runtime involves learning its APIs and managing dependencies for different hardware providers, which can be steep for simple or existing deployments without clear plug-and-play solutions.