A uniform interface to run deep learning models from multiple frameworks like TensorFlow, PyTorch, and Keras in C++ and Python.
Neuropod is a library that provides a uniform interface to run deep learning models from multiple frameworks like TensorFlow, PyTorch, Keras, and TorchScript. It solves the problem of framework lock-in by allowing researchers to build models in their preferred framework while simplifying production deployment with a consistent inference API.
Machine learning engineers and researchers who need to deploy models from various deep learning frameworks into production environments, especially those working in teams using multiple frameworks.
Developers choose Neuropod because it eliminates framework-specific inference code, enables easy model swapping, and provides tools like Problem APIs to standardize and optimize ML pipelines across different frameworks and versions.
A uniform interface to run deep learning models from multiple frameworks
Enables running TensorFlow, PyTorch, Keras, and TorchScript models with identical code, as shown in the README where TensorFlow and PyTorch addition models are run using the same inference call.
Allows defining input/output specifications for problems like 2D object detection, facilitating model swapping and shared inference pipelines without code changes, as detailed in the Problem API section.
Supports both C++ and Python, including PyTorch models without TorchScript conversion, making it versatile for production systems with mixed language requirements.
Uses out-of-process execution to run multiple framework versions concurrently, such as Torch nightly with stable releases, enabling safe experimentation alongside production models.
Only supports TensorFlow, PyTorch, Keras, TorchScript, and Ludwig; lacks integration with other popular frameworks like MXNet, JAX, or ONNX Runtime, which may limit adoption in heterogeneous environments.
Requires managing dependencies for multiple deep learning backends, increasing installation complexity and deployment footprint compared to single-framework solutions.
The uniform API layer may introduce latency and memory overhead compared to native framework APIs, especially for high-throughput or real-time inference tasks, though zero-copy operations mitigate this partially.
A toolkit for developing and comparing reinforcement learning algorithms.
The fastai deep learning library
The open source AI engineering platform for agents, LLMs, and ML models. MLflow enables teams of all sizes to debug, evaluate, monitor, and optimize production-quality AI applications while controlling costs and managing access to models and data.
MNN: A blazing-fast, lightweight inference engine battle-tested by Alibaba, powering high-performance on-device LLMs and Edge AI.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.