A dataflow compiler for quantized neural network inference on FPGAs, generating highly efficient custom accelerators.
FINN is a dataflow compiler framework for quantized neural network inference on FPGAs. It generates highly efficient, customized dataflow-style architectures to accelerate neural network inference, achieving high throughput and low latency. The framework is open-source and experimental, developed by AMD Research & Advanced Development to explore neural network implementations across software/hardware stacks.
Researchers and engineers working on FPGA-based deep learning acceleration, particularly those focused on quantized neural networks and custom hardware architectures. It's also suitable for developers exploring high-performance inference solutions with low latency requirements.
Developers choose FINN for its ability to generate highly efficient, dataflow-style FPGA accelerators tailored to specific quantized neural networks, offering superior performance and flexibility compared to generic inference frameworks. Its open-source nature enables deep customization and research across the hardware/software stack.
Dataflow compiler for QNN inference on FPGAs
Specifically targets quantized neural networks, enhancing FPGA efficiency and performance for low-precision inference as emphasized in the README.
Generates dataflow-style architectures tailored to each network, achieving high throughput and low latency through specialized hardware pipelines.
Fully open-source, enabling deep customization and research across software/hardware abstraction layers for advanced users and academia.
Uses Docker for compilation to manage complex dependencies, ensuring reproducible builds and easier setup in controlled environments.
Labeled as experimental, so it lacks the stability, regular updates, and production support of mature frameworks like TensorRT or Vitis AI.
Only supports Docker-based execution, which adds container overhead and limits flexibility for bare-metal or non-containerized deployments.
Requires FPGA development tools and hardware access, making initial setup challenging and time-consuming for those unfamiliar with FPGA workflows.
A toolkit for developing and comparing reinforcement learning algorithms.
The fastai deep learning library
The open source AI engineering platform for agents, LLMs, and ML models. MLflow enables teams of all sizes to debug, evaluate, monitor, and optimize production-quality AI applications while controlling costs and managing access to models and data.
MNN: A blazing-fast, lightweight inference engine battle-tested by Alibaba, powering high-performance on-device LLMs and Edge AI.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.