This document provides a summary of a presentation about quantized neural network inference on FPGAs using FINN and LogicNets. It discusses:
- Xilinx Research Labs in Dublin and their work quantifying machine learning applications on Xilinx devices.
- How neural network quantization can improve efficiency by reducing precision while trading off accuracy, and how this is well-suited for FPGAs.
- The FINN toolflow which includes quantization-aware training in PyTorch with Brevitas, the FINN compiler to map networks to hardware, and deployment with PYNQ.
- LogicNets which further improves efficiency by unfolding DNNs into fully pipelined datapath circuits for
Related topics: