FPGA AI Suite Handbook

ID 863373
Date 11/21/2025
Public
Document Table of Contents

2.4. The FPGA AI Suite IP

The FPGA AI Suite IP is an RTL-instantiable configurable IP with AXI interfaces that you can instantiate into a generic embedded FPGA system.

The IP is configured through parameters defined in an architecture description file (.arch), often called an architecture file. The architecture description file, along with the OpenVINO™ intermediate representation of your trained model, is compiled by the FPGA AI Suite compiler into configuration instructions for the IP.

Figure 5. High-Level Architecture of the FPGA AI Suite IP

The following diagram shows a high-level architecture of the FPGA AI Suite IP:

The primary parameters defined in an Architecture Description File cover the following properties:

  • PE array vectorization
  • Scratch pad sizing
  • External memory bus bandwidth
  • Types/vectorization of auxiliary layer blocks

The following diagram is an architecture diagram for a specific instantiation of the FPGA AI Suite IP. The blocks connected to the crossbar in this diagram are examples. The selection of blocks connected to the crossbar are determined by compile time parameters.

Figure 6. Architecture of an Example Instantiation of the FPGA AI Suite IP
Two teams are typically involved in the implementation of an AI feature:
  • An AI/software development team responsible for developing and delivering an AI model.
  • An FPGA hardware development team responsible for integrating the FPGA AI Suite IP and runtime together into a system.

Defining the IP architecture straddles the boundary between these two teams. The ML team must develop an AI model that meets the target performance in some parameterization of the configurable IP. The FPGA team must ensure it fits onto the FPGA and closes timing.

Although responsibility for defining the parameterization of the configurable architecture can lie with either team (but is a joint responsibility), it can often be easier for the ML team to define the architecture.

The team responsible for defining the IP parameterization can use the FPGA AI Suite compiler (dla_compiler command) area and performance estimator tools to guide their decisions. For more information about using the compiler, refer to Compiling Your Model with the FPGA AI Suite Compiler.

In addition to the FPGA team and the ML team, another team is likely responsible for the software integration on the host. Depending on the system details, this software is likely responsible for interfacing with OpenVINO™ and communicating (via the BSP) with the FPGA AI Suite IP. This software is likely based on the runtime system that is included with the FPGA AI Suite design examples.