Skip To Main Content
Intel logo - Return to the home page
My Tools

Select Your Language

  • Bahasa Indonesia
  • Deutsch
  • English
  • Español
  • Français
  • Português
  • Tiếng Việt
  • ไทย
  • 한국어
  • 日本語
  • 简体中文
  • 繁體中文
Sign In to access restricted content

Using Intel.com Search

You can easily search the entire Intel.com site in several ways.

  • Brand Name: Core i9
  • Document Number: 123456
  • Code Name: Emerald Rapids
  • Special Operators: “Ice Lake”, Ice AND Lake, Ice OR Lake, Ice*

Quick Links

You can also try the quick links below to see results for most popular searches.

  • Product Information
  • Support
  • Drivers & Software

Recent Searches

Sign In to access restricted content

Advanced Search

Only search in

Sign in to access restricted content.

The browser version you are using is not recommended for this site.
Please consider upgrading to the latest version of your browser by clicking one of the following links.

  • Safari
  • Chrome
  • Edge
  • Firefox

Deep Learning Inference with Intel® FPGAs

  • Summary
  • Modules
  • Videos

Summary

An FPGA provides an extremely low-latency, flexible architecture that enables deep learning acceleration in a power-efficient solution. Learn how to deploy a computer vision application on a CPU, and then accelerate the deep learning inference on the FPGA. Next, learn how to take that application and use Docker* containers to scale the application across multiple nodes in a cluster using Kubernetes*.

By the end of this course, students will have practical knowledge of:

  • What convolutional neural networks are and how they are built
  • How to build a deep learning computer vision application
  • What an FPGA is from a software developer's perspective, and why FPGAs are so well suited for accelerating real-time machine learning applications
  • The components of the Intel® FPGA Deep Learning Acceleration Suite
  • What constitutes a computer vision application that uses deep learning to extract patterns from data
  • How to use the Intel® Distribution of OpenVINO™ toolkit to target convolutional neural network (CNN) based inferencing on Intel® CPUs and FPGAs
  • How the Acceleration Stack for Intel® Xeon® CPUs with FPGAs enables higher level cloud and data center software applications to leverage the FPGA seamlessly

The course is structured around five weeks of lectures and exercises. Each week requires three hours to complete. The exercises are implemented in Python*.

Prerequisites

Python programming

Working knowledge of pandas and scikit-learn*

Basic statistics

Deep Learning Course

Machine Learning Course

Week 1

This class reviews the basics of deep learning and FPGAs. Topics include:

  • Machine learning terminology and use cases
  • Basic topologies such as feed-forward networks and AlexNet
  • An overview of FPGA architecture, advantages, and uses
Download
Week 2

This class teaches how to make computer vision applications. Topics include:

  • The essential components of computer vision software
  • How Intel® software and hardware are used to improve applications
  • The common practices, languages, tools, and libraries used for computer vision
Download
Week 3

This class teaches about the Intel Distribution of OpenVINO toolkit. Topics include:

  • An overview of the Intel Distribution of OpenVINO toolkit and how to use each component for computer vision
  • How to convert and optimize a Caffe* or TensorFlow* model into the format for the inference engine
  • Why using the inference engine for FPGA accelerator speeds up vision applications
Download
Week 4

This class explains the Intel FPGA Deep Learning Acceleration Suite. Topics include:

  • How the Intel Distribution of OpenVINO toolkit can map network topologies onto FPGA architecture
  • The different deep learning architectures available for FPGAs
  • How lower precision is handled in FPGA machine learning models
Download
Week 5

This class explains how the acceleration stack can be used for FPGAs. Topics include:

  • How to use the acceleration stack to enable FPGA clusters
  • Learning about the Open Programmable Acceleration Engine (OPAE) for application developers
  • How to set up a host application to discover an FPGA accelerator
Download

Watch the Classes

  • Company Overview
  • Contact Intel
  • Newsroom
  • Investors
  • Careers
  • Corporate Responsibility
  • Inclusion
  • Public Policy
  • © Intel Corporation
  • Terms of Use
  • *Trademarks
  • Cookies
  • Privacy
  • Supply Chain Transparency
  • Site Map
  • Recycling
  • Your Privacy Choices California Consumer Privacy Act (CCPA) Opt-Out Icon
  • Notice at Collection

Intel technologies may require enabled hardware, software or service activation. // No product or component can be absolutely secure. // Your costs and results may vary. // Performance varies by use, configuration, and other factors. Learn more at intel.com/performanceindex. // See our complete legal Notices and Disclaimers. // Intel is committed to respecting human rights and avoiding causing or contributing to adverse impacts on human rights. See Intel’s Global Human Rights Principles. Intel’s products and software are intended only to be used in applications that do not cause or contribute to adverse impacts on human rights.

Intel Footer Logo