Latest Intel® AI Reference Model Optimizations for Intel® Xeon Scalable Processors

ID 793414
Updated 11/13/2023
Version Latest
Public

author-image

By

Description 

This document provides links to step-by-step instructions on how to leverage the latest reference model docker containers to run optimized open-source Deep Learning Training and Inference workloads using PyTorch* and TensorFlow* frameworks on Intel® Xeon® Scalable processors.

Note: The containers below are finely tuned to demonstrate best performance on Intel® Extension for PyTorch* and Intel® Optimized TensorFlow*  and are not intended for use in production. 

Use cases

The tables below link to documentation on how to run each use case using docker containers. These containers were validated on a host running Linux. 

Image Recognition

Framework
Model
Precisions 
Mode
Dataset
PyTorch
FP32,BF32,BF16
Training
ImageNet 2012
PyTorch

 

FP32,BF32,BF16,INT8
Inference
ImageNet 2012
PyTorch
FP32,BF32,BF16,INT8
Inference
ImageNet 2012
TensorFlow
FP32,BF16,FP16
Training
ImageNet 2012
TensorFlow
FP32,BF32,BF16,INT8
Inference
ImageNet 2012
TensorFlow
FP32,BF32,FP16,INT8
Inference
ImageNet 2012

Image Segmentation

Framework
Model
Precisions 
Mode
Dataset
TensorFlow
FP32,BF16,INT8
Inference
BRATS 2019

Object Detection

Framework
Model
Precisions 
Mode
Dataset
PyTorch
FP32,BF32,BF16
Training
COCO 2017
PyTorch
FP32,BF32,BF16
Inference
COCO 2017
PyTorch
FP32,BF32,BF16
Training
COCO 2017
PyTorch
FP32,BF32,BF16,INT8
Inference
COCO 2017
TensorFlow
FP32,BF32,BF16
Training
COCO 2017 
TensorFlow
FP32,BF16,INT8
Inference
COCO 2017 
TensorFlow
FP32,BF32,BF16,INT8
Inference
COCO 2017

Language Modeling

Framework
Model
Precisions
Mode
Dataset
PyTorch
FP32,BF32,BF16
Training
Preprocessed Text dataset
PyTorch
FP32,BF32,BF16,INT8
Inference
SQuAD1.0
PyTorch
FP32,BF32,BF16,INT8
Inference
LibriSpeech
PyTorch
FP32,BF32,FP16
Training
LibriSpeech
PyTorch
FP32,BF32,BF16,INT8-BF16,INT8-BF32
Inference
SST-2
TensorFlow
FP32,BF16
Training
SQuAD and MRPC
TensorFlow
FP32,BF32,BF16,INT8
Inference
SQuAD

Language Translation

Framework
Model
Precisions
Mode
Dataset
TensorFlow
Transformer_LT_mlperf
FP32,BF16
Training
WMT English-German dataset
TensorFlow
Transformer_LT_mlperf
FP32,BF32,BF16,INT8
Inference
WMT English-German dataset

Recommendation

Framework
Model
Precisions
Mode
Dataset
PyTorch
FP32,BF32,BF16
Training
Criteo Terabyte
PyTorch
FP32,BF32,BF16,INT8
Inference
Criteo Terabyte
TensorFlow
FP32,BF32,BF16
Inference
DIEN dataset

Documentation and Sources 

Get Started                                     Code Sources                                              

Main GitHub*                                                         PyTorch Dockerfiles

Release Notes                                                       TensorFlow Dockerfiles

Report Issue                                                                         

License Agreement

LEGAL NOTICE: By accessing, downloading or using this software and any required dependent software (the “Software Package”), you agree to the terms and conditions of the software license agreements for the Software Package, which may also include notices, disclaimers, or license terms for third party software included with the Software Package. Please refer to the license file for additional details.

View All Containers and Solutions 🡢