Skip To Main Content
Intel logo - Return to the home page
My Tools

Select Your Language

  • Bahasa Indonesia
  • Deutsch
  • English
  • Español
  • Français
  • Português
  • Tiếng Việt
  • ไทย
  • 한국어
  • 日本語
  • 简体中文
  • 繁體中文
Sign In to access restricted content

Using Intel.com Search

You can easily search the entire Intel.com site in several ways.

  • Brand Name: Core i9
  • Document Number: 123456
  • Code Name: Emerald Rapids
  • Special Operators: “Ice Lake”, Ice AND Lake, Ice OR Lake, Ice*

Quick Links

You can also try the quick links below to see results for most popular searches.

  • Product Information
  • Support
  • Drivers & Software

Recent Searches

Sign In to access restricted content

Advanced Search

Only search in

Sign in to access restricted content.

The browser version you are using is not recommended for this site.
Please consider upgrading to the latest version of your browser by clicking one of the following links.

  • Safari
  • Chrome
  • Edge
  • Firefox

Optimize Distributed Training and Inference for Intel® Data Center GPUs

@IntelDevTools

Subscribe Now

Stay in the know on all things CODE. Updates are delivered to your inbox.

Sign Up

Overview

The complexity of deep learning models is surging, which warrants enhanced training and inference in distributed compute environments. This session focuses on the essential techniques to use with Intel® Data Center GPUs to balance distributed AI workloads and meet data center challenges to improve advances in efficiency and performance.

Explore Intel® Extension for PyTorch*, which optimizes neural network operations on Intel® hardware, and learn how DeepSpeed* can be integrated to perform training operations at scale.

Included topics:

  • Tackle model scalability in a distributed environment skillfully, handling workloads efficiently across Intel Data Center GPUs.
  • Gain familiarity with essential Intel tools to simplify operations, including PyTorch Distributed Data Parallel (DDP), Intel® oneAPI Collective Communications Library (oneCCL), and the DeepSpeed library that streamlines network training at scale.
  • Deploy practical solutions that maximize hardware efficiency and perfect strategies that ensure top performance for AI development.
  • Sample code and see benchmarking milestones, using tools such as Intel Extension for PyTorch large language models (LLM) to illustrate performance achievements.

Skill level: Any

 

Featured Software

  • oneCCL
  • Intel Extension for PyTorch (GitHub*) or from the AI Frameworks and Tools Selector
  • Intel® Extension for DeepSpeed*

Jump to:


You May Also Like
 

   

You May Also Like

Related Articles

Intel® Gaudi® 2 AI Processor for Deep Learning Gets Even Better

PyTorch Optimizations from Intel

Trusted AI in Intel® Tiber™ Developer Cloud

Related Videos

Prototype and Deploy LLM Applications on Intel® NPUs

Intel® Extension for PyTorch*: New Features on CPUs and GPUs

  • Company Overview
  • Contact Intel
  • Newsroom
  • Investors
  • Careers
  • Corporate Responsibility
  • Inclusion
  • Public Policy
  • © Intel Corporation
  • Terms of Use
  • *Trademarks
  • Cookies
  • Privacy
  • Supply Chain Transparency
  • Site Map
  • Recycling
  • Your Privacy Choices California Consumer Privacy Act (CCPA) Opt-Out Icon
  • Notice at Collection

Intel technologies may require enabled hardware, software or service activation. // No product or component can be absolutely secure. // Your costs and results may vary. // Performance varies by use, configuration, and other factors. Learn more at intel.com/performanceindex. // See our complete legal Notices and Disclaimers. // Intel is committed to respecting human rights and avoiding causing or contributing to adverse impacts on human rights. See Intel’s Global Human Rights Principles. Intel’s products and software are intended only to be used in applications that do not cause or contribute to adverse impacts on human rights.

Intel Footer Logo