Skip To Main Content
Intel logo - Return to the home page
My Tools

Select Your Language

  • Bahasa Indonesia
  • Deutsch
  • English
  • Español
  • Français
  • Português
  • Tiếng Việt
  • ไทย
  • 한국어
  • 日本語
  • 简体中文
  • 繁體中文
Sign In to access restricted content

Using Intel.com Search

You can easily search the entire Intel.com site in several ways.

  • Brand Name: Core i9
  • Document Number: 123456
  • Code Name: Emerald Rapids
  • Special Operators: “Ice Lake”, Ice AND Lake, Ice OR Lake, Ice*

Quick Links

You can also try the quick links below to see results for most popular searches.

  • Product Information
  • Support
  • Drivers & Software

Recent Searches

Sign In to access restricted content

Advanced Search

Only search in

Sign in to access restricted content.

The browser version you are using is not recommended for this site.
Please consider upgrading to the latest version of your browser by clicking one of the following links.

  • Safari
  • Chrome
  • Edge
  • Firefox

Intel® oneAPI Collective Communications Library

Scalable & Efficient Distributed Training for Deep Neural Networks

 

  • Overview
  • Download
  • Documentation & Resources

Implement Multi-Node Communication Patterns

The Intel® oneAPI Collective Communications Library (oneCCL) enables developers and researchers to more quickly train newer and deeper models. This is done by using optimized communication patterns to distribute model training across multiple nodes.

The library is designed for easy integration into deep learning frameworks, whether you are implementing them from scratch or customizing existing ones.

  • Built on top of lower-level communication middleware. Message passing interface (MPI) and libfabrics transparently support many interconnects, such as Cornelis Networks*, InfiniBand*, and Ethernet.
  • Optimized for high performance on Intel CPUs and GPUs. 
  • Allows the tradeoff of compute for communication performance to drive scalability of communication patterns.
  • Enables efficient implementations of collectives that are heavily used for neural network training, including all-gather, all-reduce, and reduce-scatter.
Download the Stand-Alone Version

A stand-alone download of oneCCL is available. You can download binaries from Intel or choose your preferred repository.

Download
Help oneCCL Evolve

oneCCL is part of the oneAPI industry standards initiative. We welcome you to participate.

 

Specification

Open Source Version (GitHub*)

Download as Part of the Toolkit

oneCCL is included as part of the Intel® oneAPI Base Toolkit, which is a core set of tools and libraries for developing high-performance, data-centric applications across diverse architectures.

Get It Now

Features

Common APIs to Support Deep Learning Frameworks

oneCCL exposes a collective API that supports:

  • Commonly used collective operations found in deep learning and machine learning workloads
  • Interoperability with SYCL* from the Khronos* Group

 

Deep Learning Optimizations

The runtime implementation enables several optimizations, including: 

  • Asynchronous progress for compute communication overlap
  • Dedication of one or more cores to ensure optimal network use
  • Message prioritization, persistence, and out-of-order execution
  • Collectives in low-precision data types

Documentation & Code Samples

Documentation

  • Get Started Guide
  • Release Notes
  • System Requirements
  • Benchmark User Guide
  • Developer Guide and Reference

 

View All Documentation

Code Samples

Learn how to access oneAPI code samples in a tool command line or IDE.
 
  • oneCCL Get Started

View All Code Samples (GitHub)

Training

Understanding oneCCL

oneAPI Collective Communications Library  [5:07]

 

Distributed AI Acceleration

Accelerate Distributed AI with a oneCCL Framework [3:24]

 

Distributed Deep Learning Optimization

Optimize a Deep Learning Recommendation Model by Using PyTorch* with a oneCCL Back End 

Efficient Model Training on Multiple CPUs 

 

🗐 View All Resources 

🗗 Training & Events Calendar

Specifications

Processors:
  • Intel® Core™ processor family
  • Intel® Xeon® processor family
  • Intel® Xeon® Scalable processor family
     
GPU:
  • Intel® Data Center GPU Max Series
Operating system:
  • Linux*
     
Languages:
  • SYCL
  • C and C++

For more information, see the system requirements.

Compilers:
  • GNU Compiler Collection (GCC)*
  • Intel® oneAPI DPC++/C++ Compiler
     
Distributed environments:
  • MPI
  • OFI

Get Help

Your success is our success. Access these forum and GitHub resources when you need assistance.

  • oneCCLForum
  • General oneAPI Support

  • Company Overview
  • Contact Intel
  • Newsroom
  • Investors
  • Careers
  • Corporate Responsibility
  • Inclusion
  • Public Policy
  • © Intel Corporation
  • Terms of Use
  • *Trademarks
  • Cookies
  • Privacy
  • Supply Chain Transparency
  • Site Map
  • Recycling
  • Your Privacy Choices California Consumer Privacy Act (CCPA) Opt-Out Icon
  • Notice at Collection

Intel technologies may require enabled hardware, software or service activation. // No product or component can be absolutely secure. // Your costs and results may vary. // Performance varies by use, configuration, and other factors. Learn more at intel.com/performanceindex. // See our complete legal Notices and Disclaimers. // Intel is committed to respecting human rights and avoiding causing or contributing to adverse impacts on human rights. See Intel’s Global Human Rights Principles. Intel’s products and software are intended only to be used in applications that do not cause or contribute to adverse impacts on human rights.

Intel Footer Logo