Skip To Main Content
Intel logo - Return to the home page
My Tools

Select Your Language

  • Bahasa Indonesia
  • Deutsch
  • English
  • Español
  • Français
  • Português
  • Tiếng Việt
  • ไทย
  • 한국어
  • 日本語
  • 简体中文
  • 繁體中文
Sign In to access restricted content

Using Intel.com Search

You can easily search the entire Intel.com site in several ways.

  • Brand Name: Core i9
  • Document Number: 123456
  • Code Name: Emerald Rapids
  • Special Operators: “Ice Lake”, Ice AND Lake, Ice OR Lake, Ice*

Quick Links

You can also try the quick links below to see results for most popular searches.

  • Product Information
  • Support
  • Drivers & Software

Recent Searches

Sign In to access restricted content

Advanced Search

Only search in

Sign in to access restricted content.

The browser version you are using is not recommended for this site.
Please consider upgrading to the latest version of your browser by clicking one of the following links.

  • Safari
  • Chrome
  • Edge
  • Firefox

🡐 AI Overview
 

AI Frameworks and Tools


Software tools at all levels of the AI stack unlock the full capabilities of your Intel® hardware. All Intel® AI tools and frameworks are built on the foundation of a standards-based, unified oneAPI programming model that helps you get the most performance from your end-to-end pipeline on all your available hardware.

 

 

 

  • Overview
  • Download
  • Documentation
This site is protected by reCAPTCHA and the Google <a href="https://policies.google.com/privacy">Privacy Policy</a> and <a href="https://policies.google.com/terms"> Terms of Service</a> apply.

All packages are for Linux only. Components have been validated with Python 3.11. Older versions of Python may not be supported. For compatibility details, refer to the System Requirements.

Offline Installer Deprecation Notice

AI Tools offline installer is deprecated in the 2025.1 release and will be removed in the future.​

  • AI Tools 2025.0 is the only version of the offline installer.​ You can continue to use it with the 2025.1 release, provided you also have 2025.0 installed.​
  • You can continue to acquire AI Tools through conda, pip, and Docker.

Accelerate your machine learning and data science pipelines with the power of open libraries optimized for Intel® architectures.

Boost the performance of your workloads, reduce model size, and improve the speed of your Deep Learning deployments on Intel® Xeon® processors with Intel® Extension for PyTorch.

Boost the performance of your workloads, reduce model size, and improve the speed of your Deep Learning deployments on Intel® Data Center GPU Max Series with Intel® Extension for PyTorch.

Boost the performance of your workloads, reduce model size, and improve the speed of your Deep Learning deployments on Intel® Xeon® processors with Intel® Extension for TensorFlow.

Boost the performance of your workloads, reduce model size, and improve the speed of your Deep Learning deployments on Intel® Data Center GPU Max Series with Intel® Extension for TensorFlow.

Reduce model size and improve the speed of your Deep Learning deployments on Intel® Xeon® processors with JAX.

 

This open source toolkit enables you to optimize a deep learning model from almost any framework and deploy it with best-in-class performance on a range of Intel processors and other hardware platforms.

A separate download is required. 

Download

Intel® Gaudi® Software

The Intel® Gaudi® AI accelerator is designed to maximize training throughput and efficiency while providing developers with optimized software and tools that scale to many workloads and systems. Intel® Gaudi® software was developed with the end user in mind, providing versatility and ease of programming to address the unique needs of users’ proprietary models, while allowing for a simple and seamless transition of their existing models over to Intel® Gaudi® technology. The Intel Gaudi software enables efficient mapping of neural network topologies onto Intel Gaudi technology.

The software suite includes a graph compiler and runtime, Tensor Processor Core (TPC) kernel library, firmware and drivers, and developer tools. Intel Gaudi software is integrated with PyTorch, and supports DeepSpeed for large language models (LLM) and performance-optimized Hugging Face models for transformer and diffusion uses.

The links below will allow you to access the PyTorch for Docker containers that contain the full Intel Gaudi software and PyTorch framework. It is recommended to use the Docker images when running models. Use the installation guide to learn how to run the Docker images or perform a manual installation on a bare metal system. Refer to the Support Matrix to see the latest versions of external software and drivers that were used in the Intel Gaudi software release.

Container

Installation Guide

Support Matrix

Intel, the Intel logo, and Intel Gaudi technology are trademarks of Intel Corporation or its subsidiaries.

Customize your tool selections for conda and pip.

Download

Register your product to receive update notifications plus hand-curated technical articles, tutorials, and training opportunities to help you optimize your code.

Size
Version
Date
SHA384

Service Currently Unavailable. Please Try Again Later
Yes, I would like to subscribe to stay connected to the latest Intel technologies and industry trends by email and telephone. I can unsubscribe at any time.
Intel strives to provide you with a great, personalized experience, and your data helps us to accomplish this.
I consent to Intel collecting and using my personal information as described below.
I agree to Intel's transfer of my personal information outside of China for processing in accordance with Intel’s Privacy and Cookie Statement and Intel's Privacy Policy Supplement for Chinese Users.
Continue as a Guest (download starts immediately) →
By submitting this form, you are confirming you are an adult 18 years or older and you agree to share your personal information with Intel to use for this business request. You also agree to subscribe to stay connected to the latest Intel technologies and industry trends by email and telephone. You may unsubscribe at any time. Intel’s web sites and communications are subject to our Privacy Notice and Terms of Use.
By submitting this form, you are confirming you are an adult 18 years or older and you agree to share your personal information with Intel to use for this business request. Intel’s web sites and communications are subject to our Privacy Notice and Terms of Use.
By downloading, you agree to our Privacy and Terms of use

Thank You

Your Download should start immediately.

Failed to submit your form.

Due to a technical difficulty, we were unable to submit the form. Please try again after a few minutes. We apologize for the inconvenience.

1. Download and Launch the Installer

a. (Optional) Download the installer package. If you have not yet downloaded it using the previous link, use the following command: 

wget https://registrationcenter-download.intel.com/akdlm/IRC_NAS/491d5c2a-67fe-48d0-884f-6aecd88f5d8a/ai-tools-2025.0.0.75_offline.sh

b. Launch the command line installer:

sh ai-tools-2025.0.0.75_offline.sh

Accept the license terms and enter the installation path.

2. Set Up Your Environment

a. Activate the AI Tools base environment:

source $HOME/intel/oneapi/intelpython/bin/activate

For detailed instructions, refer to Get Started with the AI Tools.

b. (For GPU users) Set up your system for GPU development as described in Get Started with the AI Tools: GPU Users.

3. Run a Get Started Sample

After a successful installation, to start using the installed product, see Build and Run a Sample Project.

  1. Set Up Your Environment

a. Create and activate a new conda environment using the following syntax. If needed, replace my-env with your preferred name for the environment.


conda create -n my-env
conda activate my-env

b. (Optional) To speed up running, use libmamba as the solver, which is the default in the latest conda distribution. For older conda installations, use this command to set libmamba: ​conda config --set solver libmamba​.

c. For a detailed walk-through on setting up conda, see Set Up System Before Installation: conda.

  1. Set Up Your Environment

a. Create and activate a new conda environment using the following syntax. If needed, replace my-env with your preferred name for the environment.


conda create -n my-env
conda activate my-env

b. (Optional) To speed up running, use libmamba as the solver, which is the default in the latest conda distribution. For older conda installations, use this command to set libmamba: ​conda config --set solver libmamba​. For a detailed walk-through on setting up conda, see Set Up System Before Installation: conda.

c. GPU installations have one additional step. For GPU optimizations, install the latest GPU drivers separately as described in Intel® Software for General Purpose GPU Capabilities.

  1. Set Up Your Environment

a. Prerequisite: If you do not have pip, use the Installation Instructions. After installation, make sure that you can run pip from the command line. 

b. Create and activate a virtual environment using the following syntax. Replace my-env with your preferred name for the environment: 


python3 -m venv my-env
source my-env/bin/activate

To install Python and define your environment, see Set Up System Before Installation: pip.

  1. Set Up Your Environment

a. Prerequisite: If you do not have pip, use the Installation Instructions. After installation, make sure that you can run pip from the command line. 

b. Create and activate a virtual environment using the following syntax. If needed, replace my-env with your preferred name for the environment:


python3 -m venv my-env
source my-env/bin/activate

To install Python and define your environment, see Set Up System Before Installation: pip.

Perform the following steps for GPU installation. GPU installations have additional steps.

c. Install Intel® Deep Learning Essentials for Linux 2025.0.2.

d. Install the latest GPU drivers separately, as described in Intel® Software for General Purpose GPU Capabilities. 

  1. Set Up Your Environment

Docker* Containers

Before running the containers, install Docker as described in the Docker Installation Instructions.

 2. Install with conda

 2. Install with conda

Alert: To display the command string, select the checkbox next to the software title you need. 

If applicable, disregard a “ClobberError” message associated with installation paths. This error does not impact the functionality of the installed packages.

 2. Install with pip

 2. Install with pip

Alert: To display the command string, select the checkbox next to the software title you need. PyTorch or TensorFlow must be installed first, if you are downloading those components.

2. Install with Docker

You may need to use sudo with the pull command to ensure proper permissions.

  3. Verify Installation

To verify that the Classic Machine Learning preset is properly installed, enter the following commands:

Intel® Optimization for XGBoost: python -c "import xgboost as xgb; print(xgb.__version__)"

Intel® Extension for Scikit-learn: python -c "from sklearnex import patch_sklearn; patch_sklearn()"

Modin: python -c "import modin; print(modin.__version__)"​​

  3. Verify Installation

To verify that the Deep Learning PyTorch CPU preset is properly installed, enter the following commands:

Intel® Extension for PyTorch (CPU): python -c "import torch; import intel_extension_for_pytorch as ipex; print(torch.__version__); print(ipex.__version__);" 

Intel® Neural Compressor: python -c "import neural_compressor as inc; print(inc.__version__)" 

ONNX Runtime: python -c "import onnxruntime; print(onnxruntime.__version__)"

  3. Verify Installation

To verify that the Deep Learning PyTorch GPU preset is properly installed, enter the following commands:

Intel® Extension for PyTorch (GPU): python -c "import torch; import intel_extension_for_pytorch as ipex; print(torch.__version__); print(ipex.__version__); [print(f'[{i}]: {torch.xpu.get_device_properties(i)}') for i in range(torch.xpu.device_count())];" 

Intel® Neural Compressor: python -c "import neural_compressor as inc; print(inc.__version__)"

  3. Verify Installation

To verify that the Deep Learning TensorFlow CPU preset is properly installed, enter the following commands:

Intel® Extension for TensorFlow (CPU): python -c "import intel_extension_for_tensorflow as itex; print(itex.__version__)" 

Intel® Neural Compressor: python -c "import neural_compressor as inc; print(inc.__version__)"

ONNX Runtime: python -c "import onnxruntime; print(onnxruntime.__version__)"

  3. Verify Installation

To verify that the Deep Learning TensorFlow GPU preset is properly installed, enter the following commands:

Intel® Extension for TensorFlow (GPU): python -c "from tensorflow.python.client import device_lib; print(device_lib.list_local_devices())" 

Intel® Neural Compressor: python -c "import neural_compressor as inc; print(inc.__version__)"

  3. Verify Installation

To verify that the Deep Learning JAX CPU preset is properly installed, enter the following command:

JAX: python -c "import jax; print(jax.__version__)"

  3. Verify Installation

To verify that the selected CPU-compatible components are properly installed, enter the following commands:

Intel® Extension for PyTorch (CPU): python -c "import torch; import intel_extension_for_pytorch as ipex; print(torch.__version__); print(ipex.__version__);"

Intel® Extension for TensorFlow (CPU): python -c "import intel_extension_for_tensorflow as itex; print(itex.__version__)"

Intel® Optimization for XGBoost: python -c "import xgboost as xgb; print(xgb.__version__)"

Intel® Extension for Scikit-learn: python -c "from sklearnex import patch_sklearn; patch_sklearn()"

Modin: python -c "import modin; print(modin.__version__)"

Intel® Neural Compressor: python -c "import neural_compressor as inc; print(inc.__version__)"

JAX: python -c "import jax; print(jax.__version__)"  

ONNX Runtime: python -c "import onnxruntime; print(onnxruntime.__version__)"

  3. Verify Installation

To verify that the selected GPU-compatible components are properly installed, enter the following commands:

Intel® Extension for PyTorch* (GPU): python -c "import torch; import intel_extension_for_pytorch as ipex; print(torch.__version__); print(ipex.__version__); [print(f'[{i}]: {torch.xpu.get_device_properties(i)}') for i in range(torch.xpu.device_count())];"

Intel® Extension for TensorFlow* (GPU): python -c "from tensorflow.python.client import device_lib; print(device_lib.list_local_devices())"

Intel® Optimization for XGBoost*: python -c "import xgboost as xgb; print(xgb.__version__)"

Intel® Extension for Scikit-learn*: python -c "from sklearnex import patch_sklearn; patch_sklearn()"

Modin*: python -c "import modin; print(modin.__version__)"

Intel® Neural Compressor: python -c "import neural_compressor as inc; print(inc.__version__)"

  3. Run Docker

To run a preset container and a get-started sample, follow the instructions in Intel AI Tools Selector Preset Containers. There are several options to run a container on a CPU or GPU.

  4. Run a Get Started Sample

After a successful installation, to start using the installed product, see Get Started Samples for AI Tools.

  Next Steps and Additional Resources

Next Steps

  • The Intel® AI Reference Models (formerly Model Zoo) repository contains links to pretrained models, sample scripts, best practices, and tutorials for many popular open source machine learning models optimized by Intel.
  • The Working with Preset Containers document provides more information about preset containers and instructions on how to run them.

Additional Resources

  • System Requirements 
  • AI Tools Selector Guide

Thank you for signing up!

Register

Register your product to receive updates plus hand-curated technical articles, tutorials, and training opportunities to help optimize your code.

Yes, I would like to subscribe to stay connected to the latest Intel technologies and industry trends by email and telephone. I can unsubscribe at any time.
Intel strives to provide you with a great, personalized experience, and your data helps us to accomplish this.
I consent to Intel collecting and using my personal information as described below.
I agree to Intel's transfer of my personal information outside of China for processing in accordance with Intel’s Privacy and Cookie Statement and Intel's Privacy Policy Supplement for Chinese Users.
By submitting this form, you are confirming you are an adult 18 years or older and you agree to share your personal information with Intel to use for this business request. You also agree to subscribe to stay connected to the latest Intel technologies and industry trends by email and telephone. You may unsubscribe at any time. Intel’s web sites and communications are subject to our Privacy Notice and Terms of Use.
By submitting this form, you are confirming you are an adult 18 years or older and you agree to share your personal information with Intel to use for this business request. Intel’s web sites and communications are subject to our Privacy Notice and Terms of Use.

Versions

Products have been updated to include functional and security updates. Customers should update to the latest versions as they become available.

Docker License Information

By accessing, downloading, or using this software and any required dependent software (the “Software Package”), you agree to the terms and conditions of the software license agreements for the Software Package, which may also include notices, disclaimers, or license terms for third-party software included with the Software Package. Preset containers are published under Apache License 2.0.

  

Support

Start-up support is available if there is an issue with the tool selector functionality.

Report an Issue

Feedback Welcome

Share your thoughts about this tool selector and provide suggestions for improvement. Responses are anonymous.

Start Survey
  • Company Overview
  • Contact Intel
  • Newsroom
  • Investors
  • Careers
  • Corporate Responsibility
  • Inclusion
  • Public Policy
  • © Intel Corporation
  • Terms of Use
  • *Trademarks
  • Cookies
  • Privacy
  • Supply Chain Transparency
  • Site Map
  • Recycling
  • Your Privacy Choices California Consumer Privacy Act (CCPA) Opt-Out Icon
  • Notice at Collection

Intel technologies may require enabled hardware, software or service activation. // No product or component can be absolutely secure. // Your costs and results may vary. // Performance varies by use, configuration, and other factors. Learn more at intel.com/performanceindex. // See our complete legal Notices and Disclaimers. // Intel is committed to respecting human rights and avoiding causing or contributing to adverse impacts on human rights. See Intel’s Global Human Rights Principles. Intel’s products and software are intended only to be used in applications that do not cause or contribute to adverse impacts on human rights.

Intel Footer Logo