Intel® Gaudi® 2 AI accelerator
High performance acceleration for GenAI and LLMs.
Intel® Gaudi® 2 AI accelerator
Our Intel® Gaudi® 2 AI accelerator Is Driving Improved Deep Learning Price-performance
And operational efficiency for training and running state-of-the-art models, from the largest language and multi-modal models to more basic computer vision and NLP models. Designed for efficient scalability—whether in the cloud or in your data center, Intel Gaudi 2 accelerators bring the AI industry the choice it needs—now more than ever.
Intel® Gaudi® 2 Architecture Features:
Built on the proven first-gen Intel Gaudi design foundation.
- 7nm process technology
- Heterogeneous compute
- 24 Tensor Processor Cores
- Dual matrix multiplication engines
- 24 100 Gigabit Ethernet integrated on chip
- 96 GB HBM2E memory on board
- 48 MB SRAM
- Integrated Media Control
Intel® Gaudi® 2 AI accelerator In the Cloud
Discover the performance and ease of use Intel® Gaudi® 2 AI accelerator provides on the Intel® Tiber™ Developer Cloud.
Learn more
Intel® Gaudi® 2 AI accelerator In the Data Center
Bring the price-performance advantage into your infrastructure with solutions from Supermicro and IEI.
Learn more
Intel® Gaudi® 2 AI accelerator Remains the ONLY Benchmarked Alternative to Nvidia H100 for Gen AI Performance
Based on the most recent MLPerf Training 3.1 and Inference 4.0 benchmarks, published in Dec 2023 and Mar 2024 respectively, Intel Gaudi 2 accelerator remains the only benchmarked alternative to Nvidia H100.
In addition to the MLPerf industry benchmark, Intel Gaudi 2 AI accelerator scores on other third-party evaluations.1
Intel® Gaudi® 2 AI accelerator Supports Massive, Flexible Scale Out
With 24x 100 Gigabit Ethernet (RoCEv2) ports integrated onto every Intel Gaudi 2 AI accelerator, customers benefit from flexible and cost-efficient scalability that extends performance of the Intel Gaudi 2 AI accelerator from one to thousands of accelerators.
See below image for server reference design featuring 8x Intel Gaudi 2 accelerators.
For more information on building out system scale with Intel Gaudi 2 AI accelerator, see our Networking page ›
Featured Partner
Cisco Nexus 9364D-GX2A switches are qualified to connect Intel Gaudi 2 accelerators to build a scale-out network for running large-language model (LLM) training, inference, or similar generative AI workloads.
Easily Build New or Migrate Existing Models on Intel Gaudi 2 AI accelerators
Intel Gaudi software, optimized for Intel Gaudi platform performance and ease, gives developers the documents, tools, how-to content and reference models to help them get started on Intel Gaudi software quickly and with ease.
Access and easily implement over 500K models with Habana Optimum Library on the Hugging Face hub.
For more information, see our developer site ›
Resources for Intel Gaudi 2 AI accelerator
Product and Performance Information
Visit https://habana.ai/habana-claims-validation for workloads and configurations. Results may vary https://huggingface.co/blog/habana-gaudi-2-benchmark https://huggingface.co/blog/habana-gaudi-2-bloom