• <Back to Collection
Masthead Light

Intel® Xeon Phi™ Coprocessors

Advanced performance for highly parallel workloads

What is the Intel® Xeon Phi™ coprocessor?

Intel® Xeon Phi™ coprocessors are PCI Express* form factor add-in cards that work synergistically with Intel® Xeon® processors to enable dramatic performance gains for highly parallel code—up to 1.2 double-precision teraFLOPS (floating point operations per second) per coprocessor.

Manufactured using Intel’s industry-leading 22nm technology with 3-D Tri-Gate transistors, each coprocessor features more cores, more threads, and wider vector execution units than an Intel Xeon processor. The high degree of parallelism compensates for the lower speed of each core to deliver higher aggregate performance for highly parallel workloads.

What applications can benefit from the Intel Xeon Phi coprocessor?

While a majority of applications (80 to 90 percent) will continue to achieve maximum performance on Intel Xeon processors, certain highly parallel applications will benefit dramatically by using Intel Xeon Phi coprocessors. To take full advantage of Intel Xeon Phi coprocessors, an application must scale well to over 100 software threads and either make extensive use of vectors or efficiently use more local memory bandwidth than is available on an Intel Xeon processor. Examples of segments with highly parallel applications include: animation, energy, finance, life sciences, manufacturing, medical, public sector, weather, and more. Learn more about Intel® Many Integrated Core Architecture (Intel® MIC Architecture) development >

When do I use Intel Xeon processors and Intel Xeon Phi coprocessors?

When do I use Intel Xeon® Processors and Intel Xeon Phi™ Coprocessors?

 

 

Think “reuse” rather than “recode”

Since languages, tools, and applications are compatible for both Intel Xeon processor and Intel Xeon Phi coprocessors, now you can think “reuse” rather than “recode”.

Single programming model for all your code. The Intel Xeon Phi coprocessor gives developers a hardware design point optimized for extreme parallelism, without requiring them to re-architect or rewrite their code. No need to rethink the entire problem or learn a new programming model; simply recompile and optimize existing code using familiar tools, libraries, and runtimes.

Performance multiplier. By maintaining a single source code between Intel Xeon processors and Intel Xeon Phi coprocessors, developers optimize once for parallelism but maximize performance on both processor and coprocessor.

Execution flexibility. Designed from the ground up for high-performance computing. Unlike a GPU, a coprocessor can host an operating system, be fully IP addressable, and support standards such as Message Passing Interface (MPI). Additionally, it can operate in multiple execution modes:

  • "Symmetric" mode: Workload tasks are shared between the host processor and coprocessor
  • "Native" mode: Workload resides entirely on the coprocessor, and essentially acts as a separate compute nodes
  • "Offload" mode: Workload resides on the host processor, and parts of the workload are sent out to the coprocessor as needed. NOTE: GPUs can ONLY be run in mode and often sits idle.