Introduction
This package contains the Intel® Distribution of OpenVINO™ Toolkit software version 2025.3 for Linux*, Windows* and macOS*.
Available Downloads
- Debian Linux*
- Size: 35.1 MB
- SHA256: FB0F02F103A382E2638C57D8DB61BD62A52F2652E8A1B91D0919C23DF75F6AB9
- Ubuntu 22.04 LTS*
- Size: 38.5 MB
- SHA256: 470F5C0FFBBF7C6983BD73C9CABF540FBC869043F5CC12F296E26011089464C0
- Ubuntu 22.04 LTS*
- Size: 58.9 MB
- SHA256: D701A115D3DC18088FF75B5B8E67A51FBF780022A3D40EE8EE7F2ADFBD9915E6
- Ubuntu 24.04 LTS*
- Size: 60 MB
- SHA256: DE0D5E16B161EFEA013A5C017E3B2BCE1191CA009A1947D392CCAB8ED9D0F6E4
- Red Hat Enterprise Linux 8*
- Size: 66 MB
- SHA256: 3D12347F8C02BDD86F58D5A75D7D18B7CFC60D558C14A58B525F4657D5440F31
- CentOS 7 (2003)*
- Size: 60 MB
- SHA256: 024348AE17CAE41E03F3DF495AEC78AC9C293781DB37338327AB9190ECAB795B
- macOS*
- Size: 39.7 MB
- SHA256: AEFCF28DEEDE2F6FD470C6218D0C9CCD47268C428E7B22B758FC087ECB35C184
- macOS*
- Size: 48.5 MB
- SHA256: 0CAA9758E09D7AE1F4783365DF91D4BB4DCF0DDFC9C3AD4E3AFFB8433172D41C
- Windows 11 Family*, Windows 10 Family*
- Size: 122.1 MB
- SHA256: 05685C652E85F92AD17572EC2800EA6D0B96C9B7FF645299AD2BA09D1AFB17B4
- Windows 11 Family*, Windows 10 Family*
- Size: 655.3 MB
- SHA256: C9252D1B056483275A5FDFCD663010A745C1A9DFAA5B1203C81F9E26B0956E22
Detailed Description
What's New
- More Gen AI coverage and frameworks integrations to minimize code changes
- New models supported: Phi-4-mini-reasoning, AFM-4.5B, Gemma-3-1B-it, Gemma-3-4B-it, and Gemma-3-12B.
- NPU support added for: Qwen3-1.7B, Qwen3-4B, and Qwen3-8B.
- LLMs optimized for NPU are now available on the OpenVINO Hugging Face collection.
- Preview: Intel® Core™ Ultra Processor and Windows-based AI PCs can now leverage the OpenVINO™ Execution Provider for Windows* ML for high-performance, off-the-shelf starting experience on Windows*.
- Broader LLM model support and more model compression optimization techniques
- The NPU plug-in adds support for longer contexts of up to 8K tokens, dynamic prompts, and dynamic LoRA for improved LLM performance.
- The NPU plug-in now supports dynamic batch sizes by reshaping the model to a batch size of 1 and concurrently managing multiple inference requests, enhancing performance and optimizing memory utilization.
- Accuracy improvements for GenAI models on both built-in and discrete graphics were achieved through the implementation of the key cache compression per channel technique, in addition to the existing KV cache per-token compression method.
- OpenVINO™ GenAI introduces TextRerankPipeline for improved retrieval relevance and RAG pipeline accuracy, plus Structured Output for enhanced response reliability and function calling while ensuring adherence to predefined formats.
- More portability and performance to run AI at the edge, in the cloud, or locally
- Announcing support for Intel® Arc™ Pro B-Series (B50 and B60).
- Preview: Hugging Face models that are GGUF-enabled for OpenVINO GenAI are now supported by the OpenVINO™ Model Server for popular LLM model architectures such as DeepSeek Distill, Qwen2, Qwen2.5, and Llama 3. This functionality reduces memory footprint and simplifies integration for GenAI workloads.
- With improved reliability and tool call accuracy, the OpenVINO™ Model Server boosts support for agentic AI use cases on AI PCs, while enhancing performance on Intel CPUs, built-in GPUs, and NPUs.
- int4 data-aware weights compression, now supported in the Neural Network Compression Framework (NNCF) for ONNX models, reduces memory footprint while maintaining accuracy and enables efficient deployment in resource-constrained environments.
Get all the details. See 2025.3 release notes.
Installation instructions
You can choose how to install OpenVINO™ Runtime from Archive* according to your operating system:
- Install OpenVINO Runtime on Linux*
- Install OpenVINO Runtime on Windows*
- Install OpenVINO Runtime on macOS*
What's included in the download package (Archive File)
- Offers both C/C++ and Python APIs
- Additionally includes code samples
Helpful Links
NOTE: Links open in a new window.
Disclaimers1
Product and Performance Information
Intel is in the process of removing non-inclusive language from our current documentation, user interfaces, and code. Please note that retroactive changes are not always possible, and some non-inclusive language may remain in older documentation, user interfaces, and code.