OpenVINO™ toolkit: An open source AI toolkit that makes it easier to write once, deploy anywhere.
What's New in Version 2025.2
OpenVINO toolkit 2025.2 takes your AI deployments to the next level with new features and performance enhancements. In this release, you’ll see continuous improvements for large language models (LLMs), optimized runtimes for Intel® hardware, and expanded capabilities for efficient AI deployment across edge, cloud, and local environments. Explore the latest updates and unlock new possibilities for your AI projects.
Latest Features
Easier Model Access and Conversion
Product |
Details |
---|---|
New Model Support |
New models supported on CPUs & GPUs: Phi-4, Mistral-7B-Instruct-v0.3, SD-XL Inpainting 0.1, Stable Diffusion* 3.5 Large Turbo, Phi-4-reasoning, Qwen3, and Qwen2.5-VL-3B-Instruct. Mistral-7B-Instruct-v0.3 is also supported on NPUs. |
GenAI and LLM Enhancements
Expanded model support and accelerated inference.
Feature |
Details |
---|---|
KV Cache Compression |
KV cache compression for CPUs is enabled by default for int8, providing a reduced memory footprint while maintaining accuracy compared to FP16. Additionally, it delivers substantial memory savings for LLMs with int4 support compared to int8. |
GenAI Enhancements |
|
More Portability and Performance
Develop once, deploy anywhere. OpenVINO toolkit enables developers to run AI at the edge, in the cloud, or locally.
Product |
Details |
---|---|
Intel® Hardware Support |
|
Model Server Updates |
OpenVINO™ model server now offers a streamlined C++ version for Windows* and enables improved performance for long-context models through prefix caching and a smaller Windows package that eliminates the Python dependency. Support for Hugging Face* models is now included. |
NNCF Updates |
Support for int4 data-free weights compression for ONNX* (Open Neural Network Exchange) models implemented in the Neural Network Compression Framework (NNCF). |
Sign Up for Exclusive News, Tips & Releases
Be among the first to learn about everything new with the Intel® Distribution of OpenVINO™ toolkit. By signing up, you get early access product updates and releases, exclusive invitations to webinars and events, training and tutorial resources, and other breaking news.
Resources
Community and Support
Explore ways to get involved and stay up-to-date with the latest announcements.
Get Started
Optimize, fine-tune, and run comprehensive AI inference using the included model optimizer and runtime and development tools.
The productive smart path to freedom from the economic and technical burdens of proprietary alternatives for accelerated computing.