OpenVINO™ toolkit: An open source AI toolkit that makes it easier to write once, deploy anywhere.
What's New in Version 2025.4
OpenVINO™ 2025.4 takes your AI deployments to the next level with new features and performance enhancements. In this release, you’ll see continuous improvements for large language models (LLMs), optimized runtimes for Intel® hardware, and expanded capabilities for efficient AI deployment across edge, cloud, and local environments. Explore the latest updates and unlock new possibilities for your AI projects..
Latest Features
Easier model access and conversion
We’ve made model conversion easier.
|
Topic |
Details |
|---|---|
|
New Model Support |
|
GenAI and LLM enhancements
We've expanded model support and accelerated inference.
|
Topic |
Details |
|---|---|
|
GPU Plug-in Support |
|
|
GenAI Enhancements |
|
More portability and performance
Develop once, deploy anywhere. OpenVINO enables developers to run AI at the edge, in the cloud, or locally.
|
Topic |
Details |
|---|---|
|
Windows ML* Support |
|
|
Intel® Hardware Support |
|
|
Model Server Updates |
OpenVINO Model Server (OVMS) and OpenVINO™ GenAI now extend support for Agentic AI scenarios with new features such as output parsing, improved chat templates for reliable multi-turn interactions, and preview functionality for the Qwen3-30B-A3B model. OVMS also introduces a preview for audio endpoints. |
| NNCF Updates | The Neural Network Compression Framework (NNCF) ONNX backend now supports INT8 static post-training quantization (PTQ) and INT8/INT4 weight-only compression to ensure accuracy parity with OpenVINO IR format models. SmoothQuant algorithm support was added for INT8 quantization. |
Sign Up for Exclusive News, Tips & Releases
Be among the first to learn about everything new with the Intel® Distribution of OpenVINO™ toolkit. By signing up, you get early access product updates and releases, exclusive invitations to webinars and events, training and tutorial resources, and other breaking news.
Resources
Community and Support
Explore ways to get involved and stay up-to-date with the latest announcements.
The productive smart path to freedom from the economic and technical burdens of proprietary alternatives for accelerated computing.