Get the latest tech news

Intel's OpenVINO 2025.0 Brings Support For Deepseek Models, Better AI Performance


Intel's software engineers working on the OpenVINO AI toolkit today released OpenVINO 2025.0 that brings support for the much talked about Deepseek models along with other large language models (LLMs), performance improvements to some of the existing model support, and other changes.

Intel's software engineers working on the OpenVINO AI toolkit today released OpenVINO 2025.0 that brings support for the much talked about Deepseek models along with other large language models (LLMs), performance improvements to some of the existing model support, and other changes. New model support with Intel's OpenVINO 2025.0 open-source AI toolkit include Qwen 2.5, Deepseek-R1-Distill-Llama-8B, DeepSeek-R1-Distill-Qwen-7B, and DeepSeek-R1-Distill-Qwen-1.5B, FLUX.1 Schnell, and FLUX.1 Dev. The OpenVINO 2025.0 also brings improvements for second token latency for LLMs, KV cache compression is now enabled for INT8 on CPUs, support for Core Ultra 200H "Arrow Lake H" processors, OpenVINO backend support with the Triton Inference Server, and the OpenVINO Model Server can now work natively on Windows Server deployments.

Get the Android app

Or read this on Phoronix

Read more on:

Photo of Intel

Intel

Photo of Support

Support

Photo of Performance

Performance

Related news:

News photo

Alpine Linux Seeking Support After Equinix Metal Sunsets

News photo

AMD Outsells Intel In the Datacenter For the First Time

News photo

AMD outsells Intel in the datacenter for the first time in Q4 2024