Get the latest tech news

DeepSeek-R1-671B-Q4_K_M with 1 or 2 Arc A770 on Xeon


Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discr...

This guide demonstrates how to use llama.cpp portable zip to directly run llama.cpp on Intel GPU with ipex-llm(without the need of manual installations). Open "Command Prompt" (cmd), and enter the extracted folder through cd /d PATH\TO\EXTRACTED\FOLDER To use GPU acceleration, several environment variables are required or recommended before running llama.cpp. Open a "Terminal", and enter the extracted folder through cd /PATH/TO/EXTRACTED/FOLDER To use GPU acceleration, several environment variables are required or recommended before running llama.cpp.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of ARC

ARC

Photo of xeon

xeon

Photo of s deepseek r1 671b

s deepseek r1 671b

Related news:

News photo

Intel Announces Xeon 6300 Series - Tops Out At 8-Core Xeon 6369P For $545 USD

News photo

Intel PyTorch Extension 2.6 Brings More Xeon 6 Optimizations

News photo

Intel C1 Demotion Knob Proposed For The Linux Kernel To Help Newer Xeon CPUs