Get the latest tech news

DeepSeek open source DeepEP – library for MoE training and Inference


DeepEP: an efficient expert-parallel communication library - deepseek-ai/DeepEP

For latency-sensitive inference decoding, DeepEP includes a set of low-latency kernels with pure RDMA to minimize delays. So an implicit CPU wait for GPU received count signal will be involved, as the following figure shows. With our receiving hook interface, the RDMA network traffics are happening in the background, without costing any GPU SMs from the computation part.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of inference

inference

Photo of library

library

Photo of open source

open source

Related news:

News photo

Spotify partners with ElevenLabs to expand its library of AI-narrated audiobooks

News photo

Does the 'Spirit' of Open Source Mean Much More Than a License?

News photo

Why the ‘spirit’ of open source means much more than a license