DeepEP

DeepEP

Powering DeepSeek-V3's MoE Performance

8 followers

DeepEP, from DeepSeek, is the open-source communication library powering the DeepSeek-V3 MoE model. Optimized for Hopper GPUs, NVLink, and RDMA.
DeepEP gallery image
DeepEP gallery image
DeepEP gallery image
Free
Launch Team
Wispr Flow: Dictation That Works Everywhere
Wispr Flow: Dictation That Works Everywhere
Stop typing. Start speaking. 4x faster.
Promoted

What do you think? …

Zac Zuo

Hi everyone!

Sharing DeepEP, a new open-source communication library from DeepSeek AI, and this is the technology that powers their impressive DeepSeek-V3 model! It's a highly technical project, but crucial for anyone working with Mixture-of-Experts (MoE) models at scale.

DeepEP is specifically designed to optimize the all-to-all communication that's critical for MoE training and inference. It's built for NVIDIA Hopper GPUs and takes advantage of NVLink and RDMA. It's even tailored to the "group-limited gating algorithm" used in DeepSeek-V3.

Essentially, DeepEP is how DeepSeek gets such great performance from their MoE models. They're open-sourcing the core communication technology that makes it all work.

DeepEP is the second product in DeepSeek's Open Source Week.