LPLB
// An early research stage expert-parallel load balancer for MoE models based on linear programming.
LPLB
An early research stage expert-parallel load balancer for MoE models based on linear programming.
13EmergingUnknown
What it does
LPLB is a parallel load balancer that leverages linear programming to optimize expert parallel workload distribution for MoE (Mixture-of-Experts) models. It dynamically reorders experts based on workload statistics, constructs replicas considering static topology, and solves optimal token assignments for each batch to achieve dynamic load balancing. The reordering process is facilitated by EPLB,
Getting Started
git
git clone https://github.com/deepseek-ai/LPLB