Soft MoE

From Sparse to Soft Mixture of Experts

In this post we review Google DeepMind’s paper that introduces Soft Mixture of Experts, a fully-differentiable sparse Transformer.

From Sparse to Soft Mixture of Experts Read More ยป