ml:mixture_of_expert_models

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
ml:mixture_of_expert_models [2025/05/29 07:34] – [MoE Large Language Models] jmflanigml:mixture_of_expert_models [2025/05/31 07:40] (current) – [MoE Large Language Models] jmflanig
Line 19: Line 19:
   * [[https://arxiv.org/pdf/2410.07348|Jin et al 2024 - MoE++: Accelerating Mixture-of-Experts Methods with Zero-Computation Experts]]   * [[https://arxiv.org/pdf/2410.07348|Jin et al 2024 - MoE++: Accelerating Mixture-of-Experts Methods with Zero-Computation Experts]]
   * [[https://arxiv.org/pdf/2505.21411|Tang et al 2025 - Pangu Pro MoE: Mixture of Grouped Experts for Efficient Sparsity]]   * [[https://arxiv.org/pdf/2505.21411|Tang et al 2025 - Pangu Pro MoE: Mixture of Grouped Experts for Efficient Sparsity]]
 +  * [[https://arxiv.org/pdf/2505.22323|Guo et al 2025 - Advancing Expert Specialization for Better MoE]]
  
 ===== People ===== ===== People =====
ml/mixture_of_expert_models.1748504060.txt.gz · Last modified: 2025/05/29 07:34 by jmflanig

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki