MoNE: Replacing Redundant Experts with Lightweight Novices for Structured Pruning of MoE
Paper
• 2507.00390 • Published
• 2
This repository contains a structured pruned variant of DeepSeek-V2-Lite using the MoNE (Mixture-of-Novice Experts) framework proposed in our paper.
*## Model Overview
This checkpoint replaces redundant experts with lightweight novice experts via structured pruning, aiming to reduce compute while preserving performance.
Title: MoNE: Replacing Redundant Experts with Lightweight Novices for Structured Pruning of MoE
Authors: Geng Zhang, Yuxuan Han, Yuxuan Lou, Yiqi Zhang, Wangbo Zhao, Yang You
arXiv: arXiv:2507.00390
Base model
deepseek-ai/DeepSeek-V2-Lite