Projects
You can find more of my projects on my GitHub Page.
Deep Learning
LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training [GitHub] | [Paper]
A series of open-sourced Mixture-of-Expert (MoE) models based on LLaMA 2 and SlimPajama.
LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training [GitHub] | [Paper]
A series of open-sourced Mixture-of-Expert (MoE) models based on LLaMA 3.