Projects
You can find more of my projects on my GitHub Page.
Deep Learning
LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training
A series of open-sourced Mixture-of-Expert (MoE) models based on LLaMA 2 and SlimPajama.
Published on EMNLP 2024. [Paper Link]
LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training.
A series of open-sourced Mixture-of-Expert (MoE) models based on LLaMA 3.
Paper available on arXiv. [Paper Link]
Other Stuff
A ChatGPT/Gemini/DeepSeek based personalized arXiv paper assistant bot for automatic paper filtering.
A LaTeX template for writing Personal Statements (PS) and Statements of Purpose (SoP) for graduate school applications.
And more on GitHub…