Overview
We study optimization methods for efficient training and adaptation of deep models and large language models (LLMs), aiming at better convergence, stability, and compute efficiency.
Outcomes
Publications
- Yineng Chen, Zuchao Li*, et al. Bidirectional Looking with A Novel Double Exponential Moving Average... ICML 2023.
- Weixi Song, Zuchao Li*, et al. Sparse is Enough in Fine-tuning Pre-trained Large Language Models. ICML 2024.