ToMoE: Converting Dense Large Language Models to Mixture-of-Experts through Dynamic Structural Pruning Paper • 2501.15316 • Published Jan 25 • 3
Adapting Large Language Models via Reading Comprehension Paper • 2309.09530 • Published Sep 18, 2023 • 81