Alimama Tech
Sep 12, 2023 · Artificial Intelligence
Megatron-LLaMA: High-Performance Large Language Model Training Framework
Megatron-LLaMA is an open‑source high‑performance training framework for LLaMA models, offering tensor, pipeline, and sequence parallelism, an overlapped optimizer, and near‑linear scalability, achieving up to 176% speedup on 32 GPUs and robust performance even with limited network bandwidth.
DeepSpeedGPU optimizationLLaMA
0 likes · 10 min read