by Luis Pacheco, Uday Yallapragada and Cristian Muñoz Large language models (LLMs) like Meta’s LLaMA 70B are revolutionizing natural language processing tasks, but training or fine-tuning them requires massive computational and memory resources. To address these challenges, we employ distributed training across multiple GPU nodes using DeepSpeed and Hugging Face Accelerate. This blog walks you […]
Cristian Munoz
Cristian lives in Santiago, Chile. He has nearly 15 years of experience in the world of data, specifically in work related to data engineering and architecture. He has held roles as a developer, data engineer, data architect, and technical leader.