Speaker: Beltrán Labrador Serrano.

Abstract: Large Language Models (LLMs) is transforming natural language processing and are now impacting speech processing. This talk addresses the challenge of training these massive neural networks required to follow this trend. I will present GPU parallelization strategies for scaling deep neural network training, including multi-GPU workload distribution, gradient calculation methods, and parameter precision choices to maximize GPU efficiency.