Sergi Tomàs Martínez
Degree and Master's Degree in Artificial Intelligence
Large Language Models (LLMs) have demonstrated remarkable capabilities in a wide variety of tasks, making them a central focus of current research in artificial intelligence. However, enabling these models to operate efficiently requires very high computational resources and considerable execution time. One of the main challenges is to determine the optimal parallelism strategy to maximize performance. This project aims to develop a framework capable of identifying the most appropriate parallelism configuration for a given LLM. Using AstraSim and its synthetic LLM generator, the framework systematically explores the search space to recommend optimal degrees of data, tensor, sequence, and pipeline parallelism. In order to further increase the accuracy of the simulation, the framework integrates an optimized network model that takes into account the effects of congestion and other limitations inherent to communication in distributed environments. The results of this project bring value to different professional profiles. On the one hand, artificial intelligence researchers can use them to obtain recommendations on the most efficient parallelism strategies for training LLMs. On the other hand, hardware engineers can benefit from information on design trade-offs and identify potential bottlenecks associated with different architectural choices.
Sergi Tomàs Martínez
Degree and Master's Degree in Artificial Intelligence
Host Organization
Supervisors
Sergi Abadal
UPC Supervisor
The content of this website reflects only the views of the Catedra Chip Chair UPC project.
Tomàs Gadea Alcaide
Research Support Investigator
Bernat Ibañez
Research Support Investigator
Antoni Pech Alberich
Research Support Investigator
Yilihamujiang Yimamu
Predoctoral Researcher
Xavier Querol Bassols
Research Support Investigator
Nuria Elizondo Cereza
Research Support Investigator
Mohammad Nasser
Predoctoral Researcher
Guillem Pastor Rué
Research Support Investigator