Jessica Marossero
Accelerating Transformer Inference on Heterogeneous Multi-Accelerator SoCs using ESP.
Rel. Daniele Jahier Pagliari, Alessio Burrello, Luca Carloni, Mohamed Amine Hamdi. Politecnico di Torino, Master of science program in Electronic Engineering, 2024
|
Preview |
PDF (Tesi_di_laurea)
- Thesis
Licence: Creative Commons Attribution Non-commercial No Derivatives. Download (5MB) | Preview |
Abstract
Transformers have become essential in deep learning, excelling in tasks like natural language processing and computer vision. However, they are computationally expensive, especially in their so-called attention layers, which require large-scale matrix multiplications with quadratic complexity. Therefore, coupling general purpose processors with specialized hardware accelerators is critical to efficiently deploy Transformers in embedded systems with limited resources. The Embedded Scalable Platform (ESP) is a pioneering open-source research platform that enables the design of such heterogeneous SoCs, by integrating multiple types of tiles in a 2D mesh architecture. This modular design allows for an efficient integration of third-party accelerators, enabling rapid prototyping and exploration of novel architectures.
This thesis focuses on the integration of the state-of-the-art Integer Transformer Accelerator (ITA) within ESP
Publication type
URI
![]() |
Modify record (reserved for operators) |
