Riccardo Bosio
Neural Networks hardware-specific optimization using different frameworks.
Rel. Paolo Garza. Politecnico di Torino, Corso di laurea magistrale in Data Science And Engineering, 2022
|
PDF (Tesi_di_laurea)
- Tesi
Licenza: Creative Commons Attribution Non-commercial No Derivatives. Download (3MB) | Preview |
Abstract: |
The goal of this thesis is to implement different optimization pipelines in order to speed up Neural Networks inference time on specific hardwares. To reach this goal I have tested different frameworks: I compared the new inference times obtained applying the pipelines on three AllRead Machine Learning Technologies' models, solving different Computer Vision tasks. The explored frameworks are Apache TVM, OpenVINO, tensorRT and DeepStream and the target hardwares of the experiments are Intel i5 and i7 CPUs and NVIDIA Jetson Xavier GPU. Apache TVM is an open-source end-to-end machine learning compiler framework for CPUs and GPUs that enables optimization on any hardware backend. OpenVINO is a toolkit designed for optimizing and delivering AI inference on Intel products: exploited optimization techniques include accuracy-aware quantization. Finally, DeepStream is an accelerated AI framework to build intelligent video analytics pipelines: it leverages tensorRT inference optimizing also the entire stream of data. The results show the impact of this hardware-specific optimization strategy: models are significantly faster without losing the original accuracy. In particular, it emerged that real-time inference can be reached using DeepStream and NVIDIA Jetson Xavier. |
---|---|
Relatori: | Paolo Garza |
Anno accademico: | 2022/23 |
Tipo di pubblicazione: | Elettronica |
Numero di pagine: | 62 |
Soggetti: | |
Corso di laurea: | Corso di laurea magistrale in Data Science And Engineering |
Classe di laurea: | Nuovo ordinamento > Laurea magistrale > LM-32 - INGEGNERIA INFORMATICA |
Ente in cotutela: | AllRead (SPAGNA) |
Aziende collaboratrici: | AllRead |
URI: | http://webthesis.biblio.polito.it/id/eprint/24614 |
Modifica (riservato agli operatori) |