Mattia Ottoborgo
Beyond Cross-Entropy: Custom Loss Functions for Finetuning SLMs on Structured Recipe Generation.
Rel. Paolo Garza, Daniele Rege Cambrin. Politecnico di Torino, Corso di laurea magistrale in Data Science And Engineering, 2025
|
Preview |
PDF (Tesi_di_laurea)
- Tesi
Licenza: Creative Commons Attribution Non-commercial No Derivatives. Download (11MB) | Preview |
Abstract
This thesis explores the use of custom loss functions to the finetuning of Small Language Models (SLMs) applied to recipe generation. With the exponential growth of deep learning, Large Language Models have proven themselves to have incredible text generation capabilities. Standard training frameworks, however, which employ Cross-Entropy loss, commonly disappoint in more demanding fields requiring high accuracy of facts and numbers, for instance the generation of procedural texts such as cooking recipes. This work tackles the intrinsic limitation of the standard solution that treats all the words indifferently, giving rise to the model’s inability to appropriate the important but frequently statistically scarce ingredients of a recipe.
Construction of a valid recipe presents several challenges
Relatori
Anno Accademico
Tipo di pubblicazione
Numero di pagine
Corso di laurea
Classe di laurea
Aziende collaboratrici
URI
![]() |
Modifica (riservato agli operatori) |
