Development site for the EIT FOAI CDT
When are embeddings enough?
Many practical applications rely on fine-tuning predictors on top of pre-trained embeddings, yet a rigorous theoretical framework explaining when this approach is truly optimal compared to end-to-end learning is lacking.
This project aims to develop a theoretical framework, potentially using tools from statistical learning theory and information theory, to understand the conditions under which using pre-trained embeddings is optimal. The students will analyse the interplay between the pre-training data distribution, the downstream task data distribution, the size of the fine-tuning dataset, and model architecture to derive bounds and conditions that guide the choice between a frozen embedding approach (y = f(e(x))) and end-to-end model training (y = f(x)).
Statistical learning theory, information theory, deep learning fundamentals
Theoretical analysis, generalisation theory, embedding models