Thesis title: Harnessing the capabilities of Generative Models
Generative models have experienced significant advancements in recent years, driven by the introduction of architectures such as Stable Diffusion, GPT-3, ChatGPT, and many others.
These models are designed to learn probability distributions and efficiently sample from them during inference, typically conditioned on inputs like text.
Trained on large volumes of unlabeled data, these models possess extensive knowledge that can be transferred to address specific tasks.
In this thesis, we show how they can be harnessed to address a variety of tasks across different domains, including reasoning, image processing, and music generation. In particular, we will explore diverse methodologies to guide the generation process of a learned model to better suit the task at hand.