MILO Seminar : Adaptation of foundation model: from Transformer adapters to VLM test-time prompt tuning

24 March 2026
from 13H30 to 14H30
Register
  • EN
  • No videoconference
  • Public
  • Guillaume EYNARD-BONTEMPS Guillaume EYNARD-BONTEMPS Ingénieur Calcul Scientifique

Adaptation of foundation model: from Transformer adapters to VLM test-time prompt tuning

 

With the advent of foundation models, model development has shifted from tailoring task-specific models to training large models capable of addressing a wide range of tasks. One way of using foundation models is to directly use them with the pretrained weights from a very optimized training. This has the advantage of not using any training resources, and of exploiting the pretraining. However, this method may fail when there is a large domain shift between the training and testing data. Another way of using them is to train them on a target task. This ensures a better alignment with the target task but does require a large dataset and substantial compute. Adaptation proposes a balance between these two extremes, allowing to take advantage of the pretrained weights, while allowing to adapt the weights to the target task. In this talk, we will explore two methods for adaptation: the first one is applied to Vision Transformers (ViTs) in the few-shot regime, and the second one for test-time prompt tuning for zero-shot classification of Vision Language Models (VLMs).

Register
Add to your diary 24-03-2026 13:30 24-03-2026 14:30 <p><strong>Adaptation of foundation model: from Transformer adapters to VLM test-time prompt tuning</strong></p><p>&nbsp;</p><p><em>With the advent of foundation models, model development has shifted from tailoring task-specific models to training large models capable of addressing a wide range of tasks. One way of using foundation models is to directly use them with the pretrained weights from a very optimized training. This has the advantage of not using any training resources, and of exploiting the pretraining. However, this method may fail when there is a large domain shift between the training and testing data. Another way of using them is to train them on a target task. This ensures a better alignment with the target task but does require a large dataset and substantial compute. Adaptation proposes a balance between these two extremes, allowing to take advantage of the pretrained weights, while allowing to adapt the weights to the target task. In this talk, we will explore two methods for adaptation: the first one is applied to Vision Transformers (ViTs) in the few-shot regime, and the second one for test-time prompt tuning for zero-shot classification of Vision Language Models (VLMs).</em></p>

To take part in this event, you must register.

Share :

Partage possible par tous (via Facebook, LinkedIn, Twitter, email)

You may find these events interesting, as well