Efficient Post-Processing Techniques for Foundation Models
Guardado en:
| Publicado en: | ProQuest Dissertations and Theses (2025) |
|---|---|
| Autor principal: | |
| Publicado: |
ProQuest Dissertations & Theses
|
| Materias: | |
| Acceso en línea: | Citation/Abstract Full Text - PDF |
| Etiquetas: |
Sin Etiquetas, Sea el primero en etiquetar este registro!
|
| Resumen: | Foundation models, such as large language models (LLMs) and large vision models (LVMs), are often fine-tuned for various downstream tasks. Popular fine-tuning techniques include Reinforcement Learning from Human Feedback (RLHF) and supervised fine-tuning (SFT). Fine-tuning approaches, including RLHF and SFT, involve modifying model parameters, which is computationally expensive and may lead to unintended degradation of model performance, such as overfitting to specific biases or reducing response diversity. This thesis addresses three key challenges in the robustness and efficiency of foundation models:• Learning interpretable disentangled representations (Chapter 2): We propose a computationally cheap post-processing technique to separate style and content features in LVMs, leading to representations that improve out-of-distribution (OOD) generalization.• Efficiently aligning LLMs (Chapter 3): We introduce a lightweight pipeline that generates synthetic data to train aligners and inspectors, enabling on-demand alignment of any LLM. Inspectors are lightweight BERT models used to determine when an output needs to be aligned, and aligners are small LLMs used to perform alignment with respect to a particular targeted human value. We use AlpacaEval and PairRM, two standard automatic LLM evaluators, to establish the competitive advantages of our approach against baseline LLMs and alternative alignment procedures.• Mitigating style-induced prompt brittleness in LLMs (Chapter 4): We present a novel mixture of formats (MOF) prompting strategy that incorporates diverse stylistic variations in few-shot examples, enhancing robustness across different models and task domains. MOF uses simple modifications to few-shot prompting, so it avoids the cost of post-processing of the LLM outputs that is incurred by other approaches to mitigating style-induced prompt brittleness.Each of these approaches is designed as a post-processing technique, meaning they do not require modifying the original foundation model’s parameters. As a result, they are efficient and they provide performance improvements that are competitive with or exceed that of alternative solutions to these challenges. |
|---|---|
| ISBN: | 9798280718159 |
| Fuente: | ProQuest Dissertations & Theses Global |