Efficient Post-Processing Techniques for Foundation Models

Uloženo v:
Podrobná bibliografie
Vydáno v:ProQuest Dissertations and Theses (2025)
Hlavní autor: Ngweta, Lilian
Vydáno:
ProQuest Dissertations & Theses
Témata:
On-line přístup:Citation/Abstract
Full Text - PDF
Tagy: Přidat tag
Žádné tagy, Buďte první, kdo vytvoří štítek k tomuto záznamu!

MARC

LEADER 00000nab a2200000uu 4500
001 3217287004
003 UK-CbPIL
020 |a 9798280718159 
035 |a 3217287004 
045 2 |b d20250101  |b d20251231 
084 |a 66569  |2 nlm 
100 1 |a Ngweta, Lilian 
245 1 |a Efficient Post-Processing Techniques for Foundation Models 
260 |b ProQuest Dissertations & Theses  |c 2025 
513 |a Dissertation/Thesis 
520 3 |a Foundation models, such as large language models (LLMs) and large vision models (LVMs), are often fine-tuned for various downstream tasks. Popular fine-tuning techniques include Reinforcement Learning from Human Feedback (RLHF) and supervised fine-tuning (SFT). Fine-tuning approaches, including RLHF and SFT, involve modifying model parameters, which is computationally expensive and may lead to unintended degradation of model performance, such as overfitting to specific biases or reducing response diversity. This thesis addresses three key challenges in the robustness and efficiency of foundation models:• Learning interpretable disentangled representations (Chapter 2): We propose a computationally cheap post-processing technique to separate style and content features in LVMs, leading to representations that improve out-of-distribution (OOD) generalization.• Efficiently aligning LLMs (Chapter 3): We introduce a lightweight pipeline that generates synthetic data to train aligners and inspectors, enabling on-demand alignment of any LLM. Inspectors are lightweight BERT models used to determine when an output needs to be aligned, and aligners are small LLMs used to perform alignment with respect to a particular targeted human value. We use AlpacaEval and PairRM, two standard automatic LLM evaluators, to establish the competitive advantages of our approach against baseline LLMs and alternative alignment procedures.• Mitigating style-induced prompt brittleness in LLMs (Chapter 4): We present a novel mixture of formats (MOF) prompting strategy that incorporates diverse stylistic variations in few-shot examples, enhancing robustness across different models and task domains. MOF uses simple modifications to few-shot prompting, so it avoids the cost of post-processing of the LLM outputs that is incurred by other approaches to mitigating style-induced prompt brittleness.Each of these approaches is designed as a post-processing technique, meaning they do not require modifying the original foundation model’s parameters. As a result, they are efficient and they provide performance improvements that are competitive with or exceed that of alternative solutions to these challenges. 
653 |a Artificial intelligence 
653 |a Computer engineering 
653 |a Computer science 
773 0 |t ProQuest Dissertations and Theses  |g (2025) 
786 0 |d ProQuest  |t ProQuest Dissertations & Theses Global 
856 4 1 |3 Citation/Abstract  |u https://www.proquest.com/docview/3217287004/abstract/embedded/L8HZQI7Z43R0LA5T?source=fedsrch 
856 4 0 |3 Full Text - PDF  |u https://www.proquest.com/docview/3217287004/fulltextPDF/embedded/L8HZQI7Z43R0LA5T?source=fedsrch