From Latent Knowledge Gathering to Side Information Injection in Discrete Sequential Models

Kaydedildi:
Detaylı Bibliyografya
Yayımlandı:ProQuest Dissertations and Theses (2024)
Yazar: Rezaee Taghiabadi, Mohammad Mehdi
Baskı/Yayın Bilgisi:
ProQuest Dissertations & Theses
Konular:
Online Erişim:Citation/Abstract
Full Text - PDF
Etiketler: Etiketle
Etiket eklenmemiş, İlk siz ekleyin!

MARC

LEADER 00000nab a2200000uu 4500
001 3059107600
003 UK-CbPIL
020 |a 9798382744445 
035 |a 3059107600 
045 2 |b d20240101  |b d20241231 
084 |a 66569  |2 nlm 
100 1 |a Rezaee Taghiabadi, Mohammad Mehdi 
245 1 |a From Latent Knowledge Gathering to Side Information Injection in Discrete Sequential Models 
260 |b ProQuest Dissertations & Theses  |c 2024 
513 |a Dissertation/Thesis 
520 3 |a Representation learning is crucial for processing sequential and discrete data, such as text in natural language processing (NLP). From classical methods like topic modeling to modern transformer-based architectures, the goal is to utilize data to learn richer representations. This thesis focuses on two primary strategies: Latent Knowledge Gathering, which involves using clustering techniques to extract semantic knowledge from training data, and Injecting Background Information, where structural priors such as pretrained models are employed to enhance learning.The encoding process transforms high-dimensional documents into compact, lowdimensional representations, optimized to capture vital information for various language tasks. For instance, in document classification, both encoder and decoder play critical roles, especially with limited data. Our experiments assess model capabilities across different data regimes, emphasizing efficient representation in the situation entity classification task.Thematic analysis has seen advancements; however, the extraction of word-level thematic topics and the utilization of auxiliary knowledge are often overlooked. We propose a novel approach combining topic models with recurrent neural networks (RNNs) to maintain and utilize lower-level representations, enhancing natural language generation. Comparative experiments show this method achieves state-of-the-art performance by effectively retaining and using word topic assignments.Additionally, we explore using structured, discrete, semi-supervised variational autoencoders to leverage incomplete and noisy side knowledge for guiding text representation. This method robustly handles varied observation levels of side knowledge, consistently improving performance across language modeling and classification metrics.Finally, we introduce a universal framework for integrating discrete information using the information bottleneck principle. Through extensive theoretical and empirical studies, including a case study on event modeling, this framework significantly enhances performance and provides a robust foundation for future research in integrating noisy and incomplete side knowledge. 
653 |a Computer science 
653 |a Computer engineering 
653 |a Information science 
773 0 |t ProQuest Dissertations and Theses  |g (2024) 
786 0 |d ProQuest  |t ProQuest Dissertations & Theses Global 
856 4 1 |3 Citation/Abstract  |u https://www.proquest.com/docview/3059107600/abstract/embedded/L8HZQI7Z43R0LA5T?source=fedsrch 
856 4 0 |3 Full Text - PDF  |u https://www.proquest.com/docview/3059107600/fulltextPDF/embedded/L8HZQI7Z43R0LA5T?source=fedsrch