Towards Generalized Event Understanding in Text via Generative Models

Guardat en:
Dades bibliogràfiques
Publicat a:ProQuest Dissertations and Theses (2024)
Autor principal: Hsu, I-Hung
Publicat:
ProQuest Dissertations & Theses
Matèries:
Accés en línia:Citation/Abstract
Full Text - PDF
Etiquetes: Afegir etiqueta
Sense etiquetes, Sigues el primer a etiquetar aquest registre!

MARC

LEADER 00000nab a2200000uu 4500
001 3060245289
003 UK-CbPIL
020 |a 9798382753140 
035 |a 3060245289 
045 2 |b d20240101  |b d20241231 
084 |a 66569  |2 nlm 
100 1 |a Hsu, I-Hung 
245 1 |a Towards Generalized Event Understanding in Text via Generative Models 
260 |b ProQuest Dissertations & Theses  |c 2024 
513 |a Dissertation/Thesis 
520 3 |a Human languages in the world, such as news or narratives, are structured around events. Focusing on these events allows Natural Language Processing (NLP) systems to better understand plots, infer motivations, consequences, and the dynamics of situations. Despite the rapidly evolving landscape of NLP technology, comprehending complex events, particularly those rarely encountered in training such as in niche domains or low-resource languages, remains a formidable challenge. This thesis explores methods to enhance NLP model generalizability for better adaptability to unfamiliar events and languages unseen during training.My approach includes two main strategies: (1) Model Perspective: I propose a novel generation-based event extraction framework, largely different from typical solutions that make predictions by learning to classify input tokens. This new framework utilizes indirect supervision from natural language generation, leveraging large-scale unsupervised data without requiring additional training modules dependent on limited event-specific data. Hence, it facilitates the models’ ability on understanding general event concepts. I further explore advanced methods to extend this framework for cross-lingual adaptation and to utilize cross-domain robust resources effectively. (2) Data Perspective: I develop techniques to generate pseudo-training data broaden the training scope for event understanding models. This includes translating structured event labels into other languages with higher accuracy and fidelity, and synthesizing novel events for the existing knowledge base.Overall, my work introduces a novel learning platform to the NLP community, emphasizing an innovative modeling paradigm and comprehensive data preparation to foster more generalized event understanding models. 
653 |a Artificial intelligence 
653 |a Computer science 
653 |a Information technology 
773 0 |t ProQuest Dissertations and Theses  |g (2024) 
786 0 |d ProQuest  |t ProQuest Dissertations & Theses Global 
856 4 1 |3 Citation/Abstract  |u https://www.proquest.com/docview/3060245289/abstract/embedded/L8HZQI7Z43R0LA5T?source=fedsrch 
856 4 0 |3 Full Text - PDF  |u https://www.proquest.com/docview/3060245289/fulltextPDF/embedded/L8HZQI7Z43R0LA5T?source=fedsrch