From Instance-Specific Training to Generalizable Solvers: Advancing Machine Learning for Scientific Computing

Guardado en:
Detalles Bibliográficos
Publicado en:ProQuest Dissertations and Theses (2025)
Autor principal: Song, Zezheng
Publicado:
ProQuest Dissertations & Theses
Materias:
Acceso en línea:Citation/Abstract
Full Text - PDF
Etiquetas: Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!

MARC

LEADER 00000nab a2200000uu 4500
001 3224180489
003 UK-CbPIL
020 |a 9798286430123 
035 |a 3224180489 
045 2 |b d20250101  |b d20251231 
084 |a 66569  |2 nlm 
100 1 |a Song, Zezheng 
245 1 |a From Instance-Specific Training to Generalizable Solvers: Advancing Machine Learning for Scientific Computing 
260 |b ProQuest Dissertations & Theses  |c 2025 
513 |a Dissertation/Thesis 
520 3 |a Neural networks have played a crucial role in scientific computing by providing data-driven solutions to complex mathematical and physical problems. However, traditional neural network solvers remain fundamentally limited in accuracy and lack interpretability, as they operate as black-box models with no explicit mathematical structure. To address these challenges, this thesis explores the Finite Expression Method (FEX), a symbolic regression approach designed to enhance interpretability and accuracy in scientific computing. FEX leverages deep reinforcement learning to discover interpretable mathematical expressions, offering a principled approach to solving high-dimensional partial differential equations (PDEs) and uncovering governing equations from experimental data. Despite these advantages, both neural network solvers and FEX still require retraining for each new equation or change in initial and boundary conditions, limiting their scalability and adaptability. To overcome these fundamental constraints, scientific computing is now transitioning to the second stage, characterized by foundation models inspired by large language models. These models are pretrained on diverse scientific data and employ in-context learning to generalize across a wide range of problems without requiring instance-specific retraining. In this thesis, we introduce FMint, a foundation model designed for the fast and accurate simulation of dynamical systems. FMint builds upon a decoder-only transformer architecture and functions as an error corrector for coarse simulations, significantly improving accuracy while maintaining computational efficiency. By learning from a broad set of dynamical system trajectories, FMint generalizes well to out-of-distribution dynamics, demonstrating superior performance compared to traditional neural network solvers. This shift from single-instance solvers to foundation models marks a major transformation in scientific computing. FMint exemplifies how pretrained models can serve as a foundation for more advanced PDE solvers, demonstrating the potential of leveraging large-scale pretraining and in-context learning for scientific applications. Its success highlights a broader direction for future research, where foundation models trained on diverse physical systems can enable more generalizable, efficient, and interpretable solutions across a wide range of computational science and engineering problems. 
653 |a Applied mathematics 
653 |a Computer science 
653 |a Mathematics 
653 |a Artificial intelligence 
773 0 |t ProQuest Dissertations and Theses  |g (2025) 
786 0 |d ProQuest  |t ProQuest Dissertations & Theses Global 
856 4 1 |3 Citation/Abstract  |u https://www.proquest.com/docview/3224180489/abstract/embedded/H09TXR3UUZB2ISDL?source=fedsrch 
856 4 0 |3 Full Text - PDF  |u https://www.proquest.com/docview/3224180489/fulltextPDF/embedded/H09TXR3UUZB2ISDL?source=fedsrch