Advances in Machine Learning-Enabled Resource Management in Manycore Systems: From Von Neumann to Heterogeneous Processing-in-Memory Architectures
Guardat en:
| Publicat a: | ProQuest Dissertations and Theses (2025) |
|---|---|
| Autor principal: | |
| Publicat: |
ProQuest Dissertations & Theses
|
| Matèries: | |
| Accés en línia: | Citation/Abstract Full Text - PDF |
| Etiquetes: |
Sense etiquetes, Sigues el primer a etiquetar aquest registre!
|
| Resum: | The carbon output of computing - from edge devices to the large data centers - must be dramatically reduced. In this respect, Voltage-Frequency Island (VFI) is a well-established design paradigm to create scalable and energy-efficient manycore chips (e.g., CPUs). Voltage/Frequency (V/F) knobs of the VFIs can be dynamically tuned to reduce the energy while maintaining the application’s quality of service (QoS). In the first part of this dissertation, we consider the problem of dynamic power management (DPM) in manycore SoCs and propose novel Machine-learning (ML)-enabled DPM strategies to improve the energy efficiency in von Neumann-based manycore architectures.Deep Neural Networks (DNNs) and Graph Neural Networks (GNNs) have enabled remarkable advancements in various real-world applications, including natural language processing, healthcare, molecular chemistry, etc. As the complexity of neural network models continues to grow, their intensive computing and memory requirements pose significant performance and energy efficiency challenges for the traditional von Neumann architectures. Processing-in-Memory (PIM)-based computing platforms have emerged as a promising alternative due to their ability to perform computation within the memory itself, thereby reducing data movement and improving energy efficiency. However, communication between PIM-based processing elements (PEs) in a manycore architecture remains a bottleneck. In addition, in-memory computation suffers from device and crossbar non-idealities arising due to temperature, conductance drift, etc. In this dissertation, we address these challenges and propose a design of thermally efficient dataflow-aware Network-on-Chip (NoC) to accelerate DNN inferencing. We also address the reliability, energy, and performance challenges in DNN training and propose a heterogeneous architecture that combines the benefits of multiple PIM devices in a single platform to enable energy-efficient and high-performance DNN training. Later in this dissertation, we exploit the heterogeneity in the computational kernels behind deep learning models such as DNNs, GNNs, and transformers to design high-performance, energy-efficient, and reliable heterogeneous PIM-based manycore systems for sustainable deep learning.Overall, we utilize ML to enable the design and resource management of high-performance, energy-efficient, and reliable computing systems spanning from von Neumann to heterogeneous PIM-based architectures. |
|---|---|
| ISBN: | 9798297636583 |
| Font: | ProQuest Dissertations & Theses Global |