
Dissertation Defense
Domain-specific Benchmarks and Architectures for Applications using Graph-based Data
This event is free and open to the publicAdd to Google Calendar

Hybrid Event: 3725 BBB / Zoom
Abtract: Graph-based processing enables many applications in logistics, e-commerce, social media, and more. However, graph workloads are slow: they are bottlenecked not by compute power, but by inefficient data access. As useful graphs get larger and graph-based algorithms become more complex, adding more powerful compute units like Graphics Processing Units (GPUs) or Tensor Processing Units (TPUs) cannot keep up with the increasing size and complexity of these workloads.
To address these challenges, we first introduce a novel taxonomy of graph-based algorithms: those that treat graphs (1) as data frameworks, (2) as algorithmic frameworks, or (3) as both. Each category has overlapping needs: higher memory bandwidth, better data organization, and greater thread-level parallelism. Next, we demonstrate that custom processing-in-memory (PIM) hardware accelerators are effective and energy-efficient solutions to the compute and memory bottlenecks of graph-based applications.
Specifically, we propose and evaluate three custom PIM accelerators, DREDGE (for graph-as-data-framework applications), ACRE (for graph-as-algorithmic-framework applications), and GLEAM (for graph-as-both applications), each targeting one of the three categories of graph applications. DREDGE targets dynamic graph applications by introducing a novel partitioning technique and dedicated hardware support to continuously improve data organization in memory. ACRE accelerates the training of tree-based machine learning models in a way that allows users to better understand the models’ reasoning. GLEAM targets graph neural networks, the primary machine learning models for graph-based data, accelerating the node aggregation operations that bottleneck training and inference operations. These designs offer a 2.5-14x speedup for their respective applications. GLEAM and ACRE save 77-93% of total system energy, and DREDGE adds as little as 2.2% energy consumption over PIM baselines. Each design fits within the logic area of modern 3D-stacked memory: 0.3-13% of the available logic space. Finally, we present two benchmark suites, DyGraph and BeXAI, making them publicly available to support future research in dynamic graphs processing and explainable machine learning acceleration. Together, these contributions enable efficient and scalable graph computing to handle the demands of tomorrow’s graph workloads.