GNN adaptive training paper is accepted by DAC 2024!

๐Ÿ‘ Paper title: GNNavigator: Towards Adaptive Training of Graph Neural Networks via Automatic Guideline Exploration. This work addresses the challenge of balancing training runtime, memory consumption, and accuracy in Graph Neural Networks (GNNs), which have seen significant success in various applications. GNNavigator introduces an adaptive GNN training configuration optimization framework. By leveraging a unified software-hardware co-abstraction, a novel GNN training performance model, and an effective design space exploration strategy, GNNavigator meets the diverse requirements of GNN applications. [related project]

GNN co-inference paper is accepted by DAC 2024!

๐Ÿ‘ Paper title: Graph Neural Networks Automated Design and Deployment on Device-Edge Co-Inference Systems. In this paper, we abstract the communication process in device-edge co-inference into a specific operation, creating a unified design space for GNN architecture and co-inference schemes. Using random search, we achieve joint optimization, leading to a GNN architecture that integrates partitioning schemes, enabling a trade-off between communication and computation, and outperforming SOTA methods. [related project]

SRAM-PIM architecture design paper is accepted by DAC 2024!

๐Ÿ‘ Paper title: Towards Efficient SRAM-PIM Architecture Design by Exploiting Unstructured Bit-Level Sparsity. In this paper, we propose Dyadic Block PIM (DB-PIM), a groundbreaking algorithm-architecture co-design framework to exploit the unstructured bit-level sparsity effectively. First, we propose an algorithm coupled with a distinctive sparsity pattern to preserve the random distribution of non-zero bits while improving regularity. Then, we develop a custom PIM macro that includes dyadic block multiplication units (DBMUs) and Canonical Signed Digit (CSD)-based adder trees to achieve unstructured bit-level sparsity. Results show that our proposed co-design framework achieves a remarkable speedup of up to 6.53x and energy savings of 77.50%. [related project]

PIM algorithm/architecture co-design paper is accepted by IEEE TCAD!

๐Ÿ‘ Paper title: DDC-PIM: Efficient Algorithm/Architecture Co-Design for Doubling Data Capacity of SRAM-Based Processing-in-Memory. We propose DDC-PIM, an efficient algorithm/architecture co-design methodology that effectively doubles the equivalent data capacity of SRAM. By combining qvfilter-wise complementary correlation algorithm with a customized architecture, we can exploit the intrinsic cross-coupled structure of 6T SRAM to store the bitwise complementary pair in their complementary states (Q/QB), thereby maximizing the data capacity and integration density of each SRAM cell. [related project]

GNN programming abstraction paper is accepted by IEEE Computer Architecture Letters!

๐Ÿ‘ Paper title: Architectural Implications of GNN Aggregation Programming Abstractions. This paper evaluates the architectural implications of programming abstractions for Graph Neural Network (GNN) aggregation. It introduces a taxonomy based on data organization and propagation methods and performs a comprehensive performance characterization across platforms and graph properties. Key findings include insights into abstraction selection, hardware adaptability, and the structural impact of graphs, providing valuable guidance for GNN acceleration research. [related project]

Model compression paper is accepted by ICCV 2023!

๐Ÿ‘ Paper title: Lossy and Lossless (L2) Post-training Model Size Compression. We propose a unified post-training model size compression method that combines lossy and lossless compression techniques, with a parametric weight transformation and a differentiable counter to guide optimization. Our method achieves a stable 10ร— compression ratio without accuracy loss and a 20ร— ratio with minimal accuracy degradation, all while easily controlling the global compression ratio and adapting it for different layers. [related project]

Hardware-aware GNN NAS paper is accepted by DAC 2023!

๐Ÿ‘ Paper title: Hardware-Aware Graph Neural Network Automated Design for Edge Computing Platforms. We explore a hardware-aware GNN architecture design for edge devices, leveraging the novel idea of “predicting GNNs with GNNs” to efficiently estimate the performance of candidate architectures during the NAS process. By thoroughly analyzing the impact of device heterogeneity on GNN performance and integrating hardware awareness into the exploration, our method achieves significant improvements in both accuracy and efficiency. [related project]

Reconfigurable In-Cache-MPUF system paper is accepted by IEEE TCAS-I!

๐Ÿ‘ Paper title: Reconfigurable and Dynamically Transformable In-Cache-MPUF System With True Randomness Based on the SOT-MRAM. In this paper, we present a reconfigurable Physically Unclonable Functions (PUF) based on the Spin-Orbit-Torque Magnetic Random-Access Memory (SOT-MRAM), which exploits thermal noise as the true dynamic entropy source. [related project]

Graph CC PIM architecture paper is accepted by IEEE TCAD!

๐Ÿ‘ Paper title:Accelerating Graph Connected Component Computation with Emerging Processing-In-Memory Architecture. In this article, we propose to accelerate CC computation with the emerging processing-in-memory (PIM) architecture through an algorithmโ€“architecture co-design manner. [related project]