Alert button
Picture for Chunming Hu

Chunming Hu

Alert button

Architectural Implications of GNN Aggregation Programming Abstractions

Oct 21, 2023
Yingjie Qi, Jianlei Yang, Ao Zhou, Tong Qiao, Chunming Hu

Figure 1 for Architectural Implications of GNN Aggregation Programming Abstractions
Figure 2 for Architectural Implications of GNN Aggregation Programming Abstractions
Figure 3 for Architectural Implications of GNN Aggregation Programming Abstractions
Figure 4 for Architectural Implications of GNN Aggregation Programming Abstractions

Graph neural networks (GNNs) have gained significant popularity due to the powerful capability to extract useful representations from graph data. As the need for efficient GNN computation intensifies, a variety of programming abstractions designed for optimizing GNN Aggregation have emerged to facilitate acceleration. However, there is no comprehensive evaluation and analysis upon existing abstractions, thus no clear consensus on which approach is better. In this letter, we classify existing programming abstractions for GNN Aggregation by the dimension of data organization and propagation method. By constructing these abstractions on a state-of-the-art GNN library, we perform a thorough and detailed characterization study to compare their performance and efficiency, and provide several insights on future GNN acceleration based on our analysis.

* 4 pages, to be published in IEEE Computer Architecture Letters (CAL) 
Viaarxiv icon

AdversarialWord Dilution as Text Data Augmentation in Low-Resource Regime

May 16, 2023
Junfan Chen, Richong Zhang, Zheyan Luo, Chunming Hu, Yongyi Mao

Figure 1 for AdversarialWord Dilution as Text Data Augmentation in Low-Resource Regime
Figure 2 for AdversarialWord Dilution as Text Data Augmentation in Low-Resource Regime
Figure 3 for AdversarialWord Dilution as Text Data Augmentation in Low-Resource Regime
Figure 4 for AdversarialWord Dilution as Text Data Augmentation in Low-Resource Regime

Data augmentation is widely used in text classification, especially in the low-resource regime where a few examples for each class are available during training. Despite the success, generating data augmentations as hard positive examples that may increase their effectiveness is under-explored. This paper proposes an Adversarial Word Dilution (AWD) method that can generate hard positive examples as text data augmentations to train the low-resource text classification model efficiently. Our idea of augmenting the text data is to dilute the embedding of strong positive words by weighted mixing with unknown-word embedding, making the augmented inputs hard to be recognized as positive by the classification model. We adversarially learn the dilution weights through a constrained min-max optimization process with the guidance of the labels. Empirical studies on three benchmark datasets show that AWD can generate more effective data augmentations and outperform the state-of-the-art text data augmentation methods. The additional analysis demonstrates that the data augmentations generated by AWD are interpretable and can flexibly extend to new examples without further training.

* Preprint, Accepted by AAAI 2023 
Viaarxiv icon

Impact-Oriented Contextual Scholar Profiling using Self-Citation Graphs

Apr 24, 2023
Yuankai Luo, Lei Shi, Mufan Xu, Yuwen Ji, Fengli Xiao, Chunming Hu, Zhiguang Shan

Figure 1 for Impact-Oriented Contextual Scholar Profiling using Self-Citation Graphs
Figure 2 for Impact-Oriented Contextual Scholar Profiling using Self-Citation Graphs
Figure 3 for Impact-Oriented Contextual Scholar Profiling using Self-Citation Graphs
Figure 4 for Impact-Oriented Contextual Scholar Profiling using Self-Citation Graphs

Quantitatively profiling a scholar's scientific impact is important to modern research society. Current practices with bibliometric indicators (e.g., h-index), lists, and networks perform well at scholar ranking, but do not provide structured context for scholar-centric, analytical tasks such as profile reasoning and understanding. This work presents GeneticFlow (GF), a suite of novel graph-based scholar profiles that fulfill three essential requirements: structured-context, scholar-centric, and evolution-rich. We propose a framework to compute GF over large-scale academic data sources with millions of scholars. The framework encompasses a new unsupervised advisor-advisee detection algorithm, a well-engineered citation type classifier using interpretable features, and a fine-tuned graph neural network (GNN) model. Evaluations are conducted on the real-world task of scientific award inference. Experiment outcomes show that the F1 score of best GF profile significantly outperforms alternative methods of impact indicators and bibliometric networks in all the 6 computer science fields considered. Moreover, the core GF profiles, with 63.6%-66.5% nodes and 12.5%-29.9% edges of the full profile, still significantly outrun existing methods in 5 out of 6 fields studied. Visualization of GF profiling result also reveals human explainable patterns for high-impact scholars.

Viaarxiv icon

Hardware-Aware Graph Neural Network Automated Design for Edge Computing Platforms

Mar 20, 2023
Ao Zhou, Jianlei Yang, Yingjie Qi, Yumeng Shi, Tong Qiao, Weisheng Zhao, Chunming Hu

Figure 1 for Hardware-Aware Graph Neural Network Automated Design for Edge Computing Platforms
Figure 2 for Hardware-Aware Graph Neural Network Automated Design for Edge Computing Platforms
Figure 3 for Hardware-Aware Graph Neural Network Automated Design for Edge Computing Platforms
Figure 4 for Hardware-Aware Graph Neural Network Automated Design for Edge Computing Platforms

Graph neural networks (GNNs) have emerged as a popular strategy for handling non-Euclidean data due to their state-of-the-art performance. However, most of the current GNN model designs mainly focus on task accuracy, lacking in considering hardware resources limitation and real-time requirements of edge application scenarios. Comprehensive profiling of typical GNN models indicates that their execution characteristics are significantly affected across different computing platforms, which demands hardware awareness for efficient GNN designs. In this work, HGNAS is proposed as the first Hardware-aware Graph Neural Architecture Search framework targeting resource constraint edge devices. By decoupling the GNN paradigm, HGNAS constructs a fine-grained design space and leverages an efficient multi-stage search strategy to explore optimal architectures within a few GPU hours. Moreover, HGNAS achieves hardware awareness during the GNN architecture design by leveraging a hardware performance predictor, which could balance the GNN model accuracy and efficiency corresponding to the characteristics of targeted devices. Experimental results show that HGNAS can achieve about $10.6\times$ speedup and $88.2\%$ peak memory reduction with a negligible accuracy loss compared to DGCNN on various edge devices, including Nvidia RTX3080, Jetson TX2, Intel i7-8700K and Raspberry Pi 3B+.

* Accepted by DAC'23 
Viaarxiv icon

Optimizing Memory Efficiency of Graph Neural Networks on Edge Computing Platforms

Apr 12, 2021
Ao Zhou, Jianlei Yang, Yeqi Gao, Tong Qiao, Yingjie Qi, Xiaoyi Wang, Yunli Chen, Pengcheng Dai, Weisheng Zhao, Chunming Hu

Figure 1 for Optimizing Memory Efficiency of Graph Neural Networks on Edge Computing Platforms
Figure 2 for Optimizing Memory Efficiency of Graph Neural Networks on Edge Computing Platforms
Figure 3 for Optimizing Memory Efficiency of Graph Neural Networks on Edge Computing Platforms
Figure 4 for Optimizing Memory Efficiency of Graph Neural Networks on Edge Computing Platforms

Graph neural networks (GNN) have achieved state-of-the-art performance on various industrial tasks. However, the poor efficiency of GNN inference and frequent Out-Of-Memory (OOM) problem limit the successful application of GNN on edge computing platforms. To tackle these problems, a feature decomposition approach is proposed for memory efficiency optimization of GNN inference. The proposed approach could achieve outstanding optimization on various GNN models, covering a wide range of datasets, which speeds up the inference by up to 3x. Furthermore, the proposed feature decomposition could significantly reduce the peak memory usage (up to 5x in memory efficiency improvement) and mitigate OOM problems during GNN inference.

* This paper has been accepted by RTAS 2021(brief industry track), with link to publicly available code 
Viaarxiv icon