Alert button
Picture for Lingyang Chu

Lingyang Chu

Alert button

Lumos: Heterogeneity-aware Federated Graph Learning over Decentralized Devices

Mar 01, 2023
Qiying Pan, Yifei Zhu, Lingyang Chu

Figure 1 for Lumos: Heterogeneity-aware Federated Graph Learning over Decentralized Devices
Figure 2 for Lumos: Heterogeneity-aware Federated Graph Learning over Decentralized Devices
Figure 3 for Lumos: Heterogeneity-aware Federated Graph Learning over Decentralized Devices
Figure 4 for Lumos: Heterogeneity-aware Federated Graph Learning over Decentralized Devices

Graph neural networks (GNN) have been widely deployed in real-world networked applications and systems due to their capability to handle graph-structured data. However, the growing awareness of data privacy severely challenges the traditional centralized model training paradigm, where a server holds all the graph information. Federated learning is an emerging collaborative computing paradigm that allows model training without data centralization. Existing federated GNN studies mainly focus on systems where clients hold distinctive graphs or sub-graphs. The practical node-level federated situation, where each client is only aware of its direct neighbors, has yet to be studied. In this paper, we propose the first federated GNN framework called Lumos that supports supervised and unsupervised learning with feature and degree protection on node-level federated graphs. We first design a tree constructor to improve the representation capability given the limited structural information. We further present a Monte Carlo Markov Chain-based algorithm to mitigate the workload imbalance caused by degree heterogeneity with theoretically-guaranteed performance. Based on the constructed tree for each client, a decentralized tree-based GNN trainer is proposed to support versatile training. Extensive experiments demonstrate that Lumos outperforms the baseline with significantly higher accuracy and greatly reduced communication cost and training time.

* 13 pages, 7 figures, to be published in the Proceedings of the 39th IEEE International Conference on Data Engineering (ICDE 2023) 
Viaarxiv icon

Revealing Unfair Models by Mining Interpretable Evidence

Jul 12, 2022
Mohit Bajaj, Lingyang Chu, Vittorio Romaniello, Gursimran Singh, Jian Pei, Zirui Zhou, Lanjun Wang, Yong Zhang

Figure 1 for Revealing Unfair Models by Mining Interpretable Evidence
Figure 2 for Revealing Unfair Models by Mining Interpretable Evidence
Figure 3 for Revealing Unfair Models by Mining Interpretable Evidence
Figure 4 for Revealing Unfair Models by Mining Interpretable Evidence

The popularity of machine learning has increased the risk of unfair models getting deployed in high-stake applications, such as justice system, drug/vaccination design, and medical diagnosis. Although there are effective methods to train fair models from scratch, how to automatically reveal and explain the unfairness of a trained model remains a challenging task. Revealing unfairness of machine learning models in interpretable fashion is a critical step towards fair and trustworthy AI. In this paper, we systematically tackle the novel task of revealing unfair models by mining interpretable evidence (RUMIE). The key idea is to find solid evidence in the form of a group of data instances discriminated most by the model. To make the evidence interpretable, we also find a set of human-understandable key attributes and decision rules that characterize the discriminated data instances and distinguish them from the other non-discriminated data. As demonstrated by extensive experiments on many real-world data sets, our method finds highly interpretable and solid evidence to effectively reveal the unfairness of trained models. Moreover, it is much more scalable than all of the baseline methods.

Viaarxiv icon

Mining Minority-class Examples With Uncertainty Estimates

Dec 15, 2021
Gursimran Singh, Lingyang Chu, Lanjun Wang, Jian Pei, Qi Tian, Yong Zhang

Figure 1 for Mining Minority-class Examples With Uncertainty Estimates
Figure 2 for Mining Minority-class Examples With Uncertainty Estimates
Figure 3 for Mining Minority-class Examples With Uncertainty Estimates
Figure 4 for Mining Minority-class Examples With Uncertainty Estimates

In the real world, the frequency of occurrence of objects is naturally skewed forming long-tail class distributions, which results in poor performance on the statistically rare classes. A promising solution is to mine tail-class examples to balance the training dataset. However, mining tail-class examples is a very challenging task. For instance, most of the otherwise successful uncertainty-based mining approaches struggle due to distortion of class probabilities resulting from skewness in data. In this work, we propose an effective, yet simple, approach to overcome these challenges. Our framework enhances the subdued tail-class activations and, thereafter, uses a one-class data-centric approach to effectively identify tail-class examples. We carry out an exhaustive evaluation of our framework on three datasets spanning over two computer vision tasks. Substantial improvements in the minority-class mining and fine-tuned model's performance strongly corroborate the value of our proposed solution.

Viaarxiv icon

Achieving Model Fairness in Vertical Federated Learning

Sep 25, 2021
Changxin Liu, Zirui Zhou, Yang Shi, Jian Pei, Lingyang Chu, Yong Zhang

Figure 1 for Achieving Model Fairness in Vertical Federated Learning
Figure 2 for Achieving Model Fairness in Vertical Federated Learning
Figure 3 for Achieving Model Fairness in Vertical Federated Learning

Vertical federated learning (VFL), which enables multiple enterprises possessing non-overlapped features to strengthen their machine learning models without disclosing their private data and model parameters, has received increasing attention lately. Similar to other machine learning algorithms, VFL suffers from fairness issues, i.e., the learned model may be unfairly discriminatory over the group with sensitive attributes. To tackle this problem, we propose a fair VFL framework in this work. First, we systematically formulate the problem of training fair models in VFL, where the learning task is modeled as a constrained optimization problem. To solve it in a federated manner, we consider its equivalent dual form and develop an asynchronous gradient coordinate-descent ascent algorithm, where each data party performs multiple parallelized local updates per communication round to effectively reduce the number of communication rounds. We prove that the algorithm finds a $\delta$-stationary point of the dual objective in $\mathcal{O}(\delta^{-4})$ communication rounds under mild conditions. Finally, extensive experiments on three benchmark datasets demonstrate the superior performance of our method in training fair models.

* 15 pages, 2 figures 
Viaarxiv icon

FedFair: Training Fair Models In Cross-Silo Federated Learning

Sep 13, 2021
Lingyang Chu, Lanjun Wang, Yanjie Dong, Jian Pei, Zirui Zhou, Yong Zhang

Figure 1 for FedFair: Training Fair Models In Cross-Silo Federated Learning
Figure 2 for FedFair: Training Fair Models In Cross-Silo Federated Learning
Figure 3 for FedFair: Training Fair Models In Cross-Silo Federated Learning
Figure 4 for FedFair: Training Fair Models In Cross-Silo Federated Learning

Building fair machine learning models becomes more and more important. As many powerful models are built by collaboration among multiple parties, each holding some sensitive data, it is natural to explore the feasibility of training fair models in cross-silo federated learning so that fairness, privacy and collaboration can be fully respected simultaneously. However, it is a very challenging task, since it is far from trivial to accurately estimate the fairness of a model without knowing the private data of the participating parties. In this paper, we first propose a federated estimation method to accurately estimate the fairness of a model without infringing the data privacy of any party. Then, we use the fairness estimation to formulate a novel problem of training fair models in cross-silo federated learning. We develop FedFair, a well-designed federated learning framework, which can successfully train a fair model with high performance without any data privacy infringement. Our extensive experiments on three real-world data sets demonstrate the excellent fair model training performance of our method.

Viaarxiv icon

Finding Representative Interpretations on Convolutional Neural Networks

Aug 20, 2021
Peter Cho-Ho Lam, Lingyang Chu, Maxim Torgonskiy, Jian Pei, Yong Zhang, Lanjun Wang

Figure 1 for Finding Representative Interpretations on Convolutional Neural Networks
Figure 2 for Finding Representative Interpretations on Convolutional Neural Networks
Figure 3 for Finding Representative Interpretations on Convolutional Neural Networks
Figure 4 for Finding Representative Interpretations on Convolutional Neural Networks

Interpreting the decision logic behind effective deep convolutional neural networks (CNN) on images complements the success of deep learning models. However, the existing methods can only interpret some specific decision logic on individual or a small number of images. To facilitate human understandability and generalization ability, it is important to develop representative interpretations that interpret common decision logics of a CNN on a large group of similar images, which reveal the common semantics data contributes to many closely related predictions. In this paper, we develop a novel unsupervised approach to produce a highly representative interpretation for a large number of similar images. We formulate the problem of finding representative interpretations as a co-clustering problem, and convert it into a submodular cost submodular cover problem based on a sample of the linear decision boundaries of a CNN. We also present a visualization and similarity ranking method. Our extensive experiments demonstrate the excellent performance of our method.

* Accepted by ICCV 2021 (http://iccv2021.thecvf.com/home) | A python notebook of the proposed method can be found in https://marketplace.huaweicloud.com/markets/aihub/notebook/detail/?id=8e92ea6c-2f4a-4cff-a89f-bf2905bb7ac0 
Viaarxiv icon

Robust Counterfactual Explanations on Graph Neural Networks

Jul 16, 2021
Mohit Bajaj, Lingyang Chu, Zi Yu Xue, Jian Pei, Lanjun Wang, Peter Cho-Ho Lam, Yong Zhang

Figure 1 for Robust Counterfactual Explanations on Graph Neural Networks
Figure 2 for Robust Counterfactual Explanations on Graph Neural Networks
Figure 3 for Robust Counterfactual Explanations on Graph Neural Networks
Figure 4 for Robust Counterfactual Explanations on Graph Neural Networks

Massive deployment of Graph Neural Networks (GNNs) in high-stake applications generates a strong demand for explanations that are robust to noise and align well with human intuition. Most existing methods generate explanations by identifying a subgraph of an input graph that has a strong correlation with the prediction. These explanations are not robust to noise because independently optimizing the correlation for a single input can easily overfit noise. Moreover, they do not align well with human intuition because removing an identified subgraph from an input graph does not necessarily change the prediction result. In this paper, we propose a novel method to generate robust counterfactual explanations on GNNs by explicitly modelling the common decision logic of GNNs on similar input graphs. Our explanations are naturally robust to noise because they are produced from the common decision boundaries of a GNN that govern the predictions of many similar input graphs. The explanations also align well with human intuition because removing the set of edges identified by an explanation from the input graph changes the prediction significantly. Exhaustive experiments on many public datasets demonstrate the superior performance of our method.

Viaarxiv icon