Alert button
Picture for My T. Thai

My T. Thai

Alert button

OASIS: Offsetting Active Reconstruction Attacks in Federated Learning

Nov 23, 2023
Tre' R. Jeter, Truc Nguyen, Raed Alharbi, My T. Thai

Federated Learning (FL) has garnered significant attention for its potential to protect user privacy while enhancing model training efficiency. However, recent research has demonstrated that FL protocols can be easily compromised by active reconstruction attacks executed by dishonest servers. These attacks involve the malicious modification of global model parameters, allowing the server to obtain a verbatim copy of users' private data by inverting their gradient updates. Tackling this class of attack remains a crucial challenge due to the strong threat model. In this paper, we propose OASIS, a defense mechanism based on image augmentation that effectively counteracts active reconstruction attacks while preserving model performance. We first uncover the core principle of gradient inversion that enables these attacks and theoretically identify the main conditions by which the defense can be robust regardless of the attack strategies. We then construct OASIS with image augmentation showing that it can undermine the attack principle. Comprehensive evaluations demonstrate the efficacy of OASIS highlighting its feasibility as a solution.

Viaarxiv icon

On the Communication Complexity of Decentralized Bilevel Optimization

Nov 19, 2023
Yihan Zhang, My T. Thai, Jie Wu, Hongchang Gao

Decentralized bilevel optimization has been actively studied in the past few years since it has widespread applications in machine learning. However, existing algorithms suffer from large communication complexity caused by the estimation of stochastic hypergradient, limiting their application to real-world tasks. To address this issue, we develop a novel decentralized stochastic bilevel gradient descent algorithm under the heterogeneous setting, which enjoys a small communication cost in each round and small communication rounds. As such, it can achieve a much better communication complexity than existing algorithms. Moreover, we extend our algorithm to the more challenging decentralized multi-level optimization. To the best of our knowledge, this is the first time achieving these theoretical results under the heterogeneous setting. At last, the experimental results confirm the efficacy of our algorithm.

Viaarxiv icon

When Decentralized Optimization Meets Federated Learning

Jun 05, 2023
Hongchang Gao, My T. Thai, Jie Wu

Figure 1 for When Decentralized Optimization Meets Federated Learning

Federated learning is a new learning paradigm for extracting knowledge from distributed data. Due to its favorable properties in preserving privacy and saving communication costs, it has been extensively studied and widely applied to numerous data analysis applications. However, most existing federated learning approaches concentrate on the centralized setting, which is vulnerable to a single-point failure. An alternative strategy for addressing this issue is the decentralized communication topology. In this article, we systematically investigate the challenges and opportunities when renovating decentralized optimization for federated learning. In particular, we discussed them from the model, data, and communication sides, respectively, which can deepen our understanding about decentralized federated learning.

* Accepted to IEEE Network 
Viaarxiv icon

FairDP: Certified Fairness with Differential Privacy

May 25, 2023
Khang Tran, Ferdinando Fioretto, Issa Khalil, My T. Thai, NhatHai Phan

Figure 1 for FairDP: Certified Fairness with Differential Privacy
Figure 2 for FairDP: Certified Fairness with Differential Privacy
Figure 3 for FairDP: Certified Fairness with Differential Privacy
Figure 4 for FairDP: Certified Fairness with Differential Privacy

This paper introduces FairDP, a novel mechanism designed to simultaneously ensure differential privacy (DP) and fairness. FairDP operates by independently training models for distinct individual groups, using group-specific clipping terms to assess and bound the disparate impacts of DP. Throughout the training process, the mechanism progressively integrates knowledge from group models to formulate a comprehensive model that balances privacy, utility, and fairness in downstream tasks. Extensive theoretical and empirical analyses validate the efficacy of FairDP, demonstrating improved trade-offs between model utility, privacy, and fairness compared with existing methods.

Viaarxiv icon

Linear Query Approximation Algorithms for Non-monotone Submodular Maximization under Knapsack Constraint

May 17, 2023
Canh V. Pham, Tan D. Tran, Dung T. K. Ha, My T. Thai

Figure 1 for Linear Query Approximation Algorithms for Non-monotone Submodular Maximization under Knapsack Constraint
Figure 2 for Linear Query Approximation Algorithms for Non-monotone Submodular Maximization under Knapsack Constraint
Figure 3 for Linear Query Approximation Algorithms for Non-monotone Submodular Maximization under Knapsack Constraint

This work, for the first time, introduces two constant factor approximation algorithms with linear query complexity for non-monotone submodular maximization over a ground set of size $n$ subject to a knapsack constraint, $\mathsf{DLA}$ and $\mathsf{RLA}$. $\mathsf{DLA}$ is a deterministic algorithm that provides an approximation factor of $6+\epsilon$ while $\mathsf{RLA}$ is a randomized algorithm with an approximation factor of $4+\epsilon$. Both run in $O(n \log(1/\epsilon)/\epsilon)$ query complexity. The key idea to obtain a constant approximation ratio with linear query lies in: (1) dividing the ground set into two appropriate subsets to find the near-optimal solution over these subsets with linear queries, and (2) combining a threshold greedy with properties of two disjoint sets or a random selection process to improve solution quality. In addition to the theoretical analysis, we have evaluated our proposed solutions with three applications: Revenue Maximization, Image Summarization, and Maximum Weighted Cut, showing that our algorithms not only return comparative results to state-of-the-art algorithms but also require significantly fewer queries.

Viaarxiv icon

Cultural-aware Machine Learning based Analysis of COVID-19 Vaccine Hesitancy

Apr 14, 2023
Raed Alharbi, Sylvia Chan-Olmsted, Huan Chen, My T. Thai

Figure 1 for Cultural-aware Machine Learning based Analysis of COVID-19 Vaccine Hesitancy
Figure 2 for Cultural-aware Machine Learning based Analysis of COVID-19 Vaccine Hesitancy
Figure 3 for Cultural-aware Machine Learning based Analysis of COVID-19 Vaccine Hesitancy
Figure 4 for Cultural-aware Machine Learning based Analysis of COVID-19 Vaccine Hesitancy

Understanding the COVID-19 vaccine hesitancy, such as who and why, is very crucial since a large-scale vaccine adoption remains as one of the most efficient methods of controlling the pandemic. Such an understanding also provides insights into designing successful vaccination campaigns for future pandemics. Unfortunately, there are many factors involving in deciding whether to take the vaccine, especially from the cultural point of view. To obtain these goals, we design a novel culture-aware machine learning (ML) model, based on our new data collection, for predicting vaccination willingness. We further analyze the most important features which contribute to the ML model's predictions using advanced AI explainers such as the Probabilistic Graphical Model (PGM) and Shapley Additive Explanations (SHAP). These analyses reveal the key factors that most likely impact the vaccine adoption decisions. Our findings show that Hispanic and African American are most likely impacted by cultural characteristics such as religions and ethnic affiliation, whereas the vaccine trust and approval influence the Asian communities the most. Our results also show that cultural characteristics, rumors, and political affiliation are associated with increased vaccine rejection.

* 6 pages, 5 figures 
Viaarxiv icon

Active Membership Inference Attack under Local Differential Privacy in Federated Learning

Feb 24, 2023
Truc Nguyen, Phung Lai, Khang Tran, NhatHai Phan, My T. Thai

Figure 1 for Active Membership Inference Attack under Local Differential Privacy in Federated Learning
Figure 2 for Active Membership Inference Attack under Local Differential Privacy in Federated Learning
Figure 3 for Active Membership Inference Attack under Local Differential Privacy in Federated Learning
Figure 4 for Active Membership Inference Attack under Local Differential Privacy in Federated Learning

Federated learning (FL) was originally regarded as a framework for collaborative learning among clients with data privacy protection through a coordinating server. In this paper, we propose a new active membership inference (AMI) attack carried out by a dishonest server in FL. In AMI attacks, the server crafts and embeds malicious parameters into global models to effectively infer whether a target data sample is included in a client's private training data or not. By exploiting the correlation among data features through a non-linear decision boundary, AMI attacks with a certified guarantee of success can achieve severely high success rates under rigorous local differential privacy (LDP) protection; thereby exposing clients' training data to significant privacy risk. Theoretical and experimental results on several benchmark datasets show that adding sufficient privacy-preserving noise to prevent our attack would significantly damage FL's model utility.

* To be published at AISTATS 2023 
Viaarxiv icon

LAVA: Granular Neuron-Level Explainable AI for Alzheimer's Disease Assessment from Fundus Images

Feb 06, 2023
Nooshin Yousefzadeh, Charlie Tran, Adolfo Ramirez-Zamora, Jinghua Chen, Ruogu Fang, My T. Thai

Figure 1 for LAVA: Granular Neuron-Level Explainable AI for Alzheimer's Disease Assessment from Fundus Images
Figure 2 for LAVA: Granular Neuron-Level Explainable AI for Alzheimer's Disease Assessment from Fundus Images
Figure 3 for LAVA: Granular Neuron-Level Explainable AI for Alzheimer's Disease Assessment from Fundus Images
Figure 4 for LAVA: Granular Neuron-Level Explainable AI for Alzheimer's Disease Assessment from Fundus Images

Alzheimer's Disease (AD) is a progressive neurodegenerative disease and the leading cause of dementia. Early diagnosis is critical for patients to benefit from potential intervention and treatment. The retina has been hypothesized as a diagnostic site for AD detection owing to its anatomical connection with the brain. Developed AI models for this purpose have yet to provide a rational explanation about the decision and neither infer the stage of disease's progression. Along this direction, we propose a novel model-agnostic explainable-AI framework, called Granular Neuron-level Explainer (LAVA), an interpretation prototype that probes into intermediate layers of the Convolutional Neural Network (CNN) models to assess the AD continuum directly from the retinal imaging without longitudinal or clinical evaluation. This method is applied to validate the retinal vasculature as a biomarker and diagnostic modality for Alzheimer's Disease (AD) evaluation. UK Biobank cognitive tests and vascular morphological features suggest LAVA shows strong promise and effectiveness in identifying AD stages across the progression continuum.

* 27 pages, 11 figures 
Viaarxiv icon

XRand: Differentially Private Defense against Explanation-Guided Attacks

Dec 14, 2022
Truc Nguyen, Phung Lai, NhatHai Phan, My T. Thai

Figure 1 for XRand: Differentially Private Defense against Explanation-Guided Attacks
Figure 2 for XRand: Differentially Private Defense against Explanation-Guided Attacks
Figure 3 for XRand: Differentially Private Defense against Explanation-Guided Attacks
Figure 4 for XRand: Differentially Private Defense against Explanation-Guided Attacks

Recent development in the field of explainable artificial intelligence (XAI) has helped improve trust in Machine-Learning-as-a-Service (MLaaS) systems, in which an explanation is provided together with the model prediction in response to each query. However, XAI also opens a door for adversaries to gain insights into the black-box models in MLaaS, thereby making the models more vulnerable to several attacks. For example, feature-based explanations (e.g., SHAP) could expose the top important features that a black-box model focuses on. Such disclosure has been exploited to craft effective backdoor triggers against malware classifiers. To address this trade-off, we introduce a new concept of achieving local differential privacy (LDP) in the explanations, and from that we establish a defense, called XRand, against such attacks. We show that our mechanism restricts the information that the adversary can learn about the top important features, while maintaining the faithfulness of the explanations.

* To be published at AAAI 2023 
Viaarxiv icon