Alert button
Picture for Zenglin Xu

Zenglin Xu

Alert button

Personalized Federated Learning via Amortized Bayesian Meta-Learning

Jul 05, 2023
Shiyu Liu, Shaogao Lv, Dun Zeng, Zenglin Xu, Hui Wang, Yue Yu

Figure 1 for Personalized Federated Learning via Amortized Bayesian Meta-Learning
Figure 2 for Personalized Federated Learning via Amortized Bayesian Meta-Learning
Figure 3 for Personalized Federated Learning via Amortized Bayesian Meta-Learning
Figure 4 for Personalized Federated Learning via Amortized Bayesian Meta-Learning

Federated learning is a decentralized and privacy-preserving technique that enables multiple clients to collaborate with a server to learn a global model without exposing their private data. However, the presence of statistical heterogeneity among clients poses a challenge, as the global model may struggle to perform well on each client's specific task. To address this issue, we introduce a new perspective on personalized federated learning through Amortized Bayesian Meta-Learning. Specifically, we propose a novel algorithm called \emph{FedABML}, which employs hierarchical variational inference across clients. The global prior aims to capture representations of common intrinsic structures from heterogeneous clients, which can then be transferred to their respective tasks and aid in the generation of accurate client-specific approximate posteriors through a few local updates. Our theoretical analysis provides an upper bound on the average generalization error and guarantees the generalization performance on unseen data. Finally, several empirical results are implemented to demonstrate that \emph{FedABML} outperforms several competitive baselines.

Viaarxiv icon

Practical Privacy-Preserving Gaussian Process Regression via Secret Sharing

Jun 26, 2023
Jinglong Luo, Yehong Zhang, Jiaqi Zhang, Shuang Qin, Hui Wang, Yue Yu, Zenglin Xu

Figure 1 for Practical Privacy-Preserving Gaussian Process Regression via Secret Sharing
Figure 2 for Practical Privacy-Preserving Gaussian Process Regression via Secret Sharing
Figure 3 for Practical Privacy-Preserving Gaussian Process Regression via Secret Sharing
Figure 4 for Practical Privacy-Preserving Gaussian Process Regression via Secret Sharing

Gaussian process regression (GPR) is a non-parametric model that has been used in many real-world applications that involve sensitive personal data (e.g., healthcare, finance, etc.) from multiple data owners. To fully and securely exploit the value of different data sources, this paper proposes a privacy-preserving GPR method based on secret sharing (SS), a secure multi-party computation (SMPC) technique. In contrast to existing studies that protect the data privacy of GPR via homomorphic encryption, differential privacy, or federated learning, our proposed method is more practical and can be used to preserve the data privacy of both the model inputs and outputs for various data-sharing scenarios (e.g., horizontally/vertically-partitioned data). However, it is non-trivial to directly apply SS on the conventional GPR algorithm, as it includes some operations whose accuracy and/or efficiency have not been well-enhanced in the current SMPC protocol. To address this issue, we derive a new SS-based exponentiation operation through the idea of 'confusion-correction' and construct an SS-based matrix inversion algorithm based on Cholesky decomposition. More importantly, we theoretically analyze the communication cost and the security of the proposed SS-based operations. Empirical results show that our proposed method can achieve reasonable accuracy and efficiency under the premise of preserving data privacy.

* Accepted for the 39th Conference on Uncertainty in Artificial Intelligence (UAI 2023) 
Viaarxiv icon

FedNoisy: Federated Noisy Label Learning Benchmark

Jun 20, 2023
Siqi Liang, Jintao Huang, Dun Zeng, Junyuan Hong, Jiayu Zhou, Zenglin Xu

Figure 1 for FedNoisy: Federated Noisy Label Learning Benchmark
Figure 2 for FedNoisy: Federated Noisy Label Learning Benchmark
Figure 3 for FedNoisy: Federated Noisy Label Learning Benchmark
Figure 4 for FedNoisy: Federated Noisy Label Learning Benchmark

Federated learning has gained popularity for distributed learning without aggregating sensitive data from clients. But meanwhile, the distributed and isolated nature of data isolation may be complicated by data quality, making it more vulnerable to noisy labels. Many efforts exist to defend against the negative impacts of noisy labels in centralized or federated settings. However, there is a lack of a benchmark that comprehensively considers the impact of noisy labels in a wide variety of typical FL settings. In this work, we serve the first standardized benchmark that can help researchers fully explore potential federated noisy settings. Also, we conduct comprehensive experiments to explore the characteristics of these data settings and unravel challenging scenarios on the federated noisy label learning, which may guide method development in the future. We highlight the 20 basic settings for more than 5 datasets proposed in our benchmark and standardized simulation pipeline for federated noisy label learning. We hope this benchmark can facilitate idea verification in federated learning with noisy labels. \texttt{FedNoisy} is available at \codeword{https://github.com/SMILELab-FL/FedNoisy}.

Viaarxiv icon

Enabling Intelligent Interactions between an Agent and an LLM: A Reinforcement Learning Approach

Jun 11, 2023
Bin Hu, Chenyang Zhao, Pu Zhang, Zihao Zhou, Yuanhang Yang, Zenglin Xu, Bin Liu

Figure 1 for Enabling Intelligent Interactions between an Agent and an LLM: A Reinforcement Learning Approach
Figure 2 for Enabling Intelligent Interactions between an Agent and an LLM: A Reinforcement Learning Approach
Figure 3 for Enabling Intelligent Interactions between an Agent and an LLM: A Reinforcement Learning Approach
Figure 4 for Enabling Intelligent Interactions between an Agent and an LLM: A Reinforcement Learning Approach

Large language models (LLMs) encode a vast amount of world knowledge acquired from massive text datasets. Recent studies have demonstrated that LLMs can assist an agent in solving complex sequential decision making tasks in embodied environments by providing high-level instructions. However, interacting with LLMs can be time-consuming, as in many practical scenarios, they require a significant amount of storage space that can only be deployed on remote cloud server nodes. Additionally, using commercial LLMs can be costly since they may charge based on usage frequency. In this paper, we explore how to enable intelligent cost-effective interactions between the agent and an LLM. We propose a reinforcement learning based mediator model that determines when it is necessary to consult LLMs for high-level instructions to accomplish a target task. Experiments on 4 MiniGrid environments that entail planning sub-goals demonstrate that our method can learn to solve target tasks with only a few necessary interactions with an LLM, significantly reducing interaction costs in testing environments, compared with baseline methods. Experimental results also suggest that by learning a mediator model to interact with the LLM, the agent's performance becomes more robust against partial observability of the environment. Our code is available at https://github.com/ZJLAB-AMMI/LLM4RL.

* 11 pages 
Viaarxiv icon

Tensorized Hypergraph Neural Networks

Jun 05, 2023
Maolin Wang, Yaoming Zhen, Yu Pan, Zenglin Xu, Ruocheng Guo, Xiangyu Zhao

Figure 1 for Tensorized Hypergraph Neural Networks
Figure 2 for Tensorized Hypergraph Neural Networks
Figure 3 for Tensorized Hypergraph Neural Networks
Figure 4 for Tensorized Hypergraph Neural Networks

Hypergraph neural networks (HGNN) have recently become attractive and received significant attention due to their excellent performance in various domains. However, most existing HGNNs rely on first-order approximations of hypergraph connectivity patterns, which ignores important high-order information. To address this issue, we propose a novel adjacency-tensor-based Tensorized Hypergraph Neural Network (THNN). THNN is a faithful hypergraph modeling framework through high-order outer product feature message passing and is a natural tensor extension of the adjacency-matrix-based graph neural networks. The proposed THNN is equivalent to an high-order polynomial regression scheme, which enable THNN with the ability to efficiently extract high-order information from uniform hypergraphs. Moreover, in consideration of the exponential complexity of directly processing high-order outer product features, we propose using a partially symmetric CP decomposition approach to reduce model complexity to a linear degree. Additionally, we propose two simple yet effective extensions of our method for non-uniform hypergraphs commonly found in real-world applications. Results from experiments on two widely used hypergraph datasets for 3-D visual object classification show the promising performance of the proposed THNN.

Viaarxiv icon

Revisiting Long-term Time Series Forecasting: An Investigation on Linear Mapping

May 18, 2023
Zhe Li, Shiyi Qi, Yiduo Li, Zenglin Xu

Figure 1 for Revisiting Long-term Time Series Forecasting: An Investigation on Linear Mapping
Figure 2 for Revisiting Long-term Time Series Forecasting: An Investigation on Linear Mapping
Figure 3 for Revisiting Long-term Time Series Forecasting: An Investigation on Linear Mapping
Figure 4 for Revisiting Long-term Time Series Forecasting: An Investigation on Linear Mapping

Long-term time series forecasting has gained significant attention in recent years. While there are various specialized designs for capturing temporal dependency, previous studies have demonstrated that a single linear layer can achieve competitive forecasting performance compared to other complex architectures. In this paper, we thoroughly investigate the intrinsic effectiveness of recent approaches and make three key observations: 1) linear mapping is critical to prior long-term time series forecasting efforts; 2) RevIN (reversible normalization) and CI (Channel Independent) play a vital role in improving overall forecasting performance; and 3) linear mapping can effectively capture periodic features in time series and has robustness for different periods across channels when increasing the input horizon. We provide theoretical and experimental explanations to support our findings and also discuss the limitations and future works. Our framework's code is available at \url{https://github.com/plumprc/RTSF}.

* 12 pages, 11 figures 
Viaarxiv icon

Multi-task Transformer with Relation-attention and Type-attention for Named Entity Recognition

Mar 20, 2023
Ying Mo, Hongyin Tang, Jiahao Liu, Qifan Wang, Zenglin Xu, Jingang Wang, Wei Wu, Zhoujun Li

Figure 1 for Multi-task Transformer with Relation-attention and Type-attention for Named Entity Recognition
Figure 2 for Multi-task Transformer with Relation-attention and Type-attention for Named Entity Recognition
Figure 3 for Multi-task Transformer with Relation-attention and Type-attention for Named Entity Recognition
Figure 4 for Multi-task Transformer with Relation-attention and Type-attention for Named Entity Recognition

Named entity recognition (NER) is an important research problem in natural language processing. There are three types of NER tasks, including flat, nested and discontinuous entity recognition. Most previous sequential labeling models are task-specific, while recent years have witnessed the rising of generative models due to the advantage of unifying all NER tasks into the seq2seq model framework. Although achieving promising performance, our pilot studies demonstrate that existing generative models are ineffective at detecting entity boundaries and estimating entity types. This paper proposes a multi-task Transformer, which incorporates an entity boundary detection task into the named entity recognition task. More concretely, we achieve entity boundary detection by classifying the relations between tokens within the sentence. To improve the accuracy of entity-type mapping during decoding, we adopt an external knowledge base to calculate the prior entity-type distributions and then incorporate the information into the model via the self and cross-attention mechanisms. We perform experiments on an extensive set of NER benchmarks, including two flat, three nested, and three discontinuous NER datasets. Experimental results show that our approach considerably improves the generative NER model's performance.

* 5 pages,accepted ICASSP 2023 
Viaarxiv icon

Stochastic Clustered Federated Learning

Mar 02, 2023
Dun Zeng, Xiangjing Hu, Shiyu Liu, Yue Yu, Qifan Wang, Zenglin Xu

Figure 1 for Stochastic Clustered Federated Learning
Figure 2 for Stochastic Clustered Federated Learning
Figure 3 for Stochastic Clustered Federated Learning
Figure 4 for Stochastic Clustered Federated Learning

Federated learning is a distributed learning framework that takes full advantage of private data samples kept on edge devices. In real-world federated learning systems, these data samples are often decentralized and Non-Independently Identically Distributed (Non-IID), causing divergence and performance degradation in the federated learning process. As a new solution, clustered federated learning groups federated clients with similar data distributions to impair the Non-IID effects and train a better model for every cluster. This paper proposes StoCFL, a novel clustered federated learning approach for generic Non-IID issues. In detail, StoCFL implements a flexible CFL framework that supports an arbitrary proportion of client participation and newly joined clients for a varying FL system, while maintaining a great improvement in model performance. The intensive experiments are conducted by using four basic Non-IID settings and a real-world dataset. The results show that StoCFL could obtain promising cluster results even when the number of clusters is unknown. Based on the client clustering results, models trained with StoCFL outperform baseline approaches in a variety of contexts.

Viaarxiv icon

A Survey of Trustworthy Federated Learning with Perspectives on Security, Robustness, and Privacy

Feb 21, 2023
Yifei Zhang, Dun Zeng, Jinglong Luo, Zenglin Xu, Irwin King

Figure 1 for A Survey of Trustworthy Federated Learning with Perspectives on Security, Robustness, and Privacy
Figure 2 for A Survey of Trustworthy Federated Learning with Perspectives on Security, Robustness, and Privacy
Figure 3 for A Survey of Trustworthy Federated Learning with Perspectives on Security, Robustness, and Privacy
Figure 4 for A Survey of Trustworthy Federated Learning with Perspectives on Security, Robustness, and Privacy

Trustworthy artificial intelligence (AI) technology has revolutionized daily life and greatly benefited human society. Among various AI technologies, Federated Learning (FL) stands out as a promising solution for diverse real-world scenarios, ranging from risk evaluation systems in finance to cutting-edge technologies like drug discovery in life sciences. However, challenges around data isolation and privacy threaten the trustworthiness of FL systems. Adversarial attacks against data privacy, learning algorithm stability, and system confidentiality are particularly concerning in the context of distributed training in federated learning. Therefore, it is crucial to develop FL in a trustworthy manner, with a focus on security, robustness, and privacy. In this survey, we propose a comprehensive roadmap for developing trustworthy FL systems and summarize existing efforts from three key aspects: security, robustness, and privacy. We outline the threats that pose vulnerabilities to trustworthy federated learning across different stages of development, including data processing, model training, and deployment. To guide the selection of the most appropriate defense methods, we discuss specific technical solutions for realizing each aspect of Trustworthy FL (TFL). Our approach differs from previous work that primarily discusses TFL from a legal perspective or presents FL from a high-level, non-technical viewpoint.

Viaarxiv icon

Generative Oversampling for Imbalanced Data via Majority-Guided VAE

Feb 14, 2023
Qingzhong Ai, Pengyun Wang, Lirong He, Liangjian Wen, Lujia Pan, Zenglin Xu

Figure 1 for Generative Oversampling for Imbalanced Data via Majority-Guided VAE
Figure 2 for Generative Oversampling for Imbalanced Data via Majority-Guided VAE
Figure 3 for Generative Oversampling for Imbalanced Data via Majority-Guided VAE
Figure 4 for Generative Oversampling for Imbalanced Data via Majority-Guided VAE

Learning with imbalanced data is a challenging problem in deep learning. Over-sampling is a widely used technique to re-balance the sampling distribution of training data. However, most existing over-sampling methods only use intra-class information of minority classes to augment the data but ignore the inter-class relationships with the majority ones, which is prone to overfitting, especially when the imbalance ratio is large. To address this issue, we propose a novel over-sampling model, called Majority-Guided VAE~(MGVAE), which generates new minority samples under the guidance of a majority-based prior. In this way, the newly generated minority samples can inherit the diversity and richness of the majority ones, thus mitigating overfitting in downstream tasks. Furthermore, to prevent model collapse under limited data, we first pre-train MGVAE on sufficient majority samples and then fine-tune based on minority samples with Elastic Weight Consolidation(EWC) regularization. Experimental results on benchmark image datasets and real-world tabular data show that MGVAE achieves competitive improvements over other over-sampling methods in downstream classification tasks, demonstrating the effectiveness of our method.

* AISTATS 2023 
Viaarxiv icon