Alert button
Picture for Xiaoqian Wang

Xiaoqian Wang

Alert button

To be Robust and to be Fair: Aligning Fairness with Robustness

Mar 31, 2023
Junyi Chai, Xiaoqian Wang

Figure 1 for To be Robust and to be Fair: Aligning Fairness with Robustness
Figure 2 for To be Robust and to be Fair: Aligning Fairness with Robustness
Figure 3 for To be Robust and to be Fair: Aligning Fairness with Robustness
Figure 4 for To be Robust and to be Fair: Aligning Fairness with Robustness

Adversarial training has been shown to be reliable in improving robustness against adversarial samples. However, the problem of adversarial training in terms of fairness has not yet been properly studied, and the relationship between fairness and accuracy attack still remains unclear. Can we simultaneously improve robustness w.r.t. both fairness and accuracy? To tackle this topic, in this paper, we study the problem of adversarial training and adversarial attack w.r.t. both metrics. We propose a unified structure for fairness attack which brings together common notions in group fairness, and we theoretically prove the equivalence of fairness attack against different notions. Moreover, we show the alignment of fairness and accuracy attack, and theoretically demonstrate that robustness w.r.t. one metric benefits from robustness w.r.t. the other metric. Our study suggests a novel way to unify adversarial training and attack w.r.t. fairness and accuracy, and experimental results show that our proposed method achieves better performance in terms of robustness w.r.t. both metrics.

Viaarxiv icon

SimFair: A Unified Framework for Fairness-Aware Multi-Label Classification

Feb 22, 2023
Tianci Liu, Haoyu Wang, Yaqing Wang, Xiaoqian Wang, Lu Su, Jing Gao

Figure 1 for SimFair: A Unified Framework for Fairness-Aware Multi-Label Classification
Figure 2 for SimFair: A Unified Framework for Fairness-Aware Multi-Label Classification
Figure 3 for SimFair: A Unified Framework for Fairness-Aware Multi-Label Classification
Figure 4 for SimFair: A Unified Framework for Fairness-Aware Multi-Label Classification

Recent years have witnessed increasing concerns towards unfair decisions made by machine learning algorithms. To improve fairness in model decisions, various fairness notions have been proposed and many fairness-aware methods are developed. However, most of existing definitions and methods focus only on single-label classification. Fairness for multi-label classification, where each instance is associated with more than one labels, is still yet to establish. To fill this gap, we study fairness-aware multi-label classification in this paper. We start by extending Demographic Parity (DP) and Equalized Opportunity (EOp), two popular fairness notions, to multi-label classification scenarios. Through a systematic study, we show that on multi-label data, because of unevenly distributed labels, EOp usually fails to construct a reliable estimate on labels with few instances. We then propose a new framework named Similarity $s$-induced Fairness ($s_\gamma$-SimFair). This new framework utilizes data that have similar labels when estimating fairness on a particular label group for better stability, and can unify DP and EOp. Theoretical analysis and experimental results on real-world datasets together demonstrate the advantage of over existing methods $s_\gamma$-SimFair on multi-label classification tasks.

* AAAI2023 
Viaarxiv icon

Forecast combinations: an over 50-year review

May 09, 2022
Xiaoqian Wang, Rob J Hyndman, Feng Li, Yanfei Kang

Figure 1 for Forecast combinations: an over 50-year review

Forecast combinations have flourished remarkably in the forecasting community and, in recent years, have become part of the mainstream of forecasting research and activities. Combining multiple forecasts produced from the single (target) series is now widely used to improve accuracy through the integration of information gleaned from different sources, thereby mitigating the risk of identifying a single "best" forecast. Combination schemes have evolved from simple combination methods without estimation, to sophisticated methods involving time-varying weights, nonlinear combinations, correlations among components, and cross-learning. They include combining point forecasts, and combining probabilistic forecasts. This paper provides an up-to-date review of the extensive literature on forecast combinations, together with reference to available open-source software implementations. We discuss the potential and limitations of various methods and highlight how these ideas have developed over time. Some important issues concerning the utility of forecast combinations are also surveyed. Finally, we conclude with current research gaps and potential insights for future research.

Viaarxiv icon

A Unified Study of Machine Learning Explanation Evaluation Metrics

Mar 27, 2022
Yipei Wang, Xiaoqian Wang

Figure 1 for A Unified Study of Machine Learning Explanation Evaluation Metrics
Figure 2 for A Unified Study of Machine Learning Explanation Evaluation Metrics
Figure 3 for A Unified Study of Machine Learning Explanation Evaluation Metrics
Figure 4 for A Unified Study of Machine Learning Explanation Evaluation Metrics

The growing need for trustworthy machine learning has led to the blossom of interpretability research. Numerous explanation methods have been developed to serve this purpose. However, these methods are deficiently and inappropriately evaluated. Many existing metrics for explanations are introduced by researchers as by-products of their proposed explanation techniques to demonstrate the advantages of their methods. Although widely used, they are more or less accused of problems. We claim that the lack of acknowledged and justified metrics results in chaos in benchmarking these explanation methods -- Do we really have good/bad explanation when a metric gives a high/low score? We split existing metrics into two categories and demonstrate that they are insufficient to properly evaluate explanations for multiple reasons. We propose guidelines in dealing with the problems in evaluating machine learning explanation and encourage researchers to carefully deal with these problems when developing explanation techniques and metrics.

Viaarxiv icon

Self-Interpretable Model with TransformationEquivariant Interpretation

Nov 09, 2021
Yipei Wang, Xiaoqian Wang

Figure 1 for Self-Interpretable Model with TransformationEquivariant Interpretation
Figure 2 for Self-Interpretable Model with TransformationEquivariant Interpretation
Figure 3 for Self-Interpretable Model with TransformationEquivariant Interpretation
Figure 4 for Self-Interpretable Model with TransformationEquivariant Interpretation

In this paper, we propose a self-interpretable model SITE with transformation-equivariant interpretations. We focus on the robustness and self-consistency of the interpretations of geometric transformations. Apart from the transformation equivariance, as a self-interpretable model, SITE has comparable expressive power as the benchmark black-box classifiers, while being able to present faithful and robust interpretations with high quality. It is worth noticing that although applied in most of the CNN visualization methods, the bilinear upsampling approximation is a rough approximation, which can only provide interpretations in the form of heatmaps (instead of pixel-wise). It remains an open question whether such interpretations can be direct to the input space (as shown in the MNIST experiments). Besides, we consider the translation and rotation transformations in our model. In future work, we will explore the robust interpretations under more complex transformations such as scaling and distortion. Moreover, we clarify that SITE is not limited to geometric transformation (that we used in the computer vision domain), and will explore SITEin other domains in future work.

* Accepted by NeurIPS 2021 
Viaarxiv icon

Group-Aware Threshold Adaptation for Fair Classification

Nov 08, 2021
Taeuk Jang, Pengyi Shi, Xiaoqian Wang

Figure 1 for Group-Aware Threshold Adaptation for Fair Classification
Figure 2 for Group-Aware Threshold Adaptation for Fair Classification
Figure 3 for Group-Aware Threshold Adaptation for Fair Classification
Figure 4 for Group-Aware Threshold Adaptation for Fair Classification

The fairness in machine learning is getting increasing attention, as its applications in different fields continue to expand and diversify. To mitigate the discriminated model behaviors between different demographic groups, we introduce a novel post-processing method to optimize over multiple fairness constraints through group-aware threshold adaptation. We propose to learn adaptive classification thresholds for each demographic group by optimizing the confusion matrix estimated from the probability distribution of a classification model output. As we only need an estimated probability distribution of model output instead of the classification model structure, our post-processing model can be applied to a wide range of classification models and improve fairness in a model-agnostic manner and ensure privacy. This even allows us to post-process existing fairness methods to further improve the trade-off between accuracy and fairness. Moreover, our model has low computational cost. We provide rigorous theoretical analysis on the convergence of our optimization algorithm and the trade-off between accuracy and fairness of our method. Our method theoretically enables a better upper bound in near optimality than existing method under same condition. Experimental results demonstrate that our method outperforms state-of-the-art methods and obtains the result that is closest to the theoretical accuracy-fairness trade-off boundary.

* 19 pages 1 figures 
Viaarxiv icon

Shapley Explanation Networks

Apr 06, 2021
Rui Wang, Xiaoqian Wang, David I. Inouye

Figure 1 for Shapley Explanation Networks
Figure 2 for Shapley Explanation Networks
Figure 3 for Shapley Explanation Networks
Figure 4 for Shapley Explanation Networks

Shapley values have become one of the most popular feature attribution explanation methods. However, most prior work has focused on post-hoc Shapley explanations, which can be computationally demanding due to its exponential time complexity and preclude model regularization based on Shapley explanations during training. Thus, we propose to incorporate Shapley values themselves as latent representations in deep models thereby making Shapley explanations first-class citizens in the modeling paradigm. This intrinsic explanation approach enables layer-wise explanations, explanation regularization of the model during training, and fast explanation computation at test time. We define the Shapley transform that transforms the input into a Shapley representation given a specific function. We operationalize the Shapley transform as a neural network module and construct both shallow and deep networks, called ShapNets, by composing Shapley modules. We prove that our Shallow ShapNets compute the exact Shapley values and our Deep ShapNets maintain the missingness and accuracy properties of Shapley values. We demonstrate on synthetic and real-world datasets that our ShapNets enable layer-wise Shapley explanations, novel Shapley regularizations during training, and fast computation while maintaining reasonable performance. Code is available at https://github.com/inouye-lab/ShapleyExplanationNetworks.

* 26 pages, 11 figures, accepted to ICLR 2021 
Viaarxiv icon

Multivariate Time Series Classification with Hierarchical Variational Graph Pooling

Oct 12, 2020
Haoyan Xu, Ziheng Duan, Yunsheng Bai, Yida Huang, Anni Ren, Qianru Yu, Qianru Zhang, Yueyang Wang, Xiaoqian Wang, Yizhou Sun, Wei Wang

Figure 1 for Multivariate Time Series Classification with Hierarchical Variational Graph Pooling
Figure 2 for Multivariate Time Series Classification with Hierarchical Variational Graph Pooling
Figure 3 for Multivariate Time Series Classification with Hierarchical Variational Graph Pooling
Figure 4 for Multivariate Time Series Classification with Hierarchical Variational Graph Pooling

Over the past decade, multivariate time series classification (MTSC) has received great attention with the advance of sensing techniques. Current deep learning methods for MTSC are based on convolutional and recurrent neural network, with the assumption that time series variables have the same effect to each other. Thus they cannot model the pairwise dependencies among variables explicitly. What's more, current spatial-temporal modeling methods based on GNNs are inherently flat and lack the capability of aggregating node information in a hierarchical manner. To address this limitation and attain expressive global representation of MTS, we propose a graph pooling based framework MTPool and view MTSC task as graph classification task. With graph structure learning and temporal convolution, MTS slices are converted to graphs and spatial-temporal features are extracted. Then, we propose a novel graph pooling method, which uses an ``encoder-decoder'' mechanism to generate adaptive centroids for cluster assignments. GNNs and graph pooling layers are used for joint graph representation learning and graph coarsening. With multiple graph pooling layers, the input graphs are hierachically coarsened to one node. Finally, differentiable classifier takes this coarsened one-node graph as input to get the final predicted class. Experiments on 10 benchmark datasets demonstrate MTPool outperforms state-of-the-art methods in MTSC tasks.

Viaarxiv icon

Modeling Complex Spatial Patterns with Temporal Features via Heterogenous Graph Embedding Networks

Sep 11, 2020
Yida Huang, Haoyan Xu, Ziheng Duan, Anni Ren, Jie Feng, Qianru Zhang, Xiaoqian Wang

Figure 1 for Modeling Complex Spatial Patterns with Temporal Features via Heterogenous Graph Embedding Networks
Figure 2 for Modeling Complex Spatial Patterns with Temporal Features via Heterogenous Graph Embedding Networks
Figure 3 for Modeling Complex Spatial Patterns with Temporal Features via Heterogenous Graph Embedding Networks
Figure 4 for Modeling Complex Spatial Patterns with Temporal Features via Heterogenous Graph Embedding Networks

Multivariate time series (MTS) forecasting is an important problem in many fields. Accurate forecasting results can effectively help decision-making. Variables in MTS have rich relations among each other and the value of each variable in MTS depends both on its historical values and on other variables. These rich relations can be static and predictable or dynamic and latent. Existing methods do not incorporate these rich relational information into modeling or only model certain relation among MTS variables. To jointly model rich relations among variables and temporal dependencies within the time series, a novel end-to-end deep learning model, termed Multivariate Time Series Forecasting via Heterogenous Graph Neural Networks (MTHetGNN) is proposed in this paper. To characterize rich relations among variables, a relation embedding module is introduced in our model, where each variable is regarded as a graph node and each type of edge represents a specific relationship among variables or one specific dynamic update strategy to model the latent dependency among variables. In addition, convolutional neural network (CNN) filters with different perception scales are used for time series feature extraction, which is used to generate the feature of each node. Finally, heterogenous graph neural networks are adopted to handle the complex structural information generated by temporal embedding module and relation embedding module. Three benchmark datasets from the real world are used to evaluate the proposed MTHetGNN and the comprehensive experiments show that MTHetGNN achieves state-of-the-art results in MTS forecasting task.

Viaarxiv icon