Alert button
Picture for Wenkai Xu

Wenkai Xu

Alert button

Nonlinear Causal Discovery via Kernel Anchor Regression

Oct 30, 2022
Wenqi Shi, Wenkai Xu

Figure 1 for Nonlinear Causal Discovery via Kernel Anchor Regression
Figure 2 for Nonlinear Causal Discovery via Kernel Anchor Regression
Figure 3 for Nonlinear Causal Discovery via Kernel Anchor Regression
Figure 4 for Nonlinear Causal Discovery via Kernel Anchor Regression

Learning causal relationships is a fundamental problem in science. Anchor regression has been developed to address this problem for a large class of causal graphical models, though the relationships between the variables are assumed to be linear. In this work, we tackle the nonlinear setting by proposing kernel anchor regression (KAR). Beyond the natural formulation using a classic two-stage least square estimator, we also study an improved variant that involves nonparametric regression in three separate stages. We provide convergence results for the proposed KAR estimators and the identifiability conditions for KAR to learn the nonlinear structural equation models (SEM). Experimental results demonstrate the superior performances of the proposed KAR estimators over existing baselines.

Viaarxiv icon

On RKHS Choices for Assessing Graph Generators via Kernel Stein Statistics

Oct 11, 2022
Moritz Weckbecker, Wenkai Xu, Gesine Reinert

Figure 1 for On RKHS Choices for Assessing Graph Generators via Kernel Stein Statistics
Figure 2 for On RKHS Choices for Assessing Graph Generators via Kernel Stein Statistics
Figure 3 for On RKHS Choices for Assessing Graph Generators via Kernel Stein Statistics
Figure 4 for On RKHS Choices for Assessing Graph Generators via Kernel Stein Statistics

Score-based kernelised Stein discrepancy (KSD) tests have emerged as a powerful tool for the goodness of fit tests, especially in high dimensions; however, the test performance may depend on the choice of kernels in an underlying reproducing kernel Hilbert space (RKHS). Here we assess the effect of RKHS choice for KSD tests of random networks models, developed for exponential random graph models (ERGMs) in Xu and Reinert (2021)and for synthetic graph generators in Xu and Reinert (2022). We investigate the power performance and the computational runtime of the test in different scenarios, including both dense and sparse graph regimes. Experimental results on kernel performance for model assessment tasks are shown and discussed on synthetic and real-world network applications.

Viaarxiv icon

A Kernelised Stein Statistic for Assessing Implicit Generative Models

May 31, 2022
Wenkai Xu, Gesine Reinert

Figure 1 for A Kernelised Stein Statistic for Assessing Implicit Generative Models
Figure 2 for A Kernelised Stein Statistic for Assessing Implicit Generative Models
Figure 3 for A Kernelised Stein Statistic for Assessing Implicit Generative Models
Figure 4 for A Kernelised Stein Statistic for Assessing Implicit Generative Models

Synthetic data generation has become a key ingredient for training machine learning procedures, addressing tasks such as data augmentation, analysing privacy-sensitive data, or visualising representative samples. Assessing the quality of such synthetic data generators hence has to be addressed. As (deep) generative models for synthetic data often do not admit explicit probability distributions, classical statistical procedures for assessing model goodness-of-fit may not be applicable. In this paper, we propose a principled procedure to assess the quality of a synthetic data generator. The procedure is a kernelised Stein discrepancy (KSD)-type test which is based on a non-parametric Stein operator for the synthetic data generator of interest. This operator is estimated from samples which are obtained from the synthetic data generator and hence can be applied even when the model is only implicit. In contrast to classical testing, the sample size from the synthetic data generator can be as large as desired, while the size of the observed data, which the generator aims to emulate is fixed. Experimental results on synthetic distributions and trained generative models on synthetic and real datasets illustrate that the method shows improved power performance compared to existing approaches.

Viaarxiv icon

AgraSSt: Approximate Graph Stein Statistics for Interpretable Assessment of Implicit Graph Generators

Mar 07, 2022
Wenkai Xu, Gesine Reinert

Figure 1 for AgraSSt: Approximate Graph Stein Statistics for Interpretable Assessment of Implicit Graph Generators
Figure 2 for AgraSSt: Approximate Graph Stein Statistics for Interpretable Assessment of Implicit Graph Generators
Figure 3 for AgraSSt: Approximate Graph Stein Statistics for Interpretable Assessment of Implicit Graph Generators
Figure 4 for AgraSSt: Approximate Graph Stein Statistics for Interpretable Assessment of Implicit Graph Generators

We propose and analyse a novel statistical procedure, coined AgraSSt, to assess the quality of graph generators that may not be available in explicit form. In particular, AgraSSt can be used to determine whether a learnt graph generating process is capable of generating graphs that resemble a given input graph. Inspired by Stein operators for random graphs, the key idea of AgraSSt is the construction of a kernel discrepancy based on an operator obtained from the graph generator. AgraSSt can provide interpretable criticisms for a graph generator training procedure and help identify reliable sample batches for downstream tasks. Using Stein`s method we give theoretical guarantees for a broad class of random graph models. We provide empirical results on both synthetic input graphs with known graph generation procedures, and real-world input graphs that the state-of-the-art (deep) generative models for graphs are trained on.

Viaarxiv icon

Generalised Kernel Stein Discrepancy(GKSD): A Unifying Approach for Non-parametric Goodness-of-fit Testing

Jun 23, 2021
Wenkai Xu

Figure 1 for Generalised Kernel Stein Discrepancy(GKSD): A Unifying Approach for Non-parametric Goodness-of-fit Testing
Figure 2 for Generalised Kernel Stein Discrepancy(GKSD): A Unifying Approach for Non-parametric Goodness-of-fit Testing
Figure 3 for Generalised Kernel Stein Discrepancy(GKSD): A Unifying Approach for Non-parametric Goodness-of-fit Testing
Figure 4 for Generalised Kernel Stein Discrepancy(GKSD): A Unifying Approach for Non-parametric Goodness-of-fit Testing

Non-parametric goodness-of-fit testing procedures based on kernel Stein discrepancies (KSD) are promising approaches to validate general unnormalised distributions in various scenarios. Existing works have focused on studying optimal kernel choices to boost test performances. However, the Stein operators are generally non-unique, while different choices of Stein operators can also have considerable effect on the test performances. In this work, we propose a unifying framework, the generalised kernel Stein discrepancy (GKSD), to theoretically compare and interpret different Stein operators in performing the KSD-based goodness-of-fit tests. We derive explicitly that how the proposed GKSD framework generalises existing Stein operators and their corresponding tests. In addition, we show thatGKSD framework can be used as a guide to develop kernel-based non-parametric goodness-of-fit tests for complex new data scenarios, e.g. truncated distributions or compositional data. Experimental results demonstrate that the proposed tests control type-I error well and achieve higher test power than existing approaches, including the test based on maximum-mean-discrepancy (MMD).

Viaarxiv icon

Meta Two-Sample Testing: Learning Kernels for Testing with Limited Data

Jun 14, 2021
Feng Liu, Wenkai Xu, Jie Lu, Danica J. Sutherland

Figure 1 for Meta Two-Sample Testing: Learning Kernels for Testing with Limited Data
Figure 2 for Meta Two-Sample Testing: Learning Kernels for Testing with Limited Data
Figure 3 for Meta Two-Sample Testing: Learning Kernels for Testing with Limited Data
Figure 4 for Meta Two-Sample Testing: Learning Kernels for Testing with Limited Data

Modern kernel-based two-sample tests have shown great success in distinguishing complex, high-dimensional distributions with appropriate learned kernels. Previous work has demonstrated that this kernel learning procedure succeeds, assuming a considerable number of observed samples from each distribution. In realistic scenarios with very limited numbers of data samples, however, it can be challenging to identify a kernel powerful enough to distinguish complex distributions. We address this issue by introducing the problem of meta two-sample testing (M2ST), which aims to exploit (abundant) auxiliary data on related tasks to find an algorithm that can quickly identify a powerful test on new target tasks. We propose two specific algorithms for this task: a generic scheme which improves over baselines and amore tailored approach which performs even better. We provide both theoretical justification and empirical evidence that our proposed meta-testing schemes out-perform learning kernel-based tests directly from scarce observations, and identify when such schemes will be successful.

* Code is available from https://github.com/fengliu90/MetaTesting 
Viaarxiv icon

A Stein Goodness of fit Test for Exponential Random Graph Models

Feb 28, 2021
Wenkai Xu, Gesine Reinert

Figure 1 for A Stein Goodness of fit Test for Exponential Random Graph Models
Figure 2 for A Stein Goodness of fit Test for Exponential Random Graph Models
Figure 3 for A Stein Goodness of fit Test for Exponential Random Graph Models
Figure 4 for A Stein Goodness of fit Test for Exponential Random Graph Models

We propose and analyse a novel nonparametric goodness of fit testing procedure for exchangeable exponential random graph models (ERGMs) when a single network realisation is observed. The test determines how likely it is that the observation is generated from a target unnormalised ERGM density. Our test statistics are derived from a kernel Stein discrepancy, a divergence constructed via Steins method using functions in a reproducing kernel Hilbert space, combined with a discrete Stein operator for ERGMs. The test is a Monte Carlo test based on simulated networks from the target ERGM. We show theoretical properties for the testing procedure for a class of ERGMs. Simulation studies and real network applications are presented.

* Proceedings of the 24th International Conference on Artificial Intelligence and Statistics (AISTATS) 2021  
Viaarxiv icon

A kernel test for quasi-independence

Nov 17, 2020
Tamara Fernández, Wenkai Xu, Marc Ditzhaus, Arthur Gretton

Figure 1 for A kernel test for quasi-independence
Figure 2 for A kernel test for quasi-independence
Figure 3 for A kernel test for quasi-independence
Figure 4 for A kernel test for quasi-independence

We consider settings in which the data of interest correspond to pairs of ordered times, e.g, the birth times of the first and second child, the times at which a new user creates an account and makes the first purchase on a website, and the entry and survival times of patients in a clinical trial. In these settings, the two times are not independent (the second occurs after the first), yet it is still of interest to determine whether there exists significant dependence {\em beyond} their ordering in time. We refer to this notion as "quasi-(in)dependence". For instance, in a clinical trial, to avoid biased selection, we might wish to verify that recruitment times are quasi-independent of survival times, where dependencies might arise due to seasonal effects. In this paper, we propose a nonparametric statistical test of quasi-independence. Our test considers a potentially infinite space of alternatives, making it suitable for complex data where the nature of the possible quasi-dependence is not known in advance. Standard parametric approaches are recovered as special cases, such as the classical conditional Kendall's tau, and log-rank tests. The tests apply in the right-censored setting: an essential feature in clinical trials, where patients can withdraw from the study. We provide an asymptotic analysis of our test-statistic, and demonstrate in experiments that our test obtains better power than existing approaches, while being more computationally efficient.

Viaarxiv icon

Kernelized Stein Discrepancy Tests of Goodness-of-fit for Time-to-Event Data

Aug 26, 2020
Tamara Fernandez, Nicolas Rivera, Wenkai Xu, Arthur Gretton

Figure 1 for Kernelized Stein Discrepancy Tests of Goodness-of-fit for Time-to-Event Data
Figure 2 for Kernelized Stein Discrepancy Tests of Goodness-of-fit for Time-to-Event Data
Figure 3 for Kernelized Stein Discrepancy Tests of Goodness-of-fit for Time-to-Event Data
Figure 4 for Kernelized Stein Discrepancy Tests of Goodness-of-fit for Time-to-Event Data

Survival Analysis and Reliability Theory are concerned with the analysis of time-to-event data, in which observations correspond to waiting times until an event of interest such as death from a particular disease or failure of a component in a mechanical system. This type of data is unique due to the presence of censoring, a type of missing data that occurs when we do not observe the actual time of the event of interest but, instead, we have access to an approximation for it given by random interval in which the observation is known to belong. Most traditional methods are not designed to deal with censoring, and thus we need to adapt them to censored time-to-event data. In this paper, we focus on non-parametric goodness-of-fit testing procedures based on combining the Stein's method and kernelized discrepancies. While for uncensored data, there is a natural way of implementing a kernelized Stein discrepancy test, for censored data there are several options, each of them with different advantages and disadvantages. In this paper, we propose a collection of kernelized Stein discrepancy tests for time-to-event data, and we study each of them theoretically and empirically; our experimental results show that our proposed methods perform better than existing tests, including previous tests based on a kernelized maximum mean discrepancy.

* Proceedings of the International Conference on Machine Learning, 2020 
Viaarxiv icon

Learning Deep Kernels for Non-Parametric Two-Sample Tests

Feb 21, 2020
Feng Liu, Wenkai Xu, Jie Lu, Guangquan Zhang, Arthur Gretton, D. J. Sutherland

Figure 1 for Learning Deep Kernels for Non-Parametric Two-Sample Tests
Figure 2 for Learning Deep Kernels for Non-Parametric Two-Sample Tests
Figure 3 for Learning Deep Kernels for Non-Parametric Two-Sample Tests
Figure 4 for Learning Deep Kernels for Non-Parametric Two-Sample Tests

We propose a class of kernel-based two-sample tests, which aim to determine whether two sets of samples are drawn from the same distribution. Our tests are constructed from kernels parameterized by deep neural nets, trained to maximize test power. These tests adapt to variations in distribution smoothness and shape over space, and are especially suited to high dimensions and complex data. By contrast, the simpler kernels used in prior kernel testing work are spatially homogeneous, and adaptive only in lengthscale. We explain how this scheme includes popular classifier-based two-sample tests as a special case, but improves on them in general. We provide the first proof of consistency for the proposed adaptation method, which applies both to kernels on deep features and to simpler radial basis kernels or multiple kernel learning. In experiments, we establish the superior performance of our deep kernels in hypothesis testing on benchmark and real-world data. The code of our deep-kernel-based two sample tests is available at https://github.com/fengliu90/DK-for-TST.

Viaarxiv icon