Alert button
Picture for Ke Tang

Ke Tang

Alert button

Enhancing Graph Collaborative Filtering via Uniformly Co-Clustered Intent Modeling

Sep 22, 2023
Jiahao Wu, Wenqi Fan, Shengcai Liu, Qijiong Liu, Qing Li, Ke Tang

Graph-based collaborative filtering has emerged as a powerful paradigm for delivering personalized recommendations. Despite their demonstrated effectiveness, these methods often neglect the underlying intents of users, which constitute a pivotal facet of comprehensive user interests. Consequently, a series of approaches have arisen to tackle this limitation by introducing independent intent representations. However, these approaches fail to capture the intricate relationships between intents of different users and the compatibility between user intents and item properties. To remedy the above issues, we propose a novel method, named uniformly co-clustered intent modeling. Specifically, we devise a uniformly contrastive intent modeling module to bring together the embeddings of users with similar intents and items with similar properties. This module aims to model the nuanced relations between intents of different users and properties of different items, especially those unreachable to each other on the user-item graph. To model the compatibility between user intents and item properties, we design the user-item co-clustering module, maximizing the mutual information of co-clusters of users and items. This approach is substantiated through theoretical validation, establishing its efficacy in modeling compatibility to enhance the mutual information between user and item representations. Comprehensive experiments on various real-world datasets verify the effectiveness of the proposed framework.

* In submission 
Viaarxiv icon

Neural Influence Estimator: Towards Real-time Solutions to Influence Blocking Maximization

Aug 27, 2023
Wenjie Chen, Shengcai Liu, Yew-Soon Ong, Ke Tang

Real-time solutions to the influence blocking maximization (IBM) problems are crucial for promptly containing the spread of misinformation. However, achieving this goal is non-trivial, mainly because assessing the blocked influence of an IBM problem solution typically requires plenty of expensive Monte Carlo simulations (MCSs). Although several approaches have been proposed to enhance efficiency, they still fail to achieve real-time solutions to IBM problems of practical scales. This work presents a novel approach that enables solving IBM problems with hundreds of thousands of nodes and edges in seconds. The key idea is to construct a fast-to-evaluate surrogate model, called neural influence estimator (NIE), as a substitute for the time-intensive MCSs. To this end, a learning problem is formulated to build the NIE that takes the false-and-true information instance as input, extracts features describing the topology and inter-relationship between two seed sets, and predicts the blocked influence. A well-trained NIE can generalize across different IBM problems defined on a social network, and can be readily combined with existing IBM optimization algorithms such as the greedy algorithm. The experiments on 25 IBM problems with up to millions of edges show that the NIE-based optimization method can be up to four orders of magnitude faster than MCSs-based optimization method to achieve the same solution quality. Moreover, given a real-time constraint of one minute, the NIE-based method can solve IBM problems with up to hundreds of thousands of nodes, which is at least one order of magnitude larger than what can be solved by existing methods.

Viaarxiv icon

Data-Driven Chance-Constrained Multiple-Choice Knapsack Problem: Model, Algorithms, and Applications

Jun 26, 2023
Xuanfeng Li, Shengcai Liu, Jin Wang, Xiao Chen, Yew-Soon Ong, Ke Tang

Figure 1 for Data-Driven Chance-Constrained Multiple-Choice Knapsack Problem: Model, Algorithms, and Applications
Figure 2 for Data-Driven Chance-Constrained Multiple-Choice Knapsack Problem: Model, Algorithms, and Applications
Figure 3 for Data-Driven Chance-Constrained Multiple-Choice Knapsack Problem: Model, Algorithms, and Applications
Figure 4 for Data-Driven Chance-Constrained Multiple-Choice Knapsack Problem: Model, Algorithms, and Applications

The multiple-choice knapsack problem (MCKP) is a classic NP-hard combinatorial optimization problem. Motivated by several significant practical applications, this work investigates a novel variant of MCKP called data-driven chance-constrained multiple-choice knapsack problem (DDCCMCKP), where the item weight is a random variable with unknown probability distribution. We first present the problem formulation of DDCCMCKP, and then establish two benchmark sets. The first set contains synthetic instances, and the second set is devised to simulate a real-world application scenario of a certain telecommunication company. To solve DDCCMCKP, we propose a data-driven adaptive local search (DDALS) algorithm. The main merit of DDALS lies in evaluating solutions with chance constraints by data-driven methods, under the condition of unknown distributions and only historical sample data being available. The experimental results demonstrate the effectiveness of the proposed algorithm and show that it is superior to other baselines. Additionally, ablation experiments confirm the necessity of each component in the algorithm. Our proposed algorithm can serve as the baseline for future research, and the code and benchmark sets will be open-sourced to further promote research on this challenging problem.

Viaarxiv icon

Statistical Tests for Replacing Human Decision Makers with Algorithms

Jun 20, 2023
Kai Feng, Han Hong, Ke Tang, Jingyuan Wang

Figure 1 for Statistical Tests for Replacing Human Decision Makers with Algorithms
Figure 2 for Statistical Tests for Replacing Human Decision Makers with Algorithms
Figure 3 for Statistical Tests for Replacing Human Decision Makers with Algorithms
Figure 4 for Statistical Tests for Replacing Human Decision Makers with Algorithms

This paper proposes a statistical framework with which artificial intelligence can improve human decision making. The performance of each human decision maker is first benchmarked against machine predictions; we then replace the decisions made by a subset of the decision makers with the recommendation from the proposed artificial intelligence algorithm. Using a large nationwide dataset of pregnancy outcomes and doctor diagnoses from prepregnancy checkups of reproductive age couples, we experimented with both a heuristic frequentist approach and a Bayesian posterior loss function approach with an application to abnormal birth detection. We find that our algorithm on a test dataset results in a higher overall true positive rate and a lower false positive rate than the diagnoses made by doctors only. We also find that the diagnoses of doctors from rural areas are more frequently replaceable, suggesting that artificial intelligence assisted decision making tends to improve precision more in less developed regions.

* 65 pages, 19 figures 
Viaarxiv icon

Perturbation-Based Two-Stage Multi-Domain Active Learning

Jun 19, 2023
Rui He, Zeyu Dai, Shan He, Ke Tang

Figure 1 for Perturbation-Based Two-Stage Multi-Domain Active Learning
Figure 2 for Perturbation-Based Two-Stage Multi-Domain Active Learning
Figure 3 for Perturbation-Based Two-Stage Multi-Domain Active Learning

In multi-domain learning (MDL) scenarios, high labeling effort is required due to the complexity of collecting data from various domains. Active Learning (AL) presents an encouraging solution to this issue by annotating a smaller number of highly informative instances, thereby reducing the labeling effort. Previous research has relied on conventional AL strategies for MDL scenarios, which underutilize the domain-shared information of each instance during the selection procedure. To mitigate this issue, we propose a novel perturbation-based two-stage multi-domain active learning (P2S-MDAL) method incorporated into the well-regarded ASP-MTL model. Specifically, P2S-MDAL involves allocating budgets for domains and establishing regions for diversity selection, which are further used to select the most cross-domain influential samples in each region. A perturbation metric has been introduced to evaluate the robustness of the shared feature extractor of the model, facilitating the identification of potentially cross-domain influential samples. Experiments are conducted on three real-world datasets, encompassing both texts and images. The superior performance over conventional AL strategies shows the effectiveness of the proposed strategy. Additionally, an ablation study has been carried out to demonstrate the validity of each component. Finally, we outline several intriguing potential directions for future MDAL research, thus catalyzing the field's advancement.

Viaarxiv icon

Large Language Models can be Guided to Evade AI-Generated Text Detection

May 19, 2023
Ning Lu, Shengcai Liu, Rui He, Qi Wang, Ke Tang

Figure 1 for Large Language Models can be Guided to Evade AI-Generated Text Detection
Figure 2 for Large Language Models can be Guided to Evade AI-Generated Text Detection
Figure 3 for Large Language Models can be Guided to Evade AI-Generated Text Detection
Figure 4 for Large Language Models can be Guided to Evade AI-Generated Text Detection

Large Language Models (LLMs) have demonstrated exceptional performance in a variety of tasks, including essay writing and question answering. However, it is crucial to address the potential misuse of these models, which can lead to detrimental outcomes such as plagiarism and spamming. Recently, several detectors have been proposed, including fine-tuned classifiers and various statistical methods. In this study, we reveal that with the aid of carefully crafted prompts, LLMs can effectively evade these detection systems. We propose a novel Substitution-based In-Context example Optimization method (SICO) to automatically generate such prompts. On three real-world tasks where LLMs can be misused, SICO successfully enables ChatGPT to evade six existing detectors, causing a significant 0.54 AUC drop on average. Surprisingly, in most cases these detectors perform even worse than random classifiers. These results firmly reveal the vulnerability of existing detectors. Finally, the strong performance of SICO suggests itself as a reliable evaluation protocol for any new detector in this field.

Viaarxiv icon

Multi-Domain Learning From Insufficient Annotations

May 04, 2023
Rui He, Shengcai Liu, Jiahao Wu, Shan He, Ke Tang

Figure 1 for Multi-Domain Learning From Insufficient Annotations
Figure 2 for Multi-Domain Learning From Insufficient Annotations
Figure 3 for Multi-Domain Learning From Insufficient Annotations
Figure 4 for Multi-Domain Learning From Insufficient Annotations

Multi-domain learning (MDL) refers to simultaneously constructing a model or a set of models on datasets collected from different domains. Conventional approaches emphasize domain-shared information extraction and domain-private information preservation, following the shared-private framework (SP models), which offers significant advantages over single-domain learning. However, the limited availability of annotated data in each domain considerably hinders the effectiveness of conventional supervised MDL approaches in real-world applications. In this paper, we introduce a novel method called multi-domain contrastive learning (MDCL) to alleviate the impact of insufficient annotations by capturing both semantic and structural information from both labeled and unlabeled data.Specifically, MDCL comprises two modules: inter-domain semantic alignment and intra-domain contrast. The former aims to align annotated instances of the same semantic category from distinct domains within a shared hidden space, while the latter focuses on learning a cluster structure of unlabeled instances in a private hidden space for each domain. MDCL is readily compatible with many SP models, requiring no additional model parameters and allowing for end-to-end training. Experimental results across five textual and image multi-domain datasets demonstrate that MDCL brings noticeable improvement over various SP models.Furthermore, MDCL can further be employed in multi-domain active learning (MDAL) to achieve a superior initialization, eventually leading to better overall performance.

* 8 pages, 11 figures 
Viaarxiv icon

MetaUE: Model-based Meta-learning for Underwater Image Enhancement

Mar 12, 2023
Zhenwei Zhang, Haorui Yan, Ke Tang, Yuping Duan

Figure 1 for MetaUE: Model-based Meta-learning for Underwater Image Enhancement
Figure 2 for MetaUE: Model-based Meta-learning for Underwater Image Enhancement
Figure 3 for MetaUE: Model-based Meta-learning for Underwater Image Enhancement
Figure 4 for MetaUE: Model-based Meta-learning for Underwater Image Enhancement

The challenges in recovering underwater images are the presence of diverse degradation factors and the lack of ground truth images. Although synthetic underwater image pairs can be used to overcome the problem of inadequately observing data, it may result in over-fitting and enhancement degradation. This paper proposes a model-based deep learning method for restoring clean images under various underwater scenarios, which exhibits good interpretability and generalization ability. More specifically, we build up a multi-variable convolutional neural network model to estimate the clean image, background light and transmission map, respectively. An efficient loss function is also designed to closely integrate the variables based on the underwater image model. The meta-learning strategy is used to obtain a pre-trained model on the synthetic underwater dataset, which contains different types of degradation to cover the various underwater environments. The pre-trained model is then fine-tuned on real underwater datasets to obtain a reliable underwater image enhancement model, called MetaUE. Numerical experiments demonstrate that the pre-trained model has good generalization ability, allowing it to remove the color degradation for various underwater attenuation images such as blue, green and yellow, etc. The fine-tuning makes the model able to adapt to different underwater datasets, the enhancement results of which outperform the state-of-the-art underwater image restoration methods. All our codes and data are available at \url{https://github.com/Duanlab123/MetaUE}.

Viaarxiv icon

Less is More: Understanding Word-level Textual Adversarial Attack via n-gram Frequency Descend

Feb 07, 2023
Ning Lu, Shengcai Liu, Zhirui Zhang, Qi Wang, Haifeng Liu, Ke Tang

Figure 1 for Less is More: Understanding Word-level Textual Adversarial Attack via n-gram Frequency Descend
Figure 2 for Less is More: Understanding Word-level Textual Adversarial Attack via n-gram Frequency Descend
Figure 3 for Less is More: Understanding Word-level Textual Adversarial Attack via n-gram Frequency Descend
Figure 4 for Less is More: Understanding Word-level Textual Adversarial Attack via n-gram Frequency Descend

Word-level textual adversarial attacks have achieved striking performance in fooling natural language processing models. However, the fundamental questions of why these attacks are effective, and the intrinsic properties of the adversarial examples (AEs), are still not well understood. This work attempts to interpret textual attacks through the lens of $n$-gram frequency. Specifically, it is revealed that existing word-level attacks exhibit a strong tendency toward generation of examples with $n$-gram frequency descend ($n$-FD). Intuitively, this finding suggests a natural way to improve model robustness by training the model on the $n$-FD examples. To verify this idea, we devise a model-agnostic and gradient-free AE generation approach that relies solely on the $n$-gram frequency information, and further integrate it into the recently proposed convex hull framework for adversarial training. Surprisingly, the resultant method performs quite similarly to the original gradient-based method in terms of model robustness. These findings provide a human-understandable perspective for interpreting word-level textual adversarial attacks, and a new direction to improve model robustness.

* 8 pages, 4 figures. In progress 
Viaarxiv icon

Enabling surrogate-assisted evolutionary reinforcement learning via policy embedding

Jan 31, 2023
Lan Tang, Xiaxi Li, Jinyuan Zhang, Guiying Li, Peng Yang, Ke Tang

Figure 1 for Enabling surrogate-assisted evolutionary reinforcement learning via policy embedding
Figure 2 for Enabling surrogate-assisted evolutionary reinforcement learning via policy embedding
Figure 3 for Enabling surrogate-assisted evolutionary reinforcement learning via policy embedding
Figure 4 for Enabling surrogate-assisted evolutionary reinforcement learning via policy embedding

Evolutionary Reinforcement Learning (ERL) that applying Evolutionary Algorithms (EAs) to optimize the weight parameters of Deep Neural Network (DNN) based policies has been widely regarded as an alternative to traditional reinforcement learning methods. However, the evaluation of the iteratively generated population usually requires a large amount of computational time and can be prohibitively expensive, which may potentially restrict the applicability of ERL. Surrogate is often used to reduce the computational burden of evaluation in EAs. Unfortunately, in ERL, each individual of policy usually represents millions of weights parameters of DNN. This high-dimensional representation of policy has introduced a great challenge to the application of surrogates into ERL to speed up training. This paper proposes a PE-SAERL Framework to at the first time enable surrogate-assisted evolutionary reinforcement learning via policy embedding (PE). Empirical results on 5 Atari games show that the proposed method can perform more efficiently than the four state-of-the-art algorithms. The training process is accelerated up to 7x on tested games, comparing to its counterpart without the surrogate and PE.

* This paper is submitted to bicta-2022 
Viaarxiv icon