Alert button
Picture for Chenguang Wang

Chenguang Wang

Alert button

Agent Instructs Large Language Models to be General Zero-Shot Reasoners

Oct 05, 2023
Nicholas Crispino, Kyle Montgomery, Fankun Zeng, Dawn Song, Chenguang Wang

We introduce a method to improve the zero-shot reasoning abilities of large language models on general language understanding tasks. Specifically, we build an autonomous agent to instruct the reasoning process of large language models. We show this approach further unleashes the zero-shot reasoning abilities of large language models to more tasks. We study the performance of our method on a wide set of datasets spanning generation, classification, and reasoning. We show that our method generalizes to most tasks and obtains state-of-the-art zero-shot performance on 20 of the 29 datasets that we evaluate. For instance, our method boosts the performance of state-of-the-art large language models by a large margin, including Vicuna-13b (13.3%), Llama-2-70b-chat (23.2%), and GPT-3.5 Turbo (17.0%). Compared to zero-shot chain of thought, our improvement in reasoning is striking, with an average increase of 10.5%. With our method, Llama-2-70b-chat outperforms zero-shot GPT-3.5 Turbo by 10.2%.

Viaarxiv icon

Causality-informed Rapid Post-hurricane Building Damage Detection in Large Scale from InSAR Imagery

Oct 02, 2023
Chenguang Wang, Yepeng Liu, Xiaojian Zhang, Xuechun Li, Vladimir Paramygin, Arthriya Subgranon, Peter Sheng, Xilei Zhao, Susu Xu

Timely and accurate assessment of hurricane-induced building damage is crucial for effective post-hurricane response and recovery efforts. Recently, remote sensing technologies provide large-scale optical or Interferometric Synthetic Aperture Radar (InSAR) imagery data immediately after a disastrous event, which can be readily used to conduct rapid building damage assessment. Compared to optical satellite imageries, the Synthetic Aperture Radar can penetrate cloud cover and provide more complete spatial coverage of damaged zones in various weather conditions. However, these InSAR imageries often contain highly noisy and mixed signals induced by co-occurring or co-located building damage, flood, flood/wind-induced vegetation changes, as well as anthropogenic activities, making it challenging to extract accurate building damage information. In this paper, we introduced an approach for rapid post-hurricane building damage detection from InSAR imagery. This approach encoded complex causal dependencies among wind, flood, building damage, and InSAR imagery using a holistic causal Bayesian network. Based on the causal Bayesian network, we further jointly inferred the large-scale unobserved building damage by fusing the information from InSAR imagery with prior physical models of flood and wind, without the need for ground truth labels. Furthermore, we validated our estimation results in a real-world devastating hurricane -- the 2022 Hurricane Ian. We gathered and annotated building damage ground truth data in Lee County, Florida, and compared the introduced method's estimation results with the ground truth and benchmarked it against state-of-the-art models to assess the effectiveness of our proposed method. Results show that our method achieves rapid and accurate detection of building damage, with significantly reduced processing time compared to traditional manual inspection methods.

* 6 pages, 3 figures 
Viaarxiv icon

Practical Membership Inference Attacks Against Large-Scale Multi-Modal Models: A Pilot Study

Sep 29, 2023
Myeongseob Ko, Ming Jin, Chenguang Wang, Ruoxi Jia

Figure 1 for Practical Membership Inference Attacks Against Large-Scale Multi-Modal Models: A Pilot Study
Figure 2 for Practical Membership Inference Attacks Against Large-Scale Multi-Modal Models: A Pilot Study
Figure 3 for Practical Membership Inference Attacks Against Large-Scale Multi-Modal Models: A Pilot Study
Figure 4 for Practical Membership Inference Attacks Against Large-Scale Multi-Modal Models: A Pilot Study

Membership inference attacks (MIAs) aim to infer whether a data point has been used to train a machine learning model. These attacks can be employed to identify potential privacy vulnerabilities and detect unauthorized use of personal data. While MIAs have been traditionally studied for simple classification models, recent advancements in multi-modal pre-training, such as CLIP, have demonstrated remarkable zero-shot performance across a range of computer vision tasks. However, the sheer scale of data and models presents significant computational challenges for performing the attacks. This paper takes a first step towards developing practical MIAs against large-scale multi-modal models. We introduce a simple baseline strategy by thresholding the cosine similarity between text and image features of a target point and propose further enhancing the baseline by aggregating cosine similarity across transformations of the target. We also present a new weakly supervised attack method that leverages ground-truth non-members (e.g., obtained by using the publication date of a target model and the timestamps of the open data) to further enhance the attack. Our evaluation shows that CLIP models are susceptible to our attack strategies, with our simple baseline achieving over $75\%$ membership identification accuracy. Furthermore, our enhanced attacks outperform the baseline across multiple models and datasets, with the weakly supervised attack demonstrating an average-case performance improvement of $17\%$ and being at least $7$X more effective at low false-positive rates. These findings highlight the importance of protecting the privacy of multi-modal foundational models, which were previously assumed to be less susceptible to MIAs due to less overfitting. Our code is available at https://github.com/ruoxi-jia-group/CLIP-MIA.

* International Conference on Computer Vision (ICCV) 2023 
Viaarxiv icon

Efficient Training of Multi-task Neural Solver with Multi-armed Bandits

May 10, 2023
Chenguang Wang, Tianshu Yu

Figure 1 for Efficient Training of Multi-task Neural Solver with Multi-armed Bandits
Figure 2 for Efficient Training of Multi-task Neural Solver with Multi-armed Bandits
Figure 3 for Efficient Training of Multi-task Neural Solver with Multi-armed Bandits
Figure 4 for Efficient Training of Multi-task Neural Solver with Multi-armed Bandits

Efficiently training a multi-task neural solver for various combinatorial optimization problems (COPs) has been less studied so far. In this paper, we propose a general and efficient training paradigm based on multi-armed bandits to deliver a unified multi-task neural solver. To this end, we resort to the theoretical loss decomposition for multiple tasks under an encoder-decoder framework, which enables more efficient training via proper bandit task-sampling algorithms through an intra-task influence matrix. Our method achieves much higher overall performance with either limited training budgets or the same training epochs, compared to standard training schedules, which can be promising for advising efficient training of other multi-task large models. Additionally, the influence matrix can provide empirical evidence of some common practices in the area of learning to optimize, which in turn supports the validity of our approach.

Viaarxiv icon

Targeted Analysis of High-Risk States Using an Oriented Variational Autoencoder

Mar 20, 2023
Chenguang Wang, Ensieh Sharifnia, Simon H. Tindemans, Peter Palensky

Figure 1 for Targeted Analysis of High-Risk States Using an Oriented Variational Autoencoder
Figure 2 for Targeted Analysis of High-Risk States Using an Oriented Variational Autoencoder
Figure 3 for Targeted Analysis of High-Risk States Using an Oriented Variational Autoencoder
Figure 4 for Targeted Analysis of High-Risk States Using an Oriented Variational Autoencoder

Variational autoencoder (VAE) neural networks can be trained to generate power system states that capture both marginal distribution and multivariate dependencies of historical data. The coordinates of the latent space codes of VAEs have been shown to correlate with conceptual features of the data, which can be leveraged to synthesize targeted data with desired features. However, the locations of the VAEs' latent space codes that correspond to specific properties are not constrained. Additionally, the generation of data with specific characteristics may require data with corresponding hard-to-get labels fed into the generative model for training. In this paper, to make data generation more controllable and efficient, an oriented variation autoencoder (OVAE) is proposed to constrain the link between latent space code and generated data in the form of a Spearman correlation, which provides increased control over the data synthesis process. On this basis, an importance sampling process is used to sample data in the latent space. Two cases are considered for testing the performance of the OVAE model: the data set is fully labeled with approximate information and the data set is incompletely labeled but with more accurate information. The experimental results show that, in both cases, the OVAE model correlates latent space codes with the generated data, and the efficiency of generating targeted samples is significantly improved.

* 10 pages, 10 figures, 1 table 
Viaarxiv icon

ASP: Learn a Universal Neural Solver!

Mar 01, 2023
Chenguang Wang, Zhouliang Yu, Stephen McAleer, Tianshu Yu, Yaodong Yang

Figure 1 for ASP: Learn a Universal Neural Solver!
Figure 2 for ASP: Learn a Universal Neural Solver!
Figure 3 for ASP: Learn a Universal Neural Solver!
Figure 4 for ASP: Learn a Universal Neural Solver!

Applying machine learning to combinatorial optimization problems has the potential to improve both efficiency and accuracy. However, existing learning-based solvers often struggle with generalization when faced with changes in problem distributions and scales. In this paper, we propose a new approach called ASP: Adaptive Staircase Policy Space Response Oracle to address these generalization issues and learn a universal neural solver. ASP consists of two components: Distributional Exploration, which enhances the solver's ability to handle unknown distributions using Policy Space Response Oracles, and Persistent Scale Adaption, which improves scalability through curriculum learning. We have tested ASP on several challenging COPs, including the traveling salesman problem, the vehicle routing problem, and the prize collecting TSP, as well as the real-world instances from TSPLib and CVRPLib. Our results show that even with the same model size and weak training signal, ASP can help neural solvers explore and adapt to unseen distributions and varying scales, achieving superior performance. In particular, compared with the same neural solvers under a standard training pipeline, ASP produces a remarkable decrease in terms of the optimality gap with 90.9% and 47.43% on generated instances and real-world instances for TSP, and a decrease of 19% and 45.57% for CVRP.

Viaarxiv icon

Analysis of Graph Neural Networks with Theory of Markov Chains

Nov 12, 2022
Weichen Zhao, Chenguang Wang, Congying Han, Tiande Guo

Figure 1 for Analysis of Graph Neural Networks with Theory of Markov Chains
Figure 2 for Analysis of Graph Neural Networks with Theory of Markov Chains
Figure 3 for Analysis of Graph Neural Networks with Theory of Markov Chains
Figure 4 for Analysis of Graph Neural Networks with Theory of Markov Chains

In this paper, we provide a theoretical tool for the interpretation and analysis of \emph{graph neural networks} (GNNs). We use Markov chains on graphs to mathematically model the forward propagation processes of GNNs. The graph neural networks are divided into two classes of operator-consistent and operator-inconsistent based on whether the Markov chains are time-homogeneous. Based on this, we study \emph{over-smoothing} which is an important problem in GNN research. We attribute the over-smoothing problem to the convergence of an arbitrary initial distribution to a stationary distribution. We prove the effectiveness of the previous methods for alleviating the over-smoothing problem. Further, we give the conclusion that operator-consistent GNN cannot avoid over-smoothing at an exponential rate in the Markovian sense. For operator-inconsistent GNN, we theoretically give a sufficient condition for avoiding over-smoothing. Based on this condition, we propose a regularization term which can be flexibly added to the training of the neural network. Finally, we design experiments to verify the effectiveness of this condition. Results show that our proposed sufficient condition not only improves the performance but also alleviates the over-smoothing phenomenon.

Viaarxiv icon

Benchmarking Language Models for Code Syntax Understanding

Oct 26, 2022
Da Shen, Xinyun Chen, Chenguang Wang, Koushik Sen, Dawn Song

Figure 1 for Benchmarking Language Models for Code Syntax Understanding
Figure 2 for Benchmarking Language Models for Code Syntax Understanding
Figure 3 for Benchmarking Language Models for Code Syntax Understanding
Figure 4 for Benchmarking Language Models for Code Syntax Understanding

Pre-trained language models have demonstrated impressive performance in both natural language processing and program understanding, which represent the input as a token sequence without explicitly modeling its structure. Some prior works show that pre-trained language models can capture the syntactic rules of natural languages without finetuning on syntax understanding tasks. However, there is limited understanding of how well pre-trained models understand the code structure so far. In this work, we perform the first thorough benchmarking of the state-of-the-art pre-trained models for identifying the syntactic structures of programs. Specifically, we introduce CodeSyntax, a large-scale dataset of programs annotated with the syntactic relationships in their corresponding abstract syntax trees. Our key observation is that existing language models pretrained on code still lack the understanding of code syntax. In fact, these pre-trained programming language models fail to match the performance of simple baselines based on positional offsets and keywords. We also present a natural language benchmark to highlight the differences between natural languages and programming languages in terms of syntactic structure understanding. Our findings point out key limitations of existing pre-training methods for programming languages, and suggest the importance of modeling code syntactic structures.

* Findings of EMNLP 2022 
Viaarxiv icon

IELM: An Open Information Extraction Benchmark for Pre-Trained Language Models

Oct 25, 2022
Chenguang Wang, Xiao Liu, Dawn Song

Figure 1 for IELM: An Open Information Extraction Benchmark for Pre-Trained Language Models
Figure 2 for IELM: An Open Information Extraction Benchmark for Pre-Trained Language Models
Figure 3 for IELM: An Open Information Extraction Benchmark for Pre-Trained Language Models
Figure 4 for IELM: An Open Information Extraction Benchmark for Pre-Trained Language Models

We introduce a new open information extraction (OIE) benchmark for pre-trained language models (LM). Recent studies have demonstrated that pre-trained LMs, such as BERT and GPT, may store linguistic and relational knowledge. In particular, LMs are able to answer ``fill-in-the-blank'' questions when given a pre-defined relation category. Instead of focusing on pre-defined relations, we create an OIE benchmark aiming to fully examine the open relational information present in the pre-trained LMs. We accomplish this by turning pre-trained LMs into zero-shot OIE systems. Surprisingly, pre-trained LMs are able to obtain competitive performance on both standard OIE datasets (CaRB and Re-OIE2016) and two new large-scale factual OIE datasets (TAC KBP-OIE and Wikidata-OIE) that we establish via distant supervision. For instance, the zero-shot pre-trained LMs outperform the F1 score of the state-of-the-art supervised OIE methods on our factual OIE datasets without needing to use any training sets. Our code and datasets are available at https://github.com/cgraywang/IELM

* EMNLP 2022. arXiv admin note: substantial text overlap with arXiv:2010.11967 
Viaarxiv icon

PALT: Parameter-Lite Transfer of Language Models for Knowledge Graph Completion

Oct 25, 2022
Jianhao Shen, Chenguang Wang, Ye Yuan, Jiawei Han, Heng Ji, Koushik Sen, Ming Zhang, Dawn Song

Figure 1 for PALT: Parameter-Lite Transfer of Language Models for Knowledge Graph Completion
Figure 2 for PALT: Parameter-Lite Transfer of Language Models for Knowledge Graph Completion
Figure 3 for PALT: Parameter-Lite Transfer of Language Models for Knowledge Graph Completion
Figure 4 for PALT: Parameter-Lite Transfer of Language Models for Knowledge Graph Completion

This paper presents a parameter-lite transfer learning approach of pretrained language models (LM) for knowledge graph (KG) completion. Instead of finetuning, which modifies all LM parameters, we only tune a few new parameters while keeping the original LM parameters fixed. We establish this via reformulating KG completion as a "fill-in-the-blank" task, and introducing a parameter-lite encoder on top of the original LMs. We show that, by tuning far fewer parameters than finetuning, LMs transfer non-trivially to most tasks and reach competitiveness with prior state-of-the-art approaches. For instance, we outperform the fully finetuning approaches on a KG completion benchmark by tuning only 1% of the parameters. The code and datasets are available at \url{https://github.com/yuanyehome/PALT}.

* Findings of EMNLP 2022 
Viaarxiv icon