Alert button
Picture for Qingkai Zeng

Qingkai Zeng

Alert button

Auto-Instruct: Automatic Instruction Generation and Ranking for Black-Box Language Models

Oct 19, 2023
Zhihan Zhang, Shuohang Wang, Wenhao Yu, Yichong Xu, Dan Iter, Qingkai Zeng, Yang Liu, Chenguang Zhu, Meng Jiang

Figure 1 for Auto-Instruct: Automatic Instruction Generation and Ranking for Black-Box Language Models
Figure 2 for Auto-Instruct: Automatic Instruction Generation and Ranking for Black-Box Language Models
Figure 3 for Auto-Instruct: Automatic Instruction Generation and Ranking for Black-Box Language Models
Figure 4 for Auto-Instruct: Automatic Instruction Generation and Ranking for Black-Box Language Models

Large language models (LLMs) can perform a wide range of tasks by following natural language instructions, without the necessity of task-specific fine-tuning. Unfortunately, the performance of LLMs is greatly influenced by the quality of these instructions, and manually writing effective instructions for each task is a laborious and subjective process. In this paper, we introduce Auto-Instruct, a novel method to automatically improve the quality of instructions provided to LLMs. Our method leverages the inherent generative ability of LLMs to produce diverse candidate instructions for a given task, and then ranks them using a scoring model trained on a variety of 575 existing NLP tasks. In experiments on 118 out-of-domain tasks, Auto-Instruct surpasses both human-written instructions and existing baselines of LLM-generated instructions. Furthermore, our method exhibits notable generalizability even with other LLMs that are not incorporated into its training process.

* Accepted to EMNLP 2023 Findings. Work was done before July 2023 
Viaarxiv icon

MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning

Jul 16, 2023
Zhenwen Liang, Dian Yu, Xiaoman Pan, Wenlin Yao, Qingkai Zeng, Xiangliang Zhang, Dong Yu

Figure 1 for MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning
Figure 2 for MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning
Figure 3 for MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning
Figure 4 for MinT: Boosting Generalization in Mathematical Reasoning via Multi-View Fine-Tuning

Reasoning in mathematical domains remains a significant challenge for relatively small language models (LMs). Many current methods focus on specializing LMs in mathematical reasoning and rely heavily on knowledge distillation from powerful but inefficient large LMs (LLMs). In this work, we explore a new direction that avoids over-reliance on LLM teachers, introducing a multi-view fine-tuning method that efficiently exploits existing mathematical problem datasets with diverse annotation styles. Our approach uniquely considers the various annotation formats as different "views" and leverages them in training the model. By postpending distinct instructions to input questions, models can learn to generate solutions in diverse formats in a flexible manner. Experimental results show that our strategy enables a LLaMA-7B model to outperform prior approaches that utilize knowledge distillation, as well as carefully established baselines. Additionally, the proposed method grants the models promising generalization ability across various views and datasets, and the capability to learn from inaccurate or incomplete noisy data. We hope our multi-view training paradigm could inspire future studies in other machine reasoning domains.

Viaarxiv icon

Automatic Controllable Product Copywriting for E-Commerce

Jun 21, 2022
Xiaojie Guo, Qingkai Zeng, Meng Jiang, Yun Xiao, Bo Long, Lingfei Wu

Figure 1 for Automatic Controllable Product Copywriting for E-Commerce
Figure 2 for Automatic Controllable Product Copywriting for E-Commerce
Figure 3 for Automatic Controllable Product Copywriting for E-Commerce
Figure 4 for Automatic Controllable Product Copywriting for E-Commerce

Automatic product description generation for e-commerce has witnessed significant advancement in the past decade. Product copywriting aims to attract users' interest and improve user experience by highlighting product characteristics with textual descriptions. As the services provided by e-commerce platforms become diverse, it is necessary to adapt the patterns of automatically-generated descriptions dynamically. In this paper, we report our experience in deploying an E-commerce Prefix-based Controllable Copywriting Generation (EPCCG) system into the JD.com e-commerce product recommendation platform. The development of the system contains two main components: 1) copywriting aspect extraction; 2) weakly supervised aspect labeling; 3) text generation with a prefix-based language model; 4) copywriting quality control. We conduct experiments to validate the effectiveness of the proposed EPCCG. In addition, we introduce the deployed architecture which cooperates with the EPCCG into the real-time JD.com e-commerce recommendation platform and the significant payoff since deployment.

* This paper has been accepted by KDD 2022 ADS 
Viaarxiv icon

Enhancing Taxonomy Completion with Concept Generation via Fusing Relational Representations

Jun 05, 2021
Qingkai Zeng, Jinfeng Lin, Wenhao Yu, Jane Cleland-Huang, Meng Jiang

Figure 1 for Enhancing Taxonomy Completion with Concept Generation via Fusing Relational Representations
Figure 2 for Enhancing Taxonomy Completion with Concept Generation via Fusing Relational Representations
Figure 3 for Enhancing Taxonomy Completion with Concept Generation via Fusing Relational Representations
Figure 4 for Enhancing Taxonomy Completion with Concept Generation via Fusing Relational Representations

Automatic construction of a taxonomy supports many applications in e-commerce, web search, and question answering. Existing taxonomy expansion or completion methods assume that new concepts have been accurately extracted and their embedding vectors learned from the text corpus. However, one critical and fundamental challenge in fixing the incompleteness of taxonomies is the incompleteness of the extracted concepts, especially for those whose names have multiple words and consequently low frequency in the corpus. To resolve the limitations of extraction-based methods, we propose GenTaxo to enhance taxonomy completion by identifying positions in existing taxonomies that need new concepts and then generating appropriate concept names. Instead of relying on the corpus for concept embeddings, GenTaxo learns the contextual embeddings from their surrounding graph-based and language-based relational information, and leverages the corpus for pre-training a concept name generator. Experimental results demonstrate that GenTaxo improves the completeness of taxonomies over existing methods.

Viaarxiv icon

Validating Label Consistency in NER Data Annotation

Jan 21, 2021
Qingkai Zeng, Mengxia Yu, Wenhao Yu, Tianwen Jiang, Tim Weninger, Meng Jiang

Figure 1 for Validating Label Consistency in NER Data Annotation
Figure 2 for Validating Label Consistency in NER Data Annotation
Figure 3 for Validating Label Consistency in NER Data Annotation
Figure 4 for Validating Label Consistency in NER Data Annotation

Data annotation plays a crucial role in ensuring your named entity recognition (NER) projects are trained with the right information to learn from. Producing the most accurate labels is a challenge due to the complexity involved with annotation. Label inconsistency between multiple subsets of data annotation (e.g., training set and test set, or multiple training subsets) is an indicator of label mistakes. In this work, we present an empirical method to explore the relationship between label (in-)consistency and NER model performance. It can be used to validate the label consistency (or catches the inconsistency) in multiple sets of NER data annotation. In experiments, our method identified the label inconsistency of test data in SCIERC and CoNLL03 datasets (with 26.7% and 5.4% label mistakes). It validated the consistency in the corrected version of both datasets.

Viaarxiv icon

Technical Question Answering across Tasks and Domains

Oct 19, 2020
Wenhao Yu, Lingfei Wu, Yu Deng, Qingkai Zeng, Ruchi Mahindru, Sinem Guven, Meng Jiang

Figure 1 for Technical Question Answering across Tasks and Domains
Figure 2 for Technical Question Answering across Tasks and Domains
Figure 3 for Technical Question Answering across Tasks and Domains
Figure 4 for Technical Question Answering across Tasks and Domains

Building automatic technical support system is an important yet challenge task. Conceptually, to answer a user question on a technical forum, a human expert has to first retrieve relevant documents, and then read them carefully to identify the answer snippet. Despite huge success the researchers have achieved in coping with general domain question answering (QA), much less attentions have been paid for investigating technical QA. Specifically, existing methods suffer from several unique challenges (i) the question and answer rarely overlaps substantially and (ii) very limited data size. In this paper, we propose a novel framework of deep transfer learning to effectively address technical QA across tasks and domains. To this end, we present an adjustable joint learning approach for document retrieval and reading comprehension tasks. Our experiments on the TechQA demonstrates superior performance compared with state-of-the-art methods.

* 10 pages, 6 figures 
Viaarxiv icon

Crossing Variational Autoencoders for Answer Retrieval

May 06, 2020
Wenhao Yu, Lingfei Wu, Qingkai Zeng, Yu Deng, Shu Tao, Meng Jiang

Figure 1 for Crossing Variational Autoencoders for Answer Retrieval
Figure 2 for Crossing Variational Autoencoders for Answer Retrieval
Figure 3 for Crossing Variational Autoencoders for Answer Retrieval
Figure 4 for Crossing Variational Autoencoders for Answer Retrieval

Answer retrieval is to find the most aligned answer from a large set of candidates given a question. Learning vector representations of questions/answers is the key factor. Question-answer alignment and question/answer semantics are two important signals for learning the representations. Existing methods learned semantic representations with dual encoders or dual variational auto-encoders. The semantic information was learned from language models or question-to-question (answer-to-answer) generative processes. However, the alignment and semantics were too separate to capture the aligned semantics between question and answer. In this work, we propose to cross variational auto-encoders by generating questions with aligned answers and generating answers with aligned questions. Experiments show that our method outperforms the state-of-the-art answer retrieval method on SQuAD.

* Accepted to ACL 2020 
Viaarxiv icon

Boosting Factual Correctness of Abstractive Summarization

Apr 04, 2020
Chenguang Zhu, William Hinthorn, Ruochen Xu, Qingkai Zeng, Michael Zeng, Xuedong Huang, Meng Jiang

Figure 1 for Boosting Factual Correctness of Abstractive Summarization
Figure 2 for Boosting Factual Correctness of Abstractive Summarization
Figure 3 for Boosting Factual Correctness of Abstractive Summarization
Figure 4 for Boosting Factual Correctness of Abstractive Summarization

A commonly observed problem with abstractive summarization is the distortion or fabrication of factual information in the article. This inconsistency between summary and original text has led to various concerns over its applicability. In this paper, we firstly propose a Fact-Aware Summarization model, FASum, which extracts factual relations from the article and integrates this knowledge into the decoding process via neural graph computation. Then, we propose a Factual Corrector model, FC, that can modify abstractive summaries generated by any model to improve factual correctness. Empirical results show that FASum generates summaries with significantly higher factual correctness compared with state-of-the-art abstractive summarization systems, both under an independently trained factual correctness evaluator and human evaluation. And FC improves the factual correctness of summaries generated by various models via only modifying several entity tokens.

* 17 pages, 3 figures 
Viaarxiv icon

Boosting Factual Correctness of Abstractive Summarization with Knowledge Graph

Mar 26, 2020
Chenguang Zhu, William Hinthorn, Ruochen Xu, Qingkai Zeng, Michael Zeng, Xuedong Huang, Meng Jiang

Figure 1 for Boosting Factual Correctness of Abstractive Summarization with Knowledge Graph
Figure 2 for Boosting Factual Correctness of Abstractive Summarization with Knowledge Graph
Figure 3 for Boosting Factual Correctness of Abstractive Summarization with Knowledge Graph
Figure 4 for Boosting Factual Correctness of Abstractive Summarization with Knowledge Graph

A commonly observed problem with abstractive summarization is the distortion or fabrication of factual information in the article. This inconsistency between summary and original text has led to various concerns over its applicability. In this paper, we propose to boost factual correctness of summaries via the fusion of knowledge, i.e. extracted factual relations from the article. We present a Fact-Aware Summarization model, FASum. In this model, the knowledge information can be organically integrated into the summary generation process via neural graph computation and effectively improves the factual correctness. Empirical results show that FASum generates summaries with significantly higher factual correctness compared with state-of-the-art abstractive summarization systems, both under an independently trained factual correctness evaluator and human evaluation. For example, in CNN/DailyMail dataset, FASum obtains 1.2% higher fact correctness scores than UniLM and 4.5% higher than BottomUp.

* 14 pages, 2 figures 
Viaarxiv icon