Alert button
Picture for Wai Lam

Wai Lam

Alert button

QuaSE: Sequence Editing under Quantifiable Guidance

Aug 31, 2018
Yi Liao, Lidong Bing, Piji Li, Shuming Shi, Wai Lam, Tong Zhang

Figure 1 for QuaSE: Sequence Editing under Quantifiable Guidance
Figure 2 for QuaSE: Sequence Editing under Quantifiable Guidance
Figure 3 for QuaSE: Sequence Editing under Quantifiable Guidance
Figure 4 for QuaSE: Sequence Editing under Quantifiable Guidance

We propose the task of Quantifiable Sequence Editing (QuaSE): editing an input sequence to generate an output sequence that satisfies a given numerical outcome value measuring a certain property of the sequence, with the requirement of keeping the main content of the input sequence. For example, an input sequence could be a word sequence, such as review sentence and advertisement text. For a review sentence, the outcome could be the review rating; for an advertisement, the outcome could be the click-through rate. The major challenge in performing QuaSE is how to perceive the outcome-related wordings, and only edit them to change the outcome. In this paper, the proposed framework contains two latent factors, namely, outcome factor and content factor, disentangled from the input sentence to allow convenient editing to change the outcome and keep the content. Our framework explores the pseudo-parallel sentences by modeling their content similarity and outcome differences to enable a better disentanglement of the latent factors, which allows generating an output to better satisfy the desired outcome and keep the content. The dual reconstruction structure further enhances the capability of generating expected output by exploiting the couplings of latent factors of pseudo-parallel sentences. For evaluation, we prepared a dataset of Yelp review sentences with the ratings as outcome. Extensive experimental results are reported and discussed to elaborate the peculiarities of our framework.

* accepted by emnlp2018 
Viaarxiv icon

Actor-Critic based Training Framework for Abstractive Summarization

Aug 15, 2018
Piji Li, Lidong Bing, Wai Lam

Figure 1 for Actor-Critic based Training Framework for Abstractive Summarization
Figure 2 for Actor-Critic based Training Framework for Abstractive Summarization
Figure 3 for Actor-Critic based Training Framework for Abstractive Summarization
Figure 4 for Actor-Critic based Training Framework for Abstractive Summarization

We present a training framework for neural abstractive summarization based on actor-critic approaches from reinforcement learning. In the traditional neural network based methods, the objective is only to maximize the likelihood of the predicted summaries, no other assessment constraints are considered, which may generate low-quality summaries or even incorrect sentences. To alleviate this problem, we employ an actor-critic framework to enhance the training procedure. For the actor, we employ the typical attention based sequence-to-sequence (seq2seq) framework as the policy network for summary generation. For the critic, we combine the maximum likelihood estimator with a well designed global summary quality estimator which is a neural network based binary classifier aiming to make the generated summaries indistinguishable from the human-written ones. Policy gradient method is used to conduct the parameter learning. An alternating training strategy is proposed to conduct the joint training of the actor and critic models. Extensive experiments on some benchmark datasets in different languages show that our framework achieves improvements over the state-of-the-art methods.

* 10 pages. arXiv admin note: text overlap with arXiv:1708.00625 
Viaarxiv icon

Learning Domain-Sensitive and Sentiment-Aware Word Embeddings

May 10, 2018
Bei Shi, Zihao Fu, Lidong Bing, Wai Lam

Figure 1 for Learning Domain-Sensitive and Sentiment-Aware Word Embeddings
Figure 2 for Learning Domain-Sensitive and Sentiment-Aware Word Embeddings
Figure 3 for Learning Domain-Sensitive and Sentiment-Aware Word Embeddings
Figure 4 for Learning Domain-Sensitive and Sentiment-Aware Word Embeddings

Word embeddings have been widely used in sentiment classification because of their efficacy for semantic representations of words. Given reviews from different domains, some existing methods for word embeddings exploit sentiment information, but they cannot produce domain-sensitive embeddings. On the other hand, some other existing methods can generate domain-sensitive word embeddings, but they cannot distinguish words with similar contexts but opposite sentiment polarity. We propose a new method for learning domain-sensitive and sentiment-aware embeddings that simultaneously capture the information of sentiment semantics and domain sensitivity of individual words. Our method can automatically determine and produce domain-common embeddings and domain-specific embeddings. The differentiation of domain-common and domain-specific words enables the advantage of data augmentation of common semantics from multiple domains and capture the varied semantics of specific words from different domains at the same time. Experimental results show that our model provides an effective way to learn domain-sensitive and sentiment-aware word embeddings which benefit sentiment classification at both sentence level and lexicon term level.

* ACL2018  
* 11 pages, published in ACL2018 
Viaarxiv icon

Transformation Networks for Target-Oriented Sentiment Classification

May 03, 2018
Xin Li, Lidong Bing, Wai Lam, Bei Shi

Figure 1 for Transformation Networks for Target-Oriented Sentiment Classification
Figure 2 for Transformation Networks for Target-Oriented Sentiment Classification
Figure 3 for Transformation Networks for Target-Oriented Sentiment Classification
Figure 4 for Transformation Networks for Target-Oriented Sentiment Classification

Target-oriented sentiment classification aims at classifying sentiment polarities over individual opinion targets in a sentence. RNN with attention seems a good fit for the characteristics of this task, and indeed it achieves the state-of-the-art performance. After re-examining the drawbacks of attention mechanism and the obstacles that block CNN to perform well in this classification task, we propose a new model to overcome these issues. Instead of attention, our model employs a CNN layer to extract salient features from the transformed word representations originated from a bi-directional RNN layer. Between the two layers, we propose a component to generate target-specific representations of words in the sentence, meanwhile incorporate a mechanism for preserving the original contextual information from the RNN layer. Experiments show that our model achieves a new state-of-the-art performance on a few benchmarks.

* ACL 2018 
Viaarxiv icon

Aspect Term Extraction with History Attention and Selective Transformation

May 02, 2018
Xin Li, Lidong Bing, Piji Li, Wai Lam, Zhimou Yang

Figure 1 for Aspect Term Extraction with History Attention and Selective Transformation
Figure 2 for Aspect Term Extraction with History Attention and Selective Transformation
Figure 3 for Aspect Term Extraction with History Attention and Selective Transformation
Figure 4 for Aspect Term Extraction with History Attention and Selective Transformation

Aspect Term Extraction (ATE), a key sub-task in Aspect-Based Sentiment Analysis, aims to extract explicit aspect expressions from online user reviews. We present a new framework for tackling ATE. It can exploit two useful clues, namely opinion summary and aspect detection history. Opinion summary is distilled from the whole input sentence, conditioned on each current token for aspect prediction, and thus the tailor-made summary can help aspect prediction on this token. Another clue is the information of aspect detection history, and it is distilled from the previous aspect predictions so as to leverage the coordinate structure and tagging schema constraints to upgrade the aspect prediction. Experimental results over four benchmark datasets clearly demonstrate that our framework can outperform all state-of-the-art methods.

* IJCAI 2018 
Viaarxiv icon

Reader-Aware Multi-Document Summarization: An Enhanced Model and The First Dataset

Aug 03, 2017
Piji Li, Lidong Bing, Wai Lam

Figure 1 for Reader-Aware Multi-Document Summarization: An Enhanced Model and The First Dataset
Figure 2 for Reader-Aware Multi-Document Summarization: An Enhanced Model and The First Dataset
Figure 3 for Reader-Aware Multi-Document Summarization: An Enhanced Model and The First Dataset
Figure 4 for Reader-Aware Multi-Document Summarization: An Enhanced Model and The First Dataset

We investigate the problem of reader-aware multi-document summarization (RA-MDS) and introduce a new dataset for this problem. To tackle RA-MDS, we extend a variational auto-encodes (VAEs) based MDS framework by jointly considering news documents and reader comments. To conduct evaluation for summarization performance, we prepare a new dataset. We describe the methods for data collection, aspect annotation, and summary writing as well as scrutinizing by experts. Experimental results show that reader comments can improve the summarization performance, which also demonstrates the usefulness of the proposed dataset. The annotated dataset for RA-MDS is available online.

* EMNLP 2017 Workshop on New Frontiers in Summarization; Dataset: http://www.se.cuhk.edu.hk/~textmine/dataset/ra-mds/ 
Viaarxiv icon

Deep Recurrent Generative Decoder for Abstractive Text Summarization

Aug 02, 2017
Piji Li, Wai Lam, Lidong Bing, Zihao Wang

Figure 1 for Deep Recurrent Generative Decoder for Abstractive Text Summarization
Figure 2 for Deep Recurrent Generative Decoder for Abstractive Text Summarization
Figure 3 for Deep Recurrent Generative Decoder for Abstractive Text Summarization
Figure 4 for Deep Recurrent Generative Decoder for Abstractive Text Summarization

We propose a new framework for abstractive text summarization based on a sequence-to-sequence oriented encoder-decoder model equipped with a deep recurrent generative decoder (DRGN). Latent structure information implied in the target summaries is learned based on a recurrent latent random model for improving the summarization quality. Neural variational inference is employed to address the intractable posterior inference for the recurrent latent variables. Abstractive summaries are generated based on both the generative latent variables and the discriminative deterministic states. Extensive experiments on some benchmark datasets in different languages show that DRGN achieves improvements over the state-of-the-art methods.

* 10 pages, EMNLP 2017 
Viaarxiv icon

Neural Rating Regression with Abstractive Tips Generation for Recommendation

Aug 01, 2017
Piji Li, Zihao Wang, Zhaochun Ren, Lidong Bing, Wai Lam

Figure 1 for Neural Rating Regression with Abstractive Tips Generation for Recommendation
Figure 2 for Neural Rating Regression with Abstractive Tips Generation for Recommendation
Figure 3 for Neural Rating Regression with Abstractive Tips Generation for Recommendation
Figure 4 for Neural Rating Regression with Abstractive Tips Generation for Recommendation

Recently, some E-commerce sites launch a new interaction box called Tips on their mobile apps. Users can express their experience and feelings or provide suggestions using short texts typically several words or one sentence. In essence, writing some tips and giving a numerical rating are two facets of a user's product assessment action, expressing the user experience and feelings. Jointly modeling these two facets is helpful for designing a better recommendation system. While some existing models integrate text information such as item specifications or user reviews into user and item latent factors for improving the rating prediction, no existing works consider tips for improving recommendation quality. We propose a deep learning based framework named NRT which can simultaneously predict precise ratings and generate abstractive tips with good linguistic quality simulating user experience and feelings. For abstractive tips generation, gated recurrent neural networks are employed to "translate" user and item latent representations into a concise sentence. Extensive experiments on benchmark datasets from different domains show that NRT achieves significant improvements over the state-of-the-art methods. Moreover, the generated tips can vividly predict the user experience and feelings.

* SIGIR 2017 
Viaarxiv icon

Jointly Learning Word Embeddings and Latent Topics

Jun 21, 2017
Bei Shi, Wai Lam, Shoaib Jameel, Steven Schockaert, Kwun Ping Lai

Figure 1 for Jointly Learning Word Embeddings and Latent Topics
Figure 2 for Jointly Learning Word Embeddings and Latent Topics
Figure 3 for Jointly Learning Word Embeddings and Latent Topics
Figure 4 for Jointly Learning Word Embeddings and Latent Topics

Word embedding models such as Skip-gram learn a vector-space representation for each word, based on the local word collocation patterns that are observed in a text corpus. Latent topic models, on the other hand, take a more global view, looking at the word distributions across the corpus to assign a topic to each word occurrence. These two paradigms are complementary in how they represent the meaning of word occurrences. While some previous works have already looked at using word embeddings for improving the quality of latent topics, and conversely, at using latent topics for improving word embeddings, such "two-step" methods cannot capture the mutual interaction between the two paradigms. In this paper, we propose STE, a framework which can learn word embeddings and latent topics in a unified manner. STE naturally obtains topic-specific word embeddings, and thus addresses the issue of polysemy. At the same time, it also learns the term distributions of the topics, and the topic distributions of the documents. Our experimental results demonstrate that the STE model can indeed generate useful topic-specific word embeddings and coherent latent topics in an effective and efficient way.

* 10 pagess, 2 figures, full paper. To appear in the proceedings of The 40th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR '17) 
Viaarxiv icon

Abstractive Multi-Document Summarization via Phrase Selection and Merging

Jun 05, 2015
Lidong Bing, Piji Li, Yi Liao, Wai Lam, Weiwei Guo, Rebecca J. Passonneau

Figure 1 for Abstractive Multi-Document Summarization via Phrase Selection and Merging
Figure 2 for Abstractive Multi-Document Summarization via Phrase Selection and Merging
Figure 3 for Abstractive Multi-Document Summarization via Phrase Selection and Merging
Figure 4 for Abstractive Multi-Document Summarization via Phrase Selection and Merging

We propose an abstraction-based multi-document summarization framework that can construct new sentences by exploring more fine-grained syntactic units than sentences, namely, noun/verb phrases. Different from existing abstraction-based approaches, our method first constructs a pool of concepts and facts represented by phrases from the input documents. Then new sentences are generated by selecting and merging informative phrases to maximize the salience of phrases and meanwhile satisfy the sentence construction constraints. We employ integer linear optimization for conducting phrase selection and merging simultaneously in order to achieve the global optimal solution for a summary. Experimental results on the benchmark data set TAC 2011 show that our framework outperforms the state-of-the-art models under automated pyramid evaluation metric, and achieves reasonably well results on manual linguistic quality evaluation.

* 11 pages, 1 figure, accepted as a full paper at ACL 2015 
Viaarxiv icon