Alert button
Picture for Kexin Yang

Kexin Yang

Alert button

MPPN: Multi-Resolution Periodic Pattern Network For Long-Term Time Series Forecasting

Jun 12, 2023
Xing Wang, Zhendong Wang, Kexin Yang, Junlan Feng, Zhiyan Song, Chao Deng, Lin zhu

Figure 1 for MPPN: Multi-Resolution Periodic Pattern Network For Long-Term Time Series Forecasting
Figure 2 for MPPN: Multi-Resolution Periodic Pattern Network For Long-Term Time Series Forecasting
Figure 3 for MPPN: Multi-Resolution Periodic Pattern Network For Long-Term Time Series Forecasting
Figure 4 for MPPN: Multi-Resolution Periodic Pattern Network For Long-Term Time Series Forecasting

Long-term time series forecasting plays an important role in various real-world scenarios. Recent deep learning methods for long-term series forecasting tend to capture the intricate patterns of time series by decomposition-based or sampling-based methods. However, most of the extracted patterns may include unpredictable noise and lack good interpretability. Moreover, the multivariate series forecasting methods usually ignore the individual characteristics of each variate, which may affecting the prediction accuracy. To capture the intrinsic patterns of time series, we propose a novel deep learning network architecture, named Multi-resolution Periodic Pattern Network (MPPN), for long-term series forecasting. We first construct context-aware multi-resolution semantic units of time series and employ multi-periodic pattern mining to capture the key patterns of time series. Then, we propose a channel adaptive module to capture the perceptions of multivariate towards different patterns. In addition, we present an entropy-based method for evaluating the predictability of time series and providing an upper bound on the prediction accuracy before forecasting. Our experimental evaluation on nine real-world benchmarks demonstrated that MPPN significantly outperforms the state-of-the-art Transformer-based, decomposition-based and sampling-based methods for long-term series forecasting.

* 21 pages 
Viaarxiv icon

Interactive Natural Language Processing

May 22, 2023
Zekun Wang, Ge Zhang, Kexin Yang, Ning Shi, Wangchunshu Zhou, Shaochun Hao, Guangzheng Xiong, Yizhi Li, Mong Yuan Sim, Xiuying Chen, Qingqing Zhu, Zhenzhu Yang, Adam Nik, Qi Liu, Chenghua Lin, Shi Wang, Ruibo Liu, Wenhu Chen, Ke Xu, Dayiheng Liu, Yike Guo, Jie Fu

Figure 1 for Interactive Natural Language Processing
Figure 2 for Interactive Natural Language Processing
Figure 3 for Interactive Natural Language Processing
Figure 4 for Interactive Natural Language Processing

Interactive Natural Language Processing (iNLP) has emerged as a novel paradigm within the field of NLP, aimed at addressing limitations in existing frameworks while aligning with the ultimate goals of artificial intelligence. This paradigm considers language models as agents capable of observing, acting, and receiving feedback iteratively from external entities. Specifically, language models in this context can: (1) interact with humans for better understanding and addressing user needs, personalizing responses, aligning with human values, and improving the overall user experience; (2) interact with knowledge bases for enriching language representations with factual knowledge, enhancing the contextual relevance of responses, and dynamically leveraging external information to generate more accurate and informed responses; (3) interact with models and tools for effectively decomposing and addressing complex tasks, leveraging specialized expertise for specific subtasks, and fostering the simulation of social behaviors; and (4) interact with environments for learning grounded representations of language, and effectively tackling embodied tasks such as reasoning, planning, and decision-making in response to environmental observations. This paper offers a comprehensive survey of iNLP, starting by proposing a unified definition and framework of the concept. We then provide a systematic classification of iNLP, dissecting its various components, including interactive objects, interaction interfaces, and interaction methods. We proceed to delve into the evaluation methodologies used in the field, explore its diverse applications, scrutinize its ethical and safety issues, and discuss prospective research directions. This survey serves as an entry point for researchers who are interested in this rapidly evolving area and offers a broad view of the current landscape and future trajectory of iNLP.

* 110 pages 
Viaarxiv icon

Adaptive Hybrid Spatial-Temporal Graph Neural Network for Cellular Traffic Prediction

Feb 28, 2023
Xing Wang, Kexin Yang, Zhendong Wang, Junlan Feng, Lin Zhu, Juan Zhao, Chao Deng

Figure 1 for Adaptive Hybrid Spatial-Temporal Graph Neural Network for Cellular Traffic Prediction
Figure 2 for Adaptive Hybrid Spatial-Temporal Graph Neural Network for Cellular Traffic Prediction
Figure 3 for Adaptive Hybrid Spatial-Temporal Graph Neural Network for Cellular Traffic Prediction
Figure 4 for Adaptive Hybrid Spatial-Temporal Graph Neural Network for Cellular Traffic Prediction

Cellular traffic prediction is an indispensable part for intelligent telecommunication networks. Nevertheless, due to the frequent user mobility and complex network scheduling mechanisms, cellular traffic often inherits complicated spatial-temporal patterns, making the prediction incredibly challenging. Although recent advanced algorithms such as graph-based prediction approaches have been proposed, they frequently model spatial dependencies based on static or dynamic graphs and neglect the coexisting multiple spatial correlations induced by traffic generation. Meanwhile, some works lack the consideration of the diverse cellular traffic patterns, result in suboptimal prediction results. In this paper, we propose a novel deep learning network architecture, Adaptive Hybrid Spatial-Temporal Graph Neural Network (AHSTGNN), to tackle the cellular traffic prediction problem. First, we apply adaptive hybrid graph learning to learn the compound spatial correlations among cell towers. Second, we implement a Temporal Convolution Module with multi-periodic temporal data input to capture the nonlinear temporal dependencies. In addition, we introduce an extra Spatial-Temporal Adaptive Module to conquer the heterogeneity lying in cell towers. Our experiments on two real-world cellular traffic datasets show AHSTGNN outperforms the state-of-the-art by a significant margin, illustrating the superior scalability of our method for spatial-temporal cellular traffic prediction.

* To be published in IEEE International Conference on Communications (ICC) 
Viaarxiv icon

Draft, Command, and Edit: Controllable Text Editing in E-Commerce

Aug 11, 2022
Kexin Yang, Dayiheng Liu, Wenqiang Lei, Baosong Yang, Qian Qu, Jiancheng Lv

Figure 1 for Draft, Command, and Edit: Controllable Text Editing in E-Commerce
Figure 2 for Draft, Command, and Edit: Controllable Text Editing in E-Commerce
Figure 3 for Draft, Command, and Edit: Controllable Text Editing in E-Commerce
Figure 4 for Draft, Command, and Edit: Controllable Text Editing in E-Commerce

Product description generation is a challenging and under-explored task. Most such work takes a set of product attributes as inputs then generates a description from scratch in a single pass. However, this widespread paradigm might be limited when facing the dynamic wishes of users on constraining the description, such as deleting or adding the content of a user-specified attribute based on the previous version. To address this challenge, we explore a new draft-command-edit manner in description generation, leading to the proposed new task-controllable text editing in E-commerce. More specifically, we allow systems to receive a command (deleting or adding) from the user and then generate a description by flexibly modifying the content based on the previous version. It is easier and more practical to meet the new needs by modifying previous versions than generating from scratch. Furthermore, we design a data augmentation method to remedy the low resource challenge in this task, which contains a model-based and a rule-based strategy to imitate the edit by humans. To accompany this new task, we present a human-written draft-command-edit dataset called E-cEdits and a new metric "Attribute Edit". Our experimental results show that using the new data augmentation method outperforms baselines to a greater extent in both automatic and human evaluations.

Viaarxiv icon

Tailor: A Prompt-Based Approach to Attribute-Based Controlled Text Generation

Apr 28, 2022
Kexin Yang, Dayiheng Liu, Wenqiang Lei, Baosong Yang, Mingfeng Xue, Boxing Chen, Jun Xie

Figure 1 for Tailor: A Prompt-Based Approach to Attribute-Based Controlled Text Generation
Figure 2 for Tailor: A Prompt-Based Approach to Attribute-Based Controlled Text Generation
Figure 3 for Tailor: A Prompt-Based Approach to Attribute-Based Controlled Text Generation
Figure 4 for Tailor: A Prompt-Based Approach to Attribute-Based Controlled Text Generation

Attribute-based Controlled Text Generation (CTG) refers to generating sentences that satisfy desirable attributes (e.g., emotions and topics). Existing works often utilize fine-tuning or resort to extra attribute classifiers, yet suffer from storage and inference time increases. To address these concerns, we explore attribute-based CTG in a prompt-based manner. In short, the proposed Tailor represents each attribute as a pre-trained continuous vector (i.e., single-attribute prompt) and guides the generation of a fixed PLM switch to a pre-specified attribute. We experimentally find that these prompts can be simply concatenated as a whole to multi-attribute CTG without any re-training, yet raises problems of fluency decrease and position sensitivity. To this end, Tailor provides a multi-attribute prompt mask and a re-indexing position-ids sequence to bridge the gap between the training (one prompt for each task) and testing stage (concatenating more than one prompt). To further enhance such single-attribute prompt combinations, Tailor also introduces a trainable prompt connector, which can be concatenated with any two single-attribute prompts to multi-attribute text generation. Experiments on 11 attribute-specific generation tasks demonstrate strong performances of Tailor on both single-attribute and multi-attribute CTG, with 0.08\% training parameters of a GPT-2.

Viaarxiv icon

Emergence of Theory of Mind Collaboration in Multiagent Systems

Sep 30, 2021
Luyao Yuan, Zipeng Fu, Linqi Zhou, Kexin Yang, Song-Chun Zhu

Figure 1 for Emergence of Theory of Mind Collaboration in Multiagent Systems
Figure 2 for Emergence of Theory of Mind Collaboration in Multiagent Systems
Figure 3 for Emergence of Theory of Mind Collaboration in Multiagent Systems
Figure 4 for Emergence of Theory of Mind Collaboration in Multiagent Systems

Currently, in the study of multiagent systems, the intentions of agents are usually ignored. Nonetheless, as pointed out by Theory of Mind (ToM), people regularly reason about other's mental states, including beliefs, goals, and intentions, to obtain performance advantage in competition, cooperation or coalition. However, due to its intrinsic recursion and intractable modeling of distribution over belief, integrating ToM in multiagent planning and decision making is still a challenge. In this paper, we incorporate ToM in multiagent partially observable Markov decision process (POMDP) and propose an adaptive training algorithm to develop effective collaboration between agents with ToM. We evaluate our algorithms with two games, where our algorithm surpasses all previous decentralized execution algorithms without modeling ToM.

* Emergent Communication Workshop, 33rd Conference on Neural Information Processing Systems (NeurIPS 2019)  
Viaarxiv icon

AnchiBERT: A Pre-Trained Model for Ancient ChineseLanguage Understanding and Generation

Sep 24, 2020
Huishuang Tian, Kexin Yang, Dayiheng Liu, Jiancheng Lv

Figure 1 for AnchiBERT: A Pre-Trained Model for Ancient ChineseLanguage Understanding and Generation
Figure 2 for AnchiBERT: A Pre-Trained Model for Ancient ChineseLanguage Understanding and Generation
Figure 3 for AnchiBERT: A Pre-Trained Model for Ancient ChineseLanguage Understanding and Generation
Figure 4 for AnchiBERT: A Pre-Trained Model for Ancient ChineseLanguage Understanding and Generation

Ancient Chinese is the essence of Chinese culture. There are several natural language processing tasks of ancient Chinese domain, such as ancient-modern Chinese translation, poem generation, and couplet generation. Previous studies usually use the supervised models which deeply rely on parallel data. However, it is difficult to obtain large-scale parallel data of ancient Chinese. In order to make full use of the more easily available monolingual ancient Chinese corpora, we release AnchiBERT, a pre-trained language model based on the architecture of BERT, which is trained on large-scale ancient Chinese corpora. We evaluate AnchiBERT on both language understanding and generation tasks, including poem classification, ancient-modern Chinese translation, poem generation, and couplet generation. The experimental results show that AnchiBERT outperforms BERT as well as the non-pretrained models and achieves state-of-the-art results in all cases.

* 10 pages with 3 figures 
Viaarxiv icon

Ancient-Modern Chinese Translation with a Large Training Dataset

Aug 11, 2018
Dayiheng Liu, Jiancheng Lv, Kexin Yang, Qian Qu

Figure 1 for Ancient-Modern Chinese Translation with a Large Training Dataset
Figure 2 for Ancient-Modern Chinese Translation with a Large Training Dataset
Figure 3 for Ancient-Modern Chinese Translation with a Large Training Dataset
Figure 4 for Ancient-Modern Chinese Translation with a Large Training Dataset

Ancient Chinese brings the wisdom and spirit culture of the Chinese nation. Automatically translation from ancient Chinese to modern Chinese helps to inherit and carry forward the quintessence of the ancients. In this paper, we propose an Ancient-Modern Chinese clause alignment approach and apply it to create a large scale Ancient-Modern Chinese parallel corpus which contains about 1.24M bilingual pairs. To our best knowledge, this is the first large high-quality Ancient-Modern Chinese dataset. Furthermore, we train the SMT and various NMT based models on this dataset and provide a strong baseline for this task

* 8 pages, 2 figures 
Viaarxiv icon