Alert button
Picture for Shilei Liu

Shilei Liu

Alert button

Advancements in 3D Lane Detection Using LiDAR Point Clouds: From Data Collection to Model Development

Sep 24, 2023
Runkai Zhao, Yuwen Heng, Yuanda Gao, Shilei Liu, Heng Wang, Changhao Yao, Jiawen Chen, Weidong Cai

Figure 1 for Advancements in 3D Lane Detection Using LiDAR Point Clouds: From Data Collection to Model Development
Figure 2 for Advancements in 3D Lane Detection Using LiDAR Point Clouds: From Data Collection to Model Development
Figure 3 for Advancements in 3D Lane Detection Using LiDAR Point Clouds: From Data Collection to Model Development
Figure 4 for Advancements in 3D Lane Detection Using LiDAR Point Clouds: From Data Collection to Model Development

Advanced Driver-Assistance Systems (ADAS) have successfully integrated learning-based techniques into vehicle perception and decision-making. However, their application in 3D lane detection for effective driving environment perception is hindered by the lack of comprehensive LiDAR datasets. The sparse nature of LiDAR point cloud data prevents an efficient manual annotation process. To solve this problem, we present LiSV-3DLane, a large-scale 3D lane dataset that comprises 20k frames of surround-view LiDAR point clouds with enriched semantic annotation. Unlike existing datasets confined to a frontal perspective, LiSV-3DLane provides a full 360-degree spatial panorama around the ego vehicle, capturing complex lane patterns in both urban and highway environments. We leverage the geometric traits of lane lines and the intrinsic spatial attributes of LiDAR data to design a simple yet effective automatic annotation pipeline for generating finer lane labels. To propel future research, we propose a novel LiDAR-based 3D lane detection model, LiLaDet, incorporating the spatial geometry learning of the LiDAR point cloud into Bird's Eye View (BEV) based lane identification. Experimental results indicate that LiLaDet outperforms existing camera- and LiDAR-based approaches in the 3D lane detection task on the K-Lane dataset and our LiSV-3DLane.

* 7 pages, 6 figures 
Viaarxiv icon

Binaural Rendering of Ambisonic Signals by Neural Networks

Nov 04, 2022
Yin Zhu, Qiuqiang Kong, Junjie Shi, Shilei Liu, Xuzhou Ye, Ju-chiang Wang, Junping Zhang

Figure 1 for Binaural Rendering of Ambisonic Signals by Neural Networks
Figure 2 for Binaural Rendering of Ambisonic Signals by Neural Networks
Figure 3 for Binaural Rendering of Ambisonic Signals by Neural Networks
Figure 4 for Binaural Rendering of Ambisonic Signals by Neural Networks

Binaural rendering of ambisonic signals is of broad interest to virtual reality and immersive media. Conventional methods often require manually measured Head-Related Transfer Functions (HRTFs). To address this issue, we collect a paired ambisonic-binaural dataset and propose a deep learning framework in an end-to-end manner. Experimental results show that neural networks outperform the conventional method in objective metrics and achieve comparable subjective metrics. To validate the proposed framework, we experimentally explore different settings of the input features, model structures, output features, and loss functions. Our proposed system achieves an SDR of 7.32 and MOSs of 3.83, 3.58, 3.87, 3.58 in quality, timbre, localization, and immersion dimensions.

Viaarxiv icon

An Understanding-Oriented Robust Machine Reading Comprehension Model

Jul 01, 2022
Feiliang Ren, Yongkang Liu, Bochao Li, Shilei Liu, Bingchao Wang, Jiaqi Wang, Chunchao Liu, Qi Ma

Figure 1 for An Understanding-Oriented Robust Machine Reading Comprehension Model
Figure 2 for An Understanding-Oriented Robust Machine Reading Comprehension Model
Figure 3 for An Understanding-Oriented Robust Machine Reading Comprehension Model
Figure 4 for An Understanding-Oriented Robust Machine Reading Comprehension Model

Although existing machine reading comprehension models are making rapid progress on many datasets, they are far from robust. In this paper, we propose an understanding-oriented machine reading comprehension model to address three kinds of robustness issues, which are over sensitivity, over stability and generalization. Specifically, we first use a natural language inference module to help the model understand the accurate semantic meanings of input questions so as to address the issues of over sensitivity and over stability. Then in the machine reading comprehension module, we propose a memory-guided multi-head attention method that can further well understand the semantic meanings of input questions and passages. Third, we propose a multilanguage learning mechanism to address the issue of generalization. Finally, these modules are integrated with a multi-task learning based method. We evaluate our model on three benchmark datasets that are designed to measure models robustness, including DuReader (robust) and two SQuAD-related datasets. Extensive experiments show that our model can well address the mentioned three kinds of robustness issues. And it achieves much better results than the compared state-of-the-art models on all these datasets under different evaluation metrics, even under some extreme and unfair evaluations. The source code of our work is available at: https://github.com/neukg/RobustMRC.

* Accepted by TALLIP 
Viaarxiv icon

Deep Understanding based Multi-Document Machine Reading Comprehension

Feb 25, 2022
Feiliang Ren, Yongkang Liu, Bochao Li, Zhibo Wang, Yu Guo, Shilei Liu, Huimin Wu, Jiaqi Wang, Chunchao Liu, Bingchao Wang

Figure 1 for Deep Understanding based Multi-Document Machine Reading Comprehension
Figure 2 for Deep Understanding based Multi-Document Machine Reading Comprehension
Figure 3 for Deep Understanding based Multi-Document Machine Reading Comprehension
Figure 4 for Deep Understanding based Multi-Document Machine Reading Comprehension

Most existing multi-document machine reading comprehension models mainly focus on understanding the interactions between the input question and documents, but ignore following two kinds of understandings. First, to understand the semantic meaning of words in the input question and documents from the perspective of each other. Second, to understand the supporting cues for a correct answer from the perspective of intra-document and inter-documents. Ignoring these two kinds of important understandings would make the models oversee some important information that may be helpful for inding correct answers. To overcome this deiciency, we propose a deep understanding based model for multi-document machine reading comprehension. It has three cascaded deep understanding modules which are designed to understand the accurate semantic meaning of words, the interactions between the input question and documents, and the supporting cues for the correct answer. We evaluate our model on two large scale benchmark datasets, namely TriviaQA Web and DuReader. Extensive experiments show that our model achieves state-of-the-art results on both datasets.

* TALLIP 
Viaarxiv icon

A Simple but Effective Bidirectional Framework for Relational Triple Extraction

Jan 05, 2022
Feiliang Ren, Longhui Zhang, Xiaofeng Zhao, Shujuan Yin, Shilei Liu, Bochao Li

Figure 1 for A Simple but Effective Bidirectional Framework for Relational Triple Extraction
Figure 2 for A Simple but Effective Bidirectional Framework for Relational Triple Extraction
Figure 3 for A Simple but Effective Bidirectional Framework for Relational Triple Extraction
Figure 4 for A Simple but Effective Bidirectional Framework for Relational Triple Extraction

Tagging based relational triple extraction methods are attracting growing research attention recently. However, most of these methods take a unidirectional extraction framework that first extracts all subjects and then extracts objects and relations simultaneously based on the subjects extracted. This framework has an obvious deficiency that it is too sensitive to the extraction results of subjects. To overcome this deficiency, we propose a bidirectional extraction framework based method that extracts triples based on the entity pairs extracted from two complementary directions. Concretely, we first extract all possible subject-object pairs from two paralleled directions. These two extraction directions are connected by a shared encoder component, thus the extraction features from one direction can flow to another direction and vice versa. By this way, the extractions of two directions can boost and complement each other. Next, we assign all possible relations for each entity pair by a biaffine model. During training, we observe that the share structure will lead to a convergence rate inconsistency issue which is harmful to performance. So we propose a share-aware learning mechanism to address it. We evaluate the proposed model on multiple benchmark datasets. Extensive experimental results show that the proposed model is very effective and it achieves state-of-the-art results on all of these datasets. Moreover, experiments show that both the proposed bidirectional extraction framework and the share-aware learning mechanism have good adaptability and can be used to improve the performance of other tagging based methods. The source code of our work is available at: https://github.com/neukg/BiRTE.

* WSDM2022 
Viaarxiv icon

A Simple but Effective Bidirectional Extraction Framework for Relational Triple Extraction

Dec 09, 2021
Feiliang Ren, Longhui Zhang, Xiaofeng Zhao, Shujuan Yin, Shilei Liu, Bochao Li

Figure 1 for A Simple but Effective Bidirectional Extraction Framework for Relational Triple Extraction
Figure 2 for A Simple but Effective Bidirectional Extraction Framework for Relational Triple Extraction
Figure 3 for A Simple but Effective Bidirectional Extraction Framework for Relational Triple Extraction
Figure 4 for A Simple but Effective Bidirectional Extraction Framework for Relational Triple Extraction

Tagging based relational triple extraction methods are attracting growing research attention recently. However, most of these methods take a unidirectional extraction framework that first extracts all subjects and then extracts objects and relations simultaneously based on the subjects extracted. This framework has an obvious deficiency that it is too sensitive to the extraction results of subjects. To overcome this deficiency, we propose a bidirectional extraction framework based method that extracts triples based on the entity pairs extracted from two complementary directions. Concretely, we first extract all possible subject-object pairs from two paralleled directions. These two extraction directions are connected by a shared encoder component, thus the extraction features from one direction can flow to another direction and vice versa. By this way, the extractions of two directions can boost and complement each other. Next, we assign all possible relations for each entity pair by a biaffine model. During training, we observe that the share structure will lead to a convergence rate inconsistency issue which is harmful to performance. So we propose a share-aware learning mechanism to address it. We evaluate the proposed model on multiple benchmark datasets. Extensive experimental results show that the proposed model is very effective and it achieves state-of-the-art results on all of these datasets. Moreover, experiments show that both the proposed bidirectional extraction framework and the share-aware learning mechanism have good adaptability and can be used to improve the performance of other tagging based methods. The source code of our work is available at: https://github.com/neukg/BiRTE.

* WSDM2022 
Viaarxiv icon

A Novel Global Feature-Oriented Relational Triple Extraction Model based on Table Filling

Sep 14, 2021
Feiliang Ren, Longhui Zhang, Shujuan Yin, Xiaofeng Zhao, Shilei Liu, Bochao Li, Yaduo Liu

Table filling based relational triple extraction methods are attracting growing research interests due to their promising performance and their abilities on extracting triples from complex sentences. However, this kind of methods are far from their full potential because most of them only focus on using local features but ignore the global associations of relations and of token pairs, which increases the possibility of overlooking some important information during triple extraction. To overcome this deficiency, we propose a global feature-oriented triple extraction model that makes full use of the mentioned two kinds of global associations. Specifically, we first generate a table feature for each relation. Then two kinds of global associations are mined from the generated table features. Next, the mined global associations are integrated into the table feature of each relation. This "generate-mine-integrate" process is performed multiple times so that the table feature of each relation is refined step by step. Finally, each relation's table is filled based on its refined table feature, and all triples linked to this relation are extracted based on its filled table. We evaluate the proposed model on three benchmark datasets. Experimental results show our model is effective and it achieves state-of-the-art results on all of these datasets. The source code of our work is available at: https://github.com/neukg/GRTE.

* EMNLP2021 
Viaarxiv icon

A Three-Stage Learning Framework for Low-Resource Knowledge-Grounded Dialogue Generation

Sep 09, 2021
Shilei Liu, Xiaofeng Zhao, Bochao Li, Feiliang Ren, Longhui Zhang, Shujuan Yin

Figure 1 for A Three-Stage Learning Framework for Low-Resource Knowledge-Grounded Dialogue Generation
Figure 2 for A Three-Stage Learning Framework for Low-Resource Knowledge-Grounded Dialogue Generation
Figure 3 for A Three-Stage Learning Framework for Low-Resource Knowledge-Grounded Dialogue Generation
Figure 4 for A Three-Stage Learning Framework for Low-Resource Knowledge-Grounded Dialogue Generation

Neural conversation models have shown great potentials towards generating fluent and informative responses by introducing external background knowledge. Nevertheless, it is laborious to construct such knowledge-grounded dialogues, and existing models usually perform poorly when transfer to new domains with limited training samples. Therefore, building a knowledge-grounded dialogue system under the low-resource setting is a still crucial issue. In this paper, we propose a novel three-stage learning framework based on weakly supervised learning which benefits from large scale ungrounded dialogues and unstructured knowledge base. To better cooperate with this framework, we devise a variant of Transformer with decoupled decoder which facilitates the disentangled learning of response generation and knowledge incorporation. Evaluation results on two benchmarks indicate that our approach can outperform other state-of-the-art methods with less training data, and even in zero-resource scenario, our approach still performs well.

* Accepted by EMNLP 2021 main conference 
Viaarxiv icon

Knowledge-Grounded Dialogue with Reward-Driven Knowledge Selection

Aug 31, 2021
Shilei Liu, Xiaofeng Zhao, Bochao Li, Feiliang Ren

Figure 1 for Knowledge-Grounded Dialogue with Reward-Driven Knowledge Selection
Figure 2 for Knowledge-Grounded Dialogue with Reward-Driven Knowledge Selection
Figure 3 for Knowledge-Grounded Dialogue with Reward-Driven Knowledge Selection
Figure 4 for Knowledge-Grounded Dialogue with Reward-Driven Knowledge Selection

Knowledge-grounded dialogue is a task of generating a fluent and informative response based on both conversation context and a collection of external knowledge, in which knowledge selection plays an important role and attracts more and more research interest. However, most existing models either select only one knowledge or use all knowledge for responses generation. The former may lose valuable information in discarded knowledge, while the latter may bring a lot of noise. At the same time, many approaches need to train the knowledge selector with knowledge labels that indicate ground-truth knowledge, but these labels are difficult to obtain and require a large number of manual annotations. Motivated by these issues, we propose Knoformer, a dialogue response generation model based on reinforcement learning, which can automatically select one or more related knowledge from the knowledge pool and does not need knowledge labels during training. Knoformer is evaluated on two knowledge-guided conversation datasets, and achieves state-of-the-art performance.

* NLPCC 2021 
Viaarxiv icon

A Conditional Cascade Model for Relational Triple Extraction

Aug 20, 2021
Feiliang Ren, Longhui Zhang, Shujuan Yin, Xiaofeng Zhao, Shilei Liu, Bochao Li

Figure 1 for A Conditional Cascade Model for Relational Triple Extraction
Figure 2 for A Conditional Cascade Model for Relational Triple Extraction
Figure 3 for A Conditional Cascade Model for Relational Triple Extraction
Figure 4 for A Conditional Cascade Model for Relational Triple Extraction

Tagging based methods are one of the mainstream methods in relational triple extraction. However, most of them suffer from the class imbalance issue greatly. Here we propose a novel tagging based model that addresses this issue from following two aspects. First, at the model level, we propose a three-step extraction framework that can reduce the total number of samples greatly, which implicitly decreases the severity of the mentioned issue. Second, at the intra-model level, we propose a confidence threshold based cross entropy loss that can directly neglect some samples in the major classes. We evaluate the proposed model on NYT and WebNLG. Extensive experiments show that it can address the mentioned issue effectively and achieves state-of-the-art results on both datasets. The source code of our model is available at: https://github.com/neukg/ConCasRTE.

* CIKM2021-Short 
Viaarxiv icon