Alert button
Picture for Lei Li

Lei Li

Alert button

Not All Errors are Equal: Learning Text Generation Metrics using Stratified Error Synthesis

Oct 10, 2022
Wenda Xu, Yilin Tuan, Yujie Lu, Michael Saxon, Lei Li, William Yang Wang

Figure 1 for Not All Errors are Equal: Learning Text Generation Metrics using Stratified Error Synthesis
Figure 2 for Not All Errors are Equal: Learning Text Generation Metrics using Stratified Error Synthesis
Figure 3 for Not All Errors are Equal: Learning Text Generation Metrics using Stratified Error Synthesis
Figure 4 for Not All Errors are Equal: Learning Text Generation Metrics using Stratified Error Synthesis

Is it possible to build a general and automatic natural language generation (NLG) evaluation metric? Existing learned metrics either perform unsatisfactorily or are restricted to tasks where large human rating data is already available. We introduce SESCORE, a model-based metric that is highly correlated with human judgements without requiring human annotation, by utilizing a novel, iterative error synthesis and severity scoring pipeline. This pipeline applies a series of plausible errors to raw text and assigns severity labels by simulating human judgements with entailment. We evaluate SESCORE against existing metrics by comparing how their scores correlate with human ratings. SESCORE outperforms all prior unsupervised metrics on multiple diverse NLG tasks including machine translation, image captioning, and WebNLG text generation. For WMT 20/21 En-De and Zh-En, SESCORE improve the average Kendall correlation with human judgement from 0.154 to 0.195. SESCORE even achieves comparable performance to the best supervised metric COMET, despite receiving no human-annotated training data.

Viaarxiv icon

PARAGEN : A Parallel Generation Toolkit

Oct 07, 2022
Jiangtao Feng, Yi Zhou, Jun Zhang, Xian Qian, Liwei Wu, Zhexi Zhang, Yanming Liu, Mingxuan Wang, Lei Li, Hao Zhou

Figure 1 for PARAGEN : A Parallel Generation Toolkit
Figure 2 for PARAGEN : A Parallel Generation Toolkit
Figure 3 for PARAGEN : A Parallel Generation Toolkit
Figure 4 for PARAGEN : A Parallel Generation Toolkit

PARAGEN is a PyTorch-based NLP toolkit for further development on parallel generation. PARAGEN provides thirteen types of customizable plugins, helping users to experiment quickly with novel ideas across model architectures, optimization, and learning strategies. We implement various features, such as unlimited data loading and automatic model selection, to enhance its industrial usage. ParaGen is now deployed to support various research and industry applications at ByteDance. PARAGEN is available at https://github.com/bytedance/ParaGen.

* 9 pages, 1 figure, 6 tables 
Viaarxiv icon

Calibrating Factual Knowledge in Pretrained Language Models

Oct 07, 2022
Qingxiu Dong, Damai Dai, Yifan Song, Jingjing Xu, Zhifang Sui, Lei Li

Figure 1 for Calibrating Factual Knowledge in Pretrained Language Models
Figure 2 for Calibrating Factual Knowledge in Pretrained Language Models
Figure 3 for Calibrating Factual Knowledge in Pretrained Language Models
Figure 4 for Calibrating Factual Knowledge in Pretrained Language Models

Previous literature has proved that Pretrained Language Models (PLMs) can store factual knowledge. However, we find that facts stored in the PLMs are not always correct. It motivates us to explore a fundamental question: How do we calibrate factual knowledge in PLMs without re-training from scratch? In this work, we propose a simple and lightweight method CaliNet to achieve this goal. To be specific, we first detect whether PLMs can learn the right facts via a contrastive score between right and fake facts. If not, we then use a lightweight method to add and adapt new parameters to specific factual texts. Experiments on the knowledge probing task show the calibration effectiveness and efficiency. In addition, through closed-book question answering, we find that the calibrated PLM possesses knowledge generalization ability after fine-tuning. Beyond the calibration performance, we further investigate and visualize the knowledge calibration mechanism.

* Accepted by Findings of EMNLP 2022 
Viaarxiv icon

Distillation-Resistant Watermarking for Model Protection in NLP

Oct 07, 2022
Xuandong Zhao, Lei Li, Yu-Xiang Wang

Figure 1 for Distillation-Resistant Watermarking for Model Protection in NLP
Figure 2 for Distillation-Resistant Watermarking for Model Protection in NLP
Figure 3 for Distillation-Resistant Watermarking for Model Protection in NLP
Figure 4 for Distillation-Resistant Watermarking for Model Protection in NLP

How can we protect the intellectual property of trained NLP models? Modern NLP models are prone to stealing by querying and distilling from their publicly exposed APIs. However, existing protection methods such as watermarking only work for images but are not applicable to text. We propose Distillation-Resistant Watermarking (DRW), a novel technique to protect NLP models from being stolen via distillation. DRW protects a model by injecting watermarks into the victim's prediction probability corresponding to a secret key and is able to detect such a key by probing a suspect model. We prove that a protected model still retains the original accuracy within a certain bound. We evaluate DRW on a diverse set of NLP tasks including text classification, part-of-speech tagging, and named entity recognition. Experiments show that DRW protects the original model and detects stealing suspects at 100% mean average precision for all four tasks while the prior method fails on two.

Viaarxiv icon

Just ClozE! A Fast and Simple Method for Evaluating the Factual Consistency in Abstractive Summarization

Oct 06, 2022
Yiyang Li, Lei Li, Qing Yang, Marina Litvak, Natalia Vanetik, Dingxin Hu, Yuze Li, Yanquan Zhou, Dongliang Xu, Xuanyu Zhang

Figure 1 for Just ClozE! A Fast and Simple Method for Evaluating the Factual Consistency in Abstractive Summarization
Figure 2 for Just ClozE! A Fast and Simple Method for Evaluating the Factual Consistency in Abstractive Summarization
Figure 3 for Just ClozE! A Fast and Simple Method for Evaluating the Factual Consistency in Abstractive Summarization
Figure 4 for Just ClozE! A Fast and Simple Method for Evaluating the Factual Consistency in Abstractive Summarization

The issue of factual consistency in abstractive summarization has attracted much attention in recent years, and the evaluation of factual consistency between summary and document has become an important and urgent task. Most of the current evaluation metrics are adopted from the question answering (QA). However, the application of QA-based metrics is extremely time-consuming in practice, causing the iteration cycle of abstractive summarization research to be severely prolonged. In this paper, we propose a new method called ClozE to evaluate factual consistency by cloze model, instantiated based on masked language model(MLM), with strong interpretability and substantially higher speed. We demonstrate that ClozE can reduce the evaluation time by nearly 96$\%$ relative to QA-based metrics while retaining their interpretability and performance through experiments on six human-annotated datasets and a meta-evaluation benchmark GO FIGURE \citep{gabriel2020go}. We also implement experiments to further demonstrate more characteristics of ClozE in terms of performance and speed. In addition, we conduct an experimental analysis of the limitations of ClozE, which suggests future research directions. The code and models for ClozE will be released upon the paper acceptance.

* The manuscript for EMNLP 2022 
Viaarxiv icon

Block-Structured Optimization for Subgraph Detection in Interdependent Networks

Oct 06, 2022
Fei Jie, Chunpai Wang, Feng Chen, Lei Li, Xindong Wu

Figure 1 for Block-Structured Optimization for Subgraph Detection in Interdependent Networks
Figure 2 for Block-Structured Optimization for Subgraph Detection in Interdependent Networks
Figure 3 for Block-Structured Optimization for Subgraph Detection in Interdependent Networks
Figure 4 for Block-Structured Optimization for Subgraph Detection in Interdependent Networks

We propose a generalized framework for block-structured nonconvex optimization, which can be applied to structured subgraph detection in interdependent networks, such as multi-layer networks, temporal networks, networks of networks, and many others. Specifically, we design an effective, efficient, and parallelizable projection algorithm, namely Graph Block-structured Gradient Projection (GBGP), to optimize a general non-linear function subject to graph-structured constraints. We prove that our algorithm: 1) runs in nearly-linear time on the network size; 2) enjoys a theoretical approximation guarantee. Moreover, we demonstrate how our framework can be applied to two very practical applications and conduct comprehensive experiments to show the effectiveness and efficiency of our proposed algorithm.

* Accepted by ICDM-2019 
Viaarxiv icon

Multimodal Analogical Reasoning over Knowledge Graphs

Oct 01, 2022
Ningyu Zhang, Lei Li, Xiang Chen, Xiaozhuan Liang, Shumin Deng, Huajun Chen

Figure 1 for Multimodal Analogical Reasoning over Knowledge Graphs
Figure 2 for Multimodal Analogical Reasoning over Knowledge Graphs
Figure 3 for Multimodal Analogical Reasoning over Knowledge Graphs
Figure 4 for Multimodal Analogical Reasoning over Knowledge Graphs

Analogical reasoning is fundamental to human cognition and holds an important place in various fields. However, previous studies mainly focus on single-modal analogical reasoning and ignore taking advantage of structure knowledge. Notably, the research in cognitive psychology has demonstrated that information from multimodal sources always brings more powerful cognitive transfer than single modality sources. To this end, we introduce the new task of multimodal analogical reasoning over knowledge graphs, which requires multimodal reasoning ability with the help of background knowledge. Specifically, we construct a Multimodal Analogical Reasoning dataSet (MARS) and a multimodal knowledge graph MarKG. We evaluate with multimodal knowledge graph embedding and pre-trained Transformer baselines, illustrating the potential challenges of the proposed task. We further propose a novel model-agnostic Multimodal analogical reasoning framework with Transformer (MarT) motivated by the structure mapping theory, which can obtain better performance.

* Work in progress 
Viaarxiv icon

SRFeat: Learning Locally Accurate and Globally Consistent Non-Rigid Shape Correspondence

Sep 16, 2022
Lei Li, Souhaib Attaiki, Maks Ovsjanikov

Figure 1 for SRFeat: Learning Locally Accurate and Globally Consistent Non-Rigid Shape Correspondence
Figure 2 for SRFeat: Learning Locally Accurate and Globally Consistent Non-Rigid Shape Correspondence
Figure 3 for SRFeat: Learning Locally Accurate and Globally Consistent Non-Rigid Shape Correspondence
Figure 4 for SRFeat: Learning Locally Accurate and Globally Consistent Non-Rigid Shape Correspondence

In this work, we present a novel learning-based framework that combines the local accuracy of contrastive learning with the global consistency of geometric approaches, for robust non-rigid matching. We first observe that while contrastive learning can lead to powerful point-wise features, the learned correspondences commonly lack smoothness and consistency, owing to the purely combinatorial nature of the standard contrastive losses. To overcome this limitation we propose to boost contrastive feature learning with two types of smoothness regularization that inject geometric information into correspondence learning. With this novel combination in hand, the resulting features are both highly discriminative across individual points, and, at the same time, lead to robust and consistent correspondences, through simple proximity queries. Our framework is general and is applicable to local feature learning in both the 3D and 2D domains. We demonstrate the superiority of our approach through extensive experiments on a wide range of challenging matching benchmarks, including 3D non-rigid shape correspondence and 2D image keypoint matching.

* 3DV 2022. Code and data: https://github.com/craigleili/SRFeat 
Viaarxiv icon

NU-net: An Unpretentious Nested U-net for Breast Tumor Segmentation

Sep 15, 2022
Gong-Ping Chen, Lei Li, Yu Dai, Jian-Xun Zhang

Figure 1 for NU-net: An Unpretentious Nested U-net for Breast Tumor Segmentation
Figure 2 for NU-net: An Unpretentious Nested U-net for Breast Tumor Segmentation
Figure 3 for NU-net: An Unpretentious Nested U-net for Breast Tumor Segmentation
Figure 4 for NU-net: An Unpretentious Nested U-net for Breast Tumor Segmentation

Breast tumor segmentation is one of the key steps that helps us characterize and localize tumor regions. However, variable tumor morphology, blurred boundary, and similar intensity distributions bring challenges for accurate segmentation of breast tumors. Recently, many U-net variants have been proposed and widely used for breast tumors segmentation. However, these architectures suffer from two limitations: (1) Ignoring the characterize ability of the benchmark networks, and (2) Introducing extra complex operations increases the difficulty of understanding and reproducing the network. To alleviate these challenges, this paper proposes a simple yet powerful nested U-net (NU-net) for accurate segmentation of breast tumors. The key idea is to utilize U-Nets with different depths and shared weights to achieve robust characterization of breast tumors. NU-net mainly has the following advantages: (1) Improving network adaptability and robustness to breast tumors with different scales, (2) This method is easy to reproduce and execute, and (3) The extra operations increase network parameters without significantly increasing computational cost. Extensive experimental results with twelve state-of-the-art segmentation methods on three public breast ultrasound datasets demonstrate that NU-net has more competitive segmentation performance on breast tumors. Furthermore, the robustness of NU-net is further illustrated on the segmentation of renal ultrasound images. The source code is publicly available on https://github.com/CGPzy/NU-net.

Viaarxiv icon

3D Textured Shape Recovery with Learned Geometric Priors

Sep 07, 2022
Lei Li, Zhizheng Liu, Weining Ren, Liudi Yang, Fangjinhua Wang, Marc Pollefeys, Songyou Peng

Figure 1 for 3D Textured Shape Recovery with Learned Geometric Priors
Figure 2 for 3D Textured Shape Recovery with Learned Geometric Priors
Figure 3 for 3D Textured Shape Recovery with Learned Geometric Priors
Figure 4 for 3D Textured Shape Recovery with Learned Geometric Priors

3D textured shape recovery from partial scans is crucial for many real-world applications. Existing approaches have demonstrated the efficacy of implicit function representation, but they suffer from partial inputs with severe occlusions and varying object types, which greatly hinders their application value in the real world. This technical report presents our approach to address these limitations by incorporating learned geometric priors. To this end, we generate a SMPL model from learned pose prediction and fuse it into the partial input to add prior knowledge of human bodies. We also propose a novel completeness-aware bounding box adaptation for handling different levels of scales and partialness of partial scans.

* 5 pages, 3 figures, 2 tables 
Viaarxiv icon