Alert button
Picture for Shanshan Liu

Shanshan Liu

Alert button

Concurrent Classifier Error Detection (CCED) in Large Scale Machine Learning Systems

Jun 02, 2023
Pedro Reviriego, Ziheng Wang, Alvaro Alonso, Zhen Gao, Farzad Niknia, Shanshan Liu, Fabrizio Lombardi

Figure 1 for Concurrent Classifier Error Detection (CCED) in Large Scale Machine Learning Systems
Figure 2 for Concurrent Classifier Error Detection (CCED) in Large Scale Machine Learning Systems
Figure 3 for Concurrent Classifier Error Detection (CCED) in Large Scale Machine Learning Systems
Figure 4 for Concurrent Classifier Error Detection (CCED) in Large Scale Machine Learning Systems

The complexity of Machine Learning (ML) systems increases each year, with current implementations of large language models or text-to-image generators having billions of parameters and requiring billions of arithmetic operations. As these systems are widely utilized, ensuring their reliable operation is becoming a design requirement. Traditional error detection mechanisms introduce circuit or time redundancy that significantly impacts system performance. An alternative is the use of Concurrent Error Detection (CED) schemes that operate in parallel with the system and exploit their properties to detect errors. CED is attractive for large ML systems because it can potentially reduce the cost of error detection. In this paper, we introduce Concurrent Classifier Error Detection (CCED), a scheme to implement CED in ML systems using a concurrent ML classifier to detect errors. CCED identifies a set of check signals in the main ML system and feeds them to the concurrent ML classifier that is trained to detect errors. The proposed CCED scheme has been implemented and evaluated on two widely used large-scale ML models: Contrastive Language Image Pretraining (CLIP) used for image classification and Bidirectional Encoder Representations from Transformers (BERT) used for natural language applications. The results show that more than 95 percent of the errors are detected when using a simple Random Forest classifier that is order of magnitude simpler than CLIP or BERT. These results illustrate the potential of CCED to implement error detection in large-scale ML models.

Viaarxiv icon

Long-lead forecasts of wintertime air stagnation index in southern China using oceanic memory effects

May 16, 2023
Chenhong Zhou, Xiaorui Zhang, Meng Gao, Shanshan Liu, Yike Guo, Jie Chen

Figure 1 for Long-lead forecasts of wintertime air stagnation index in southern China using oceanic memory effects
Figure 2 for Long-lead forecasts of wintertime air stagnation index in southern China using oceanic memory effects
Figure 3 for Long-lead forecasts of wintertime air stagnation index in southern China using oceanic memory effects

Stagnant weather condition is one of the major contributors to air pollution as it is favorable for the formation and accumulation of pollutants. To measure the atmosphere's ability to dilute air pollutants, Air Stagnation Index (ASI) has been introduced as an important meteorological index. Therefore, making long-lead ASI forecasts is vital to make plans in advance for air quality management. In this study, we found that autumn Ni\~no indices derived from sea surface temperature (SST) anomalies show a negative correlation with wintertime ASI in southern China, offering prospects for a prewinter forecast. We developed an LSTM-based model to predict the future wintertime ASI. Results demonstrated that multivariate inputs (past ASI and Ni\~no indices) achieve better forecast performance than univariate input (only past ASI). The model achieves a correlation coefficient of 0.778 between the actual and predicted ASI, exhibiting a high degree of consistency.

* ICLR 2023 Tackling Climate Change with Machine Learning Workshop 
Viaarxiv icon

MaskOCR: Text Recognition with Masked Encoder-Decoder Pretraining

Jun 01, 2022
Pengyuan Lyu, Chengquan Zhang, Shanshan Liu, Meina Qiao, Yangliu Xu, Liang Wu, Kun Yao, Junyu Han, Errui Ding, Jingdong Wang

Figure 1 for MaskOCR: Text Recognition with Masked Encoder-Decoder Pretraining
Figure 2 for MaskOCR: Text Recognition with Masked Encoder-Decoder Pretraining
Figure 3 for MaskOCR: Text Recognition with Masked Encoder-Decoder Pretraining
Figure 4 for MaskOCR: Text Recognition with Masked Encoder-Decoder Pretraining

In this paper, we present a model pretraining technique, named MaskOCR, for text recognition. Our text recognition architecture is an encoder-decoder transformer: the encoder extracts the patch-level representations, and the decoder recognizes the text from the representations. Our approach pretrains both the encoder and the decoder in a sequential manner. (i) We pretrain the encoder in a self-supervised manner over a large set of unlabeled real text images. We adopt the masked image modeling approach, which shows the effectiveness for general images, expecting that the representations take on semantics. (ii) We pretrain the decoder over a large set of synthesized text images in a supervised manner and enhance the language modeling capability of the decoder by randomly masking some text image patches occupied by characters input to the encoder and accordingly the representations input to the decoder. Experiments show that the proposed MaskOCR approach achieves superior results on the benchmark datasets, including Chinese and English text images.

Viaarxiv icon

PGNet: Real-time Arbitrarily-Shaped Text Spotting with Point Gathering Network

Apr 12, 2021
Pengfei Wang, Chengquan Zhang, Fei Qi, Shanshan Liu, Xiaoqiang Zhang, Pengyuan Lyu, Junyu Han, Jingtuo Liu, Errui Ding, Guangming Shi

Figure 1 for PGNet: Real-time Arbitrarily-Shaped Text Spotting with Point Gathering Network
Figure 2 for PGNet: Real-time Arbitrarily-Shaped Text Spotting with Point Gathering Network
Figure 3 for PGNet: Real-time Arbitrarily-Shaped Text Spotting with Point Gathering Network
Figure 4 for PGNet: Real-time Arbitrarily-Shaped Text Spotting with Point Gathering Network

The reading of arbitrarily-shaped text has received increasing research attention. However, existing text spotters are mostly built on two-stage frameworks or character-based methods, which suffer from either Non-Maximum Suppression (NMS), Region-of-Interest (RoI) operations, or character-level annotations. In this paper, to address the above problems, we propose a novel fully convolutional Point Gathering Network (PGNet) for reading arbitrarily-shaped text in real-time. The PGNet is a single-shot text spotter, where the pixel-level character classification map is learned with proposed PG-CTC loss avoiding the usage of character-level annotations. With a PG-CTC decoder, we gather high-level character classification vectors from two-dimensional space and decode them into text symbols without NMS and RoI operations involved, which guarantees high efficiency. Additionally, reasoning the relations between each character and its neighbors, a graph refinement module (GRM) is proposed to optimize the coarse recognition and improve the end-to-end performance. Experiments prove that the proposed method achieves competitive accuracy, meanwhile significantly improving the running speed. In particular, in Total-Text, it runs at 46.7 FPS, surpassing the previous spotters with a large margin.

* 10 pages, 8 figures, AAAI 2021 
Viaarxiv icon

Neural Machine Reading Comprehension: Methods and Trends

Jul 15, 2019
Shanshan Liu, Xin Zhang, Sheng Zhang, Hui Wang, Weiming Zhang

Figure 1 for Neural Machine Reading Comprehension: Methods and Trends
Figure 2 for Neural Machine Reading Comprehension: Methods and Trends
Figure 3 for Neural Machine Reading Comprehension: Methods and Trends
Figure 4 for Neural Machine Reading Comprehension: Methods and Trends

Machine Reading Comprehension (MRC), which requires the machine to answer questions based on the given context, has gained increasingly wide attention with the incorporation of various deep learning techniques over the past few years. Although the research of MRC based on deep learning is flourishing, there remains a lack of a comprehensive survey to summarize existing approaches and recent trends, which motivates our work presented in this article. Specifically, we give a thorough review of this research field, covering different aspects including (1) typical MRC tasks: their definitions, differences and representative datasets; (2) general architecture of neural MRC: the main modules and prevalent approaches to each of them; and (3) new trends: some emerging focuses in neural MRC as well as the corresponding challenges. Last but not least, in retrospect of what has been achieved so far, the survey also envisages what the future may hold by discussing the open issues left to be addressed.

* 39 pages 
Viaarxiv icon