Alert button
Picture for Lei Huang

Lei Huang

Alert button

corresponding author

P-ROCKET: Pruning Random Convolution Kernels for Time Series Classification

Sep 15, 2023
Shaowu Chen, Weize Sun, Lei Huang, Xiaopeng Li, Qingyuan Wang, Deepu John

In recent years, two time series classification models, ROCKET and MINIROCKET, have attracted much attention for their low training cost and state-of-the-art accuracy. Utilizing random 1-D convolutional kernels without training, ROCKET and MINIROCKET can rapidly extract features from time series data, allowing for the efficient fitting of linear classifiers. However, to comprehensively capture useful features, a large number of random kernels are required, which is incompatible for resource-constrained devices. Therefore, a heuristic evolutionary algorithm named S-ROCKET is devised to recognize and prune redundant kernels. Nevertheless, the inherent nature of evolutionary algorithms renders the evaluation of kernels within S-ROCKET an unacceptable time-consuming process. In this paper, diverging from S-ROCKET, which directly evaluates random kernels with nonsignificant differences, we remove kernels from a feature selection perspective by eliminating associating connections in the sequential classification layer. To this end, we start by formulating the pruning challenge as a Group Elastic Net classification problem and employ the ADMM method to arrive at a solution. Sequentially, we accelerate the aforementioned time-consuming solving process by bifurcating the $l_{2,1}$ and $l_2$ regularizations into two sequential stages and solve them separately, which ultimately forms our core algorithm, named P-ROCKET. Stage 1 of P-ROCKET employs group-wise regularization similarly to our initial ADMM-based Algorithm, but introduces dynamically varying penalties to greatly accelerate the process. To mitigate overfitting, Stage 2 of P-ROCKET implements element-wise regularization to refit a linear classifier, utilizing the retained features.

Viaarxiv icon

High-rate discretely-modulated continuous-variable quantum key distribution using quantum machine learning

Aug 07, 2023
Qin Liao, Jieyu Liu, Anqi Huang, Lei Huang, Zhuoying Fei, Xiquan Fu

Figure 1 for High-rate discretely-modulated continuous-variable quantum key distribution using quantum machine learning
Figure 2 for High-rate discretely-modulated continuous-variable quantum key distribution using quantum machine learning
Figure 3 for High-rate discretely-modulated continuous-variable quantum key distribution using quantum machine learning
Figure 4 for High-rate discretely-modulated continuous-variable quantum key distribution using quantum machine learning

We propose a high-rate scheme for discretely-modulated continuous-variable quantum key distribution (DM CVQKD) using quantum machine learning technologies, which divides the whole CVQKD system into three parts, i.e., the initialization part that is used for training and estimating quantum classifier, the prediction part that is used for generating highly correlated raw keys, and the data-postprocessing part that generates the final secret key string shared by Alice and Bob. To this end, a low-complexity quantum k-nearest neighbor (QkNN) classifier is designed for predicting the lossy discretely-modulated coherent states (DMCSs) at Bob's side. The performance of the proposed QkNN-based CVQKD especially in terms of machine learning metrics and complexity is analyzed, and its theoretical security is proved by using semi-definite program (SDP) method. Numerical simulation shows that the secret key rate of our proposed scheme is explicitly superior to the existing DM CVQKD protocols, and it can be further enhanced with the increase of modulation variance.

* 18 pages, 17 figures 
Viaarxiv icon

DiffDTM: A conditional structure-free framework for bioactive molecules generation targeted for dual proteins

Jun 24, 2023
Lei Huang, Zheng Yuan, Huihui Yan, Rong Sheng, Linjing Liu, Fuzhou Wang, Weidun Xie, Nanjun Chen, Fei Huang, Songfang Huang, Ka-Chun Wong, Yaoyun Zhang

Figure 1 for DiffDTM: A conditional structure-free framework for bioactive molecules generation targeted for dual proteins
Figure 2 for DiffDTM: A conditional structure-free framework for bioactive molecules generation targeted for dual proteins
Figure 3 for DiffDTM: A conditional structure-free framework for bioactive molecules generation targeted for dual proteins
Figure 4 for DiffDTM: A conditional structure-free framework for bioactive molecules generation targeted for dual proteins

Advances in deep generative models shed light on de novo molecule generation with desired properties. However, molecule generation targeted for dual protein targets still faces formidable challenges including protein 3D structure data requisition for model training, auto-regressive sampling, and model generalization for unseen targets. Here, we proposed DiffDTM, a novel conditional structure-free deep generative model based on a diffusion model for dual targets based molecule generation to address the above issues. Specifically, DiffDTM receives protein sequences and molecular graphs as inputs instead of protein and molecular conformations and incorporates an information fusion module to achieve conditional generation in a one-shot manner. We have conducted comprehensive multi-view experiments to demonstrate that DiffDTM can generate drug-like, synthesis-accessible, novel, and high-binding affinity molecules targeting specific dual proteins, outperforming the state-of-the-art (SOTA) models in terms of multiple evaluation metrics. Furthermore, we utilized DiffDTM to generate molecules towards dopamine receptor D2 and 5-hydroxytryptamine receptor 1A as new antipsychotics. The experimental results indicate that DiffDTM can be easily plugged into unseen dual targets to generate bioactive molecules, addressing the issues of requiring insufficient active molecule data for training as well as the need to retrain when encountering new targets.

Viaarxiv icon

One-Bit Spectrum Sensing for Cognitive Radio

Jun 23, 2023
Pei-Wen Wu, Lei Huang, David Ramírez, Yu-Hang Xiao, Hing Cheung So

Figure 1 for One-Bit Spectrum Sensing for Cognitive Radio
Figure 2 for One-Bit Spectrum Sensing for Cognitive Radio
Figure 3 for One-Bit Spectrum Sensing for Cognitive Radio
Figure 4 for One-Bit Spectrum Sensing for Cognitive Radio

Spectrum sensing in cognitive radio necessitates effective monitoring of wide bandwidths, which requires high-rate sampling. Traditional spectrum sensing methods employing high-precision analog-to-digital converters (ADCs) result in increased power consumption and expensive hardware costs. In this paper, we explore blind spectrum sensing utilizing one-bit ADCs. We derive a closed-form detector based on Rao's test and demonstrate its equivalence with the second-order eigenvalue-moment-ratio test. Furthermore, a near-exact distribution based on the moment-based method, and an approximate distribution in the low signal-to-noise ratio (SNR) regime with the use of the central limit theorem, are obtained. Theoretical analysis is then performed and our results show that the performance loss of the proposed detector is approximately $2$ dB ($\pi/2$) compared to detectors employing $\infty$-bit ADCs when SNR is low. This loss can be compensated for by using approximately $2.47$ ($\pi^2/4$) times more samples. In addition, we unveil that the efficiency of incoherent accumulation in one-bit detection is the square root of that of coherent accumulation. Simulation results corroborate the correctness of our theoretical calculations.

Viaarxiv icon

Modulate Your Spectrum in Self-Supervised Learning

May 26, 2023
Xi Weng, Yunhao Ni, Tengwei Song, Jie Luo, Rao Muhammad Anwer, Salman Khan, Fahad Shahbaz Khan, Lei Huang

Figure 1 for Modulate Your Spectrum in Self-Supervised Learning
Figure 2 for Modulate Your Spectrum in Self-Supervised Learning
Figure 3 for Modulate Your Spectrum in Self-Supervised Learning
Figure 4 for Modulate Your Spectrum in Self-Supervised Learning

Whitening loss provides theoretical guarantee in avoiding feature collapse for self-supervised learning (SSL) using joint embedding architectures. One typical implementation of whitening loss is hard whitening that designs whitening transformation over embedding and imposes the loss on the whitened output. In this paper, we propose spectral transformation (ST) framework to map the spectrum of embedding to a desired distribution during forward pass, and to modulate the spectrum of embedding by implicit gradient update during backward pass. We show that whitening transformation is a special instance of ST by definition, and there exist other instances that can avoid collapse by our empirical investigation. Furthermore, we propose a new instance of ST, called IterNorm with trace loss (INTL). We theoretically prove that INTL can avoid collapse and modulate the spectrum of embedding towards an equal-eigenvalue distribution during the course of optimization. Moreover, INTL achieves 76.6% top-1 accuracy in linear evaluation on ImageNet using ResNet-50, which exceeds the performance of the supervised baseline, and this result is obtained by using a batch size of only 256. Comprehensive experiments show that INTL is a promising SSL method in practice. The code is available at https://github.com/winci-ai/intl.

Viaarxiv icon

A Lightweight Recurrent Learning Network for Sustainable Compressed Sensing

Apr 23, 2023
Yu Zhou, Yu Chen, Xiao Zhang, Pan Lai, Lei Huang, Jianmin Jiang

Figure 1 for A Lightweight Recurrent Learning Network for Sustainable Compressed Sensing
Figure 2 for A Lightweight Recurrent Learning Network for Sustainable Compressed Sensing
Figure 3 for A Lightweight Recurrent Learning Network for Sustainable Compressed Sensing
Figure 4 for A Lightweight Recurrent Learning Network for Sustainable Compressed Sensing

Recently, deep learning-based compressed sensing (CS) has achieved great success in reducing the sampling and computational cost of sensing systems and improving the reconstruction quality. These approaches, however, largely overlook the issue of the computational cost; they rely on complex structures and task-specific operator designs, resulting in extensive storage and high energy consumption in CS imaging systems. In this paper, we propose a lightweight but effective deep neural network based on recurrent learning to achieve a sustainable CS system; it requires a smaller number of parameters but obtains high-quality reconstructions. Specifically, our proposed network consists of an initial reconstruction sub-network and a residual reconstruction sub-network. While the initial reconstruction sub-network has a hierarchical structure to progressively recover the image, reducing the number of parameters, the residual reconstruction sub-network facilitates recurrent residual feature extraction via recurrent learning to perform both feature fusion and deep reconstructions across different scales. In addition, we also demonstrate that, after the initial reconstruction, feature maps with reduced sizes are sufficient to recover the residual information, and thus we achieved a significant reduction in the amount of memory required. Extensive experiments illustrate that our proposed model can achieve a better reconstruction quality than existing state-of-the-art CS algorithms, and it also has a smaller number of network parameters than these algorithms. Our source codes are available at: https://github.com/C66YU/CSRN.

* has been accepted to IEEE TETCI 
Viaarxiv icon

One-Bit Covariance Reconstruction with Non-zero Thresholds: Algorithm and Performance Analysis

Mar 29, 2023
Yu-Hang Xiao, Lei Huang, David Ramírez, Cheng Qian, Hing Cheung So

Figure 1 for One-Bit Covariance Reconstruction with Non-zero Thresholds: Algorithm and Performance Analysis
Figure 2 for One-Bit Covariance Reconstruction with Non-zero Thresholds: Algorithm and Performance Analysis
Figure 3 for One-Bit Covariance Reconstruction with Non-zero Thresholds: Algorithm and Performance Analysis
Figure 4 for One-Bit Covariance Reconstruction with Non-zero Thresholds: Algorithm and Performance Analysis

Covariance matrix reconstruction is a topic of great significance in the field of one-bit signal processing and has numerous practical applications. Despite its importance, the conventional arcsine law with zero threshold is incapable of recovering the diagonal elements of the covariance matrix. To address this limitation, recent studies have proposed the use of non-zero clipping thresholds. However, the relationship between the estimation error and the sampling threshold is not yet known. In this paper, we undertake an analysis of the mean squared error by computing the Fisher information matrix for a given threshold. Our results reveal that the optimal threshold can vary considerably, depending on the variances and correlation coefficients. As a result, it is inappropriate to use a constant threshold to encompass parameters that vary widely. To mitigate this issue, we present a recovery scheme that incorporates time-varying thresholds. Our approach differs from existing methods in that it utilizes the exact values of the threshold, rather than its statistical properties, to enhance the estimation performance. Our simulations, including the direction-of-arrival estimation problem, demonstrate the efficacy of the developed scheme, especially in complex scenarios where the covariance elements are widely separated.

Viaarxiv icon

WHC: Weighted Hybrid Criterion for Filter Pruning on Convolutional Neural Networks

Feb 16, 2023
Shaowu Chen, Weize Sun, Lei Huang

Figure 1 for WHC: Weighted Hybrid Criterion for Filter Pruning on Convolutional Neural Networks
Figure 2 for WHC: Weighted Hybrid Criterion for Filter Pruning on Convolutional Neural Networks
Figure 3 for WHC: Weighted Hybrid Criterion for Filter Pruning on Convolutional Neural Networks
Figure 4 for WHC: Weighted Hybrid Criterion for Filter Pruning on Convolutional Neural Networks

Filter pruning has attracted increasing attention in recent years for its capacity in compressing and accelerating convolutional neural networks. Various data-independent criteria, including norm-based and relationship-based ones, were proposed to prune the most unimportant filters. However, these state-of-the-art criteria fail to fully consider the dissimilarity of filters, and thus might lead to performance degradation. In this paper, we first analyze the limitation of relationship-based criteria with examples, and then introduce a new data-independent criterion, Weighted Hybrid Criterion (WHC), to tackle the problems of both norm-based and relationship-based criteria. By taking the magnitude of each filter and the linear dependence between filters into consideration, WHC can robustly recognize the most redundant filters, which can be safely pruned without introducing severe performance degradation to networks. Extensive pruning experiments in a simple one-shot manner demonstrate the effectiveness of the proposed WHC. In particular, WHC can prune ResNet-50 on ImageNet with more than 42% of floating point operations reduced without any performance loss in top-5 accuracy.

* Accepted by ICASSP 2023 
Viaarxiv icon

Scale-Semantic Joint Decoupling Network for Image-text Retrieval in Remote Sensing

Dec 12, 2022
Chengyu Zheng, Ning song, Ruoyu Zhang, Lei Huang, Zhiqiang Wei, Jie Nie

Figure 1 for Scale-Semantic Joint Decoupling Network for Image-text Retrieval in Remote Sensing
Figure 2 for Scale-Semantic Joint Decoupling Network for Image-text Retrieval in Remote Sensing
Figure 3 for Scale-Semantic Joint Decoupling Network for Image-text Retrieval in Remote Sensing
Figure 4 for Scale-Semantic Joint Decoupling Network for Image-text Retrieval in Remote Sensing

Image-text retrieval in remote sensing aims to provide flexible information for data analysis and application. In recent years, state-of-the-art methods are dedicated to ``scale decoupling'' and ``semantic decoupling'' strategies to further enhance the capability of representation. However, these previous approaches focus on either the disentangling scale or semantics but ignore merging these two ideas in a union model, which extremely limits the performance of cross-modal retrieval models. To address these issues, we propose a novel Scale-Semantic Joint Decoupling Network (SSJDN) for remote sensing image-text retrieval. Specifically, we design the Bidirectional Scale Decoupling (BSD) module, which exploits Salience Feature Extraction (SFE) and Salience-Guided Suppression (SGS) units to adaptively extract potential features and suppress cumbersome features at other scales in a bidirectional pattern to yield different scale clues. Besides, we design the Label-supervised Semantic Decoupling (LSD) module by leveraging the category semantic labels as prior knowledge to supervise images and texts probing significant semantic-related information. Finally, we design a Semantic-guided Triple Loss (STL), which adaptively generates a constant to adjust the loss function to improve the probability of matching the same semantic image and text and shorten the convergence time of the retrieval model. Our proposed SSJDN outperforms state-of-the-art approaches in numerical experiments conducted on four benchmark remote sensing datasets.

Viaarxiv icon

FIXED: Frustratingly Easy Domain Generalization with Mixup

Nov 07, 2022
Wang Lu, Jindong Wang, Han Yu, Lei Huang, Xiang Zhang, Yiqiang Chen, Xing Xie

Figure 1 for FIXED: Frustratingly Easy Domain Generalization with Mixup
Figure 2 for FIXED: Frustratingly Easy Domain Generalization with Mixup
Figure 3 for FIXED: Frustratingly Easy Domain Generalization with Mixup
Figure 4 for FIXED: Frustratingly Easy Domain Generalization with Mixup

Domain generalization (DG) aims to learn a generalizable model from multiple training domains such that it can perform well on unseen target domains. A popular strategy is to augment training data to benefit generalization through methods such as Mixup~\cite{zhang2018mixup}. While the vanilla Mixup can be directly applied, theoretical and empirical investigations uncover several shortcomings that limit its performance. Firstly, Mixup cannot effectively identify the domain and class information that can be used for learning invariant representations. Secondly, Mixup may introduce synthetic noisy data points via random interpolation, which lowers its discrimination capability. Based on the analysis, we propose a simple yet effective enhancement for Mixup-based DG, namely domain-invariant Feature mIXup (FIX). It learns domain-invariant representations for Mixup. To further enhance discrimination, we leverage existing techniques to enlarge margins among classes to further propose the domain-invariant Feature MIXup with Enhanced Discrimination (FIXED) approach. We present theoretical insights about guarantees on its effectiveness. Extensive experiments on seven public datasets across two modalities including image classification (Digits-DG, PACS, Office-Home) and time series (DSADS, PAMAP2, UCI-HAR, and USC-HAD) demonstrate that our approach significantly outperforms nine state-of-the-art related methods, beating the best performing baseline by 6.5\% on average in terms of test accuracy.

* Technical report; code for DG at: https://github.com/jindongwang/transferlearning/tree/master/code/DeepDG 
Viaarxiv icon