Alert button
Picture for Ngai Wong

Ngai Wong

Alert button

A Spectral Perspective towards Understanding and Improving Adversarial Robustness

Jun 25, 2023
Binxiao Huang, Rui Lin, Chaofan Tao, Ngai Wong

Figure 1 for A Spectral Perspective towards Understanding and Improving Adversarial Robustness
Figure 2 for A Spectral Perspective towards Understanding and Improving Adversarial Robustness
Figure 3 for A Spectral Perspective towards Understanding and Improving Adversarial Robustness
Figure 4 for A Spectral Perspective towards Understanding and Improving Adversarial Robustness

Deep neural networks (DNNs) are incredibly vulnerable to crafted, imperceptible adversarial perturbations. While adversarial training (AT) has proven to be an effective defense approach, the AT mechanism for robustness improvement is not fully understood. This work investigates AT from a spectral perspective, adding new insights to the design of effective defenses. In particular, we show that AT induces the deep model to focus more on the low-frequency region, which retains the shape-biased representations, to gain robustness. Further, we find that the spectrum of a white-box attack is primarily distributed in regions the model focuses on, and the perturbation attacks the spectral bands where the model is vulnerable. Based on this observation, to train a model tolerant to frequency-varying perturbation, we propose a spectral alignment regularization (SAR) such that the spectral output inferred by an attacked adversarial input stays as close as possible to its natural input counterpart. Experiments demonstrate that SAR and its weight averaging (WA) extension could significantly improve the robust accuracy by 1.14% ~ 3.87% relative to the standard AT, across multiple datasets (CIFAR-10, CIFAR-100 and Tiny ImageNet), and various attacks (PGD, C&W and Autoattack), without any extra data.

Viaarxiv icon

Overcoming Beam Squint in Dual-Wideband mmWave MIMO Channel Estimation: A Bayesian Multi-Band Sparsity Approach

Jun 19, 2023
Le Xu, Lei Cheng, Ngai Wong, Yik-Chung Wu, H. Vincent Poor

Figure 1 for Overcoming Beam Squint in Dual-Wideband mmWave MIMO Channel Estimation: A Bayesian Multi-Band Sparsity Approach
Figure 2 for Overcoming Beam Squint in Dual-Wideband mmWave MIMO Channel Estimation: A Bayesian Multi-Band Sparsity Approach
Figure 3 for Overcoming Beam Squint in Dual-Wideband mmWave MIMO Channel Estimation: A Bayesian Multi-Band Sparsity Approach
Figure 4 for Overcoming Beam Squint in Dual-Wideband mmWave MIMO Channel Estimation: A Bayesian Multi-Band Sparsity Approach

The beam squint effect, which manifests in different steering matrices in different sub-bands, has been widely considered a challenge in millimeter wave (mmWave) multiinput multi-output (MIMO) channel estimation. Existing methods either require specific forms of the precoding/combining matrix, which restrict their general practicality, or simply ignore the beam squint effect by only making use of a single sub-band for channel estimation. Recognizing that different steering matrices are coupled by the same set of unknown channel parameters, this paper proposes to exploit the common sparsity structure of the virtual channel model so that signals from different subbands can be jointly utilized to enhance the performance of channel estimation. A probabilistic model is built to induce the common sparsity in the spatial domain, and the first-order Taylor expansion is adopted to get rid of the grid mismatch in the dictionaries. To learn the model parameters, a variational expectation-maximization (EM) algorithm is derived, which automatically obtains the balance between the likelihood function and the common sparsity prior information, and is applicable to arbitrary forms of precoding/combining matrices. Simulation results show the superior estimation accuracy of the proposed algorithm over existing methods under different noise powers and system configurations.

Viaarxiv icon

To Fold or Not to Fold: Graph Regularized Tensor Train for Visual Data Completion

Jun 19, 2023
Le Xu, Lei Cheng, Ngai Wong, Yik-Chung Wu

Figure 1 for To Fold or Not to Fold: Graph Regularized Tensor Train for Visual Data Completion
Figure 2 for To Fold or Not to Fold: Graph Regularized Tensor Train for Visual Data Completion
Figure 3 for To Fold or Not to Fold: Graph Regularized Tensor Train for Visual Data Completion
Figure 4 for To Fold or Not to Fold: Graph Regularized Tensor Train for Visual Data Completion

Tensor train (TT) representation has achieved tremendous success in visual data completion tasks, especially when it is combined with tensor folding. However, folding an image or video tensor breaks the original data structure, leading to local information loss as nearby pixels may be assigned into different dimensions and become far away from each other. In this paper, to fully preserve the local information of the original visual data, we explore not folding the data tensor, and at the same time adopt graph information to regularize local similarity between nearby entries. To overcome the high computational complexity introduced by the graph-based regularization in the TT completion problem, we propose to break the original problem into multiple sub-problems with respect to each TT core fiber, instead of each TT core as in traditional methods. Furthermore, to avoid heavy parameter tuning, a sparsity promoting probabilistic model is built based on the generalized inverse Gaussian (GIG) prior, and an inference algorithm is derived under the mean-field approximation. Experiments on both synthetic data and real-world visual data show the superiority of the proposed methods.

Viaarxiv icon

Weight-Inherited Distillation for Task-Agnostic BERT Compression

May 16, 2023
Taiqiang Wu, Cheng Hou, Zhe Zhao, Shanshan Lao, Jiayi Li, Ngai Wong, Yujiu Yang

Figure 1 for Weight-Inherited Distillation for Task-Agnostic BERT Compression
Figure 2 for Weight-Inherited Distillation for Task-Agnostic BERT Compression
Figure 3 for Weight-Inherited Distillation for Task-Agnostic BERT Compression
Figure 4 for Weight-Inherited Distillation for Task-Agnostic BERT Compression

Knowledge Distillation (KD) is a predominant approach for BERT compression. Previous KD-based methods focus on designing extra alignment losses for the student model to mimic the behavior of the teacher model. These methods transfer the knowledge in an indirect way. In this paper, we propose a novel Weight-Inherited Distillation (WID), which directly transfers knowledge from the teacher. WID does not require any additional alignment loss and trains a compact student by inheriting the weights, showing a new perspective of knowledge distillation. Specifically, we design the row compactors and column compactors as mappings and then compress the weights via structural re-parameterization. Experimental results on the GLUE and SQuAD benchmarks show that WID outperforms previous state-of-the-art KD-based baselines. Further analysis indicates that WID can also learn the attention patterns from the teacher model without any alignment loss on attention distributions.

* 9 pages, 4 figures 
Viaarxiv icon

Edge-free but Structure-aware: Prototype-Guided Knowledge Distillation from GNNs to MLPs

Mar 27, 2023
Taiqiang Wu, Zhe Zhao, Jiahao Wang, Xingyu Bai, Lei Wang, Ngai Wong, Yujiu Yang

Figure 1 for Edge-free but Structure-aware: Prototype-Guided Knowledge Distillation from GNNs to MLPs
Figure 2 for Edge-free but Structure-aware: Prototype-Guided Knowledge Distillation from GNNs to MLPs
Figure 3 for Edge-free but Structure-aware: Prototype-Guided Knowledge Distillation from GNNs to MLPs
Figure 4 for Edge-free but Structure-aware: Prototype-Guided Knowledge Distillation from GNNs to MLPs

Distilling high-accuracy Graph Neural Networks~(GNNs) to low-latency multilayer perceptrons~(MLPs) on graph tasks has become a hot research topic. However, MLPs rely exclusively on the node features and fail to capture the graph structural information. Previous methods address this issue by processing graph edges into extra inputs for MLPs, but such graph structures may be unavailable for various scenarios. To this end, we propose a Prototype-Guided Knowledge Distillation~(PGKD) method, which does not require graph edges~(edge-free) yet learns structure-aware MLPs. Specifically, we analyze the graph structural information in GNN teachers, and distill such information from GNNs to MLPs via prototypes in an edge-free setting. Experimental results on popular graph benchmarks demonstrate the effectiveness and robustness of the proposed PGKD.

* 8 pages, 4 figures, 9 tables 
Viaarxiv icon

Context-Aware Transformer for 3D Point Cloud Automatic Annotation

Mar 27, 2023
Xiaoyan Qian, Chang Liu, Xiaojuan Qi, Siew-Chong Tan, Edmund Lam, Ngai Wong

Figure 1 for Context-Aware Transformer for 3D Point Cloud Automatic Annotation
Figure 2 for Context-Aware Transformer for 3D Point Cloud Automatic Annotation
Figure 3 for Context-Aware Transformer for 3D Point Cloud Automatic Annotation
Figure 4 for Context-Aware Transformer for 3D Point Cloud Automatic Annotation

3D automatic annotation has received increased attention since manually annotating 3D point clouds is laborious. However, existing methods are usually complicated, e.g., pipelined training for 3D foreground/background segmentation, cylindrical object proposals, and point completion. Furthermore, they often overlook the inter-object feature relation that is particularly informative to hard samples for 3D annotation. To this end, we propose a simple yet effective end-to-end Context-Aware Transformer (CAT) as an automated 3D-box labeler to generate precise 3D box annotations from 2D boxes, trained with a small number of human annotations. We adopt the general encoder-decoder architecture, where the CAT encoder consists of an intra-object encoder (local) and an inter-object encoder (global), performing self-attention along the sequence and batch dimensions, respectively. The former models intra-object interactions among points, and the latter extracts feature relations among different objects, thus boosting scene-level understanding. Via local and global encoders, CAT can generate high-quality 3D box annotations with a streamlined workflow, allowing it to outperform existing state-of-the-art by up to 1.79% 3D AP on the hard task of the KITTI test set.

Viaarxiv icon

DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural Network Inference

Feb 24, 2023
Jiajun Zhou, Jiajun Wu, Yizhao Gao, Yuhao Ding, Chaofan Tao, Boyu Li, Fengbin Tu, Kwang-Ting Cheng, Hayden Kwok-Hay So, Ngai Wong

Figure 1 for DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural Network Inference
Figure 2 for DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural Network Inference
Figure 3 for DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural Network Inference
Figure 4 for DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural Network Inference

To accelerate the inference of deep neural networks (DNNs), quantization with low-bitwidth numbers is actively researched. A prominent challenge is to quantize the DNN models into low-bitwidth numbers without significant accuracy degradation, especially at very low bitwidths (< 8 bits). This work targets an adaptive data representation with variable-length encoding called DyBit. DyBit can dynamically adjust the precision and range of separate bit-field to be adapted to the DNN weights/activations distribution. We also propose a hardware-aware quantization framework with a mixed-precision accelerator to trade-off the inference accuracy and speedup. Experimental results demonstrate that the inference accuracy via DyBit is 1.997% higher than the state-of-the-art at 4-bit quantization, and the proposed framework can achieve up to 8.1x speedup compared with the original model.

Viaarxiv icon

Frequency Regularization for Improving Adversarial Robustness

Dec 24, 2022
Binxiao Huang, Chaofan Tao, Rui Lin, Ngai Wong

Figure 1 for Frequency Regularization for Improving Adversarial Robustness
Figure 2 for Frequency Regularization for Improving Adversarial Robustness
Figure 3 for Frequency Regularization for Improving Adversarial Robustness
Figure 4 for Frequency Regularization for Improving Adversarial Robustness

Deep neural networks are incredibly vulnerable to crafted, human-imperceptible adversarial perturbations. Although adversarial training (AT) has proven to be an effective defense approach, we find that the AT-trained models heavily rely on the input low-frequency content for judgment, accounting for the low standard accuracy. To close the large gap between the standard and robust accuracies during AT, we investigate the frequency difference between clean and adversarial inputs, and propose a frequency regularization (FR) to align the output difference in the spectral domain. Besides, we find Stochastic Weight Averaging (SWA), by smoothing the kernels over epochs, further improves the robustness. Among various defense schemes, our method achieves the strongest robustness against attacks by PGD-20, C\&W and Autoattack, on a WideResNet trained on CIFAR-10 without any extra data.

* accepted by AAAI 2023 workshop 
Viaarxiv icon

ODG-Q: Robust Quantization via Online Domain Generalization

Oct 17, 2022
Chaofan Tao, Ngai Wong

Figure 1 for ODG-Q: Robust Quantization via Online Domain Generalization
Figure 2 for ODG-Q: Robust Quantization via Online Domain Generalization
Figure 3 for ODG-Q: Robust Quantization via Online Domain Generalization
Figure 4 for ODG-Q: Robust Quantization via Online Domain Generalization

Quantizing neural networks to low-bitwidth is important for model deployment on resource-limited edge hardware. Although a quantized network has a smaller model size and memory footprint, it is fragile to adversarial attacks. However, few methods study the robustness and training efficiency of quantized networks. To this end, we propose a new method by recasting robust quantization as an online domain generalization problem, termed ODG-Q, which generates diverse adversarial data at a low cost during training. ODG-Q consistently outperforms existing works against various adversarial attacks. For example, on CIFAR-10 dataset, ODG-Q achieves 49.2% average improvements under five common white-box attacks and 21.7% average improvements under five common black-box attacks, with a training cost similar to that of natural training (viz. without adversaries). To our best knowledge, this work is the first work that trains both quantized and binary neural networks on ImageNet that consistently improve robustness under different attacks. We also provide a theoretical insight of ODG-Q that accounts for the bound of model risk on attacked data.

Viaarxiv icon

PECAN: A Product-Quantized Content Addressable Memory Network

Aug 13, 2022
Jie Ran, Rui Lin, Jason Chun Lok Li, Jiajun Zhou, Ngai Wong

Figure 1 for PECAN: A Product-Quantized Content Addressable Memory Network
Figure 2 for PECAN: A Product-Quantized Content Addressable Memory Network
Figure 3 for PECAN: A Product-Quantized Content Addressable Memory Network
Figure 4 for PECAN: A Product-Quantized Content Addressable Memory Network

A novel deep neural network (DNN) architecture is proposed wherein the filtering and linear transform are realized solely with product quantization (PQ). This results in a natural implementation via content addressable memory (CAM), which transcends regular DNN layer operations and requires only simple table lookup. Two schemes are developed for the end-to-end PQ prototype training, namely, through angle- and distance-based similarities, which differ in their multiplicative and additive natures with different complexity-accuracy tradeoffs. Even more, the distance-based scheme constitutes a truly multiplier-free DNN solution. Experiments confirm the feasibility of such Product-Quantized Content Addressable Memory Network (PECAN), which has strong implication on hardware-efficient deployments especially for in-memory computing.

Viaarxiv icon