Picture for Liqiang Wang

Liqiang Wang

Trace-Norm Adversarial Examples

Add code
Jul 02, 2020
Figure 1 for Trace-Norm Adversarial Examples
Figure 2 for Trace-Norm Adversarial Examples
Figure 3 for Trace-Norm Adversarial Examples
Figure 4 for Trace-Norm Adversarial Examples
Viaarxiv icon

Neural Networks Are More Productive Teachers Than Human Raters: Active Mixup for Data-Efficient Knowledge Distillation from a Blackbox Model

Add code
Mar 31, 2020
Figure 1 for Neural Networks Are More Productive Teachers Than Human Raters: Active Mixup for Data-Efficient Knowledge Distillation from a Blackbox Model
Figure 2 for Neural Networks Are More Productive Teachers Than Human Raters: Active Mixup for Data-Efficient Knowledge Distillation from a Blackbox Model
Figure 3 for Neural Networks Are More Productive Teachers Than Human Raters: Active Mixup for Data-Efficient Knowledge Distillation from a Blackbox Model
Figure 4 for Neural Networks Are More Productive Teachers Than Human Raters: Active Mixup for Data-Efficient Knowledge Distillation from a Blackbox Model
Viaarxiv icon

BachGAN: High-Resolution Image Synthesis from Salient Object Layout

Add code
Mar 27, 2020
Figure 1 for BachGAN: High-Resolution Image Synthesis from Salient Object Layout
Figure 2 for BachGAN: High-Resolution Image Synthesis from Salient Object Layout
Figure 3 for BachGAN: High-Resolution Image Synthesis from Salient Object Layout
Figure 4 for BachGAN: High-Resolution Image Synthesis from Salient Object Layout
Viaarxiv icon

Rethinking Class-Balanced Methods for Long-Tailed Visual Recognition from a Domain Adaptation Perspective

Add code
Mar 24, 2020
Figure 1 for Rethinking Class-Balanced Methods for Long-Tailed Visual Recognition from a Domain Adaptation Perspective
Figure 2 for Rethinking Class-Balanced Methods for Long-Tailed Visual Recognition from a Domain Adaptation Perspective
Figure 3 for Rethinking Class-Balanced Methods for Long-Tailed Visual Recognition from a Domain Adaptation Perspective
Figure 4 for Rethinking Class-Balanced Methods for Long-Tailed Visual Recognition from a Domain Adaptation Perspective
Viaarxiv icon

Self-supervised learning for audio-visual speaker diarization

Add code
Feb 13, 2020
Figure 1 for Self-supervised learning for audio-visual speaker diarization
Figure 2 for Self-supervised learning for audio-visual speaker diarization
Figure 3 for Self-supervised learning for audio-visual speaker diarization
Figure 4 for Self-supervised learning for audio-visual speaker diarization
Viaarxiv icon

AdaFilter: Adaptive Filter Fine-tuning for Deep Transfer Learning

Add code
Dec 09, 2019
Figure 1 for AdaFilter: Adaptive Filter Fine-tuning for Deep Transfer Learning
Figure 2 for AdaFilter: Adaptive Filter Fine-tuning for Deep Transfer Learning
Figure 3 for AdaFilter: Adaptive Filter Fine-tuning for Deep Transfer Learning
Figure 4 for AdaFilter: Adaptive Filter Fine-tuning for Deep Transfer Learning
Viaarxiv icon

Defending Against Adversarial Attacks Using Random Forests

Add code
Jun 16, 2019
Figure 1 for Defending Against Adversarial Attacks Using Random Forests
Figure 2 for Defending Against Adversarial Attacks Using Random Forests
Figure 3 for Defending Against Adversarial Attacks Using Random Forests
Figure 4 for Defending Against Adversarial Attacks Using Random Forests
Viaarxiv icon

Robust Sparse Regularization: Simultaneously Optimizing Neural Network Robustness and Compactness

Add code
May 30, 2019
Figure 1 for Robust Sparse Regularization: Simultaneously Optimizing Neural Network Robustness and Compactness
Figure 2 for Robust Sparse Regularization: Simultaneously Optimizing Neural Network Robustness and Compactness
Figure 3 for Robust Sparse Regularization: Simultaneously Optimizing Neural Network Robustness and Compactness
Figure 4 for Robust Sparse Regularization: Simultaneously Optimizing Neural Network Robustness and Compactness
Viaarxiv icon

NATTACK: Learning the Distributions of Adversarial Examples for an Improved Black-Box Attack on Deep Neural Networks

Add code
May 13, 2019
Figure 1 for NATTACK: Learning the Distributions of Adversarial Examples for an Improved Black-Box Attack on Deep Neural Networks
Figure 2 for NATTACK: Learning the Distributions of Adversarial Examples for an Improved Black-Box Attack on Deep Neural Networks
Figure 3 for NATTACK: Learning the Distributions of Adversarial Examples for an Improved Black-Box Attack on Deep Neural Networks
Viaarxiv icon

Frame-Recurrent Video Inpainting by Robust Optical Flow Inference

Add code
May 08, 2019
Figure 1 for Frame-Recurrent Video Inpainting by Robust Optical Flow Inference
Figure 2 for Frame-Recurrent Video Inpainting by Robust Optical Flow Inference
Figure 3 for Frame-Recurrent Video Inpainting by Robust Optical Flow Inference
Figure 4 for Frame-Recurrent Video Inpainting by Robust Optical Flow Inference
Viaarxiv icon