Alert button
Picture for Weijie Lyu

Weijie Lyu

Alert button

Continual Learning in Open-vocabulary Classification with Complementary Memory Systems

Jul 04, 2023
Zhen Zhu, Weijie Lyu, Yao Xiao, Derek Hoiem

Figure 1 for Continual Learning in Open-vocabulary Classification with Complementary Memory Systems
Figure 2 for Continual Learning in Open-vocabulary Classification with Complementary Memory Systems
Figure 3 for Continual Learning in Open-vocabulary Classification with Complementary Memory Systems
Figure 4 for Continual Learning in Open-vocabulary Classification with Complementary Memory Systems

We introduce a method for flexible continual learning in open-vocabulary image classification, drawing inspiration from the complementary learning systems observed in human cognition. We propose a "tree probe" method, an adaption of lazy learning principles, which enables fast learning from new examples with competitive accuracy to batch-trained linear models. Further, we propose a method to combine predictions from a CLIP zero-shot model and the exemplar-based model, using the zero-shot estimated probability that a sample's class is within any of the exemplar classes. We test in data incremental, class incremental, and task incremental settings, as well as ability to perform flexible inference on varying subsets of zero-shot and learned categories. Our proposed method achieves a good balance of learning speed, target task effectiveness, and zero-shot effectiveness.

* In review 
Viaarxiv icon

Consistent Multimodal Generation via A Unified GAN Framework

Jul 04, 2023
Zhen Zhu, Yijun Li, Weijie Lyu, Krishna Kumar Singh, Zhixin Shu, Soeren Pirk, Derek Hoiem

Figure 1 for Consistent Multimodal Generation via A Unified GAN Framework
Figure 2 for Consistent Multimodal Generation via A Unified GAN Framework
Figure 3 for Consistent Multimodal Generation via A Unified GAN Framework
Figure 4 for Consistent Multimodal Generation via A Unified GAN Framework

We investigate how to generate multimodal image outputs, such as RGB, depth, and surface normals, with a single generative model. The challenge is to produce outputs that are realistic, and also consistent with each other. Our solution builds on the StyleGAN3 architecture, with a shared backbone and modality-specific branches in the last layers of the synthesis network, and we propose per-modality fidelity discriminators and a cross-modality consistency discriminator. In experiments on the Stanford2D3D dataset, we demonstrate realistic and consistent generation of RGB, depth, and normal images. We also show a training recipe to easily extend our pretrained model on a new domain, even with a few pairwise data. We further evaluate the use of synthetically generated RGB and depth pairs for training or fine-tuning depth estimators. Code will be available at https://github.com/jessemelpolio/MultimodalGAN.

* In review 
Viaarxiv icon

SafeBench: A Benchmarking Platform for Safety Evaluation of Autonomous Vehicles

Jun 20, 2022
Chejian Xu, Wenhao Ding, Weijie Lyu, Zuxin Liu, Shuai Wang, Yihan He, Hanjiang Hu, Ding Zhao, Bo Li

Figure 1 for SafeBench: A Benchmarking Platform for Safety Evaluation of Autonomous Vehicles
Figure 2 for SafeBench: A Benchmarking Platform for Safety Evaluation of Autonomous Vehicles
Figure 3 for SafeBench: A Benchmarking Platform for Safety Evaluation of Autonomous Vehicles
Figure 4 for SafeBench: A Benchmarking Platform for Safety Evaluation of Autonomous Vehicles

As shown by recent studies, machine intelligence-enabled systems are vulnerable to test cases resulting from either adversarial manipulation or natural distribution shifts. This has raised great concerns about deploying machine learning algorithms for real-world applications, especially in the safety-critical domains such as autonomous driving (AD). On the other hand, traditional AD testing on naturalistic scenarios requires hundreds of millions of driving miles due to the high dimensionality and rareness of the safety-critical scenarios in the real world. As a result, several approaches for autonomous driving evaluation have been explored, which are usually, however, based on different simulation platforms, types of safety-critical scenarios, scenario generation algorithms, and driving route variations. Thus, despite a large amount of effort in autonomous driving testing, it is still challenging to compare and understand the effectiveness and efficiency of different testing scenario generation algorithms and testing mechanisms under similar conditions. In this paper, we aim to provide the first unified platform SafeBench to integrate different types of safety-critical testing scenarios, scenario generation algorithms, and other variations such as driving routes and environments. Meanwhile, we implement 4 deep reinforcement learning-based AD algorithms with 4 types of input (e.g., bird's-eye view, camera) to perform fair comparisons on SafeBench. We find our generated testing scenarios are indeed more challenging and observe the trade-off between the performance of AD agents under benign and safety-critical testing scenarios. We believe our unified platform SafeBench for large-scale and effective autonomous driving testing will motivate the development of new testing scenario generation and safe AD algorithms. SafeBench is available at https://safebench.github.io.

Viaarxiv icon

CircuitBot: Learning to Survive with Robotic Circuit Drawing

Nov 10, 2020
Xianglong Tan, Weijie Lyu, Andre Rosendo

Figure 1 for CircuitBot: Learning to Survive with Robotic Circuit Drawing
Figure 2 for CircuitBot: Learning to Survive with Robotic Circuit Drawing
Figure 3 for CircuitBot: Learning to Survive with Robotic Circuit Drawing
Figure 4 for CircuitBot: Learning to Survive with Robotic Circuit Drawing

Robots with the ability to actively acquire power from surroundings will be greatly beneficial for long-term autonomy, and to survive in dynamic, uncertain environments. In this work, a scenario is presented where a robot has limited energy, and the only way to survive is to access the energy from a power source. With no cables or wires available, the robot learns to construct an electrical path and avoid potential obstacles during the connection. We present this robot, capable of drawing connected circuit patterns with graphene-based conductive ink. A state-of-the-art Mix-Variable Bayesian Optimization is adopted to optimize the placement of conductive shapes to maximize the power this robot receives. Our results show that, within a small number of trials, the robot learns to build parallel circuits to maximize the voltage received and avoid obstacles which steal energy from the robot.

Viaarxiv icon