Alert button
Picture for Bogdan Vlahov

Bogdan Vlahov

Alert button

A Multi-step Dynamics Modeling Framework For Autonomous Driving In Multiple Environments

May 03, 2023
Jason Gibson, Bogdan Vlahov, David Fan, Patrick Spieler, Daniel Pastor, Ali-akbar Agha-mohammadi, Evangelos A. Theodorou

Figure 1 for A Multi-step Dynamics Modeling Framework For Autonomous Driving In Multiple Environments
Figure 2 for A Multi-step Dynamics Modeling Framework For Autonomous Driving In Multiple Environments
Figure 3 for A Multi-step Dynamics Modeling Framework For Autonomous Driving In Multiple Environments
Figure 4 for A Multi-step Dynamics Modeling Framework For Autonomous Driving In Multiple Environments

Modeling dynamics is often the first step to making a vehicle autonomous. While on-road autonomous vehicles have been extensively studied, off-road vehicles pose many challenging modeling problems. An off-road vehicle encounters highly complex and difficult-to-model terrain/vehicle interactions, as well as having complex vehicle dynamics of its own. These complexities can create challenges for effective high-speed control and planning. In this paper, we introduce a framework for multistep dynamics prediction that explicitly handles the accumulation of modeling error and remains scalable for sampling-based controllers. Our method uses a specially-initialized Long Short-Term Memory (LSTM) over a limited time horizon as the learned component in a hybrid model to predict the dynamics of a 4-person seating all-terrain vehicle (Polaris S4 1000 RZR) in two distinct environments. By only having the LSTM predict over a fixed time horizon, we negate the need for long term stability that is often a challenge when training recurrent neural networks. Our framework is flexible as it only requires odometry information for labels. Through extensive experimentation, we show that our method is able to predict millions of possible trajectories in real-time, with a time horizon of five seconds in challenging off road driving scenarios.

Viaarxiv icon

Variational Inference MPC using Tsallis Divergence

Apr 01, 2021
Ziyi Wang, Oswin So, Jason Gibson, Bogdan Vlahov, Manan S. Gandhi, Guan-Horng Liu, Evangelos A. Theodorou

Figure 1 for Variational Inference MPC using Tsallis Divergence
Figure 2 for Variational Inference MPC using Tsallis Divergence
Figure 3 for Variational Inference MPC using Tsallis Divergence
Figure 4 for Variational Inference MPC using Tsallis Divergence

In this paper, we provide a generalized framework for Variational Inference-Stochastic Optimal Control by using thenon-extensive Tsallis divergence. By incorporating the deformed exponential function into the optimality likelihood function, a novel Tsallis Variational Inference-Model Predictive Control algorithm is derived, which includes prior works such as Variational Inference-Model Predictive Control, Model Predictive PathIntegral Control, Cross Entropy Method, and Stein VariationalInference Model Predictive Control as special cases. The proposed algorithm allows for effective control of the cost/reward transform and is characterized by superior performance in terms of mean and variance reduction of the associated cost. The aforementioned features are supported by a theoretical and numerical analysis on the level of risk sensitivity of the proposed algorithm as well as simulation experiments on 5 different robotic systems with 3 different policy parameterizations.

Viaarxiv icon

Approximate Inverse Reinforcement Learning from Vision-based Imitation Learning

Apr 17, 2020
Keuntaek Lee, Bogdan Vlahov, Jason Gibson, James M. Rehg, Evangelos A. Theodorou

Figure 1 for Approximate Inverse Reinforcement Learning from Vision-based Imitation Learning
Figure 2 for Approximate Inverse Reinforcement Learning from Vision-based Imitation Learning
Figure 3 for Approximate Inverse Reinforcement Learning from Vision-based Imitation Learning
Figure 4 for Approximate Inverse Reinforcement Learning from Vision-based Imitation Learning

In this work, we present a method for obtaining an implicit objective function for vision-based navigation. The proposed methodology relies on Imitation Learning, Model Predictive Control (MPC), and Deep Learning. We use Imitation Learning as a means to do Inverse Reinforcement Learning in order to create an approximate costmap generator for a visual navigation challenge. The resulting costmap is used in conjunction with a Model Predictive Controller for real-time control and outperforms other state-of-the-art costmap generators combined with MPC in novel environments. The proposed process allows for simple training and robustness to out-of-sample data. We apply our method to the task of vision-based autonomous driving in multiple real and simulated environments using the same weights for the costmap predictor in all environments.

Viaarxiv icon

Online Center of Mass Estimation for a Humanoid Wheeled Inverted Pendulum Robot

Oct 07, 2018
Munzir Zafar, Akash Patel, Bogdan Vlahov, Nathaniel Glaser, Sergio Aguillera, Seth Hutchinson

Figure 1 for Online Center of Mass Estimation for a Humanoid Wheeled Inverted Pendulum Robot
Figure 2 for Online Center of Mass Estimation for a Humanoid Wheeled Inverted Pendulum Robot
Figure 3 for Online Center of Mass Estimation for a Humanoid Wheeled Inverted Pendulum Robot
Figure 4 for Online Center of Mass Estimation for a Humanoid Wheeled Inverted Pendulum Robot

We present a novel application of robust control and online learning for the balancing of a n Degree of Freedom (DoF), Wheeled Inverted Pendulum (WIP) humanoid robot. Our technique condenses the inaccuracies of a mass model into a Center of Mass (CoM) error, balances despite this error, and uses online learning to update the mass model for a better CoM estimate. Using a simulated model of our robot, we meta-learn a set of excitory joint poses that makes our gradient descent algorithm quickly converge to an accurate (CoM) estimate. This simulated pipeline executes in a fully online fashion, using active disturbance rejection to address the mass errors that result from a steadily evolving mass model. Experiments were performed on a 19 DoF WIP, in which we manually acquired the data for the learned set of poses and show that the mass model produced by a gradient descent produces a CoM estimate that improves overall control and efficiency. This work contributes to a greater corpus of whole body control on the Golem Krang humanoid robot.

Viaarxiv icon