Alert button
Picture for Jack Saunders

Jack Saunders

Alert button

FACTS: Facial Animation Creation using the Transfer of Styles

Jul 18, 2023
Jack Saunders, Steven Caulkin, Vinay Namboodiri

Figure 1 for FACTS: Facial Animation Creation using the Transfer of Styles
Figure 2 for FACTS: Facial Animation Creation using the Transfer of Styles
Figure 3 for FACTS: Facial Animation Creation using the Transfer of Styles
Figure 4 for FACTS: Facial Animation Creation using the Transfer of Styles

The ability to accurately capture and express emotions is a critical aspect of creating believable characters in video games and other forms of entertainment. Traditionally, this animation has been achieved with artistic effort or performance capture, both requiring costs in time and labor. More recently, audio-driven models have seen success, however, these often lack expressiveness in areas not correlated to the audio signal. In this paper, we present a novel approach to facial animation by taking existing animations and allowing for the modification of style characteristics. Specifically, we explore the use of a StarGAN to enable the conversion of 3D facial animations into different emotions and person-specific styles. We are able to maintain the lip-sync of the animations with this method thanks to the use of a novel viseme-preserving loss.

Viaarxiv icon

Resource-Constrained Station-Keeping for Helium Balloons using Reinforcement Learning

Mar 02, 2023
Jack Saunders, Loïc Prenevost, Özgür Şimşek, Alan Hunter, Wenbin Li

Figure 1 for Resource-Constrained Station-Keeping for Helium Balloons using Reinforcement Learning
Figure 2 for Resource-Constrained Station-Keeping for Helium Balloons using Reinforcement Learning
Figure 3 for Resource-Constrained Station-Keeping for Helium Balloons using Reinforcement Learning
Figure 4 for Resource-Constrained Station-Keeping for Helium Balloons using Reinforcement Learning

High altitude balloons have proved useful for ecological aerial surveys, atmospheric monitoring, and communication relays. However, due to weight and power constraints, there is a need to investigate alternate modes of propulsion to navigate in the stratosphere. Very recently, reinforcement learning has been proposed as a control scheme to maintain the balloon in the region of a fixed location, facilitated through diverse opposing wind-fields at different altitudes. Although air-pump based station keeping has been explored, there is no research on the control problem for venting and ballasting actuated balloons, which is commonly used as a low-cost alternative. We show how reinforcement learning can be used for this type of balloon. Specifically, we use the soft actor-critic algorithm, which on average is able to station-keep within 50\;km for 25\% of the flight, consistent with state-of-the-art. Furthermore, we show that the proposed controller effectively minimises the consumption of resources, thereby supporting long duration flights. We frame the controller as a continuous control reinforcement learning problem, which allows for a more diverse range of trajectories, as opposed to current state-of-the-art work, which uses discrete action spaces. Furthermore, through continuous control, we can make use of larger ascent rates which are not possible using air-pumps. The desired ascent-rate is decoupled into desired altitude and time-factor to provide a more transparent policy, compared to low-level control commands used in previous works. Finally, by applying the equations of motion, we establish appropriate thresholds for venting and ballasting to prevent the agent from exploiting the environment. More specifically, we ensure actions are physically feasible by enforcing constraints on venting and ballasting.

Viaarxiv icon

READ Avatars: Realistic Emotion-controllable Audio Driven Avatars

Mar 01, 2023
Jack Saunders, Vinay Namboodiri

Figure 1 for READ Avatars: Realistic Emotion-controllable Audio Driven Avatars
Figure 2 for READ Avatars: Realistic Emotion-controllable Audio Driven Avatars
Figure 3 for READ Avatars: Realistic Emotion-controllable Audio Driven Avatars
Figure 4 for READ Avatars: Realistic Emotion-controllable Audio Driven Avatars

We present READ Avatars, a 3D-based approach for generating 2D avatars that are driven by audio input with direct and granular control over the emotion. Previous methods are unable to achieve realistic animation due to the many-to-many nature of audio to expression mappings. We alleviate this issue by introducing an adversarial loss in the audio-to-expression generation process. This removes the smoothing effect of regression-based models and helps to improve the realism and expressiveness of the generated avatars. We note furthermore, that audio should be directly utilized when generating mouth interiors and that other 3D-based methods do not attempt this. We address this with audio-conditioned neural textures, which are resolution-independent. To evaluate the performance of our method, we perform quantitative and qualitative experiments, including a user study. We also propose a new metric for comparing how well an actor's emotion is reconstructed in the generated avatar. Our results show that our approach outperforms state of the art audio-driven avatar generation methods across several metrics. A demo video can be found at \url{https://youtu.be/QSyMl3vV0pA}

* 13 Pages, 8 Figures For demo video see https://youtu.be/QSyMl3vV0pA 
Viaarxiv icon

Parallel Reinforcement Learning Simulation for Visual Quadrotor Navigation

Sep 22, 2022
Jack Saunders, Sajad Saeedi, Wenbin Li

Figure 1 for Parallel Reinforcement Learning Simulation for Visual Quadrotor Navigation
Figure 2 for Parallel Reinforcement Learning Simulation for Visual Quadrotor Navigation
Figure 3 for Parallel Reinforcement Learning Simulation for Visual Quadrotor Navigation
Figure 4 for Parallel Reinforcement Learning Simulation for Visual Quadrotor Navigation

Reinforcement learning (RL) is an agent-based approach for teaching robots to navigate within the physical world. Gathering data for RL is known to be a laborious task, and real-world experiments can be risky. Simulators facilitate the collection of training data in a quicker and more cost-effective manner. However, RL frequently requires a significant number of simulation steps for an agent to become skilful at simple tasks. This is a prevalent issue within the field of RL-based visual quadrotor navigation where state dimensions are typically very large and dynamic models are complex. Furthermore, rendering images and obtaining physical properties of the agent can be computationally expensive. To solve this, we present a simulation framework, built on AirSim, which provides efficient parallel training. Building on this framework, Ape-X is modified to incorporate decentralised training of AirSim environments to make use of numerous networked computers. Through experiments we were able to achieve a reduction in training time from 3.9 hours to 11 minutes using the aforementioned framework and a total of 74 agents and two networked computers. Further details including a github repo and videos about our project, PRL4AirSim, can be found at https://sites.google.com/view/prl4airsim/home

* This work has been submitted to the IEEE International Conference on Robotics and Automation (ICRA) for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessible 
Viaarxiv icon

Autonomous Aerial Delivery Vehicles, a Survey of Techniques on how Aerial Package Delivery is Achieved

Oct 06, 2021
Jack Saunders, Sajad Saeedi, Wenbin Li

Figure 1 for Autonomous Aerial Delivery Vehicles, a Survey of Techniques on how Aerial Package Delivery is Achieved
Figure 2 for Autonomous Aerial Delivery Vehicles, a Survey of Techniques on how Aerial Package Delivery is Achieved
Figure 3 for Autonomous Aerial Delivery Vehicles, a Survey of Techniques on how Aerial Package Delivery is Achieved
Figure 4 for Autonomous Aerial Delivery Vehicles, a Survey of Techniques on how Aerial Package Delivery is Achieved

Autonomous aerial delivery vehicles have gained significant interest in the last decade. This has been enabled by technological advancements in aerial manipulators and novel grippers with enhanced force to weight ratios. Furthermore, improved control schemes and vehicle dynamics are better able to model the payload and improved perception algorithms to detect key features within the unmanned aerial vehicle's (UAV) environment. In this survey, a systematic review of the technological advancements and open research problems of autonomous aerial delivery vehicles is conducted. First, various types of manipulators and grippers are discussed in detail, along with dynamic modelling and control methods. Then, landing on static and dynamic platforms is discussed. Subsequently, risks such as weather conditions, state estimation and collision avoidance to ensure safe transit is considered. Finally, delivery UAV routing is investigated which categorises the topic into two areas: drone operations and drone-truck collaborative operations.

* Submitted for review in the Journal of Field Robotics 
Viaarxiv icon