Alert button
Picture for André Biedenkapp

André Biedenkapp

Alert button

DeepCAVE: An Interactive Analysis Tool for Automated Machine Learning

Jun 07, 2022
René Sass, Eddie Bergman, André Biedenkapp, Frank Hutter, Marius Lindauer

Figure 1 for DeepCAVE: An Interactive Analysis Tool for Automated Machine Learning
Figure 2 for DeepCAVE: An Interactive Analysis Tool for Automated Machine Learning

Automated Machine Learning (AutoML) is used more than ever before to support users in determining efficient hyperparameters, neural architectures, or even full machine learning pipelines. However, users tend to mistrust the optimization process and its results due to a lack of transparency, making manual tuning still widespread. We introduce DeepCAVE, an interactive framework to analyze and monitor state-of-the-art optimization procedures for AutoML easily and ad hoc. By aiming for full and accessible transparency, DeepCAVE builds a bridge between users and AutoML and contributes to establishing trust. Our framework's modular and easy-to-extend nature provides users with automatically generated text, tables, and graphic visualizations. We show the value of DeepCAVE in an exemplary use-case of outlier detection, in which our framework makes it easy to identify problems, compare multiple runs and interpret optimization processes. The package is freely available on GitHub https://github.com/automl/DeepCAVE.

Viaarxiv icon

Automated Dynamic Algorithm Configuration

May 27, 2022
Steven Adriaensen, André Biedenkapp, Gresa Shala, Noor Awad, Theresa Eimer, Marius Lindauer, Frank Hutter

Figure 1 for Automated Dynamic Algorithm Configuration
Figure 2 for Automated Dynamic Algorithm Configuration
Figure 3 for Automated Dynamic Algorithm Configuration
Figure 4 for Automated Dynamic Algorithm Configuration

The performance of an algorithm often critically depends on its parameter configuration. While a variety of automated algorithm configuration methods have been proposed to relieve users from the tedious and error-prone task of manually tuning parameters, there is still a lot of untapped potential as the learned configuration is static, i.e., parameter settings remain fixed throughout the run. However, it has been shown that some algorithm parameters are best adjusted dynamically during execution, e.g., to adapt to the current part of the optimization landscape. Thus far, this is most commonly achieved through hand-crafted heuristics. A promising recent alternative is to automatically learn such dynamic parameter adaptation policies from data. In this article, we give the first comprehensive account of this new field of automated dynamic algorithm configuration (DAC), present a series of recent advances, and provide a solid foundation for future research in this field. Specifically, we (i) situate DAC in the broader historical context of AI research; (ii) formalize DAC as a computational problem; (iii) identify the methods used in prior-art to tackle this problem; (iv) conduct empirical case studies for using DAC in evolutionary optimization, AI planning, and machine learning.

Viaarxiv icon

Contextualize Me -- The Case for Context in Reinforcement Learning

Feb 09, 2022
Carolin Benjamins, Theresa Eimer, Frederik Schubert, Aditya Mohan, André Biedenkapp, Bodo Rosenhahn, Frank Hutter, Marius Lindauer

Figure 1 for Contextualize Me -- The Case for Context in Reinforcement Learning
Figure 2 for Contextualize Me -- The Case for Context in Reinforcement Learning
Figure 3 for Contextualize Me -- The Case for Context in Reinforcement Learning
Figure 4 for Contextualize Me -- The Case for Context in Reinforcement Learning

While Reinforcement Learning (RL) has made great strides towards solving increasingly complicated problems, many algorithms are still brittle to even slight changes in environments. Contextual Reinforcement Learning (cRL) provides a theoretical framework to model such changes in a principled manner, thereby enabling flexible, precise and interpretable task specification and generation. Thus, cRL is an important formalization for studying generalization in RL. In this work, we reason about solving cRL in theory and practice. We show that theoretically optimal behavior in contextual Markov Decision Processes requires explicit context information. In addition, we empirically explore context-based task generation, utilizing context information in training and propose cGate, our state-modulating policy architecture. To this end, we introduce the first benchmark library designed for generalization based on cRL extensions of popular benchmarks, CARL. In short: Context matters!

* arXiv admin note: substantial text overlap with arXiv:2110.02102 
Viaarxiv icon

Theory-inspired Parameter Control Benchmarks for Dynamic Algorithm Configuration

Feb 07, 2022
André Biedenkapp, Nguyen Dang, Martin S. Krejca, Frank Hutter, Carola Doerr

Figure 1 for Theory-inspired Parameter Control Benchmarks for Dynamic Algorithm Configuration
Figure 2 for Theory-inspired Parameter Control Benchmarks for Dynamic Algorithm Configuration
Figure 3 for Theory-inspired Parameter Control Benchmarks for Dynamic Algorithm Configuration
Figure 4 for Theory-inspired Parameter Control Benchmarks for Dynamic Algorithm Configuration

It has long been observed that the performance of evolutionary algorithms and other randomized search heuristics can benefit from a non-static choice of the parameters that steer their optimization behavior. Mechanisms that identify suitable configurations on the fly ("parameter control") or via a dedicated training process ("dynamic algorithm configuration") are therefore an important component of modern evolutionary computation frameworks. Several approaches to address the dynamic parameter setting problem exist, but we barely understand which ones to prefer for which applications. As in classical benchmarking, problem collections with a known ground truth can offer very meaningful insights in this context. Unfortunately, settings with well-understood control policies are very rare. One of the few exceptions for which we know which parameter settings minimize the expected runtime is the LeadingOnes problem. We extend this benchmark by analyzing optimal control policies that can select the parameters only from a given portfolio of possible values. This also allows us to compute optimal parameter portfolios of a given size. We demonstrate the usefulness of our benchmarks by analyzing the behavior of the DDQN reinforcement learning approach for dynamic algorithm configuration.

Viaarxiv icon

Automated Reinforcement Learning (AutoRL): A Survey and Open Problems

Jan 11, 2022
Jack Parker-Holder, Raghu Rajan, Xingyou Song, André Biedenkapp, Yingjie Miao, Theresa Eimer, Baohe Zhang, Vu Nguyen, Roberto Calandra, Aleksandra Faust, Frank Hutter, Marius Lindauer

Figure 1 for Automated Reinforcement Learning (AutoRL): A Survey and Open Problems
Figure 2 for Automated Reinforcement Learning (AutoRL): A Survey and Open Problems
Figure 3 for Automated Reinforcement Learning (AutoRL): A Survey and Open Problems
Figure 4 for Automated Reinforcement Learning (AutoRL): A Survey and Open Problems

The combination of Reinforcement Learning (RL) with deep learning has led to a series of impressive feats, with many believing (deep) RL provides a path towards generally capable agents. However, the success of RL agents is often highly sensitive to design choices in the training process, which may require tedious and error-prone manual tuning. This makes it challenging to use RL for new problems, while also limits its full potential. In many other areas of machine learning, AutoML has shown it is possible to automate such design choices and has also yielded promising initial results when applied to RL. However, Automated Reinforcement Learning (AutoRL) involves not only standard applications of AutoML but also includes additional challenges unique to RL, that naturally produce a different set of methods. As such, AutoRL has been emerging as an important area of research in RL, providing promise in a variety of applications from RNA design to playing games such as Go. Given the diversity of methods and environments considered in RL, much of the research has been conducted in distinct subfields, ranging from meta-learning to evolution. In this survey we seek to unify the field of AutoRL, we provide a common taxonomy, discuss each area in detail and pose open problems which would be of interest to researchers going forward.

Viaarxiv icon

CARL: A Benchmark for Contextual and Adaptive Reinforcement Learning

Oct 11, 2021
Carolin Benjamins, Theresa Eimer, Frederik Schubert, André Biedenkapp, Bodo Rosenhahn, Frank Hutter, Marius Lindauer

Figure 1 for CARL: A Benchmark for Contextual and Adaptive Reinforcement Learning
Figure 2 for CARL: A Benchmark for Contextual and Adaptive Reinforcement Learning
Figure 3 for CARL: A Benchmark for Contextual and Adaptive Reinforcement Learning
Figure 4 for CARL: A Benchmark for Contextual and Adaptive Reinforcement Learning

While Reinforcement Learning has made great strides towards solving ever more complicated tasks, many algorithms are still brittle to even slight changes in their environment. This is a limiting factor for real-world applications of RL. Although the research community continuously aims at improving both robustness and generalization of RL algorithms, unfortunately it still lacks an open-source set of well-defined benchmark problems based on a consistent theoretical framework, which allows comparing different approaches in a fair, reliable and reproducibleway. To fill this gap, we propose CARL, a collection of well-known RL environments extended to contextual RL problems to study generalization. We show the urgent need of such benchmarks by demonstrating that even simple toy environments become challenging for commonly used approaches if different contextual instances of this task have to be considered. Furthermore, CARL allows us to provide first evidence that disentangling representation learning of the states from the policy learning with the context facilitates better generalization. By providing variations of diverse benchmarks from classic control, physical simulations, games and a real-world application of RNA design, CARL will allow the community to derive many more such insights on a solid empirical foundation.

Viaarxiv icon

SMAC3: A Versatile Bayesian Optimization Package for Hyperparameter Optimization

Sep 20, 2021
Marius Lindauer, Katharina Eggensperger, Matthias Feurer, André Biedenkapp, Difan Deng, Carolin Benjamins, René Sass, Frank Hutter

Figure 1 for SMAC3: A Versatile Bayesian Optimization Package for Hyperparameter Optimization
Figure 2 for SMAC3: A Versatile Bayesian Optimization Package for Hyperparameter Optimization

Algorithm parameters, in particular hyperparameters of machine learning algorithms, can substantially impact their performance. To support users in determining well-performing hyperparameter configurations for their algorithms, datasets and applications at hand, SMAC3 offers a robust and flexible framework for Bayesian Optimization, which can improve performance within a few evaluations. It offers several facades and pre-sets for typical use cases, such as optimizing hyperparameters, solving low dimensional continuous (artificial) global optimization problems and configuring algorithms to perform well across multiple problem instances. The SMAC3 package is available under a permissive BSD-license at https://github.com/automl/SMAC3.

Viaarxiv icon

TempoRL: Learning When to Act

Jun 09, 2021
André Biedenkapp, Raghu Rajan, Frank Hutter, Marius Lindauer

Figure 1 for TempoRL: Learning When to Act
Figure 2 for TempoRL: Learning When to Act
Figure 3 for TempoRL: Learning When to Act
Figure 4 for TempoRL: Learning When to Act

Reinforcement learning is a powerful approach to learn behaviour through interactions with an environment. However, behaviours are usually learned in a purely reactive fashion, where an appropriate action is selected based on an observation. In this form, it is challenging to learn when it is necessary to execute new decisions. This makes learning inefficient, especially in environments that need various degrees of fine and coarse control. To address this, we propose a proactive setting in which the agent not only selects an action in a state but also for how long to commit to that action. Our TempoRL approach introduces skip connections between states and learns a skip-policy for repeating the same action along these skips. We demonstrate the effectiveness of TempoRL on a variety of traditional and deep RL environments, showing that our approach is capable of learning successful policies up to an order of magnitude faster than vanilla Q-learning.

* Accepted at ICML'21 
Viaarxiv icon

Self-Paced Context Evaluation for Contextual Reinforcement Learning

Jun 09, 2021
Theresa Eimer, André Biedenkapp, Frank Hutter, Marius Lindauer

Figure 1 for Self-Paced Context Evaluation for Contextual Reinforcement Learning
Figure 2 for Self-Paced Context Evaluation for Contextual Reinforcement Learning
Figure 3 for Self-Paced Context Evaluation for Contextual Reinforcement Learning
Figure 4 for Self-Paced Context Evaluation for Contextual Reinforcement Learning

Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach.

Viaarxiv icon

DACBench: A Benchmark Library for Dynamic Algorithm Configuration

May 18, 2021
Theresa Eimer, André Biedenkapp, Maximilian Reimer, Steven Adriaensen, Frank Hutter, Marius Lindauer

Figure 1 for DACBench: A Benchmark Library for Dynamic Algorithm Configuration
Figure 2 for DACBench: A Benchmark Library for Dynamic Algorithm Configuration
Figure 3 for DACBench: A Benchmark Library for Dynamic Algorithm Configuration
Figure 4 for DACBench: A Benchmark Library for Dynamic Algorithm Configuration

Dynamic Algorithm Configuration (DAC) aims to dynamically control a target algorithm's hyperparameters in order to improve its performance. Several theoretical and empirical results have demonstrated the benefits of dynamically controlling hyperparameters in domains like evolutionary computation, AI Planning or deep learning. Replicating these results, as well as studying new methods for DAC, however, is difficult since existing benchmarks are often specialized and incompatible with the same interfaces. To facilitate benchmarking and thus research on DAC, we propose DACBench, a benchmark library that seeks to collect and standardize existing DAC benchmarks from different AI domains, as well as provide a template for new ones. For the design of DACBench, we focused on important desiderata, such as (i) flexibility, (ii) reproducibility, (iii) extensibility and (iv) automatic documentation and visualization. To show the potential, broad applicability and challenges of DAC, we explore how a set of six initial benchmarks compare in several dimensions of difficulty.

* Accepted at IJCAI 2021 
Viaarxiv icon