Alert button
Picture for Oleg Szehr

Oleg Szehr

Alert button

Hierarchical Multi-Agent Reinforcement Learning for Air Combat Maneuvering

Add code
Bookmark button
Alert button
Sep 20, 2023
Ardian Selmonaj, Oleg Szehr, Giacomo Del Rio, Alessandro Antonucci, Adrian Schneider, Michael Rüegsegger

Figure 1 for Hierarchical Multi-Agent Reinforcement Learning for Air Combat Maneuvering
Figure 2 for Hierarchical Multi-Agent Reinforcement Learning for Air Combat Maneuvering
Figure 3 for Hierarchical Multi-Agent Reinforcement Learning for Air Combat Maneuvering
Figure 4 for Hierarchical Multi-Agent Reinforcement Learning for Air Combat Maneuvering
Viaarxiv icon

Hedging of Financial Derivative Contracts via Monte Carlo Tree Search

Add code
Bookmark button
Alert button
Mar 03, 2021
Oleg Szehr

Figure 1 for Hedging of Financial Derivative Contracts via Monte Carlo Tree Search
Figure 2 for Hedging of Financial Derivative Contracts via Monte Carlo Tree Search
Figure 3 for Hedging of Financial Derivative Contracts via Monte Carlo Tree Search
Viaarxiv icon

An exact kernel framework for spatio-temporal dynamics

Add code
Bookmark button
Alert button
Nov 13, 2020
Oleg Szehr, Dario Azzimonti, Laura Azzimonti

Figure 1 for An exact kernel framework for spatio-temporal dynamics
Figure 2 for An exact kernel framework for spatio-temporal dynamics
Figure 3 for An exact kernel framework for spatio-temporal dynamics
Figure 4 for An exact kernel framework for spatio-temporal dynamics
Viaarxiv icon

Hedging using reinforcement learning: Contextual $k$-Armed Bandit versus $Q$-learning

Add code
Bookmark button
Alert button
Jul 03, 2020
Loris Cannelli, Giuseppe Nuti, Marzio Sala, Oleg Szehr

Figure 1 for Hedging using reinforcement learning: Contextual $k$-Armed Bandit versus $Q$-learning
Figure 2 for Hedging using reinforcement learning: Contextual $k$-Armed Bandit versus $Q$-learning
Figure 3 for Hedging using reinforcement learning: Contextual $k$-Armed Bandit versus $Q$-learning
Figure 4 for Hedging using reinforcement learning: Contextual $k$-Armed Bandit versus $Q$-learning
Viaarxiv icon