Alert button
Picture for Antti Oulasvirta

Antti Oulasvirta

Alert button

Department of Communications and Networking, Aalto University, Finland

Modeling human road crossing decisions as reward maximization with visual perception limitations

Jan 27, 2023
Yueyang Wang, Aravinda Ramakrishnan Srinivasan, Jussi P. P. Jokinen, Antti Oulasvirta, Gustav Markkula

Figure 1 for Modeling human road crossing decisions as reward maximization with visual perception limitations
Figure 2 for Modeling human road crossing decisions as reward maximization with visual perception limitations
Figure 3 for Modeling human road crossing decisions as reward maximization with visual perception limitations
Figure 4 for Modeling human road crossing decisions as reward maximization with visual perception limitations

Understanding the interaction between different road users is critical for road safety and automated vehicles (AVs). Existing mathematical models on this topic have been proposed based mostly on either cognitive or machine learning (ML) approaches. However, current cognitive models are incapable of simulating road user trajectories in general scenarios, and ML models lack a focus on the mechanisms generating the behavior and take a high-level perspective which can cause failures to capture important human-like behaviors. Here, we develop a model of human pedestrian crossing decisions based on computational rationality, an approach using deep reinforcement learning (RL) to learn boundedly optimal behavior policies given human constraints, in our case a model of the limited human visual system. We show that the proposed combined cognitive-RL model captures human-like patterns of gap acceptance and crossing initiation time. Interestingly, our model's decisions are sensitive to not only the time gap, but also the speed of the approaching vehicle, something which has been described as a "bias" in human gap acceptance behavior. However, our results suggest that this is instead a rational adaption to human perceptual limitations. Moreover, we demonstrate an approach to accounting for individual differences in computational rationality models, by conditioning the RL policy on the parameters of the human constraints. Our results demonstrate the feasibility of generating more human-like road user behavior by combining RL with cognitive models.

* 6 pages, 5 figures,1 table, manuscript created for consideration at IEEE IV 2023 conference 
Viaarxiv icon

Investigating Positive and Negative Qualities of Human-in-the-Loop Optimization for Designing Interaction Techniques

Apr 15, 2022
Liwei Chan, Yi-Chi Liao, George B. Mo, John J. Dudley, Chun-Lien Cheng, Per Ola Kristensson, Antti Oulasvirta

Figure 1 for Investigating Positive and Negative Qualities of Human-in-the-Loop Optimization for Designing Interaction Techniques
Figure 2 for Investigating Positive and Negative Qualities of Human-in-the-Loop Optimization for Designing Interaction Techniques
Figure 3 for Investigating Positive and Negative Qualities of Human-in-the-Loop Optimization for Designing Interaction Techniques
Figure 4 for Investigating Positive and Negative Qualities of Human-in-the-Loop Optimization for Designing Interaction Techniques

Designers reportedly struggle with design optimization tasks where they are asked to find a combination of design parameters that maximizes a given set of objectives. In HCI, design optimization problems are often exceedingly complex, involving multiple objectives and expensive empirical evaluations. Model-based computational design algorithms assist designers by generating design examples during design, however they assume a model of the interaction domain. Black box methods for assistance, on the other hand, can work with any design problem. However, virtually all empirical studies of this human-in-the-loop approach have been carried out by either researchers or end-users. The question stands out if such methods can help designers in realistic tasks. In this paper, we study Bayesian optimization as an algorithmic method to guide the design optimization process. It operates by proposing to a designer which design candidate to try next, given previous observations. We report observations from a comparative study with 40 novice designers who were tasked to optimize a complex 3D touch interaction technique. The optimizer helped designers explore larger proportions of the design space and arrive at a better solution, however they reported lower agency and expressiveness. Designers guided by an optimizer reported lower mental effort but also felt less creative and less in charge of the progress. We conclude that human-in-the-loop optimization can support novice designers in cases where agency is not critical.

* CHI 2022 
Viaarxiv icon

Rediscovering Affordance: A Reinforcement Learning Perspective

Jan 07, 2022
Yi-Chi Liao, Kashyap Todi, Aditya Acharya, Antti Keurulainen, Andrew Howes, Antti Oulasvirta

Figure 1 for Rediscovering Affordance: A Reinforcement Learning Perspective
Figure 2 for Rediscovering Affordance: A Reinforcement Learning Perspective
Figure 3 for Rediscovering Affordance: A Reinforcement Learning Perspective
Figure 4 for Rediscovering Affordance: A Reinforcement Learning Perspective

Affordance refers to the perception of possible actions allowed by an object. Despite its relevance to human-computer interaction, no existing theory explains the mechanisms that underpin affordance-formation; that is, how affordances are discovered and adapted via interaction. We propose an integrative theory of affordance-formation based on the theory of reinforcement learning in cognitive sciences. The key assumption is that users learn to associate promising motor actions to percepts via experience when reinforcement signals (success/failure) are present. They also learn to categorize actions (e.g., "rotating" a dial), giving them the ability to name and reason about affordance. Upon encountering novel widgets, their ability to generalize these actions determines their ability to perceive affordances. We implement this theory in a virtual robot model, which demonstrates human-like adaptation of affordance in interactive widgets tasks. While its predictions align with trends in human data, humans are able to adapt affordances faster, suggesting the existence of additional mechanisms.

* 15 pages, In proceedings of the ACM CHI 2022 
Viaarxiv icon

Toward AI Assistants That Let Designers Design

Jul 22, 2021
Sebastiaan De Peuter, Antti Oulasvirta, Samuel Kaski

Figure 1 for Toward AI Assistants That Let Designers Design
Figure 2 for Toward AI Assistants That Let Designers Design
Figure 3 for Toward AI Assistants That Let Designers Design

AI for supporting designers needs to be rethought. It should aim to cooperate, not automate, by supporting and leveraging the creativity and problem-solving of designers. The challenge for such AI is how to infer designers' goals and then help them without being needlessly disruptive. We present AI-assisted design: a framework for creating such AI, built around generative user models which enable reasoning about designers' goals, reasoning, and capabilities.

* 9 pages, 3 figures, submitted to IEEE Computer magazine 
Viaarxiv icon

Adapting User Interfaces with Model-based Reinforcement Learning

Mar 11, 2021
Kashyap Todi, Gilles Bailly, Luis A. Leiva, Antti Oulasvirta

Figure 1 for Adapting User Interfaces with Model-based Reinforcement Learning
Figure 2 for Adapting User Interfaces with Model-based Reinforcement Learning
Figure 3 for Adapting User Interfaces with Model-based Reinforcement Learning
Figure 4 for Adapting User Interfaces with Model-based Reinforcement Learning

Adapting an interface requires taking into account both the positive and negative effects that changes may have on the user. A carelessly picked adaptation may impose high costs to the user -- for example, due to surprise or relearning effort -- or "trap" the process to a suboptimal design immaturely. However, effects on users are hard to predict as they depend on factors that are latent and evolve over the course of interaction. We propose a novel approach for adaptive user interfaces that yields a conservative adaptation policy: It finds beneficial changes when there are such and avoids changes when there are none. Our model-based reinforcement learning method plans sequences of adaptations and consults predictive HCI models to estimate their effects. We present empirical and simulation results from the case of adaptive menus, showing that the method outperforms both a non-adaptive and a frequency-based policy.

* 13 pages, 10 figures, ACM CHI 2021 Full Paper 
Viaarxiv icon

Improving Artificial Teachers by Considering How People Learn and Forget

Feb 19, 2021
Aurélien Nioche, Pierre-Alexandre Murena, Carlos de la Torre-Ortiz, Antti Oulasvirta

Figure 1 for Improving Artificial Teachers by Considering How People Learn and Forget
Figure 2 for Improving Artificial Teachers by Considering How People Learn and Forget
Figure 3 for Improving Artificial Teachers by Considering How People Learn and Forget
Figure 4 for Improving Artificial Teachers by Considering How People Learn and Forget

The paper presents a novel model-based method for intelligent tutoring, with particular emphasis on the problem of selecting teaching interventions in interaction with humans. Whereas previous work has focused on either personalization of teaching or optimization of teaching intervention sequences, the proposed individualized model-based planning approach represents convergence of these two lines of research. Model-based planning picks the best interventions via interactive learning of a user memory model's parameters. The approach is novel in its use of a cognitive model that can account for several key individual- and material-specific characteristics related to recall/forgetting, along with a planning technique that considers users' practice schedules. Taking a rule-based approach as a baseline, the authors evaluated the method's benefits in a controlled study of artificial teaching in second-language vocabulary learning (N=53).

* 15 pages, 5 figures, to be published in IUI'21 
Viaarxiv icon