Iolanda Leite’s research while affiliated with KTH Royal Institute of Technology and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (171)


FLoRA: Sample-Efficient Preference-based RL via Low-Rank Style Adaptation of Reward Functions
  • Preprint

April 2025

·

3 Reads

·

Simon Holk

·

·

[...]

·

Iolanda Leite

Preference-based reinforcement learning (PbRL) is a suitable approach for style adaptation of pre-trained robotic behavior: adapting the robot's policy to follow human user preferences while still being able to perform the original task. However, collecting preferences for the adaptation process in robotics is often challenging and time-consuming. In this work we explore the adaptation of pre-trained robots in the low-preference-data regime. We show that, in this regime, recent adaptation approaches suffer from catastrophic reward forgetting (CRF), where the updated reward model overfits to the new preferences, leading the agent to become unable to perform the original task. To mitigate CRF, we propose to enhance the original reward model with a small number of parameters (low-rank matrices) responsible for modeling the preference adaptation. Our evaluation shows that our method can efficiently and effectively adjust robotic behavior to human preferences across simulation benchmark tasks and multiple real-world robotic tasks.


Fig. 3: Tablet interface participants' used to control the robot (utilising the co-designed action space) during group activity sessions. Users' had to select the desired action and could identify an individual target GM where appropriate. Selecting an action only (with no individual target) resulted in an instance of the group-targeting version of that action.
Ice-Breakers, Turn-Takers and Fun-Makers: Exploring Robots for Groups with Teenagers
  • Preprint
  • File available

April 2025

·

9 Reads

Successful, enjoyable group interactions are important in public and personal contexts, especially for teenagers whose peer groups are important for self-identity and self-esteem. Social robots seemingly have the potential to positively shape group interactions, but it seems difficult to effect such impact by designing robot behaviors solely based on related (human interaction) literature. In this article, we take a user-centered approach to explore how teenagers envisage a social robot "group assistant". We engaged 16 teenagers in focus groups, interviews, and robot testing to capture their views and reflections about robots for groups. Over the course of a two-week summer school, participants co-designed the action space for such a robot and experienced working with/wizarding it for 10+ hours. This experience further altered and deepened their insights into using robots as group assistants. We report results regarding teenagers' views on the applicability and use of a robot group assistant, how these expectations evolved throughout the study, and their repeat interactions with the robot. Our results indicate that each group moves on a spectrum of need for the robot, reflected in use of the robot more (or less) for ice-breaking, turn-taking, and fun-making as the situation demanded.

Download

Figure 2: Experiment flow for the priming study.
Figure 3: EmCorp results for the priming study. Figure 4: GAToRS results for the priming study.
Figure 7: Explanation satisfaction results for the main study.
Video design for the priming study.
Video design for the main study.
Expectations, Explanations, and Embodiment: Attempts at Robot Failure Recovery

April 2025

·

15 Reads

Expectations critically shape how people form judgments about robots, influencing whether they view failures as minor technical glitches or deal-breaking flaws. This work explores how high and low expectations, induced through brief video priming, affect user perceptions of robot failures and the utility of explanations in HRI. We conducted two online studies (N=600 total participants); each replicated two robots with different embodiments, Furhat and Pepper. In our first study, grounded in expectation theory, participants were divided into two groups, one primed with positive and the other with negative expectations regarding the robot's performance, establishing distinct expectation frameworks. This validation study aimed to verify whether the videos could reliably establish low and high-expectation profiles. In the second study, participants were primed using the validated videos and then viewed a new scenario in which the robot failed at a task. Half viewed a version where the robot explained its failure, while the other half received no explanation. We found that explanations significantly improved user perceptions of Furhat, especially when participants were primed to have lower expectations. Explanations boosted satisfaction and enhanced the robot's perceived expressiveness, indicating that effectively communicating the cause of errors can help repair user trust. By contrast, Pepper's explanations produced minimal impact on user attitudes, suggesting that a robot's embodiment and style of interaction could determine whether explanations can successfully offset negative impressions. Together, these findings underscore the need to consider users' expectations when tailoring explanation strategies in HRI. When expectations are initially low, a cogent explanation can make the difference between dismissing a failure and appreciating the robot's transparency and effort to communicate.


Fig. 2. Simplified Behavior Tree of the BT-ACTION System.
Fig. 3. Sketch of the Experiment setup.
Fig. 4. Picture from the experiment setup.
Fig. 5. Mean and standard error of the number of mistakes for the two conditions.
Fig. 6. Mean and standard error for MDMT Capacity Trust for the two conditions.
BT-ACTION: A Test-Driven Approach for Modular Understanding of User Instruction Leveraging Behaviour Trees and LLMs

April 2025

·

15 Reads

Natural language instructions are often abstract and complex, requiring robots to execute multiple subtasks even for seemingly simple queries. For example, when a user asks a robot to prepare avocado toast, the task involves several sequential steps. Moreover, such instructions can be ambiguous or infeasible for the robot or may exceed the robot's existing knowledge. While Large Language Models (LLMs) offer strong language reasoning capabilities to handle these challenges, effectively integrating them into robotic systems remains a key challenge. To address this, we propose BT-ACTION, a test-driven approach that combines the modular structure of Behavior Trees (BT) with LLMs to generate coherent sequences of robot actions for following complex user instructions, specifically in the context of preparing recipes in a kitchen-assistance setting. We evaluated BT-ACTION in a comprehensive user study with 45 participants, comparing its performance to direct LLM prompting. Results demonstrate that the modular design of BT-ACTION helped the robot make fewer mistakes and increased user trust, and participants showed a significant preference for the robot leveraging BT-ACTION. The code is publicly available at https://github.com/1Eggbert7/BT_LLM.


Fig. 1. The interaction between the robot and three participants from two different perspectives. The robot moderates the group discussion and needs to decide when to move on to the next topic.
Let's move on: Topic Change in Robot-Facilitated Group Discussions

April 2025

·

5 Reads

Robot-moderated group discussions have the potential to facilitate engaging and productive interactions among human participants. Previous work on topic management in conversational agents has predominantly focused on human engagement and topic personalization, with the agent having an active role in the discussion. Also, studies have shown the usefulness of including robots in groups, yet further exploration is still needed for robots to learn when to change the topic while facilitating discussions. Accordingly, our work investigates the suitability of machine-learning models and audiovisual non-verbal features in predicting appropriate topic changes. We utilized interactions between a robot moderator and human participants, which we annotated and used for extracting acoustic and body language-related features. We provide a detailed analysis of the performance of machine learning approaches using sequential and non-sequential data with different sets of features. The results indicate promising performance in classifying inappropriate topic changes, outperforming rule-based approaches. Additionally, acoustic features exhibited comparable performance and robustness compared to the complete set of multimodal features. Our annotated data is publicly available at https://github.com/ghadj/topic-change-robot-discussions-data-2024.


Fig. 1. Overview of the method. Starting from a 3D map and its 3D scene graph representation, our approach computes a preferred based trajectory which is socially aware of the human presence in the scene.
Long-Term Planning Around Humans in Domestic Environments with 3D Scene Graphs

March 2025

·

8 Reads

Long-term planning for robots operating in domestic environments poses unique challenges due to the interactions between humans, objects, and spaces. Recent advancements in trajectory planning have leveraged vision-language models (VLMs) to extract contextual information for robots operating in real-world environments. While these methods achieve satisfying performance, they do not explicitly model human activities. Such activities influence surrounding objects and reshape spatial constraints. This paper presents a novel approach to trajectory planning that integrates human preferences, activities, and spatial context through an enriched 3D scene graph (3DSG) representation. By incorporating activity-based relationships, our method captures the spatial impact of human actions, leading to more context-sensitive trajectory adaptation. Preliminary results demonstrate that our approach effectively assigns costs to spaces influenced by human activities, ensuring that the robot trajectory remains contextually appropriate and sensitive to the ongoing environment. This balance between task efficiency and social appropriateness enhances context-aware human-robot interactions in domestic settings. Future work includes implementing a full planning pipeline and conducting user studies to evaluate trajectory acceptability.


Fig. 3: User rankings (S3) of three modalities, namely Virtual Reality (VR), 2D Top-Down (2D-TD), and 2D First-Person View (2D-FPV), based on perceived usefulness, ease of use, and intention to use. Each bar represents the percentage of participants who assigned first, second, and third ranks to each modality. VR is predominantly ranked highest in usefulness and intention to use.
The Impact of VR and 2D Interfaces on Human Feedback in Preference-Based Robot Learning

March 2025

·

15 Reads

Aligning robot navigation with human preferences is essential for ensuring comfortable and predictable robot movement in shared spaces, facilitating seamless human-robot coexistence. While preference-based learning methods, such as reinforcement learning from human feedback (RLHF), enable this alignment, the choice of the preference collection interface may influence the process. Traditional 2D interfaces provide structured views but lack spatial depth, whereas immersive VR offers richer perception, potentially affecting preference articulation. This study systematically examines how the interface modality impacts human preference collection and navigation policy alignment. We introduce a novel dataset of 2,325 human preference queries collected through both VR and 2D interfaces, revealing significant differences in user experience, preference consistency, and policy outcomes. Our findings highlight the trade-offs between immersion, perception, and preference reliability, emphasizing the importance of interface selection in preference-based robot learning. The dataset will be publicly released to support future research.





Citations (69)


... As a result, adaptive conversation design has gained increasing attention, driven by advances in generative AI that support more dynamic MPIs [13]- [15]. Recent approaches, such as the 'Template and Graph-Based Modeling'-framework [16], propose a structured decision making model that decouples 'who to address' from 'what to communicate', using graph neural networks for adaptive small group interactions. Moreover, recent progress in large language models (LLMs), such as GPT-4 (OpenAI, USA, [17]) and LLaMA 3 (Meta, USA, [18]), opens new possibilities for group-sensitive CAI. ...

Reference:

Are We Generalizing from the Exception? An In-the-Wild Study on Group-Sensitive Conversation Design in Human-Agent Interactions
Templates and Graph Neural Networks for Social Robots Interacting in Small Groups of Varying Sizes
  • Citing Conference Paper
  • March 2025

... Promoting the advancement of trustworthy facial affect analysis can have a wide impact across diverse fields such as healthcare [23], education [21], affective computing [1], [12], social robots deployment and human-robot interaction (HRI) [9], [25], [10] and collaboration [15]. Given the innately high-stakes and sensitive nature of the above use cases, it is essential that researchers ensure the trustworthiness of these systems and adopt measures aligned with ethical guidelines [14], [13], [24]. ...

HRI Wasn’t Built In a Day: A Call To Action For Responsible HRI Research
  • Citing Conference Paper
  • August 2024

... Recent advances in preference-based learning, including reinforcement learning from human feedback (RLHF) [1], demonstrate the potential of human-in-the-loop methods to shape robot behavior in alignment with user expectations. In fact, preferences have been leveraged in robot learning across various settings, including multi-task learning [2], collaborative tasks [3], language-based tasks [4], [5], and social navigation [6]. ...

Shielding for Socially Appropriate Robot Listening Behaviors
  • Citing Conference Paper
  • August 2024

... The effectiveness of learning is also contingent on the quality of queries presented to humans. To enhance query informativeness, various PbRL active learning techniques have been developed, leveraging policy ensembles [16] or unsupervised learning [17], [18]. ...

SEQUEL: Semi-Supervised Preference-based RL with Query Synthesis via Latent Interpolation
  • Citing Conference Paper
  • May 2024

... Another challenge in SocialNav is the lack of realism in configurations [15]. Current approaches typically simplify the environment to only include a robot and surrounding humans, neglecting the complexity of the scene itself [16], [17], [18]. Moreover, the solutions often assume the robot has access to global information, such as real-time human positions or a full map of the environment [8]. ...

POLITE: Preferences Combined with Highlights in Reinforcement Learning
  • Citing Conference Paper
  • May 2024

... Previous studies utilizing similar experimental protocols have examined the influence of politeness behaviors of artificial agents on the routes participants take when joining small, free-standing groups of robots or virtual characters [43], [44]. Additionally, they have explored the interplay between politeness and embodiment [45] and various modalities of communication [46]. However, they have not addressed how cultural differences impact perceptions of politeness and interactions with agents during group joining, which is the focus of this study. ...

Join Me Here if You Will: Investigating Embodiment and Politeness Behaviors When Joining Small Groups of Humans, Robots, and Virtual Characters

... Robots must adopt similar strategies by offering transparent, interpretable behaviours and accessible feedback, allowing users to understand the robot's goals and intentions. Such transparency in a robot's actions fosters trust and positions robots as empathetic and effective collaborators, particularly in therapeutic settings [119,155]. However, explainability extends beyond mere transparency. ...

Explainability for Human-Robot Collaboration
  • Citing Conference Paper
  • March 2024

... Human preferences in robot learning Preference-based RL [4] has become a suitable approach for formulating complex objectives into reward functions. In robot learning, preferences have been used in multi-task settings [19], collaborative tasks [20], and alongside language [21]. Nonetheless, the need for large amounts of human feedback in these applications has revealed constraints in their feasibility for real-world robotics and other sophisticated environments [22], [23], [24], [15]. ...

PREDILECT: Preferences Delineated with Zero-Shot Language-based Reasoning in Reinforcement Learning
  • Citing Conference Paper
  • March 2024

... As human-robot teams in construction environments may be complex and dynamic, future work needs to carefully consider this aspect. One potential way to address this is to design robot behaviours which positively shape the interactions between multiple people [4] or can better support individual users' experiences [2]. ...

Interaction-Shaping Robotics: Robots That Influence Interactions between Other Agents
  • Citing Article
  • February 2024

ACM Transactions on Human-Robot Interaction

... Sound design that facilitates human-robot interaction is still a new and emerging research area, which may yield many benefits for continued exploration (Orthmann, Leite, Bresin, & Torre, 2023;Pelikan, Robinson, Keevallik, Velonaki, Broth, & Bown, 2021). ...

Sounding Robots: Design and Evaluation of Auditory Displays for Unintentional Human-robot Interaction

ACM Transactions on Human-Robot Interaction