Christoph Gebhardt’s research while affiliated with ETH Zurich and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (14)


Continual Human-in-the-Loop Optimization
  • Conference Paper

April 2025

·

2 Reads

·

Paul Streli

·

Zhipeng Li

·

[...]

·


Figure 8: Estimated typing performance for different keyboard configurations according to the BNN population model with a growing user base. Note that the design parameters are normalized to the range [0, 1], while predicted performance is normalized to [−5, 5].
Adjustable parameters for the keyboard optimization.
Participant characteristics.
Continual Human-in-the-Loop Optimization
  • Preprint
  • File available

March 2025

·

17 Reads

Optimal input settings vary across users due to differences in motor abilities and personal preferences, which are typically addressed by manual tuning or calibration. Although human-in-the-loop optimization has the potential to identify optimal settings during use, it is rarely applied due to its long optimization process. A more efficient approach would continually leverage data from previous users to accelerate optimization, exploiting shared traits while adapting to individual characteristics. We introduce the concept of Continual Human-in-the-Loop Optimization and a Bayesian optimization-based method that leverages a Bayesian-neural-network surrogate model to capture population-level characteristics while adapting to new users. We propose a generative replay strategy to mitigate catastrophic forgetting. We demonstrate our method by optimizing virtual reality keyboard parameters for text entry using direct touch, showing reduced adaptation times with a growing user base. Our method opens the door for next-generation personalized input systems that improve with accumulated experience.

Download

Regressor-Guided Image Editing Regulates Emotional Response to Reduce Online Engagement

January 2025

·

8 Reads

Emotions are known to mediate the relationship between users' content consumption and their online engagement, with heightened emotional intensity leading to increased engagement. Building on this insight, we propose three regressor-guided image editing approaches aimed at diminishing the emotional impact of images. These include (i) a parameter optimization approach based on global image transformations known to influence emotions, (ii) an optimization approach targeting the style latent space of a generative adversarial network, and (iii) a diffusion-based approach employing classifier guidance and classifier-free guidance. Our findings demonstrate that approaches can effectively alter the emotional properties of images while maintaining high visual quality. Optimization-based methods primarily adjust low-level properties like color hues and brightness, whereas the diffusion-based approach introduces semantic changes, such as altering appearance or facial expressions. Notably, results from a behavioral study reveal that only the diffusion-based approach successfully elicits changes in viewers' emotional responses while preserving high perceived image quality. In future work, we will investigate the impact of these image adaptations on internet user behavior.




SituationAdapt: Contextual UI Optimization in Mixed Reality with Situation Awareness via LLM Reasoning

September 2024

·

56 Reads

Mixed Reality is increasingly used in mobile settings beyond controlled home and office spaces. This mobility introduces the need for user interface layouts that adapt to varying contexts. However, existing adaptive systems are designed only for static environments. In this paper, we introduce SituationAdapt, a system that adjusts Mixed Reality UIs to real-world surroundings by considering environmental and social cues in shared settings. Our system consists of perception, reasoning, and optimization modules for UI adaptation. Our perception module identifies objects and individuals around the user, while our reasoning module leverages a Vision-and-Language Model to assess the placement of interactive UI elements. This ensures that adapted layouts do not obstruct relevant environmental cues or interfere with social norms. Our optimization module then generates Mixed Reality interfaces that account for these considerations as well as temporal constraints. For evaluation, we first validate our reasoning module's capability of assessing UI contexts in comparison to human expert users. In an online user study, we then establish SituationAdapt's capability of producing context-aware layouts for Mixed Reality, where it outperformed previous adaptive layout methods. We conclude with a series of applications and scenarios to demonstrate SituationAdapt's versatility.


MARLUI: Multi-Agent Reinforcement Learning for Adaptive Point-and-Click UIs

June 2024

·

13 Reads

·

6 Citations

Proceedings of the ACM on Human-Computer Interaction

As the number of selectable items increases, point-and-click interfaces rapidly become complex, leading to a decrease in usability. Adaptive user interfaces can reduce this complexity by automatically adjusting an interface to only display the most relevant items. A core challenge for developing adaptive interfaces is to infer user intent and chose adaptations accordingly. Current methods rely on tediously hand-crafted rules or carefully collected user data. Furthermore, heuristics need to be recrafted and data regathered for every new task and interface. To address this issue, we formulate interface adaptation as a multi-agent reinforcement learning problem. Our approach learns adaptation policies without relying on heuristics or real user data, facilitating the development of adaptive interfaces across various tasks with minimal adjustments needed. In our formulation, a user agent mimics a real user and learns to interact with an interface via point-and-click actions. Simultaneously, an interface agent learns interface adaptations, to maximize the user agent's efficiency, by observing the user agent's behavior. For our evaluation, we substituted the simulated user agent with actual users. Our study involved twelve participants and concentrated on automatic toolbar item assignment. The results show that the policies we developed in simulation effectively apply to real users. These users were able to complete tasks with fewer actions and in similar times compared to methods trained with real data. Additionally, we demonstrated our method's efficiency and generalizability across four different interfaces and tasks.


Figure 2: In a 5x5 grid environment with lava, (a) the expert trajectory is characterized by noisy data that passes through lava without resulting in death. (c) GAIL, (d) AIRL and (e) IQLearn learn to imitate the expert's path precisely, leading them to either get stuck near the lava or enter it and perish. (b) RILe avoids the noisy data, better mimics the expert in later stages, and successfully reaches the goal. Subfigures (f-i) display the value tables for RILe, GAIL, AIRL, and IQLearn respectively. The optimal path, derived from the reward of the teacher or discriminator, is highlighted with green lines.
Test and validation results on seven LocoMujoco tasks.
RILe: Reinforced Imitation Learning

June 2024

·

61 Reads

Reinforcement Learning has achieved significant success in generating complex behavior but often requires extensive reward function engineering. Adversarial variants of Imitation Learning and Inverse Reinforcement Learning offer an alternative by learning policies from expert demonstrations via a discriminator. Employing discriminators increases their data- and computational efficiency over the standard approaches; however, results in sensitivity to imperfections in expert data. We propose RILe, a teacher-student system that achieves both robustness to imperfect data and efficiency. In RILe, the student learns an action policy while the teacher dynamically adjusts a reward function based on the student's performance and its alignment with expert demonstrations. By tailoring the reward function to both performance of the student and expert similarity, our system reduces dependence on the discriminator and, hence, increases robustness against data imperfections. Experiments show that RILe outperforms existing methods by 2x in settings with limited or noisy expert data.


Detecting Users' Emotional States during Passive Social Media Use

May 2024

·

55 Reads

·

1 Citation

Proceedings of the ACM on Interactive Mobile Wearable and Ubiquitous Technologies

The widespread use of social media significantly impacts users' emotions. Negative emotions, in particular, are frequently produced, which can drastically affect mental health. Recognizing these emotional states is essential for implementing effective warning systems for social networks. However, detecting emotions during passive social media use---the predominant mode of engagement---is challenging. We introduce the first predictive model that estimates user emotions during passive social media consumption alone. We conducted a study with 29 participants who interacted with a controlled social media feed. Our apparatus captured participants' behavior and their physiological signals while they browsed the feed and filled out self-reports from two validated emotion models. Using this data for supervised training, our emotion classifier robustly detected up to 8 emotional states and achieved 83% peak accuracy to classify affect. Our analysis shows that behavioral features were sufficient to robustly recognize participants' emotions. It further highlights that within 8 seconds following a change in media content, objective features reveal a participant's new emotional state. We show that grounding labels in a componential emotion model outperforms dimensional models in higher-resolutional state detection. Our findings also demonstrate that using emotional properties of images, predicted by a deep learning model, further improves emotion recognition.



Citations (8)


... In contrast to them, we address the prob-lem in its more challenging, single-image setting. Jiang et al. [19] use biomechanical constraints for more accurate 3D pose estimation, but their work adopts the SMPL model, making the output incompatible with biomechanical simulations [10]. Moreover, there is progress with the datasets for biomechanics. ...

Reference:

Reconstructing Humans with a Biomechanically Accurate Skeleton
MANIKIN: Biomechanically Accurate Neural Inverse Kinematics for Human Motion Estimation
  • Citing Chapter
  • October 2024

... The expectation that AI systems can act in situ and comprehend their surrounding environment has prompted extensive evaluation of their situational (or contextual) awareness [25,[73][74][75]. Empirically, LLMs not only reject user requests that violate safety criteria [76], but can also reversely infer the precise context they are in-solely from abstract rules, without being given specific tasks or examples [77]. ...

Reference:

AI Awareness
SituationAdapt: Contextual UI Optimization in Mixed Reality with Situation Awareness via LLM Reasoning
  • Citing Conference Paper
  • October 2024

... Previous work has shown that defining suitable rewards for RL in UI adaptation is nontrivial, especially in scenarios where user goals are ambiguous or the outcomes of actions are not immediately clear. For instance, Langerak et al. (2022Langerak et al. ( , 2024 introduced a multi-agent RL framework to address these challenges, where a user agent learns to interact with the UI, while an interface agent learns to adapt it. Vidmanov and Alfimtsev (2024) expanded this paradigm by integrating a usability reward model into their multi-agent RL framework, allowing agents to cooperatively adapt UI elements, such as widgets and content, to improve user experience. ...

MARLUI: Multi-Agent Reinforcement Learning for Adaptive Point-and-Click UIs
  • Citing Article
  • June 2024

Proceedings of the ACM on Human-Computer Interaction

... Physiological sensing, in particular, has been extensively applied to understand affective states, especially in high-stress contexts. Wearable devices such as heart rate monitors, electrodermal activity (EDA) bands, and EEG headbands capture emotional workload and stress responses, offering insights into real-time emotional regulation [32,64,87] and adaptive education environment [115]. In social environments, these sensors track interpersonal synchrony, capturing shared engagement between individuals and groups, such as actors and audiences in live performances [113]. ...

Detecting Users' Emotional States during Passive Social Media Use
  • Citing Article
  • May 2024

Proceedings of the ACM on Interactive Mobile Wearable and Ubiquitous Technologies

... To realize such adaptive MR UI behaviors, recent research formulates the problem as multi-objective optimization [9,10,16,17,22,25,26,32]. User's goals are formulated as a set of objective functions and placements are selected that maximize/minimize these objectives. ...

InteractionAdapt: Interaction-driven Workspace Adaptation for Situated Virtual Reality Environments
  • Citing Conference Paper
  • October 2023

... The COVID-19 pandemic and the resulting need for social distancing [1], [2], [3], [4] significantly accelerated the adoption of social Virtual Reality (VR) as a medium for virtual communication. Applications such as virtual meetings and negotiations [5], [6], [7], [8] have demonstrated the potential of VR to improve convenience and productivity in human interaction. However, these virtual environments often struggle to effectively convey social intentions [9], [10], particularly non-verbal behaviors, which are estimated to have five times the impact of verbal communication on expressing emotional connections [11]. ...

ViGather: Inclusive Virtual Conferencing with a Joint Experience Across Traditional Screen Devices and Mixed Reality Headsets
  • Citing Article
  • September 2023

Proceedings of the ACM on Human-Computer Interaction

... Modern AUIs [27,75,76,82] leverage machine learning (ML) methods that find correlations between user input, user intention, and adaptation. Such approaches significantly lower the development effort and improve the usability of AUIs over their rule-based predecessors [28]. However, their reliance on user data introduces three significant limitations. ...

Optimal Control to Support High-Level User Goals in Human-Computer Interaction
  • Citing Chapter
  • November 2021

... Affordable yet highly sophisticated drone devices and operating systems are increasingly available on the consumer market. Drones are now being used for the aerial photography and scanning of static objects [23,24,45,56], for interactive applications with environments, objects, and humans, including disaster investigations and rescue [44], product delivery [7], remote repairs [81], haptic proxy for VR [1], outdoor navigation [37], and communication agents [3,83]. For such scenarios that require drone operators' real-time judgment, interactive drone-piloting approaches have become dominant over the recently advanced autopiloting technologies [68,73]. ...

Optimization-based User Support for Cinematographic Quadrotor Camera Target Framing
  • Citing Conference Paper
  • May 2021