Dennis Becker’s research while affiliated with Hamburg University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (9)


Influence of Robots’ Voice Naturalness on Trust and Compliance
  • Article
  • Full-text available

January 2025

·

67 Reads

·

1 Citation

ACM Transactions on Human-Robot Interaction

Dennis Becker

·

Lukas Braach

·

Lennart Clasmeier

·

[...]

·

With the increasing performance of text-to-speech systems and their generated voices indistinguishable from natural human speech, the use of these systems for robots raises ethical and safety concerns. A robot with a natural voice could increase trust, which might result in over-reliance despite evidence for robot unreliability. To estimate the influence of a robot's voice on trust and compliance, we design a study that consists of two experiments. In a pre-study ( N1=60N_{1}=60 ) the most suitable natural and mechanical voice for the main study are estimated and selected for the main study. Afterward, in the main study ( N2=68N_{2}=68 ), the influence of a robot's voice on trust and compliance is evaluated in a cooperative game of Battleship with a robot as an assistant. During the experiment, the acceptance of the robot's advice and response time are measured, which indicate trust and compliance respectively. The results show that participants expect robots to sound human-like and that a robot with a natural voice is perceived as safer. Additionally, a natural voice can affect compliance. Despite repeated incorrect advice, the participants are more likely to rely on the robot with the natural voice. The results do not show a direct effect on trust. Natural voices provide increased intelligibility, and while they can increase compliance with the robot, the results indicate that natural voices might not lead to over-reliance. The results highlight the importance of incorporating voices into the design of social robots to improve communication, avoid adverse effects, and increase acceptance and adoption in society.

Download

The Emotional Dilemma: Influence of a Human-like Robot on Trust and Cooperation

August 2023

·

47 Reads

·

2 Citations

Increasing anthropomorphic robot behavioral design could affect trust and cooperation positively. However, studies have shown contradicting results and suggest a task-dependent relationship between robots that display emotions and trust. Therefore, this study analyzes the effect of robots that display human-like emotions on trust, cooperation, and participants' emotions. In the between-group study, participants play the coin entrustment game with an emotional and a non-emotional robot. The results show that the robot that displays emotions induces more anxiety than the neutral robot. Accordingly, the participants trust the emotional robot less and are less likely to cooperate. Furthermore, the perceived intelligence of a robot increases trust, while a desire to outcompete the robot can reduce trust and cooperation. Thus, the design of robots expressing emotions should be task dependent to avoid adverse effects that reduce trust and cooperation.


The Emotional Dilemma: Influence of a Human-like Robot on Trust and Cooperation

July 2023

·

93 Reads

Increasing anthropomorphic robot behavioral design could affect trust and cooperation positively. However, studies have shown contradicting results and suggest a task-dependent relationship between robots that display emotions and trust. Therefore, this study analyzes the effect of robots that display human-like emotions on trust, cooperation, and participants' emotions. In the between-group study, participants play the coin entrustment game with an emotional and a non-emotional robot. The results show that the robot that displays emotions induces more anxiety than the neutral robot. Accordingly, the participants trust the emotional robot less and are less likely to cooperate. Furthermore, the perceived intelligence of a robot increases trust, while a desire to outcompete the robot can reduce trust and cooperation. Thus, the design of robots expressing emotions should be task dependent to avoid adverse effects that reduce trust and cooperation.


Integrating Uncertainty into Neural Network-based Speech Enhancement

May 2023

·

3 Reads

Supervised masking approaches in the time-frequency domain aim to employ deep neural networks to estimate a multiplicative mask to extract clean speech. This leads to a single estimate for each input without any guarantees or measures of reliability. In this paper, we study the benefits of modeling uncertainty in clean speech estimation. Prediction uncertainty is typically categorized into aleatoric uncertainty and epistemic uncertainty. The former refers to inherent randomness in data, while the latter describes uncertainty in the model parameters. In this work, we propose a framework to jointly model aleatoric and epistemic uncertainties in neural network-based speech enhancement. The proposed approach captures aleatoric uncertainty by estimating the statistical moments of the speech posterior distribution and explicitly incorporates the uncertainty estimate to further improve clean speech estimation. For epistemic uncertainty, we investigate two Bayesian deep learning approaches: Monte Carlo dropout and Deep ensembles to quantify the uncertainty of the neural network parameters. Our analyses show that the proposed framework promotes capturing practical and reliable uncertainty, while combining different sources of uncertainties yields more reliable predictive uncertainty estimates. Furthermore, we demonstrate the benefits of modeling uncertainty on speech enhancement performance by evaluating the framework on different datasets, exhibiting notable improvement over comparable models that fail to account for uncertainty.


Fig. 1: Block diagram of the proposed neural network-based aleatoric uncertainty estimation.
Fig. 2: Input-output characteristics of the AMAP estimator W AMAP ft and Wiener filter W WF ft (setting σ 2 s,ft =σ 2 n,ft =1 in this example).
Fig. 8: Sparsification plots of aleatoric˜λaleatoric˜ aleatoric˜λ, epistemic Σ , and overall predictive uncertainty Σ (i.e., aleatoric & epistemic) on the DNS test dataset. Note that Oracle aleatoric and Oracle aleatoric & epistemic are overlapping.
Fig. 9: Performance improvement on the dataset with speech from WSJ0 and noise from CHiME3. PESQi denotes PESQ improvement with respect to noisy mixtures. ESTOIi and SI-SDRi are defined similarly. Markers and vertical bars indicate the mean and 95% confidence interval.
Integrating Uncertainty Into Neural Network-Based Speech Enhancement

April 2023

·

163 Reads

·

5 Citations

IEEE/ACM Transactions on Audio Speech and Language Processing

Supervised masking approaches in the time-frequency domain aim to employ deep neural networks to estimate a multiplicative mask to extract clean speech. This leads to a single estimate for each input without any guarantees or measures of reliability. In this paper, we study the benefits of modeling uncertainty in clean speech estimation. Prediction uncertainty is typically categorized into aleatoric uncertainty and epistemic uncertainty . The former refers to inherent randomness in data, while the latter describes uncertainty in the model parameters. In this work, we propose a framework to jointly model aleatoric and epistemic uncertainties in neural network-based speech enhancement. The proposed approach captures aleatoric uncertainty by estimating the statistical moments of the speech posterior distribution and explicitly incorporates the uncertainty estimate to further improve clean speech estimation. For epistemic uncertainty, we investigate two Bayesian deep learning approaches: Monte Carlo dropout and Deep ensembles to quantify the uncertainty of the neural network parameters. Our analyses show that the proposed framework promotes capturing practical and reliable uncertainty, while combining different sources of uncertainties yields more reliable predictive uncertainty estimates. Furthermore, we demonstrate the benefits of modeling uncertainty on speech enhancement performance by evaluating the framework on different datasets, exhibiting notable improvement over comparable models that fail to account for uncertainty.


Word-by-Word Generation of Visual Dialog Using Reinforcement Learning

September 2022

·

4 Reads

Lecture Notes in Computer Science

The task of visual dialog generation requires an agent holding a conversation referencing question history, putting the current question into context, and processing visual content. While previous research focused on arranging questions to form dialog, we tackle the more challenging task of arranging questions from words, and dialog from questions. We develop our model in a simple “Guess which?” game scenario where the agent needs to predict an image region that has been selected by an oracle by asking questions to the oracle. As a result, the reinforcement learning agent arranges words to refer to the image features strategically to acquire the required information from the oracle, memorizing it and giving the correct prediction with an accuracy well above 80%. Imposing costs on the number of questions asked to the oracle leads to a strategy using few questions, while imposing costs on the number of words used leads to more but shorter questions. Our results are a step towards making goal-directed dialog fully generic by assembling it from words, elementary constituents of language.Keywords Visual dialog generation Deep reinforcement learning Compositionality


Fig. 1 NICO, the humanoid robot, assists a person. NICO uses nonverbal communication to signal to have an issue with a received request; it can explain the issue on a symbolic level and, when prompted, can also visualize elements of the underlying neural processing. The visualization is accompanied by a verbal explanation
Fig. 5 Two humanoid robots play a game of UTTT against each other. Screenshot of a video sequence that was provided to the participants
Fig. 6 Top row: Images from the synthetic training set. Bottom row: Images from the synthetic test set with bounding boxes and classification found by the trained RetinaNet
Fig. 8 Top: NICO collects samples in a virtual environment by placing an object and recording its corresponding joint values. Bottom row: Image part of the collected samples without and with distractor objects in different colors (no distractor; green, gray and red distractors)
Fig. 9 Grad-CAM visualization of the neural end-to-end grasp approach for the first shoulder joint shows which part of the input image (right side) is relevant to generate the output. The blue and red parts of the grasp object are visible in both Grad-CAM visualizations. The lower Grad-CAM explains why red distractor objects are detrimental for grasping; they show up prominently in the Grad-CAM visualization; the network has not learned to ignore these distractors
What’s on Your Mind, NICO?: XHRI: A Framework for eXplainable Human-Robot Interaction

August 2022

·

547 Reads

·

7 Citations

KI - Ku_nstliche Intelligenz

Explainable AI has become an important field of research on neural machine learning models. However, most existing methods are designed as tools that provide expert users with additional insights into their models. In contrast, in human-robot interaction scenarios, non-expert users are frequently confronted with complex, embodied AI systems whose inner workings are unknown. Therefore, eXplainable Human-Robot Interaction (XHRI) should leverage the user’s intuitive ability to collaborate and to use efficient communication. Using NICO, the Neuro-Inspired COmpanion, as a use-case study, we propose an XHRI framework and show how different types of explanations enhance the interaction experience. These explanations range from (a) non-verbal cues for simple and intuitive feedback of inner states via (b) comprehensive verbal explanations of the robot’s intentions, knowledge and reasoning to (c) multimodal explanations using visualizations, speech and text. We revisit past HRI-related studies conducted with NICO and analyze them with the proposed framework. Furthermore, we present two novel XHRI approaches to extract suitable verbal and multimodal explanations from neural network modules in an HRI scenario.


Explain yourself! Effects of Explanations in Human-Robot Interaction

June 2022

·

62 Reads

·

24 Citations

Recent developments in explainable artificial intelligence promise the potential to transform human-robot interaction: Explanations of robot decisions could affect user perceptions, justify their reliability, and increase trust. However, the effects on human perceptions of robots that explain their decisions have not been studied thoroughly. To analyze the effect of explainable robots, we conduct a study in which two simulated robots play a competitive board game. While one robot explains its moves, the other robot only announces them. Providing explanations for its actions was not sufficient to change the perceived competence, intelligence, likeability or safety ratings of the robot. However, the results show that the robot that explains its moves is perceived as more lively and human-like. This study demonstrates the need for and potential of explainable human-robot interaction and the wider assessment of its effects as a novel research direction.


Explain yourself! Effects of Explanations in Human-Robot Interaction

April 2022

·

64 Reads

Recent developments in explainable artificial intelligence promise the potential to transform human-robot interaction: Explanations of robot decisions could affect user perceptions, justify their reliability, and increase trust. However, the effects on human perceptions of robots that explain their decisions have not been studied thoroughly. To analyze the effect of explainable robots, we conduct a study in which two simulated robots play a competitive board game. While one robot explains its moves, the other robot only announces them. Providing explanations for its actions was not sufficient to change the perceived competence, intelligence, likeability or safety ratings of the robot. However, the results show that the robot that explains its moves is perceived as more lively and human-like. This study demonstrates the need for and potential of explainable human-robot interaction and the wider assessment of its effects as a novel research direction.

Citations (4)


... Transparent AI voice systems that clarify their artificial nature tend to foster more trust than those that obscure it. Becker et al. [30] suggest that the naturalness of an AI-generated voice plays a pivotal Table 3. Ability to rewrite the news piece from news agencies into radio report according to the criteria of Slovak "radio language" by the five most used generative AI tools in Slovakia. role in compliance and trust. ...

Reference:

Perspective Chapter: Artificial Intelligence in Slovak Radio Industry - The Present and the Future of Broadcasting
Influence of Robots’ Voice Naturalness on Trust and Compliance

ACM Transactions on Human-Robot Interaction

... Previous research has shown that imparting a degree of social intelligence to household communication systems can increase the system acceptability and induce social behavior from users toward the system [54]. Because the perceived intelligence of robots is related to trust in robots [55], utterances regarding user values may also lead to increased trust in robots. Furthermore, it is not surprising that no significant differences were observed in the anthropomorphism and likability of the robot's impression. ...

The Emotional Dilemma: Influence of a Human-like Robot on Trust and Cooperation
  • Citing Conference Paper
  • August 2023

... While the explainable agency did not significantly affect children's game efficiency, it positively impacted their perceived difficulty, suggesting potential improvements in the learning process through reflection on the robot's explainable agency. Various explanation types have also been study in a eXplainable Human-Robot Interaction (XHRI) framework (Kerzel et al., 2022). The framework incorporates various explanation types, from nonverbal cues to comprehensive verbal and multimodal explanations. ...

What’s on Your Mind, NICO?: XHRI: A Framework for eXplainable Human-Robot Interaction

KI - Ku_nstliche Intelligenz

... While this form of transparency diminished trust between the children and the robot, it did not affect the children's affection towards the robot [26]. Ambsdorf et al. [55] indicated that even though the explanations for decisions provided by robots didn't significantly improve their perceived reliability and cognitive capabilities, it made them appear more "human-like." Furthermore, excessive transparency might create the illusion that the robot is more "intelligent" or "autonomous" than it truly is, potentially misleading users and inducing inappropriate trust or reliance [56]. ...

Explain yourself! Effects of Explanations in Human-Robot Interaction
  • Citing Conference Paper
  • June 2022