| Literature DB >> 34336933 |
Finn Rietz1, Alexander Sutherland1, Suna Bensch2, Stefan Wermter1, Thomas Hellström2.
Abstract
Wizard-of-Oz experiments play a vital role in Human-Robot Interaction (HRI), as they allow for quick and simple hypothesis testing. Still, a publicly available general tool to conduct such experiments is currently not available in the research community, and researchers often develop and implement their own tools, customized for each individual experiment. Besides being inefficient in terms of programming efforts, this also makes it harder for non-technical researchers to conduct Wizard-of-Oz experiments. In this paper, we present a general and easy-to-use tool for the Pepper robot, one of the most commonly used robots in this context. While we provide the concrete interface for Pepper robots only, the system architecture is independent of the type of robot and can be adapted for other robots. A configuration file, which saves experiment-specific parameters, enables a quick setup for reproducible and repeatable Wizard-of-Oz experiments. A central server provides a graphical interface via a browser while handling the mapping of user input to actions on the robot. In our interface, keyboard shortcuts may be assigned to phrases, gestures, and composite behaviors to simplify and speed up control of the robot. The interface is lightweight and independent of the operating system. Our initial tests confirm that the system is functional, flexible, and easy to use. The interface, including source code, is made commonly available, and we hope that it will be useful for researchers with any background who want to conduct HRI experiments.Entities:
Keywords: HRI experiments; Human-Robot interaction; open-source software; pepper robot; tele-operation interface; wizard-of-oz
Year: 2021 PMID: 34336933 PMCID: PMC8317264 DOI: 10.3389/frobt.2021.668057
Source DB: PubMed Journal: Front Robot AI ISSN: 2296-9144
FIGURE 1SoftBank’s Pepper robot interacting with humans in different environments (images courtesy of SoftBank Robotics).
A summary of desired robot functionalities found in the literature, how this is provided in WoZ4U, and the corresponding part number introduced in Figure 3.
| Desired functionality | Literature | Implementation in WoZ4U | Part Nr |
|---|---|---|---|
| Documentation of experiments |
| Recording of data from cameras and microphones | (1) |
| Supervision of experiments |
| Real-time monitoring of cameras and microphones | (1) |
| Accepting non-verbal input |
| Touch interface for the tablet | (1) |
| Ability to move in the environment |
| Control of translation and rotation | (2) |
| Social cues through facial expressions |
| Control of eyes and LEDs | (6) (7) |
| Social cues through movements |
| Gesture control | (8) |
| Ability to speak in a natural way |
| Generation of animated speech | (5) |
| Ability to speak using different voices |
| Setting of speech parameters | (5) |
| Presenting audio-based information |
| Play audio files in loudspeakers | (5) |
| Presenting image-based information |
| Display images and videos on tablet | (4) |
| Varying levels of automated behavior |
| Autonomy settings | (3) |
FIGURE 3The WoZ4U GUI for Pepper robots, comprising parts (1–8) with the following functionalities: (1) Connection to robot and monitoring, (2) Motion and rotation of head and base, (3) Autonomy configuration, (4) Tablet control, (5) Speech and audio control, (6) and (7) Eye control, and (8) Gesture control.
FIGURE 2The complete WoZ4U system in the form of an extended deployment UML diagram showing the connected and interacting components, starting with the wizard’s input, and leading to the execution of a command on the robot. The WoZ4U server component maps requests from the frontend to robot API calls and keeps the frontend up-to-date by continuously fetching data from the robot. The frontend is populated with experiment-specific UI elements defined in the configuration file.
Tags that can be included in sentences to be spoken by the robot. The range and default values are standardized scales provided by SoftBank.
| Tag | Function | Range | Default |
|---|---|---|---|
| \\vct = value\\ | Changes the pitch of the voice | 50–200 | 100 |
| \\rspd = value\\ | Changes the speaking rate | 50–400 | 100 |
| \\pau = value\\ | Pauses speech for value msec | — | — |
| \\vol = value\\ | Changes the volume for speech | 0–1007 | 80 |
| \\rst\\ | Resets control sequences to default | — | — |
Control sequences for execution of animations that may be mixed with regular text in sentences.
| Control sequence | Function |
|---|---|
| ^run(A) | Suspend speech, run animation A, resume speech |
| ^start(A) | Start animation A (speech is uninterrupted) |
| ^stop(A) | Stop animation A |
| ^wait(A) | Suspend speech, run animation A until it is finished, then resume speech |
FIGURE 4Illustration of how items in the configuration file control the appearance of the GUI.
Results from questionnaire answered by four users U1-U4.
| Question | U1 | U2 | U3 | U4 |
|---|---|---|---|---|
| Q1: It didn’t take me long to install the system | 2 | 5 | 3 | 2 |
| Q2: I think people with limited technical know-how could follow the installation guide | 2 | 5 | 2 | 4 |
| Q3: The visual structure of the interface is coherent and supportive with respect to operating Pepper | 4 | 5 | 4 | 4 |
| Q4: It didn’t need a lot of practice to use the tool | 4 | 5 | 5 | 4 |
| Q5: I prefer using WoZ4U over alternative systems to control Pepper | 4 | 5 | 3 | 3 |
| Q6: I quickly understood how to edit the configuration file to change the elements in the interface | 4 | 5 | 3 | 5 |
| Q7: The buttons and other controls produced predictable results on the robot | 3 | 5 | 4 | 4 |
| Q8: Rate the overall usability (“ease of use”) of the WoZ4U system | 7 | 9 | 6 | 8 |