Examining the legibility of humanoid robot arm movements in a pointing task (codes)
Creators
Description
Human–robot interaction requires robots whose actions are legible, allowing humans to interpret, predict, and feel safe around them. This study investigates the legibility of humanoid robot arm movements in a pointing task, aiming to understand how humans predict robot intentions from truncated movements and bodily cues. We designed an experiment using the NICO humanoid robot, where participants observed its arm movements towards targets on a touchscreen. Robot cues varied across conditions: gaze, pointing, as well as congruent or incongruent gaze-pointing. Arm trajectories were stopped at 60% or 80% of their full length, and participants predicted the final target. We tested the multimodal superiority and ocular primacy hypotheses, both of which were confirmed by the experiment.
Files
paper_lucny_icsr_code.zip
Files
(101.6 kB)
Name | Size | Download all |
---|---|---|
md5:db21f39235f92e185ef5683759623864
|
101.6 kB | Preview Download |
Additional details
Related works
- Is supplement to
- Publication: 10.48550/arXiv.2508.05104 (DOI)
Funding
Software
- Repository URL
- https://github.com/andylucny/nico2/tree/main/experiment
- Programming language
- Python