This paper studies non-physical feedback mechanisms to guide human workers toward ergonomic body postures. Specifically, the focus is to solve the tasks that involve no direct physical interaction between the human and the robotic system, therefore tactile guidance by the robot b
...
This paper studies non-physical feedback mechanisms to guide human workers toward ergonomic body postures. Specifically, the focus is to solve the tasks that involve no direct physical interaction between the human and the robotic system, therefore tactile guidance by the robot body is not feasible. We propose a multi-modal ergonomic posture guidance system that comprises visual feedback and speech-based audio feedback. We hypothesise that the proposed multi-modal system leads to better performance compared to uni-modal feedback systems when trying to guide users from one pose to another. To test the hypothesis we conducted an experiment that compared conditions with only audio feedback, only visual feedback and multi-modal feedback. In addition, we examined speech-based audio guidance in joint space and in endpoint space. The results showed that the speech-based feedback in joint space came out as the preferred audio feedback due to its ability to allow users to carry out efficient and coordinated inter-joint movements, especially in cases of high redundancy. Furthermore, the proposed multi-modal feedback system was superior compared to the other feedback modalities both in terms of objective measures and subjective measures.
@en