Interactive sonification of collaborative ar-based planning tasks for enhancing joint attention
Abstract
This paper introduces a novel sonification-based interaction support
for cooperating users in an Augmented Reality setting. When
using head-mounted AR displays, the field of view is limited
which causes users to miss important activities such as object
interactions or deictic references of their interaction partner to
(re-)establish joint attention. We introduce an interactive sonification
which makes object manipulations of both interaction partners
mutually transparent by sounds that convey information about the
kind of activity, and can optionally even identify the object itself.
In this paper we focus on the sonification method, interaction design
and sound design, and we furthermore render the sonification
both from sensor data (e.g. object tracking) and manual annotations.
As a spin-off of our approach we propose this method further
for the enhancement of interaction observation, data analysis,
and multimodal annotation in interactional linguistics and conversation
analysis.