Most research to date on spoken language interaction has focused on supporting dialog with single users in limited domains and contexts. Efforts in this space have led to significant progress, including wide-scale deployments of telephony-based systems, multimodal voice search applications and voice-enabled personal assistants. At the same time, numerous and important challenges in the realm of physically situated, open-world interaction (e.g., Xbox avatars in the living room, in-car systems, robots in a public space) have remained largely unaddressed.
In this talk, I will give an overview of the Situated Interaction project at Microsoft Research, which aims to address these challenges, and enable a new generation of interactive systems that reason more deeply about their surroundings and embed spoken interaction into the natural flow of everyday tasks, activities and collaborations. Specifically, I will outline a core set of computational models required for supporting physically situated spoken language interactions in open-world environments (e.g. engagement, turn-taking, interaction planning), and discuss our recent research efforts in this space.
Dan Bohus is a researcher in the Adaptive Systems and Interaction Group at Microsoft Research. His research agenda is focused on physically situated, open-world spoken language interaction. Before joining Microsoft Research, Dan has received his Ph.D. degree (2007) in Computer Science from Carnegie Mellon University.