Visual priming to improve keyword detection in free speech dialogue

Qu, Chao, Willem-Paul Brinkman, Wiggers, Pascal, Heynderickx, Ingrid. Visual priming to improve keyword detection in free speech dialogue. In Proceedings of the 28th Annual European Conference on Cognitive Ergonomics. ECCE '10, pages 337-338, ACM, New York, NY, USA, 2010. [doi]

Abstract

Motivation – Talking out loud with synthetic characters in a virtual world is currently considered as a treatment for social phobic patients. The use of keyword detection, instead of full speech recognition will make the system more robust. Important therefore is the need to increase the chance that users use specific keywords during their conversation.

Research approach – A two by two experiment, in which participants (n = 20) were asked to answer a number of open questions. Prior to the session participants watched priming videos or unrelated videos. Furthermore, during the session they could see priming pictures or unrelated pictures on a whiteboard behind the person who asked the questions.

Findings/Design – Initial results suggest that participants more often mention specific keywords in their answers when they see priming pictures or videos instead of unrelated pictures or videos.

Research limitations/Implications – If visual priming in the background can increase the chance that people use specific keywords in their discussion with a dialogue partner, it might be possible to create dialogues in a virtual environment which users perceive as natural.

Take away message – Visual priming might be able to steer people’s answers in a dialogue.

Technical report

The following is a technical report version of this publication:

Predecessors

The following are older variants of this publication: