Designing SpeechActs: issues in speech user interfaces
CHI '95 Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Training Personal Robots Using Natural Language Instruction
IEEE Intelligent Systems
The reliability of a dialogue structure coding scheme
Computational Linguistics
Natural behavior of a listening agent
Lecture Notes in Computer Science
Incremental Multimodal Feedback for Conversational Agents
IVA '07 Proceedings of the 7th international conference on Intelligent Virtual Agents
Predicting Listener Backchannels: A Probabilistic Multimodal Approach
IVA '08 Proceedings of the 8th international conference on Intelligent Virtual Agents
ICMI '08 Proceedings of the 10th international conference on Multimodal interfaces
Learning a model of speaker head nods using gesture corpora
Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 1
First Steps Towards the Automatic Construction of Argument-Diagrams from Real Discussions
Proceedings of the 2006 conference on Computational Models of Argument: Proceedings of COMMA 2006
A probabilistic multimodal approach for predicting listener backchannels
Autonomous Agents and Multi-Agent Systems
Proceedings of the Workshop on Use of Context in Vision Processing
Modeling embodied feedback with virtual humans
ZiF'06 Proceedings of the Embodied communication in humans and machines, 2nd ZiF research group international conference on Modeling communication with robots and virtual humans
Latent mixture of discriminative experts for multimodal prediction modeling
COLING '10 Proceedings of the 23rd International Conference on Computational Linguistics
Differences in listener responses between procedural and narrative task
Proceedings of the 2nd international workshop on Social signal processing
Concensus of self-features for nonverbal behavior analysis
HBU'10 Proceedings of the First international conference on Human behavior understanding
Backchannel strategies for artificial listeners
IVA'10 Proceedings of the 10th international conference on Intelligent virtual agents
Learning and evaluating response prediction models using parallel listener consensus
International Conference on Multimodal Interfaces and the Workshop on Machine Learning for Multimodal Interaction
Coherent back-channel feedback tagging of in-car spoken dialogue corpus
SIGDIAL '10 Proceedings of the 11th Annual Meeting of the Special Interest Group on Discourse and Dialogue
The multiLis corpus - dealing with individual differences in nonverbal listening behavior
Proceedings of the Third COST 2102 international training school conference on Toward autonomous, adaptive, and context-aware multimodal interfaces: theoretical and practical issues
Turn-taking cues in task-oriented dialogue
Computer Speech and Language
Modeling wisdom of crowds using latent mixture of discriminative experts
HLT '11 Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies: short papers - Volume 2
Contrasting multi-lingual prosodic cues to predict verbal feedback for rapport
HLT '11 Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies: short papers - Volume 2
Backchannels: quantity, type and timing matters
IVA'11 Proceedings of the 10th international conference on Intelligent virtual agents
When do we smile? analysis and modeling of the nonverbal context of listener smiles in conversation
ACII'11 Proceedings of the 4th international conference on Affective computing and intelligent interaction - Volume Part I
Affirmative cue words in task-oriented dialogue
Computational Linguistics
Online behavior evaluation with the switching wizard of oz
IVA'12 Proceedings of the 12th international conference on Intelligent Virtual Agents
Speaker-adaptive multimodal prediction model for listener responses
Proceedings of the 15th ACM on International conference on multimodal interaction
Hi-index | 0.00 |
Spoken dialogue systems would be more acceptable if they were able to produce backchannel continuers such as mm-hmm in naturalistic locations during the user's utterances. Using the HCRC Map Task Corpus as our data source, we describe models for predicting these locations using only limited processing and features of the user's speech that are commonly available, and which therefore could be used as a low-cost improvement for current systems. The baseline model inserts continuers after a predetermined number of words. One further model correlates back-channel continuers with pause duration, while a second predicts their occurrence using trigram POS frequencies. Combining these two models gives the best results.