The relative impact of student affect on performance models in a spoken dialogue tutoring system

  • Authors:
  • Kate Forbes-Riley;Mihai Rotaru;Diane J. Litman

  • Affiliations:
  • Learning Research and Development Center, University of Pittsburgh, Pittsburgh, USA 15260;Computer Science Department, University of Pittsburgh, Pittsburgh, USA 15260;Learning Research and Development Center, University of Pittsburgh, Pittsburgh, USA 15260 and Computer Science Department, University of Pittsburgh, Pittsburgh, USA 15260

  • Venue:
  • User Modeling and User-Adapted Interaction
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

We hypothesize that student affect is a useful predictor of spoken dialogue system performance, relative to other parameters. We test this hypothesis in the context of our spoken dialogue tutoring system, where student learning is the primary performance metric. We first present our system and corpora, which have been annotated with several student affective states, student correctness and discourse structure. We then discuss unigram and bigram parameters derived from these annotations. The unigram parameters represent each annotation type individually, as well as system-generic features. The bigram parameters represent annotation combinations, including student state sequences and student states in the discourse structure context. We then use these parameters to build learning models. First, we build simple models based on correlations between each of our parameters and learning. Our results suggest that our affect parameters are among our most useful predictors of learning, particularly in specific discourse structure contexts. Next, we use the PARADISE framework (multiple linear regression) to build complex learning models containing only the most useful subset of parameters. Our approach is a value-added one; we perform a number of model-building experiments, both with and without including our affect parameters, and then compare the performance of the models on the training and the test sets. Our results show that when included as inputs, our affect parameters are selected as predictors in most models, and many of these models show high generalizability in testing. Our results also show that overall, the affect-included models significantly outperform the affect-excluded models.