The Effect of Model Granularity on Student Performance Prediction Using Bayesian Networks

  • Authors:
  • Zachary A. Pardos;Neil T. Heffernan;Brigham Anderson;Cristina L. Heffernan

  • Affiliations:
  • Worcester Polytechnic Institute, Carnegie Mellon University, Worcester Public Schools,;Worcester Polytechnic Institute, Carnegie Mellon University, Worcester Public Schools,;Worcester Polytechnic Institute, Carnegie Mellon University, Worcester Public Schools,;Worcester Polytechnic Institute, Carnegie Mellon University, Worcester Public Schools,

  • Venue:
  • UM '07 Proceedings of the 11th international conference on User Modeling
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

A standing question in the field of Intelligent Tutoring Systems and User Modeling in general is what is the appropriate level of model granularity (how many skills to model) and how is that granularity derived? In this paper we will explore models with varying levels of skill generality (1, 5, 39 and 106 skill models) and measure the accuracy of these models by predicting student performance within our tutoring system called ASSISTment as well as their performance on a state standardized test. We employ the use of Bayes nets to model user knowledge and to use for prediction of student responses. Our results show that the finer the granularity of the skill model, the better we can predict student performance for our online data. However, for the standardized test data we received, it was the 39 skill model that performed the best. We view this as support for fine-grained skill models despite the finest grain model not predicting the state test scores the best.