Multimodal interactive transcription of text images

  • Authors:
  • Alejandro H. Toselli;Verónica Romero;Moisés Pastor;Enrique Vidal

  • Affiliations:
  • Instituto Tecnológico de Informática Universidad Politécnica de Valencia Camino de Vera s/n, 46071 Valencia, Spain;Instituto Tecnológico de Informática Universidad Politécnica de Valencia Camino de Vera s/n, 46071 Valencia, Spain;Instituto Tecnológico de Informática Universidad Politécnica de Valencia Camino de Vera s/n, 46071 Valencia, Spain;Instituto Tecnológico de Informática Universidad Politécnica de Valencia Camino de Vera s/n, 46071 Valencia, Spain

  • Venue:
  • Pattern Recognition
  • Year:
  • 2010

Quantified Score

Hi-index 0.01

Visualization

Abstract

To date, automatic handwriting recognition systems are far from being perfect and heavy human intervention is often required to check and correct the results of such systems. This ''post-editing'' process is both inefficient and uncomfortable to the user. An example is the transcription of historic documents: state-of-the-art handwritten text recognition technology is not suitable to perform this task automatically and expensive paleography expert work is needed to achieve correct transcriptions. As an alternative to fully manual transcription and post-editing, a multimodal interactive approach is proposed here where user feedback is provided by means of touchscreen pen strokes and/or more traditional keyboard and mouse operation. User's feedback directly allows to improve system accuracy, while multimodality increases system ergonomy and user acceptability. Multimodal interaction is approached in such a way that both the main and the feedback data streams help each-other to optimize overall performance and usability. Empirical tests on three cursive handwritten tasks suggest that, using this approach, considerable amounts of user effort can be saved with respect to both pure manual work and non-interactive, post-editing processing.