SpeechPlay: composing and sharing expressive speech through visually augmented text

  • Authors:
  • Kian Peen Yeo;Suranga Nanayakkara

  • Affiliations:
  • Singapore University of Technology and Design, Singapore;Singapore University of Technology and Design, Singapore

  • Venue:
  • Proceedings of the 25th Australian Computer-Human Interaction Conference: Augmentation, Application, Innovation, Collaboration
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

SpeechPlay allows users to create and share expressive synthetic voices in a fun and interactive manner. It promotes a new level of self-expression and public communication by adding expressiveness to a plain text. Control of prosody information in synthesized speech output is based on the visual appearance of the text, which can be manipulated with touch gestures. Users could create/modify contents using their mobile phone (SpeechPlay Mobile application) and publish/share their work on a large screen (SpeechPlay Surface). Initial user reactions suggest that the correlation between the visual appearance of a text phrase and the resulting audio was intuitive. While it is possible to make the speech output more expressive, users could easily distort the naturalness of the voice in a fun manner. This could also be a useful tool for music composers and for training new musicians.