Learning linearly separable languages

  • Authors:
  • Leonid Kontorovich;Corinna Cortes;Mehryar Mohri

  • Affiliations:
  • Carnegie Mellon University, Pittsburgh, PA;Google Research, New York, NY;Google Research, New York, NY

  • Venue:
  • ALT'06 Proceedings of the 17th international conference on Algorithmic Learning Theory
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents a novel paradigm for learning languages that consists of mapping strings to an appropriate high-dimensional feature space and learning a separating hyperplane in that space. It initiates the study of the linear separability of automata and languages by examining the rich class of piecewise-testable languages. It introduces a high-dimensional feature map and proves piecewise-testable languages to be linearly separable in that space. The proof makes use of word combinatorial results relating to subsequences. It also shows that the positive definite kernel associated to this embedding can be computed in quadratic time. It examines the use of support vector machines in combination with this kernel to determine a separating hyperplane and the corresponding learning guarantees. It also proves that all languages linearly separable under a regular finite cover embedding, a generalization of the embedding we used, are regular.