Natural language analysis by stochastic optimization: a progress report on project APRIL
Journal of Experimental & Theoretical Artificial Intelligence
Copycat: a computer model of high-level perception and conceptual slippage in analogy-making
Copycat: a computer model of high-level perception and conceptual slippage in analogy-making
Tabletop: an emergent, stochastic computer model of analogy-making
Tabletop: an emergent, stochastic computer model of analogy-making
A stochastic finite-state word-segmentation algorithm for Chinese
Computational Linguistics
Integrating word boundary identification with sentence understanding
ACL '93 Proceedings of the 31st annual meeting on Association for Computational Linguistics
A stochastic finite-state word-segmentation algorithm for Chinese
ACL '94 Proceedings of the 32nd annual meeting on Association for Computational Linguistics
Word identification for Mandarin Chinese sentences
COLING '92 Proceedings of the 14th conference on Computational linguistics - Volume 1
Chinese text retrieval without using a dictionary
Proceedings of the 20th annual international ACM SIGIR conference on Research and development in information retrieval
Automatic generation of English/Chinese thesaurus based on a parallel corpus in laws
Journal of the American Society for Information Science and Technology
Critical tokenization and its properties
Computational Linguistics
COLING '98 Proceedings of the 17th international conference on Computational linguistics - Volume 1
Learning case-based knowledge for disambiguating Chinese word segmentation: a preliminary study
SIGHAN '02 Proceedings of the first SIGHAN workshop on Chinese language processing - Volume 18
Chinese word segmentation as LMR tagging
SIGHAN '03 Proceedings of the second SIGHAN workshop on Chinese language processing - Volume 17
Hi-index | 0.00 |
This paper proposes that the process of language understanding can be modeled as a collective phenomenon that emerges from a myriad of microscopic and diverse activities. The process is analogous to the crystallization process in chemistry. The essential features of this model are: asynchronous parallelism; temperature-controlled randomness; and statistically emergent active symbols. A computer program that tests this model on the task of capturing the effect of context on the perception of ambiguous word boundaries in Chinese sentences is presented. The program adopts a holistic approach in which word identification forms an integral component of sentence analysis. Various types of knowledge, from statistics to linguistics, are seamlessly integrated for the tasks of word boundary disambiguation as well as sentential analysis. Our experimental results showed that the model is able to address the word boundary ambiguity problems effectively.