Evaluating usability evaluation techniques
ACM Computing Surveys (CSUR) - Special issue: position statements on strategic directions in computing research
A Practical Guide to Usability Testing
A Practical Guide to Usability Testing
BLEU: a method for automatic evaluation of machine translation
ACL '02 Proceedings of the 40th Annual Meeting on Association for Computational Linguistics
Which comes first, usability or utility?
Proceedings of the 14th IEEE Visualization 2003 (VIS'03)
BLANC: learning evaluation metrics for MT
HLT '05 Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing
Metrics for evaluating human information interaction systems
Interacting with Computers
Some Improvements over the BLEU Metric for Measuring Translation Quality for Hindi
ICCTA '07 Proceedings of the International Conference on Computing: Theory and Applications
Automatic evaluation of machine translation quality using n-gram co-occurrence statistics
HLT '02 Proceedings of the second international conference on Human Language Technology Research
Journal of Field Robotics
Dependency-based automatic evaluation for machine translation
SSST '07 Proceedings of the NAACL-HLT 2007/AMTA Workshop on Syntax and Structure in Statistical Translation
PerMIS '08 Proceedings of the 8th Workshop on Performance Metrics for Intelligent Systems
Evaluating speech translation systems: applying SCORE to TRANSTAC technologies
PerMIS '09 Proceedings of the 9th Workshop on Performance Metrics for Intelligent Systems
PerMIS '09 Proceedings of the 9th Workshop on Performance Metrics for Intelligent Systems
Probability of successful transfer of low-level concepts via machine translation: a meta-evaluation
PerMIS '09 Proceedings of the 9th Workshop on Performance Metrics for Intelligent Systems
Utility assessment in TRANSTAC: using a set of complementary methods
PerMIS '09 Proceedings of the 9th Workshop on Performance Metrics for Intelligent Systems
Stereo hidden Markov modeling for noise robust speech recognition
Computer Speech and Language
Hi-index | 0.00 |
One of the most difficult challenges that military personnel face when operating in foreign countries is clear and successful communication with the local population. To address this issue, the Defense Advanced Research Projects Agency (DARPA) is funding academic institutions and industrial organizations through the Spoken Language Communication and Translation System for Tactical Use (TRANSTAC) program to develop practical machine translation systems. The goal of the TRANSTAC program is to demonstrate capabilities to rapidly develop and field free-form, two-way, speech-to-speech translation systems that enable speakers of different languages to communicate with one another in real-world tactical situations without an interpreter. Evaluations of these technologies are a significant part of the program and DARPA has asked the National Institute of Standards and Technology (NIST) to lead this effort. This article presents the experimental design of the TRANSTAC evaluations and the metrics, both quantitative and qualitative, that were used to comprehensively assess the systems' performance.