Proceedings of the 6th International Workshop on Traceability in Emerging Forms of Software Engineering
On integrating orthogonal information retrieval methods to improve traceability recovery
ICSM '11 Proceedings of the 2011 27th IEEE International Conference on Software Maintenance
A comparative evaluation of two user feedback techniques for requirements trace retrieval
Proceedings of the 27th Annual ACM Symposium on Applied Computing
Proceedings of the 34th International Conference on Software Engineering
Proceedings of the 34th International Conference on Software Engineering
Hi-index | 0.00 |
Software Engineering challenges and contests are becoming increasingly popular for focusing researchers' efforts on particular problems. Such contests tend to follow either an exploratory model, in which the contest holders provide data and ask the contestants to discover ``interesting things'' they can do with it, or task-oriented contests in which contestants must perform a specific task on a provided dataset. Only occasionally do contests provide more rigorous evaluation mechanisms that precisely specify the task to be performed and the metrics that will be used to evaluate the results. In this paper, we propose actionable and crowd-sourced contests: actionable because the contest describes a precise task, datasets, and evaluation metrics, and also provides a downloadable operating environment for the contest; and crowd-sourced because providing these features creates accessibility to Information Technology hobbyists and students who are attracted by the challenge. Our proposed approach is illustrated using research challenges from the software traceability area as well as an experimental workbench named TraceLab.