The TRY system -or- how to avoid testing student programs
SIGCSE '89 Proceedings of the twentieth SIGCSE technical symposium on Computer science education
Grading student programs using ASSYST
SIGCSE '97 Proceedings of the twenty-eighth SIGCSE technical symposium on Computer science education
Experience with an automatically assessed course
Proceedings of the 5th annual SIGCSE/SIGCUE ITiCSEconference on Innovation and technology in computer science education
Grading student programs - a software testing approach
CCSC '00 Proceedings of the fourteenth annual consortium on Small Colleges Southeastern conference
Fully automatic assessment of programming exercises
Proceedings of the 6th annual conference on Innovation and technology in computer science education
Experiences in automatic assessment on mass courses and issues for designing virtual courses
Proceedings of the 7th annual conference on Innovation and technology in computer science education
GAME: A Generic Automated Marking Environment for Programming Assessment
ITCC '04 Proceedings of the International Conference on Information Technology: Coding and Computing (ITCC'04) Volume 2 - Volume 2
An experimental analysis of GAME: a generic automated marking environment
Proceedings of the 9th annual SIGCSE conference on Innovation and technology in computer science education
Automated assessment of GUI programs using JEWL
Proceedings of the 9th annual SIGCSE conference on Innovation and technology in computer science education
Ability-training-oriented automated assessment in introductory programming course
Computers & Education
Software verification and graph similarity for automated evaluation of students' assignments
Information and Software Technology
Hi-index | 0.00 |
This paper describes the Generic Automated Marking Environment (GAME) and provides a detailed analysis of its performance in assessing student programming projects and exercises. GAME has been designed to automatically assess programming assignments written in a variety of languages based on the ''structure'' of the source code and the correctness of the program's output. Currently, the system is able to mark programs written in Java, C++ and the C language. To use the system, instructors are required to provide a simple ''marking schema'' for each given assessment item, which includes pertinent information such as the location of files and the model solution. In this research, GAME has been tested on a number of student programming exercises and assignments and its performance has been compared against that of a human marker. An in-depth statistical analysis of the comparison is presented, providing encouraging results and directions for employing GAME as a tool for teaching and learning.