Grading student programs using ASSYST
SIGCSE '97 Proceedings of the twenty-eighth SIGCSE technical symposium on Computer science education
A gimmick to integrate software testing throughout the curriculum
SIGCSE '02 Proceedings of the 33rd SIGCSE technical symposium on Computer science education
Rethinking computer science education from a test-first perspective
OOPSLA '03 Companion of the 18th annual ACM SIGPLAN conference on Object-oriented programming, systems, languages, and applications
Testing first: emphasizing testing in early programming courses
ITiCSE '05 Proceedings of the 10th annual SIGCSE conference on Innovation and technology in computer science education
MuJava: an automated class mutation system: Research Articles
Software Testing, Verification & Reliability
Helping students appreciate test-driven development (TDD)
Companion to the 21st ACM SIGPLAN symposium on Object-oriented programming systems, languages, and applications
Bug Hunt: Making Early Software Testing Lessons Engaging and Affordable
ICSE '07 Proceedings of the 29th international conference on Software Engineering
The JavaFest: a collaborative learning technique for Java programming courses
Proceedings of the 6th international symposium on Principles and practice of programming in Java
Companion to the 23rd ACM SIGPLAN conference on Object-oriented programming systems languages and applications
Efficient mutation testing by checking invariant violations
Proceedings of the eighteenth international symposium on Software testing and analysis
Javalanche: efficient mutation testing for Java
Proceedings of the the 7th joint meeting of the European software engineering conference and the ACM SIGSOFT symposium on The foundations of software engineering
Running students' software tests against each others' code: new life for an old "gimmick"
Proceedings of the 43rd ACM technical symposium on Computer Science Education
Toward practical mutation analysis for evaluating the quality of student-written software tests
Proceedings of the ninth annual international ACM conference on International computing education research
Automatically assessing the quality of student-written tests
Proceedings of the ninth annual international ACM conference on International computing education research
Automated assessment of students' testing skills for improving correctness of their code
Proceedings of the 2013 companion publication for conference on Systems, programming, & applications: software for humanity
Hi-index | 0.00 |
Learning to program should include learning about proper software testing. Some automatic assessment systems, e.g. Web-CAT, allow assessing student-generated test suites using coverage metrics. While this encourages testing, we have observed that sometimes students can get rewarded from high coverage although their tests are of poor quality. Exploring alternative methods of assessment, we have tested mutation analysis to evaluate students' solutions. Initial results from applying mutation analysis to real course submissions indicate that mutation analysis could be used to fix some problems of code coverage in the assessment. Combining both metrics is likely to give more accurate feedback.