Semi-automated tool for providing effective feedback on programming assignments

Human grading of introductory programming assignments is tedious and error-prone, hence researchers have attempted to develop tools that support automatic assessment of programming code. However, most such efforts often focus only on scoring solutions, rather than assessing whether students correctl...

Full description

Saved in:
Bibliographic Details
Main Authors: BEH, Min Yan, GOTTIPATI, Swapna, LO, David, SHANKARARAMAN, Venky
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2016
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/3748
https://ink.library.smu.edu.sg/context/sis_research/article/4750/viewcontent/Semi_AutomatedTool_Feedback_ICCE_2016_pv.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
Description
Summary:Human grading of introductory programming assignments is tedious and error-prone, hence researchers have attempted to develop tools that support automatic assessment of programming code. However, most such efforts often focus only on scoring solutions, rather than assessing whether students correctly understand the problems. To aid the students improve programming skills, effective feedback on programming assignments plays an important role. Individual feedback generation is tedious and painstaking process. We present a tool that not only automatically generates the static and dynamic program analysis outcomes, but also clusters similar code submissions to provide scalable and effective feedback to the students. We studied our tool on data from introductory Java programming assignments of year 1 course in School of Information Systems. In this paper, we share the details of our tool and findings of our experiments on 261 code submissions.