AUTOMATED SHORT ANSWER SCORING USING SEMANTIC SIMILARITY BASED ON WORD EMBEDDINGS

Short-anwertests are one of the ways used to assess students which is an important part of learning. It is considered capable of measuring students' abilities that are complex such as the ability to arrange ideas and arguments related to the questions asked. However, the short-anwer appraisal p...

全面介紹

Saved in:
書目詳細資料
主要作者: Mutaqin
格式: Theses
語言:Indonesia
在線閱讀:https://digilib.itb.ac.id/gdl/view/42747
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
機構: Institut Teknologi Bandung
語言: Indonesia
實物特徵
總結:Short-anwertests are one of the ways used to assess students which is an important part of learning. It is considered capable of measuring students' abilities that are complex such as the ability to arrange ideas and arguments related to the questions asked. However, the short-anwer appraisal process has a disadvantage that is it requires a long time to assess and there is subjectivity between assessors so that it is possible to have different assessment results. Therefore, automatedshort-answer scoring is needed to help the assessment process to be faster and more objective. Research related to automated short-answer scoring continues to evolve. The main problem raised in the research related to automated essay answer scoring is how to improve the accuracy of the assessment so that it approaches the results of human scoring. The process of evaluating essay answers is basically comparing student answers with the correct answer reference. Student answers are considered correct if they have similarities with reference answer semantically and there are no opposite meanings. This study proposes improving the accuracy of short-anwer scoring by utilizing word-embedding semantic similarity measurements and syntactic analysis. Utilization of the word-embedding model is used to anticipate the diversity of students' answers because the model is able to map words that are adjacent to meanings. Meanwhile, syntactic analysis is used to detect meaning of opposing sentences by utilizing part-of speech tags and dependency relationships in sentences. The results show a good correlation between automatic essay answer evaluation with human assessment manually with a correlation coefficient of 0.7085. The level of accuracy of the automatic essay answer assessment as measured by the average absolute error also shows better results than previous studies.