Abstract

Manual grading of essays by humans is time-consuming and likely to be susceptible to inconsistencies and inaccuracies. In recent years, an abundance of research has been done to automate essay evaluation processes, yet little has been done to take into consideration the syntax, semantic coherence and sentiments of the essay's text together. Our proposed system incorporates not just the rule-based grammar and surface level coherence check but also includes the semantic similarity of the sentences. We propose to use Graph-based relationships within the essay's content and polarity of opinion expressions. Semantic similarity is determined between each statement of the essay to form these Graph-based spatial relationships and novel features are obtained from it. Our algorithm uses 23 salient features with high predictive power, which is less than the current systems while considering every aspect to cover the dimensions that a human grader focuses on. Fewer features help us get rid of the redundancies of the data so that the predictions are based on more representative features and are robust to noisy data. The prediction of the scores is done with neural networks using the data released by the ASAP competition held by Kaggle. The resulting agreement between human grader's score and the system's prediction is measured using Quadratic Weighted Kappa (QWK). Our system produces a QWK of 0.793.

Highlights

  • Essay writing is used in many academic disciplines as a form of evaluation

  • We propose an automated essay scoring (AES) system which uses a fewer number of high-quality, independent variables and provides the essence of the essay which is used to accurately predict the score

  • We propose to use semantics based graph based relationships within the essay content and polarity of opinion expressions

Read more

Summary

INTRODUCTION

Essay writing is used in many academic disciplines as a form of evaluation. Generally, a human grader assesses and assigns a score to an essay submission which is written concerning an essay’s prompt. Previous study [1] has shown that when AES is compared with human graders about crucial characteristics of a good essay, the top responses are about the analysis of how the essay revolves around the question prompt, how well structured and sleek the information flow is, quality of grammar used, length, spellings, and punctuation. With respect to these responses, features are extracted from the essays and . Different prediction models are tested to find the one which works best for predicting the essay scores

RELATED WORK
Findings
EVALUATION METRIC
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call