key: cord-0046972-cm4ftw3o authors: Uto, Masaki; Uchida, Yuto title: Automated Short-Answer Grading Using Deep Neural Networks and Item Response Theory date: 2020-06-10 journal: Artificial Intelligence in Education DOI: 10.1007/978-3-030-52240-7_61 sha: 62cd32bbbb34ba3c48958245702c15ff4de520b2 doc_id: 46972 cord_uid: cm4ftw3o Automated short-answer grading (ASAG) methods using deep neural networks (DNN) have achieved state-of-the-art accuracy. However, further improvement is required for high-stakes and large-scale examinations because even a small scoring error will affect many test-takers. To improve scoring accuracy, we propose a new ASAG method that combines a conventional DNN-ASAG model and an item response theory (IRT) model. Our method uses an IRT model to estimate the test-taker’s ability from his/her true-false responses to objective questions that are offered with a target short-answer question in the same test. Then, the target short-answer score is predicted by jointly using the ability value and a distributed short-answer representation, which is obtained from an intermediate layer of a DNN-ASAG model. Short-answer questions are widely used to evaluate the higher abilities of testtakers, such as logical thinking and expressive ability. World-wide large-scale tests, such as the Test of English as a Foreign Language and the Graduate Management Admission Test, incorporate short-answer questions. However, the introduction of this type of question to these large-scale tests has prompted concerns related to scoring accuracy, time complexity, and monetary cost. Automated short-answer grading (ASAG) methods have attracted much attention as a way to alleviate these concerns [1, 2] . Conventional ASAG methods have relied on manually tuned features, which are laborious to develop [3, [10] [11] [12] . However, many deep neural network (DNN) methods, which obviate the need for feature engineering, have been proposed [5, [7] [8] [9] 13] . DNN methods automatically extract effective features for score prediction using a dataset of graded short answers, and have achieved state-ofthe-art scoring accuracy [5, [7] [8] [9] 13] . However, further improvement of the accuracy of these methods is required, especially for high-stakes and large-scale examinations because even a slight scoring error will have a large effect on many test-takers. To improve scoring accuracy, we propose a new ASAG method that combines a conventional DNN model and an item response theory (IRT) model [6] . We focus on short-answer questions given as a part of a test including objective questions. Because a test measures a particular ability, we can assume that short-answer questions and objective questions on the same test measure similar abilities. Thus, estimating the test-takers' ability from the objective questions should be useful for short-answer grading. Based on this assumption, our method incorporates the test-taker's ability, which is estimated using an IRT model from his/her true-false responses for objective questions, into a DNN-ASAG model. Our method is formulated as a DNN framework that predicts a target shortanswer score by jointly using the IRT-based ability estimate and a distributed representation of the short-answer text as obtained from an intermediate layer of a DNN-ASAG model. Although the proposed method is suitable for any DNN-ASAG model, we implement it with the most standard long short-term memory (LSTM) ASAG model [9] . The effectiveness of our model is evaluated by using data from an actual experiment. To our knowledge, this is a new approach that focuses on using responses to objective questions to grade short answers. The architecture of the proposed method is shown in Fig. 1 . The method transforms the word sequence in a given short answer to a fixed-length hidden vector M through a lookup table layer, a LSTM layer, and a temporal mean layer, as in the conventional LSTM ASAG model [9] . Here, the lookup table layer transforms each word in a given short answer to a word embedding representation, the LSTM layer transforms the embedded word sequence to a sequence of hidden vectors that capture the long-distance dependencies of the words at each time step, and the temporal mean layer averages the outputs of the LSTM layer to produce a fixed-length hidden vector M , which can be regarded as a distributed representation of a given short-answer text. The concatenation block, a newly added component in this method, concatenates the distributed text representation M and an IRT-based test-taker's ability θ which is estimated from his/her true-false responses to objective questions offered together with the short-answer question during the same examination. We use the two-parameter logistic IRT model that defines the probability of a testtaker answering correctly for objective question i as ( where θ is the test-taker's ability, and α i and β i are discrimination and difficulty parameters of question i. The fully connected (dense) layer projects the concatenated vector M = [M , θ] to a lower-dimensional hidden vector using a fully connected feedforward neural network. This layer is also newly added in this study to capture the non-linear relation between the test-takers' abilities and short-answer scores. Finally, the linear layer projects the output of the fully connected layer to a scalar value in the range [0, 1] by using the sigmoid function σ (W M + b) , where W is the weight matrix and b is the bias. The model training is conducted by back-propagation with the mean squared error loss function using the training dataset, in which the scores are normalized to the [0, 1] scale. During the prediction phase, the predicted scores are rescaled to the original score range. For the IRT parameter estimation, we use a Markov chain Monte Carlo algorithm [14, 15] . This section demonstrates the effectiveness of the proposed method by using real data. For this experiment, we used response data from a Japanese reading comprehension test developed by Benesse Educational Research and Development Institute, Japan. This dataset comprises responses given by 511 test-takers (Japanese university students) to three short-answer questions and true-false responses for 44 objective questions. Scores for the short answers were provided by expert raters using three rating categories for two evaluation viewpoints. The total score of the two evaluation viewpoints was also given. Using the data, we conducted five-fold cross validation to evaluate the Pearson's correlation between the true scores and predicted scores for each evaluation viewpoint and the total score. For model training, the dimensions of the word embedding, the LSTM layer, and the fully connected layer were set to 50, 300, and 50, respectively. The mini-batch size and maximum epochs were 32 and 50, respectively. The dropout probabilities for the lookup table layer and the temporal mean layer were 0.5. The recurrent dropout probability for the LSTM layer was set to 0.1. This experiment was conducted for the proposed method and the conventional method. Furthermore, to evaluate effectiveness of the fully connected (dense) layer, we also conducted the experiment for the proposed method without the dense layer and the conventional method with the dense layer. Table 1 shows the results. The Score 1 and Score 2 columns indicate the results for the two evaluation viewpoints in each question; the Total column indicates the results for the sum of the two viewpoints' scores; and the Avg. column shows the averaged performance for each method. * indicates that the averaged performance of the method is higher than that of the conventional method at the 1% significance level by the paired t-test. The table shows that the proposed method has better performance than the conventional method in almost all cases, and the averaged performance of the proposed method is also significantly higher. These results suggest that the proposed method is effective in improving the scoring accuracy. The table also shows that the performance tends to decrease when the dense layer is omitted from the proposed method. Moreover, when the dense layer is added to the conventional method, the performance tends to increase. These results suggest that the incorporation of the fully connected dense layer improves the accuracy. Comparing the proposed method and the conventional method with the dense layer shows that the proposed method provides higher performance in all cases except for Question 3, validating the effectiveness of incorporating the IRT-based ability. The drop in performance for Question 3 might be caused by disagreement between the distribution of IRT ability and that of the observed score. We confirmed that Question 3 has a strongly skewed score distribution in which the highest score category is overused, whereas the IRT ability follows a normal distribution [4] . Note that test items with strongly skewed score distributions are generally inappropriate because they do not distinguish the ability of test-takers well. Thus, we conclude that incorporating ability values improves the scoring accuracy when target short-answer questions measure ability well. This study proposed a new DNN-ASAG method that integrates the ability of test-takers estimated from true-false responses for objective questions using IRT. An experiment using real data suggested that incorporating ability improves scoring accuracy when a target short-answer question can measure ability well. In future work, we plan to examine the behavior of the proposed method in more detail by applying it to various datasets. We will also examine the potential for scoring bias that might arise from the use of true-false responses. The eras and trends of automatic short answer grading Balancing human efforts and performance of student response analyzer in dialog-based tutors ETS: domain adaptation and stacking for short answer scoring Handbook of Item Response Theory Automatic short answer grading via multiway attention networks Applications of Item Response Theory to Practical Testing Problems Multiple data augmentation strategies for improving performance on automatic short answer scoring Analytic score prediction and justification identification in automated short answer scoring Investigating neural architectures for short answer scoring Sentence level or token level features for automatic short answer grading?: Use both Effective feature integration for automated short answer scoring Fast and easy short answer grading with high accuracy Improving short answer grading using transformer-based pre-training Rater-effect IRT model integrating supervised LDA for accurate measurement of essay writing ability Item response theory for peer assessment Acknowledgment. This work was supported by JSPS KAKENHI 17H04726 and 17K20024. We thank Yuki Doka and Yoshihiro Kato at Benesse Educational Research and Development Institute for permission to use the actual data.