TY - GEN
T1 - Indonesian Automated Short Answer Scoring Using Sentence Transformers and Siamese LSTM
AU - Chamidah, Nurul
AU - Budi, Indra
AU - Aji, Rizal Fathoni
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - The objective of automated short-answer scoring is to provide automated scores that are as close as possible to human scores. However, the lexicon-based approach using a bag of words has not yielded satisfactory score predictions due to the possibility that sentences with identical meanings can be conveyed using different words. To address the limitation, this study proposes the use of semantic features derived from pretrained Sentence BERT (SBERT) embedding vectors, which are then fed into the Siamese LSTM. The proposed model was evaluated using an Indonesian automated short answer scoring benchmark dataset. Furthermore, cosine similarity and Manhattan distance were evaluated as output functions on the Siamese LSTM. The experimental results obtained from the test data yielded an optimal Pearson's correlation of 0.920 and a Root Mean Squared Error (RMSE) of 10.019 when SBERT and Siamese LSTM were fine-tuned using Manhattan distance as the output function. These values show a correlation of 0.77 and an RMSE of 10.019, which are 0.77 and 7.154 higher, respectively, than the values reported in the previous study. The test result for cosine similarity as the output function yielded the optimal RMSE of 11.494 and correlation of 0.892, which represent improvements of 5.679 and 0.049, respectively, in RMSE and correlation, respectively, over the previous study. These results illustrate that the integration of sentence embeddings and LSTM in this study has produced superior results in Indonesian automated short answer scoring on the benchmark dataset using lexicon-based techniques.
AB - The objective of automated short-answer scoring is to provide automated scores that are as close as possible to human scores. However, the lexicon-based approach using a bag of words has not yielded satisfactory score predictions due to the possibility that sentences with identical meanings can be conveyed using different words. To address the limitation, this study proposes the use of semantic features derived from pretrained Sentence BERT (SBERT) embedding vectors, which are then fed into the Siamese LSTM. The proposed model was evaluated using an Indonesian automated short answer scoring benchmark dataset. Furthermore, cosine similarity and Manhattan distance were evaluated as output functions on the Siamese LSTM. The experimental results obtained from the test data yielded an optimal Pearson's correlation of 0.920 and a Root Mean Squared Error (RMSE) of 10.019 when SBERT and Siamese LSTM were fine-tuned using Manhattan distance as the output function. These values show a correlation of 0.77 and an RMSE of 10.019, which are 0.77 and 7.154 higher, respectively, than the values reported in the previous study. The test result for cosine similarity as the output function yielded the optimal RMSE of 11.494 and correlation of 0.892, which represent improvements of 5.679 and 0.049, respectively, in RMSE and correlation, respectively, over the previous study. These results illustrate that the integration of sentence embeddings and LSTM in this study has produced superior results in Indonesian automated short answer scoring on the benchmark dataset using lexicon-based techniques.
KW - automated short answer scoring
KW - SBERT
KW - sentence transformers
KW - Siamese LSTM
UR - http://www.scopus.com/inward/record.url?scp=105004581727&partnerID=8YFLogxK
U2 - 10.1109/ICIC64337.2024.10957673
DO - 10.1109/ICIC64337.2024.10957673
M3 - Conference contribution
AN - SCOPUS:105004581727
T3 - 2024 9th International Conference on Informatics and Computing, ICIC 2024
BT - 2024 9th International Conference on Informatics and Computing, ICIC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 9th International Conference on Informatics and Computing, ICIC 2024
Y2 - 24 October 2024 through 25 October 2024
ER -