Call for Paper - November 2022 Edition
IJCA solicits original research papers for the November 2022 Edition. Last date of manuscript submission is October 20, 2022. Read More

Automated Essay Scoring using Word2vec and Support Vector Machine

Print
PDF
International Journal of Computer Applications
Foundation of Computer Science (FCS), NY, USA
Year of Publication: 2019
Authors:
A. E. E. Elalfi, A. F. Elgamal, N. A. Amasha
10.5120/ijca2019919707

A E E Elalfi, A F Elgamal and N A Amasha. Automated Essay Scoring using Word2vec and Support Vector Machine. International Journal of Computer Applications 177(25):20-29, December 2019. BibTeX

@article{10.5120/ijca2019919707,
	author = {A. E. E. Elalfi and A. F. Elgamal and N. A. Amasha},
	title = {Automated Essay Scoring using Word2vec and Support Vector Machine},
	journal = {International Journal of Computer Applications},
	issue_date = {December 2019},
	volume = {177},
	number = {25},
	month = {Dec},
	year = {2019},
	issn = {0975-8887},
	pages = {20-29},
	numpages = {10},
	url = {http://www.ijcaonline.org/archives/volume177/number25/31053-2019919707},
	doi = {10.5120/ijca2019919707},
	publisher = {Foundation of Computer Science (FCS), NY, USA},
	address = {New York, USA}
}

Abstract

Essay scoring is one of the most important tools for evaluating and assessing the level of achievement of educational goals. It aims to innovate performance, arrange, integrate ideas, and connect them by using the vocabulary of the particular subjects. Human essay scoring consumes a lot of time and effort, this leads to mistakes. Automated Essay Scoring (AES) solve to great extent problems. A new approach for AES is presented. It is based on Natural Language Processing (NLP) which is used to unify linguistic answers, word2vec model which converts words into features and synonyms in semantic space, Support Vector Machine(SVM) is used to classify students answers and estimate score levels. The system stages consist of preprocessing, feature extraction, classification and similarity algorithm. The results of proposed method reaches high precision (94%) relative to human resident scores.

References

  1. Salvatore Valenti, Francesca Neri and Alessandro Cucchiarelli, "An Overview of Current Research on Automated Essay Grading", Journal of Information Technology Education, 2003
  2. Mark D. Shermis, Jill Burstein and el, "Automated Essay Scoring: Writing Assessment and Instruction,chapter Automated Essay Scoring".
  3. Semire Dikli," An Overview of Automated Scoring of Essays", The Journal of Technology, Learning, and Assessment, Volume 5, Number1, August 2006.
  4. John Palmer, Robert Williams & Heinz Dreher," Automated Essay Grading System Applied to a First Year University Subject – How Can We do it Better?", InSITE - “Where Parallels Intersect”, 2002.
  5. Arshad Arafat, Mohammed Raihanuzzaman," Automated Essay Grading with Recommendation", Bachelor of Science In Computer Science And Engineering, Department of Computer Science & Engineering, BRAC University, 2016.
  6. L. Streeter, J. Pstoka, D. Laham and D. MacCuish, "The credible grading machine: Automated essay scoring in the dod", 2003.
  7. Govinnage R. Perera, Deenuka N. Perera, A. R. Weerasinghe, "A Dynamic Semantic Space Modelling Approach for Short Essay Grading", International Conference on Advances in ICT for Emerging Regions (ICTer), 2015.
  8. André Smolentzov, “ Automated Essay Scoring”, Scoring Essays in Swedish, 2012.
  9. Parag A. Guruji and el, “ Evaluation Of Subjective Answers using Glsa Enhanced with Contextual Synonymy”, International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015.
  10. Hongbo Chen and Ben He,” Automated Essay Scoring by Maximizing Human-machine greement”, Conference on Empirical Methods in Natural Language Processing, Seattle, Washington,201
  11. Ketul Barot, Jialing Zhang, Seung Woo Son, “Using Natural Language Processing Modelsfor Understanding Network Anomalies”, IEEE, 2016.
  12. Isabelle Guyon and Andr´e Elisseeff, “An Introduction to Feature Extraction”, 2006.
  13. T. Mikolov, K. Chen, G. Corrado, J. Dean, “Efficient Estimation of Word Representations in Vector Space,” Proc. Workshop at ICLR, 2013.
  14. T. Mikolov, I. Sutskever, K. Chen, G. Corrado, J. Dean, “Distributed Representations of Words and Phrases and their Compositionality,” Proc. NIPS, 2013.
  15. T. Mikolov, W.Yih, G. Zweig, “Linguistic Regularities in Continuous Space Word Representations,” Proc. NAACL HLT, 2013.
  16. Long Ma, Yanqing Zhang, “Using Word2Vec to Process Big Text Data”, IEEE International Conference on Big Data, 2015.
  17. Rajni Jindal, Ruchika Malhotra and etc, “Techniques for text classification: Literature review and current trends”, Webology, Volume 12, Number 2, December, 2015.
  18. Joseph Lilleberg, Yun Zhu, Yanqing Zhang , “Support Vector Machines and Word2vec for Text
  19. Classification with Semantic Features”, Proc. 1IEE 14th Inl'l Coni. on Cognitive Inlormatics & Cognitive Computing, 2015.
  20. Dhanashree K. Thakur, Babasaheb Ambedkar, " Natural Language Processing Basic Tasks for Text Mining ", International Journal of Emerging Technologies in Computational and Applied Sc, 2014.Open NLP (http://opennlp.apache.org).
  21. C. D. Manning, Surdeanu, Mihai, Bauer, John, Finkel, Jenny, Bethard, J. Steven, McClosky and David, "The Stanford CoreNLP Natural Language Processing Toolkit.," 2014.
  22. "The Art of Tokenization", available : https://www.ibm.com/developerworks/community/blogs/nlp/entry/tokenization?lang=en, 2013.
  23. https://nlp.stanford.edu/software/tokenizer.shtml.
  24. Abhimanyu Chopra, Abhinav Prashar, Chandresh Sain, " Natural Language Processing", International Journal of Technology Enhancements and Emerging Engineering Research, Vol 1, Issue 4, 2013.
  25. Eduard Heindl, Ajay Singh, " Topic- Stemming Technology", " Course- Master Of Sciences- Business Consulting".
  26. Anjali Ganesh Jivani," A Comparative Study of Stemming Algorithms", IJCTA ,NOV-DEC 2011.
  27. Jaideepsinh K. Raulji , Jatinderkumar R. Saini, "Stop-Word Removal Algorithm and its Implementation for Sanskrit Language ", International Journal of Computer Applications , 2016.
  28. http://examples.yourdictionary.com/examples-of-adverbs.html.
  29. https://en.wikipedia.org/wiki/Word2vec.
  30. “What are the continuous bag of words and skip-gram architectures?” [Online]. Available: https://www.quora.com/What-are-the-continuous-bag-of-words-and-skip-gram-architectures.
  31. Alabhya Farkiya, Prashant Saini and etc. , “Natural Language Processing using NLTK and WordNet” , International Journal of Computer Science and Information Technologies, Vol. 6 (6) , 2015, 5465-5469.
  32. C. Chang and C.-J. Lin, \LIBSVM: A library for support vector machines,"ACM Transactions on Intelligent Systems and Technology, vol. 2, pp. 27:1{27:27, 2011.
  33. Suresha M, Shilpa N.A, “ Apples Grading based on SVM Classifier”, International Journal of Computer Applications (0975 – 8878) on National Conference on Advanced Computing and Communications - NCACC, April 2012.
  34. Grigori Sidorov, Alexander Gelbukh, Helena G´omez-Adorno, “Soft Similarity and Soft Cosine Measure: Similarity of Features in Vector Space Model”, Computación y Sistemas Vol. 18, No. 3, 2014 pp. 491–504.
  35. https://www.dataschool.io/simple-guide-to-confusion-matrix-terminology/
  36. Md. Monjurul Islam, A. S. M. Latiful Hoque, "Automated Essay Scoring Using Generalized Latent Semantic Analysis", Journal of Computers, Vol. 7, No. 3, March 2012.

Keywords

Automated Essay Scoring, Word2vec, Support Vector Machine, Natural Language Processing