CFP last date
20 May 2024
Reseach Article

A Comparative Study of Ensemble Methods for Students' Performance Modeling

by Mrinal Pandey, S Taruna
International Journal of Computer Applications
Foundation of Computer Science (FCS), NY, USA
Volume 103 - Number 8
Year of Publication: 2014
Authors: Mrinal Pandey, S Taruna
10.5120/18095-9151

Mrinal Pandey, S Taruna . A Comparative Study of Ensemble Methods for Students' Performance Modeling. International Journal of Computer Applications. 103, 8 ( October 2014), 26-32. DOI=10.5120/18095-9151

@article{ 10.5120/18095-9151,
author = { Mrinal Pandey, S Taruna },
title = { A Comparative Study of Ensemble Methods for Students' Performance Modeling },
journal = { International Journal of Computer Applications },
issue_date = { October 2014 },
volume = { 103 },
number = { 8 },
month = { October },
year = { 2014 },
issn = { 0975-8887 },
pages = { 26-32 },
numpages = {9},
url = { https://ijcaonline.org/archives/volume103/number8/18095-9151/ },
doi = { 10.5120/18095-9151 },
publisher = {Foundation of Computer Science (FCS), NY, USA},
address = {New York, USA}
}
%0 Journal Article
%1 2024-02-06T22:34:02.340484+05:30
%A Mrinal Pandey
%A S Taruna
%T A Comparative Study of Ensemble Methods for Students' Performance Modeling
%J International Journal of Computer Applications
%@ 0975-8887
%V 103
%N 8
%P 26-32
%D 2014
%I Foundation of Computer Science (FCS), NY, USA
Abstract

Student performance prediction is a great area of concern for educational institutions to prevent their students from failure by providing necessary support and counseling to complete their degree successfully. The scope of this research is to examine the accuracy of the ensemble techniques for predicting the student's academic performance, particularly for four year engineering graduate program. To this end, five ensemble techniques based on four representative learning algorithms, namely Adaboost, Bagging, Random Forest and Rotation Forest have been used to construct and combine different number of ensembles. These four algorithms have been compared for the same number (ten) of base classifiers and the Rotation Forest is found to be the best ensemble classifiers for predicting the student performance at the initial stages of the degree program.

References
  1. Dietterich TG. (2000). Ensemble methods in machine learning. In: Proceedings of Multiple Classifier System. vol. 1857. Springer;( 2000). pp. 1–15.
  2. Bauer, E. and Kohavi, R. 1999. An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants. Machine Learning, 35: 1-38.
  3. Domingos, P. 1996. Using Partitioning to Speed Up Specific-to-General Rule Induction. In Proceedings of the AAAI-96 Workshop on Integrating Multiple Learned Models, pp. 29-34, AAAI Press. .
  4. Opitz, D. and Maclin, R. 1999. Popular Ensemble Methods: An Empirical Study, Journal of Artificial Research, 11: 169-198, 1999.
  5. Quinlan, J. R. 1996. Bagging , Boosting, and C4. 5. In Proceedings of the Thirteenth National Conference on Artificial Intelligence, pages 725-730.
  6. Breiman L. 1996. Bagging predictors, Machine Learning, 24(2):123-140.
  7. Efron, B. and Tibshirani, R. J. 1993. An Introduction to the Bootstrap, Chapman & Hall, New York.
  8. Freund, Y. and Schapire, R. Experiments with a New BoostingAlgorithm, Proceedings: ICML'96, 148–156.
  9. Breiman, L. : Random forests. 2001. Mach. Learn. 45(1), 5–32 .
  10. Rodrguez, J. J. , Kuncheva , L. I. and Alonso, C. J. 2006. Rotation forest: A new classifier ensemble method,IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, no. 10, pp. 1619-1630, 2006.
  11. Tanner, T. and Toivonen H. 2010. Predicting and preventing student failure – using the k-nearest neighbour method to predict student performance in an online course environment,International Journal of Learning Technology(IJLT) ,Vol. 5, pages 356-377.
  12. Wilhelmiina, H. and Vinni, M. 2006. Comparison of machine learning methods for intelligent tutoring systems,Intelligent Tutoring Systems, Vol. 4053, of Lecture Notes in Computer Science, pages 525-534.
  13. Kalles, D. and Pierrakeas C. 2004. Analyzing student performance in distance learning with genetic algorithms and decision trees, Laboratory of Educational Material and Educational Methodology Hellenic Open University, Patras, Greece.
  14. Dekker,G. , Pechenizkiy, M. , and Vleeshouwers , J. 2009. Predicting students drop out: a casestudy. Proceedings of the 2nd International Conference on Educational Data Mining, pages 41-50.
  15. Kotsiantis,S. , Patriarcheas, K. , Xenos, M. 2010. A combinational incremental ensemble of classifiers as a technique for predicting students' performance in distance education. ScienceDirect ,Knowledge-Based Systems 23 (2010) Pages 529–535, journal homepage: www. elsevier. com/ locate/knosys.
  16. Kumar , S. Anupama and Vijayalakshmi, M. N . 2011. Efficiency of Decision Trees in Predicting Student,Academic Performance. Computer Science & Information Technology 02, pp. 335–343,2011.
  17. Sharaf, Ahmed. , Malaka, ElDen. , Moustafa, A. , Harb, Hany M. , Emara, Abdel H. (2013). Adaboost Ensemble with Simple Genetic Algorithm for Student Prediction Model, International Journal of Computer Science & Information Technology (IJCSIT) Vol 5, No 2, April 2013.
  18. Essa, Alfred. and Ayad, Hanan. 2012. Improving student success using predictive models and data Visualizations, Research in Learning Technology Supplement: ALT-C 2012 Conference Proceedings, ISBN 978-91-977071-4-5 (print), 978-91-977071-5-2(online), http://dx. doi. org/10. 3402/rlt. v20i0. 19191.
  19. M. , Bahador B. Nooraei. , and Heffernan, Neil T. 2011. Ensembling Predictions of Student Knowledge within Intelligent Tutoring Systems, Joseph A. Konstan et al. (Eds. ): UMAP 2011, LNCS 6787, pp. 13–24, 2011. © Springer-Verlag Berlin Heidelberg .
  20. Weka, University of Waikato, New Zealand,http://www. cs. waikato. ac. nz/ml/weka/.
  21. Galar, Mikel. , Fern´andez, Alberto. , Barrenechea, Edurne. , Bustince Humberto, and Herrera, Francisco . 2011. A Review on Ensembles for the Class Imbalance Problem: Bagging-, Boosting-, and Hybrid-Based Approaches. IEEE Transcations on systems, Man and Cybernetics—Part –c:Applications and Reviews, Digital Object Identifier 10. 1109/TSMCC. 2011. 2161285.
  22. Han, J. , Kamber M. and Pie. 2006. Data Mining Concepts and Techniques 2nd edition, Morgan Kaufmann Publishers 2006.
  23. Qasem A. Al-Radaideh, Emad M. Al-Shawakfa, and Mustafa I. Al-Najjar. 2006. Mining Student Data Using DecisionTrees. Published in proceedings of International Arab Conference on Information Technology, 2006
Index Terms

Computer Science
Information Sciences

Keywords

Prediction Efficiency Ensembles Performance Learning Algorithms