CFP last date
20 June 2024
Reseach Article

Composing Sequential Test Items with Multipart Criteria in Adaptive Testing

Published on December 2012 by Kavitha Rajamani, Vijaya Kathiravan
EGovernance and Cloud Computing Services - 2012
Foundation of Computer Science USA
EGOV - Number 3
December 2012
Authors: Kavitha Rajamani, Vijaya Kathiravan
63958b81-1aa9-4d02-bb9a-c6ccf45ebb1e

Kavitha Rajamani, Vijaya Kathiravan . Composing Sequential Test Items with Multipart Criteria in Adaptive Testing. EGovernance and Cloud Computing Services - 2012. EGOV, 3 (December 2012), 3-7.

@article{
author = { Kavitha Rajamani, Vijaya Kathiravan },
title = { Composing Sequential Test Items with Multipart Criteria in Adaptive Testing },
journal = { EGovernance and Cloud Computing Services - 2012 },
issue_date = { December 2012 },
volume = { EGOV },
number = { 3 },
month = { December },
year = { 2012 },
issn = 0975-8887,
pages = { 3-7 },
numpages = 5,
url = { /proceedings/egov/number3/9496-1021/ },
publisher = {Foundation of Computer Science (FCS), NY, USA},
address = {New York, USA}
}
%0 Proceeding Article
%1 EGovernance and Cloud Computing Services - 2012
%A Kavitha Rajamani
%A Vijaya Kathiravan
%T Composing Sequential Test Items with Multipart Criteria in Adaptive Testing
%J EGovernance and Cloud Computing Services - 2012
%@ 0975-8887
%V EGOV
%N 3
%P 3-7
%D 2012
%I International Journal of Computer Applications
Abstract

The traditional learning environment is being rapidly supplemented by an E-Learning environment, particularly Computer Assisted Instruction (CAI). Each learner has different learning status and therefore should use different test items in their evaluation. The Computerized Adaptive Test (CAT) can adjust the degree of difficulty of test items dynamically depending on their ability. A good test will not only help the instructor evaluate the learning status of the students, but also facilitate the diagnosis of the problems embedded in the students' learning process. One of the most important and challenging issues in conducting a good test is the construction of test sheets that can meet various criteria. Therefore, several measures have been proposed to represent the quality of each test item, such as degree of difficulty and discrimination. However, the quality of a test not only depends on the quality of the item bank, but also relates to the way the assessment sheet is constructed. Selection of appropriate test items is important when constructing an assessment sheet that meets multi-criteria assessment requirements, such as expected difficulty degree, expected discrimination degree, number of the test items, estimated testing time and the specified distribution of relevant concept weights. Dynamic question generation is proposed which uses the novel approach of Particle Swarm Optimization. This approach will improve the efficiency of composing near optimal serial test items to meet multiple assessment criteria. The proposed approach can be compared with some existing means in terms of efficiency.

References
  1. C. Chou, "Constructing a computer-assisted testing and evaluation system on the world wide web-the CATES experience," IEEE Trans. Educ. , vol. 43, no. 3, pp. 266–272, Aug. 2000.
  2. S. Piramuthu, "Knowledge-based web-enabled agents and intelligent tutoring systems," IEEE Trans. Educ. , vol. 48, no. 4, pp. 750–756, Nov. 2005.
  3. E. Guzm´an and R. Conejo, "Self-assessment in a feasible, adaptive webbased testing system," IEEE Trans. Educ. , vol. 48, no. 4, pp. 688–695, Nov. 2005.
  4. A. Kav?ci?c, "Fuzzy user modeling for adaptation in educational hypermedia," IEEE Trans. Syst. , Man, Cybern. C, Appl. Rev. , vol. 34, no. 4, pp. 439–449, Nov. 2004.
  5. G. -J. Hwang, B. M. T. Lin, H. -H. Tseng, and T. -L. Lin, "On the development of a computer-assisted testing system with genetic test sheet generating approach," IEEE Trans. Syst. , Man, Cybern. C, Appl. Rev. ,vol. 35, no. 4, pp. 590–594, Nov. 2005.
  6. Xiao-Min Hu, Jun Zhang, "An Intelligent Testing System Embedded with an Ant-Colony optimization based Test composition Method", IEEE Trans. Syst. , Man, Cybern. C, Appl. Rev. , vol. 39, no. 6, pp. 659–669, Nov. 2009.
  7. Marks, A. M. , & Cronje, j. C. "Randomised items in computer-based tests: Russian roulette in assessment?". Journal of educationaltechnology & society, 11 (4), 2008.
  8. Betül Erdo?du, "Computer based testing evaluation of question classification for Computer Adaptive testing", A Master Thesis (2009).
  9. Yin, P. -Y. , Chang, K. -C. , Hwang, G. -J. , Hwang, G. -H. , & Chan, Y. , "A Particle Swarm Optimization Approach to Composing Serial Test Sheets", Journal of Educational Technology & Society, 9 (3), 3-15.
  10. Chi-Keung Leung, Hua-Hua Chang, and Kit-Tai Hau, "Computerized Adaptive Testing: A Comparison of Three Content Balancing Methods", The Journal of Technology, Learning, and Assessment, Volume 2, Number 5, December 2003.
  11. Kennedy, J. , & Eberhart, R. C. (1995). Particle swarm optimization. In Proceedings of the IEEE international conference on neural networks 4, pp. 1942–1948.
Index Terms

Computer Science
Information Sciences

Keywords

Computer Adaptive Testing Assessment Sheet Generation Intelligent Testing Intelligent Tutoring