A Comparative Study of Modified Principal Component Analysis and Generalized Latent Semantic Analysis Approach to Automated Marking of Theoretical-Based Exams

Examination is a means by which individual’s ability on an acquired skill or knowledge is evaluated. Some examiners useMultiple Choice Questions (MCQ), while others use the theory-based Questions. This study performed a comparative study of two techniques: Modified Principal Component Analysis and Generalized Latent Semantic Analysis used in grading students’ responses in theory-based examination. Softcopy of students’ responses and examiner’s marking scheme were captured in an electronic format of .txt. file. The inherent stopwords and stemming in the .txt document were pre-processed to address morphological variations using standard stopwords list and porters stemmer algorithm, respectively. N-gram terms were derived for each student’s response and the marking schemes (MS) using the vector space model. A Document Term Matrix (DTM) was generated with n-gram terms of MS and students response representing columns and rows, respectively. MPCA and GLSA algorithms were used to reduce the sparseness of the DTM to obtain a vector representation of the students’ answers and the marking scheme. The reduced vector representation of the students’ answers were graded according to the mark assigned to each question in the marking scheme using cosine similarity measure. The developed Automated Theory-Based Marking System (ATBMS) was implemented in Matrix Laboratory 8.1 (R2013b). Performance of MPCA was compared with GLSA to determine the effectiveness of ATBMS on the grading of students’ answers in COM 317 and COM 325 courses, in terms of Pearson correlation coefficient(r) and coefficient of determination (R 2 ).These performance evaluation shows that MPCA is a better feature extraction techniques.

[1]  Abejide OluAde-Ibijola,et al.  An Expert System for Automated Essay Scoring (AES) in Computing using Shallow NLP Techniques for Inferencing , 2012 .

[2]  Chad W. Buckendahl,et al.  A Review of Strategies for Validating Computer-Automated Scoring , 2002 .

[3]  Md. Monjurul Islam,et al.  Automated essay scoring using Generalized Latent Semantic Analysis , 2010, CIT 2010.

[4]  Lawrence M. Rudner,et al.  An Overview of Three Approaches to Scoring Written Essays by Computer. ERIC Digest. , 2001 .

[5]  Elijah Olusayo,et al.  Evaluation of N-gram Text Representations for Automated Essay-Type Grading Systems , 2015 .

[6]  Claudia Leacock,et al.  Automated evaluation of essays and short answers , 2001 .

[7]  R. Rao From unstructured data to actionable intelligence , 2003 .

[8]  Omidiora Elijah Olusayo,et al.  Development of a Feature Extraction Technique for Online Character Recognition System , 2012 .

[9]  Lawrence M. Rudner,et al.  Automated Essay Scoring Using Bayes' Theorem , 2002 .

[10]  O. Mason,et al.  Automated free text marking with Paperless School , 2002 .

[11]  Ian T. Jolliffe,et al.  Principal Component Analysis , 2002, International Encyclopedia of Statistical Science.

[12]  Shaidah Jusoh,et al.  International Conference on Intelligent and Advanced Systems 2007 Natural Language Interface for Online Sales Systems , 2022 .

[13]  Elijah Olusayo Omidiora,et al.  A modified principal component analysis approach to automated essay-type grading , 2016, 2016 Future Technologies Conference (FTC).

[14]  Gerard Salton,et al.  A vector space model for automatic indexing , 1975, CACM.

[15]  Nirmalie Wiratunga,et al.  Learning to Author Text with textual CBR , 2010, ECAI.

[16]  Oya Kalipsiz,et al.  Advanced Information Extraction with n-gram based LSI , 2008 .

[17]  Jill Burstein,et al.  AUTOMATED ESSAY SCORING WITH E‐RATER® V.2.0 , 2004 .

[18]  William Wresch,et al.  The Imminence of Grading Essays by Computer-25 Years Later , 1993 .

[19]  Ibrahim Adeyanju Generating Weather Forecast Texts with Case Based Reasoning , 2015, ArXiv.

[20]  M. J. Allen,et al.  Assessing Academic Programs in Higher Education , 2003 .