An Appraisal Model Based on a Synthetic Feature Selection Approach for Students' Academic Achievement

Obtaining necessary information (and even extracting hidden messages) from existing big data, and then transforming them into knowledge, is an important skill. Data mining technology has received increased attention in various fields in recent years because it can be used to find historical patterns and employ machine learning to aid in decision-making. When we find unexpected rules or patterns from the data, they are likely to be of high value. This paper proposes a synthetic feature selection approach (SFSA), which is combined with a support vector machine (SVM) to extract patterns and find the key features that influence students’ academic achievement. For verifying the proposed model, two databases, namely, “Student Profile” and “Tutorship Record”, were collected from an elementary school in Taiwan, and were concatenated into an integrated dataset based on students’ names as a research dataset. The results indicate the following: (1) the accuracy of the proposed feature selection approach is better than that of the Minimum-Redundancy-Maximum-Relevance (mRMR) approach; (2) the proposed model is better than the listing methods when the six least influential features have been deleted; and (3) the proposed model can enhance the accuracy and facilitate the interpretation of the pattern from a hybrid-type dataset of students’ academic achievement.

[1]  James C. Bezdek,et al.  A comparison of neural network and fuzzy clustering techniques in segmenting magnetic resonance images of the brain , 1992, IEEE Trans. Neural Networks.

[2]  J. Coleman,et al.  Social Capital in the Creation of Human Capital , 1988, American Journal of Sociology.

[3]  Judea Pearl,et al.  Bayesian Networks , 1998, Encyclopedia of Social Network Analysis and Mining. 2nd Ed..

[4]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[5]  J. Friedman Regularized Discriminant Analysis , 1989 .

[6]  Robert M Malina,et al.  Effect of physical education and activity levels on academic achievement in children. , 2006, Medicine and science in sports and exercise.

[7]  Isabelle Guyon,et al.  An Introduction to Variable and Feature Selection , 2003, J. Mach. Learn. Res..

[8]  Peter M. Blau,et al.  The American Occupational Structure , 1967 .

[9]  Mark A. Hall,et al.  Correlation-based Feature Selection for Machine Learning , 2003 .

[10]  Jenq-Neng Hwang,et al.  The cascade-correlation learning: a projection pursuit learning perspective , 1996, IEEE Trans. Neural Networks.

[11]  Bernd Eggers,et al.  Classroom Management Strategies Gaining And Maintaining Students Cooperation , 2016 .

[12]  Christian Lebiere,et al.  The Cascade-Correlation Learning Architecture , 1989, NIPS.

[13]  F. Agakov,et al.  Application of high-dimensional feature selection: evaluation for genomic prediction in man , 2015, Scientific Reports.

[14]  Hermann Bondi,et al.  Theory and Practice of Regressive Education , 1993 .

[15]  Markus Höhfeld,et al.  Learning with limited numerical precision using the cascade-correlation algorithm , 1992, IEEE Trans. Neural Networks.

[16]  Gregory Piatetsky-Shapiro,et al.  The KDD process for extracting useful knowledge from volumes of data , 1996, CACM.

[17]  D. G. Morrison On the Interpretation of Discriminant Analysis , 1969 .

[18]  Heekuck Oh,et al.  Neural Networks for Pattern Recognition , 1993, Adv. Comput..

[19]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[20]  Qiang Ji,et al.  Information fusion with Bayesian networks for monitoring human fatigue , 2002, Proceedings of the Fifth International Conference on Information Fusion. FUSION 2002. (IEEE Cat.No.02EX5997).

[21]  E. Humby,et al.  Programs from decision tables , 1973 .

[22]  Shang-Liang Chen,et al.  Orthogonal least squares learning algorithm for radial basis function networks , 1991, IEEE Trans. Neural Networks.

[23]  Robert P. W. Duin,et al.  Multiclass Linear Dimension Reduction by Weighted Pairwise Fisher Criteria , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[24]  Megan Tschannen-Moran,et al.  Fostering Student Learning: The Relationship of Collective Teacher Efficacy and Student Achievement , 2004 .

[25]  Sankar K. Pal,et al.  Multilayer perceptron, fuzzy sets, and classification , 1992, IEEE Trans. Neural Networks.

[26]  Jonathan Solity,et al.  Classroom Management: Principles to Practice , 1987 .

[27]  P. Bickel,et al.  Some theory for Fisher''s linear discriminant function , 2004 .

[28]  Finn V. Jensen,et al.  Bayesian Networks and Decision Graphs , 2001, Statistics for Engineering and Information Science.

[29]  Finn Verner Jensen,et al.  Introduction to Bayesian Networks , 2008, Innovations in Bayesian Networks.

[30]  Ralph A. Smith,et al.  Aesthetics and arts education , 1991 .

[31]  M. Posner Foundations of cognitive science , 1989 .

[32]  K. Hacker,et al.  Is there a relationship between physical fitness and academic achievement? Positive results from public school children in the northeastern United States. , 2009, The Journal of school health.

[33]  Oded Maimon Knowledge Discovery and Data Mining : The Info-Fuzzy Network (IFN) Methodology , 2000 .

[34]  I. Friedman,et al.  Teacher self-efficacy: a classroom-organization conceptualization , 2002 .

[35]  J. Chall Stages of reading development , 1983 .

[36]  Padhraic Smyth Breaking out of the Black-Box: Research Challenges in Data Mining , 2001, DMKD.

[37]  Larry S. Davis,et al.  Human expression recognition from motion using a radial basis function network architecture , 1996, IEEE Trans. Neural Networks.

[38]  Francesco Piazza,et al.  On the complex backpropagation algorithm , 1992, IEEE Trans. Signal Process..

[39]  Igor V. Tetko,et al.  Neural Network Studies. 3. Variable Selection in the Cascade-Correlation Learning Architecture , 1998, J. Chem. Inf. Comput. Sci..

[40]  Claudio Moraga,et al.  The Influence of the Sigmoid Function Parameters on the Speed of Backpropagation Learning , 1995, IWANN.

[41]  Irina Rish,et al.  An empirical study of the naive Bayes classifier , 2001 .

[42]  Stephen A. Billings,et al.  Radial basis function network configuration using genetic algorithms , 1995, Neural Networks.

[43]  N. Astone,et al.  Family structure, parental practices and high school completion. , 1991 .

[44]  A. W. Hoy,et al.  Collective Teacher Efficacy: Its Meaning, Measure, and Impact on Student Achievement , 2000 .

[45]  Ferat Sahin,et al.  Learning from experience using a decision-theoretic intelligent agent in multi-agent systems , 2001, SMCia/01. Proceedings of the 2001 IEEE Mountain Workshop on Soft Computing in Industrial Applications (Cat. No.01EX504).

[46]  Paul R. Burden Classroom Management and Discipline Methods to Facilitate Cooperation and Instruction , 1995 .

[47]  G. Lewicki,et al.  Approximation by Superpositions of a Sigmoidal Function , 2003 .

[48]  Geoffrey I. Webb,et al.  Advances in Knowledge Discovery and Data Mining , 2018, Lecture Notes in Computer Science.

[49]  Ramón Díaz-Uriarte,et al.  Gene selection and classification of microarray data using random forest , 2006, BMC Bioinformatics.

[50]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[51]  Vladimir Vapnik,et al.  The Nature of Statistical Learning , 1995 .

[52]  Lloyd H. Barrow,et al.  Building bridges between science and reading , 1984 .

[53]  Fuhui Long,et al.  Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy , 2003, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[54]  Larry D. Yore A Preliminary Exploration of Grade Five Students' Science Achievement and Ability to Read Science Textbooks as a Function of Gender, Reading Vocabulary, and Reading Comprehension. , 1987 .

[55]  Norman Edward Gronlund,et al.  Assessment of student achievement , 1997 .

[56]  Ron Kohavi,et al.  The Power of Decision Tables , 1995, ECML.

[57]  Ramakant Nevatia,et al.  Bayesian framework for video surveillance application , 2000, Proceedings 15th International Conference on Pattern Recognition. ICPR-2000.

[58]  Christos Faloutsos,et al.  Data-driven evolution of data mining algorithms , 2002, CACM.

[59]  Mark E. Oxley,et al.  Comparative Analysis of Backpropagation and the Extended Kalman Filter for Training Multilayer Perceptrons , 1992, IEEE Trans. Pattern Anal. Mach. Intell..

[60]  F. Barron,et al.  Creativity, Intelligence, and Personality , 1981 .

[61]  Theofanis Sapatinas,et al.  Discriminant Analysis and Statistical Pattern Recognition , 2005 .

[62]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[63]  Scott E. Fahlman,et al.  The Recurrent Cascade-Correlation Architecture , 1990, NIPS.

[64]  David S. Broomhead,et al.  Multivariable Functional Interpolation and Adaptive Networks , 1988, Complex Syst..

[65]  J Nisbet,et al.  Family environment and intelligence. , 1953, The Eugenics review.

[66]  D. Hand,et al.  Idiot's Bayes—Not So Stupid After All? , 2001 .

[67]  A. V. Olgac,et al.  Performance Analysis of Various Activation Functions in Generalized MLP Architectures of Neural Networks , 2011 .