The Discriminant Analysis Approach for Evaluating Effectiveness of Learning in an Instructor-Led Virtual Classroom

## Publications

/ Export Citation / / / Text size:

#### International Journal on Smart Sensing and Intelligent Systems

Exeley Inc. (New York)

Subject: Computational Science & Engineering, Engineering, Electrical & Electronic

eISSN: 1178-5608

41
103
Visit(s)
0
Comment(s)
0
Share(s)

SEARCH WITHIN CONTENT

FIND ARTICLE

Volume / Issue / page

Archive
Volume 14 (2021)
Volume 13 (2020)
Volume 12 (2019)
Volume 11 (2018)
Volume 10 (2017)
Volume 9 (2016)
Volume 8 (2015)
Volume 7 (2014)
Volume 6 (2013)
Volume 5 (2012)
Volume 4 (2011)
Volume 3 (2010)
Volume 2 (2009)
Volume 1 (2008)
Related articles

VOLUME 13 , ISSUE 1 (Jan 2020) > List of articles

### The Discriminant Analysis Approach for Evaluating Effectiveness of Learning in an Instructor-Led Virtual Classroom

Citation Information : International Journal on Smart Sensing and Intelligent Systems. Volume 13, Issue 1, Pages 1-15, DOI: https://doi.org/10.21307/ijssis-2020-035

Received Date : 04-January-2018 / Published Online: 30-December-2020

### ARTICLE

#### ABSTRACT

The effective learning requires putting down various associations of new ideas to old ones to integrate some innovative thoughts. The learners must change the associations among the things they already know, or even reject some long-held attitude about the world. The choice to the essential reformation is to deform the new information to fit their old ideas or to reject the new information entirely. Learners come to the classroom with their own ideas, some may be correct and some may not be, concerning roughly each topic they are expected to come across. If their perception and misunderstanding are unnoticed or discharged out of control, it affects the learning of a learner. The learners must be encouraged to build up new observation by seeing how such observation helps them make better sense of the world. The objective of this research paper is to put down the fundamentals of learning that promotes effective learning in an instructor-led virtual classroom and to analyze the learners’ learning performance using the Discriminant Analysis, a data mining technique. The Discriminant Analysis uses statistically significant determinants to predict learners’ learning in a classroom.

#### Graphical ABSTRACT

Educational psychology affords many academic values to be applied in the growth and evaluation of computer-assisted instructional technology. Milheim and Martin (1991) identified learner control as a significant variable in increasing the pedagogy of software in studying learner control motivation, attribution, and informational processing theory. It is advantageous for making best use of learner control as it enhances the relevance of learning, expectations for success, and general contentment, leading to sharp impulse (Gagne, 1985). In recent years, the educational development is focused on smart education in which the technology plays a vital role in smart learning carried out in the classroom (Kankaanranta and Makela, 2014). A smart learning environment will lead to new knowledge as every learner is unique.

Educational psychology is a very important regulation that is added to the learning of instructors and learners. The main concern in an educational psychology is to understand the learning process, i.e., the procedures and strategies that the learners use to acquire new information. Learning, still, does not take place in a vacuum. Understanding the best circumstances in which learning can occur is essential. Learning deals with the issues in the surrounding. The learning situation comes in the middle between the learner and the instructor. Topics like classroom environment and group-active techniques and aids that facilitate learning, evaluation techniques, and practices, guidance, and counseling, help in the smooth functioning of the teaching learning process.

Furthermore, learning is typically exaggerated by learning styles, and if learners make use of several learning styles, learning speed is higher (Galit et al., 2007). In recent years, interest has grown in the learning-centered paradigm (McManus, 2001) because it places learners at the center of the experience, empowers, and motivates them to assume responsibility in their learning, and approves teaching and learning approaches designed to encourage students to see themselves as active thinkers and problem-solvers. A learning-style preference study to native speakers (NNS) and cognitive styles were studied by Reid (1987).

## Literature review

The environment of a learner could be distinguished by synchronous and asynchronous learning, also the level of communication between learner-learners, learner-instructor, learner-content, and learner-instructional media (Hiltz, 1994; Marks et al., 2005; Oliver et al., 2009). The partial access to time and place that is a fixed schedule in a specific classroom is the disadvantage of synchronous learning. The face-to-face interactions between learner and other learners, instructor, and content that are supposed to induce attention in learning and to highlight knowledge-acquisition process are the advantages of synchronous learning. The physical closeness may generate an intelligence of belonging and exchange of emotional support (Karacapilidis, 2010). However, the common atmosphere of a classroom may support a tendency of group conventionality to the attitude expressed by the faculty. A student with individual ingenuity may have to slow down his/her progress to the speed of the class.

Asynchronous learning has an advantage in offering unlimited access to class contents or materials anytime and anywhere. A responsive plan of course deliverance permits a learner to improve at his/her own speed. The learner not getting immediate answers to their questions on a certain topic is the drawback of asynchronous learning. Also, the learner does not get immediate feedback on their response to class materials and instructional delivery to strengthen their learning (Keller and Knopp, 1987). Alternatively, the real-time online feedback becomes possible with instructors and learners with modern growth of internet-based communication skills, and even creates a virtual learning community (Bower and Richards, 2005; Karacapilidis, 2010) that studied the impact of virtual classroom laboratories that are supportive mostly in the field of computer science education. Also, the research by Ferm and Naughton (2002) discusses the positive role of video-conferencing technologies in online education. This study primarily focuses on the creation of brainstorming-style discussions and small group meetings that are fundamental to many of modern educational techniques. Magdalene Delighta Angeline et al. (2017) analyzed the learners’ problem in learning, and also introduced a new method of teaching to improve the learners’ performance in academics. Angel et al. (2011) identified that North Carolina Virtual Public Schools (NCVPS) offered students improved flexibility and responsibilities, expanded opportunities, and individualized instruction and support. However, some problems existed for learner eagerness. Learners did not all the time have the practical skills or resources for online learning and many lacked self-direction. Han and Kamber (2006) describe the data mining software that allows the clients to examine the information in diverse aspects, classify it, and summarize the relationships, which are identified during the mining process. Galit et al. (2007) gave a case study that is to analyze the learners’ learning behavior to predict the results and to warn learners at risk before their final exams. Minaei and Punch (2003) used genetic algorithms with a combination of multiple classifiers to predict students’ final grades. Magdalene Delighta Angeline et al. (2015) classify the categories of learner’s performance in their academic qualification by reducing the failure ratio by taking appropriate steps at the right time to improve the quality of education using the class association rule. Kotsiantis et al. (2004) and Nebot et al. (2006) applied several machine-learning and classification techniques to predict the students’ final score, and the significance of each feature is also assessed. Etchells et al. 2006) carried out this work using artificial neural networks to predict students’ final grades. AI-Radaideh et al. (2006) applied the data mining approaches, principally categorization to assist in raising the quality of the educational system by evaluating learners’ data. He also applied a Decision Tree model to predict the final grade of learners who studied the C++ course at Yarmouk University, Jordan, in the year 2005. Different classification methods, namely ID3, C4.5, and the Naive Bayes, were used. The upshot of their consequences indicated that the Decision Tree model had better prediction than other models.

New York’s Smart School program emphasizes the role of technology integrated into the classroom by enhancing student achievement and practice students to take part in the 21st-century financial system (New York Smart Schools Commission Report, 2014). In recent times, many investigations instigate to pay concentration to the significance and requirement of genuine activities in which learners work with problems in the real world (Hwang et al., 2008). So, as to locate students in genuine learning environments, it is imperative to propose learning that combines both real and virtual learning environments. This paper discusses about smart learning in an instructor-led virtual classroom and also the effective prediction of learners’ performance with Discriminant Analysis.

## Learning in the instructor-led virtual classroom

The instructor-led virtual classroom is a smart way of teaching and learning environment where learners can work jointly, exchange a few words, monitor and argue the presentation, and employ with learning resources working in a group. The learner’s concentration strengthens in instructor-led virtual classrooms as its environment is interactive. The smart instructor-led virtual classrooms provide

• Immediate feedback.

• Face-to-face interaction.

In smart instructor-led virtual classrooms, there is an instructor in the environment with learners when doing or looking at activities. Instructor-led virtual classrooms can also be more instructive, provided that a supplementary provides effective knowledge. The class notes are recorded and the important details of the topic are pointed out in an instructor-led virtual classroom. The interaction is made through content-related discussion topics by the learner. The learner uses text features to discuss the topics with the group. The discussion group helps the learner to get an answer for the posted question. The learner is allowed to take part in the discussion and share their own creative ideas. The learner has the facility to view the lecture on a topic in powerpoint slides. The learner uses programs independently to create an innovative product for the course in applications such as word and excel. The instructor uses multimedia resource that helps the learner to access and view to acquire more ideas. The real-time feedback is provided in such a smart educational environment to engage the learner into effective, proficient, and meaningful learning. This instructor-led virtual classrooms use a cognitive approach further to enhance the withstanding ability of any learner. The analysis of learner’s information, such as assessment of academics, attitudes, and behavioral pattern, helps the educational institutes to predict failure rate measuring to reduce the same and to check whether they are using their resources in the right places and producing the right results.

### Cognitive approach

Cognition is a new topic in the field of cognitive science. The basic argument is about the significance of physical experience in sense-making and learning (Nunez et al., 1999). The cognitive approach is a specialized learning method that the learners use to learn more effectively. It includes recurrence, organizing new language, reviewing meaning, speculating meaning from context, and using imagery for memorization. All of these approaches involve purposeful manipulation of language to improve learning. The cognitive approach refers to the means an individual imagines and practices information. It is essential for a learner to improve their cognitive skills to produce a good outcome. The learner should possess:

• The ability to sustain concentration on a particular object, action, or thought.

• The ability to manipulate objects.

• The ability to visualize images and scenarios.

• Abilities that facilitate goal-oriented behavior, such as the tendency to plan and execute a goal.

• The ability to withstand distraction and internal advises.

• The ability to identify and manage one’s own emotions for good performance.

Learning in an instructor-led virtual classroom is done with the help of various methods such as project-based approach, interactive learning approach, exposition learning approach, contingent assignment, and imaginative empowerment approach. Each approach identifies the learner with different cognitive skills and learning is made with the respective identified approach. Identifying the cognitive skills of each learner and providing learning based on the skill is very difficult. In the instructor-led virtual classroom approach, learning is afforded by observing the learner’s cognitive skills. The different cognitive skills are given in Figure 1. The cognitive skills of each learner are recognized with their associated behavior. The associated behavior comprises solving problems, making use of a new method, defining new examples, summarizing, memorizing topics, judging the value of methods, pointing out, and recalling. The instructor-led virtual classrooms implemented with the cognitive approach of learning create a new learning environment and increase the learning speed of each learner. The monotonous learners enjoy the new situation of learning and the outcome of the learner is improved in a short span of time.

##### Figure 1:

Cognitive skills and related behaviors.

The observation made from each learner is summarized as follows:

• A monotonous learner’s keenness is augmented.

• An innovative thought-provoked outcome is yielded.

• Learners engage themselves in group activities voluntarily.

• Idea sharing and discussion are done.

• The upshot of learning is enhanced.

Later, the prediction and analysis are made with the Discriminant analysis that enhances the performance of each learner.

## Learning approaches

### Project-based approach

In a project-based method, the learners are grouped into seven teams with four members in each team. All the team members are instructed to select a topic from microprocessor and microcontroller subjects. The teams were given one month’s time to get ready with their own topics. Each team used different materials to explain their concepts. The materials used by the teams are charts, newspaper content, information from internet sources, cardboard work, and real-life examples to explain the microprocessor working along with the instructions and instruction set. The cognitive skills observed at the end of the work from each team are synthesis, intelligence, reasoning, evaluation, and application. The percentage of skills observed from each team is analyzed and grouped under a grade.

If-then rules were formed as follows:

If skill percentage> = 90 then Grade = ‘A’

If skill percentage> = 80 then Grade = ‘B’

If skill percentage> = 70 then Grade = ‘C’

If skill percentage< = 50 then Grade = ‘D’

Grade ‘A’ has the highest percentage and the maximum team members came out with a good explanation on their topics given in Figure 2. The viewers of this approach learned well and were able to grasp the basic concepts.

### Team-based work

In team-based work, the learners are grouped and given different topics for each team in software-testing subject. The time duration for preparing the topic is 45 minutes and the presentation is 15 minutes to each team. In total, there were ten teams with two or three members in each team. The learners enjoyed preparing, discussing among them, and working together preparing materials. The presentation of each team was more creative and realistic. The outcomes from the students are as follows:

• Increased interest.

• Cleared doubts.

• Equally shared in doing.

The cognitive skills are observed and analysis was made for each team.

### Creative empowerment approach

In a creative empowerment approach, the students were grouped into four teams. Each team was provided with a specialized topic on “how a final product undergoes various testing before releasing it.” The learners were very much enthusiastic in proceeding with their work. They themselves implemented their own ideas in various levels of testing. The upshot of this method was:

• Increased creative skills.

• Came out with their own ideas.

• Different implementation methods.

• Diverse styles in presenting.

• Sound understanding in the technological approach.

### Interactive learning approach

In this approach, the learners were made to act, listen, view, and ask question immediately. This approach was applied to learners for object-oriented programming and computer network subjects. The findings of this approach were:

• Active participation.

• Improved listening capability.

• Knowing the models effortlessly.

• Questioning the abilities enhanced.

• Better communication between the learner and facilitator.

### Presentation-learning approach

In a presentation-learning approach, the learner as well as the facilitator were made to present content using powerpoint slides or multimedia presentation softwares. The learner was able to study about the working principles and their simulation in a real way. The simulation and presentation were for mobile communication and programming paradigm subjects. The learners’ individual cognitive skills were analyzed. The learners were educated to be aware of the concept in detail. The learner’s cognitive processes influence the nature of what is learned. People learn new information more easily when they can relate it to something they already know.

Figure 3 shows the analysis chart of various approaches for different cognitive skills. The creative empowerment approach produces good effect among learners with 92.2%.

##### Figure 3:

Analysis of learning with various approaches.

It was found that the cognitive skill named evaluation has 90.4% in Figure 4, which was the highest when compared with other skills. The analysis was made with the different subjects for the same group of learner in each approach that modernizes and enhances the learning of the learners. Intelligence encompasses a number of mental abilities such as reasoning, planning, and problem-solving. Cognitive theory also highlights student-centered instruction, supportive learning groups, multiple presentations of key thoughts, and energetic, investigative learning.

##### Figure 4:

Overall percentage of cognitive skills when applied with various learning approaches in an instructor-led virtual classroom.

## Data mining algorithm

There are various data mining algorithms such as decision tree, Naive Bayes, Rule induction, Supervised Learning, Apriori algorithm, and Association rule analysis. This paper scrutinizes the Discriminant analysis that is a very powerful data mining technique.

### Data collection

The learners’ data sets collected in the current research study pertain to the different subjects pursued by the learners of engineering graduates from Dr. G. U. Pope College of Engineering. Learners’ performances in the respective subject prerequisites were collected from the departmental records of result summaries pertaining to three passed-out batches of computer science and engineering discipline.

### Discriminant analysis

Discriminant analysis is a statistical technique to classify objects based on a set of measurable object’s feature (Michael et al., 2000). The linear combination for a discriminant analysis, also known as the Discriminant function, is derived from an equation:

$(1)Z=W1X1+W2X2+W3X3+…+WnXn$
(1)where Z = Discriminant score Wi = Discriminant weight for variable i Xi = Independent variable i Fisher’s Discriminant Algorithm is given as follows:Estimate class means, covariance matrices, and prior probabilities.Compute the pooled covariance matrix and invert it.Compute the discriminant vector.Apply the discriminant using equation Y = WT x.

Fisher’s Linear Discriminant Analysis is based on the idea of searching for a linear combination of variables that best separates the target classes. Normally, we seek a direction w such that:

$(2)JF=|wT(m1−m2)|2wTSWw$
(2)is a maximum, where m1 and m2 are the group means and SW is the pooled within-class sample covariance matrix, in its bias-corrected form given by:
$(3)1n−2(n1∑1+n2∑2)$

“Pooled” is the total, all points contributing. The effectiveness of the discrimination is assessed by calculating the Mahalanobis distance between two groups. If the distance is greater than 3, then the probability of misclassification is reasonably undersized:

$(4)Δ2=βT(μ1−μ2)$

Last, a new point is classified by projecting it onto the maximally separating direction and classifying it as C1 if:

$(5)βT(X−(μ1+μ22)>logP(c1)P(c2)$

The variances appear along the diagonal and covariances appear in the off-diagonal elements given as follows:

$(6)V=[∑X12N∑X1X2N⋯∑X1X2N∑X1X2N∑X22N⋯∑X1X2N∑X1X2N∑X1X2N⋯∑Xc2N]$
(6)whereV is a c x c variance-covariance matrix.N is the number of scores in each of the c data sets. X i is a deviation score from the ith data set.ΣX i 2/N is the variance of elements from the ith data set.ΣX i X j/N is the covariance for elements from the ith and jth data sets.

The variance-covariance matrix is created as follows:

Suppose X is an n x k matrix holding ordered sets of raw data:

• Start with the raw data of matrix X, create a variance-covariance matrix to show the variance within each column and the covariance between columns.

• Transform the raw scores from matrix X into deviation scores for matrix x.

$(7)x=X−11’X(1/n)$
(7)where1 is an n x 1 column vector of one’s.x is an n x k matrix of deviation scores: x11, x12, … , xnk.X is an n x k matrix of raw scores: X11, X12, … , Xnk.
• Compute x’x, the k x k deviation sums of squares and cross-product matrix for x.

• Then, divide each term in the deviation sums of squares and cross-product matrix by n to create the variance-covariance matrix. That is:

$(8)V=x’x(1/n)$
(8)whereV is a k x k variance-covariance matrix.x’x is the deviation sums of squares and cross-product matrix.n is the number of scores in each column of the original matrix X.

## Experimental results

The learner’s cognitive skills are observed during the lecture time and learning in the instructor-led virtual classroom is done using cognitive methods. The learner is assessed using the internal assessment test and the attendance in the classroom. Table 1 describes about the discriminant analysis for performance. The mean for the input data internal 1, internal 2, attendance is estimated for the performance. The performance is categorized into three: average, good, and poor.

##### Table 1.

Discriminant analysis for performance.

Discriminant function analysis undergoes two steps: (1) testing the significance of a set of discriminant functions and (2) classification. The first step is computationally identical to MANOVA. There is a matrix of total variances and covariances. Similarly, there is a matrix of pooled within-group variances and covariances. The two matrices are compared via multivariate F tests in order to determine whether or not there are any significant differences between groups. First, it performs the multivariate test, and, if statistically important, proceeds to see which of the variables have considerably different means across the groups. Once group means are found to be statistically considerable, classification of variables is undertaken. DA automatically determines some best possible combination of variables so that the first function provides the most overall discrimination between groups, the second provides second most, and so on. Once the discriminant functions are determined, groups are differentiated; the utility of these functions can be examined via their ability to correctly classify each data point to their a priori groups. Classification functions are derived from the linear discriminant functions to accomplish this use. Different classification functions are used and equations exist that are best suited for equal or unequal samples in each group.

Table 2 gives the classification matrix that is an important tool for assessing the results of prediction. The rows in the matrix represent the predicted values for the model, whereas the columns represent the actual values. The categories used in analysis are false positive, true positive, false negative, and true negative.

##### Table 2.

Classification matrix.

Table 3 provides the covariance matrix of whole observations by treating all observations as from a single sample. The inputs were evaluated and grouped under three classes: average, good, and poor. The mean and covariance are estimated for each class. A variance-covariance matrix is a square matrix that contains the variances and covariances associated with several variables. The diagonal elements of the matrix contain the variances of the variables and the off-diagonal elements contain the covariances between all possible pairs of variables.

##### Table 3.

Matrix of variance and covariance.

In Table 3, the variances are displayed in bold along the diagonal. The variances of average class with the attributes are Internal1, Internal2, and attendances are 126.5, 472.7, and 100.5, respectively. The covariance between Internal1 and Internal2 is 243.5. The pooled matrix is calculated, which is the weighted sum of the group covariance matrix. The variance-covariance matrix is symmetric because the covariance between Internal1 and Internal2 is the same as the covariance between Internal2 and Internal1. Therefore, the covariance for each pair of variables is displayed twice in the matrix: the covariance between the ith and jth variables is displayed at positions (i, j) and (j, i).

Table 4 gives the classification summary. The percentage of all observations correctly classified is 81%. The percentage that would be correctly classified by classifying all observations in the largest category is 52.4%. The percentage of error gap filled is 60%. In Figure 5, the statistical distance used is the Mahalanobis distances from each of group means to the observation. The observation is classified to the group to which it is nearby, i.e., the distance value is the smallest. The statistical distance of each observation to the mean vector of poor, good, and average is calculated.

##### Table 4.

Summary classification.

##### Figure 5:

Statistical distance of each observation to the mean vector.

The learners’ performance in the instructor-led virtual classroom is analyzed using Discriminant analysis and the classification is made with Mahalanobis distances.

Figure 6 shows the periodical assessment outcome along with the final outcome of the learner. Initially, the outcome of a learner is poor in the PA1. Then it gradually increases after applying different learning approaches in a smart instructor-led virtual classroom with cognitive skills and learners gained increased attention. Finally, the learners’ outcome reached a maximum level stating that the learning approaches with a continuous analysis with data mining algorithms and training work well for each learner.

##### Figure 6:

Comparison of the final outcome with periodical assessment.

From the implementation result, it is identified that learning with cognitive skills produces a good outcome with 96.4%. The dropout ratio of the learner gets reduced with this enhanced learning approach.

Tables 5 and 6 describe about the discriminant analysis for performance. The mean for the input data internal1, internal2, and attendance is estimated for the performance. The performance is categorized into three: average, good, and poor.

##### Table 5.

Summary statistics.

##### Table 6.

Summary statistics (validation).

##### Table 7.

Sum of weights and prior probabilities for each class.

##### Table 8.

Mahalanobis distances.

##### Table 9.

Generalized squared distances.

##### Table 10.

Fisher distances.

##### Table 11.

P values for Fisher distances.

In discriminant analysis, Wilk’s lambda (Table 12) tests how well each level of independent variable contributes to the model. The scale ranges from 0 to 1, where 0 means total discrimination and 1 means no discrimination. The lambda value for the results obtained is 0.000, therefore, it implies total discrimination. Each independent variable is tested by putting it into the model and then taking it out generating a Λ statistic. The significance of the change in Λ is measured with an F test. As the F value (observed value) is greater than the critical value, the variable is kept in the model. The table also provides a chi-square statistic to test the significance of Wilk’s lambda. Since the p value is 0.001, which is less than 0.05, it can be concluded that the corresponding function explains the group membership well.

##### Table 12.

Wilks’ Lambda test (Rao’s approximation).

The test interpretation is H0: the mean vectors of the 3 classes are equal, Ha: at least one of the mean vectors is different from another. As the computed p value is lower than the significance-level alpha = 0.05, one should reject the null hypothesis H0, and accept the alternative hypothesis Ha. The risk to reject the null hypothesis H0 while it is true is lower than 0.09% in Table 12.

The test interpretation is H0: the mean vectors of the 3 classes are equal, Ha: at least one of the mean vectors is different from another. As the computed p value is lower than the significance-level alpha = 0.05, one should reject the null hypothesis H0, and accept the alternative hypothesis Ha. The risk to reject the null hypothesis H0 while it is true is lower than 0.02% in Table 13.

##### Table 13.

Pillai’s trace.

The test interpretation is H0: the mean vectors of the 3 classes are equal, Ha: at least one of the mean vectors is different from another. As the computed p value is lower than the significance-level alpha = 0.05, one should reject the null hypothesis H0, and accept the alternative hypothesis Ha. The risk to reject the null hypothesis H0 while it is true is lower than 1.14% in Table 14.

##### Table 14.

Hotelling–Lawley trace.

The test interpretation is H0: the mean vectors of the 3 classes are equal, Ha: at least one of the mean vectors is different from another. As the computed p value is lower than the significance-level alpha = 0.05, one should reject the null hypothesis H0, and accept the alternative hypothesis Ha. The risk to reject the null hypothesis H0 while it is true is lower than 0.17%.

##### Table 15.

Roy’s greatest root.

Eigenvalue.

##### Table 17.

Bartlett’s test for eigenvalue significance.

Eigenvalue is a ratio between the explained and unexplained variation in a model. For a good model, the eigenvalue must be more than one. In discriminant analysis, there is one eigenvalue for each discriminant function.

The bigger the eigenvalue, the stronger is the discriminating power of the function. The eigenvalue for the obtained results is 426.213 that has the strongest discriminating power of the function (Tables 7-17). Figure 7 shows the chart of eigenvalues that has the strongest discriminating function. The canonical correlation is a measure of the association between the groups in the dependent variable and the discriminant function in the table. As it has a high value given in Table 18, it implies a high level of association between the two and vice versa.

##### Table 18.

Canonical correlations.

##### Figure 7:

Chart of the eigenvalue.

Group centroids are the mean discriminant scores for each group in the dependent variable for each of the discriminant functions specified in Table 19. For two groups in the dependent variable, there is a single discriminant function. The centroids are in a one-dimensional space, one center for each group. For three groups in the dependent variable, there are two discriminant functions. Hence, the centroids are in a two-dimensional space. By connecting the centroids, a canonical plot can be created depicting a discriminant function space. The group centroids for the candidates are average (−1.441, 17.951), good (27.354, −8.465), and poor (−23.159, −10.085) given in Figure 8.

##### Table 19.

Functions at the centroids.

##### Figure 8:

Outcome of predicted performance with Bartlett’s test.

In Figures 8 and 9, the range in the vertical axis is small. Hence, F2 does make much difference. Only F1, the horizontal axis, is important for differentiation. The classification matrix is a simple cross-tabulation of the observed and predicted memberships. For a good prediction, the values in the diagonal must be high and the values off the diagonal must be close to 0 (Figure 10).

##### Figure 9:

Observations (axes F1 and F2: 100.00%).

Tables 20–22 show the confusion matrix of the training sample, validation sample, and cross-validation results used in the Discriminant Analysis approach for three classes: good, average, and poor. The smart learning with the instructor-led virtual classroom produces good learners with increasing intellectual capability providing excellent feedback every time this is necessary for the learners to cope with the various learning approaches. The discriminant analysis works well with the dataset covering all the groups of data and provides a better prediction. The assessment of learners is performed during the learning period and evaluation after the assessment is made with discriminant analysis to reduce the dropout ratio in the final outcomes. The implemented smart approach with the instructor-led virtual classroom produced 96.4% outcome in the final exam, which implies that the proposed approach changes the traditional way of learning in the classroom with smart learning.

##### Table 20.

Confusion matrix for the training sample.

##### Table 21.

Confusion matrix for the validation sample.

##### Table 22.

Confusion matrix for the cross-validation results.

## Conclusions

The proposed work identifies cognitive skills of each learner with their associated behavior and learning is made in the instructor-led virtual classroom. The learners’ learning skills are improved and the thinking capacity of each learner is increased. The different views of the learner make every learner to easily understand the concept by improving the concentration of the learner. The performance measure of each learner is predicted using the discriminant analysis. The information obtained subsequent to the execution of the data mining technique probably will help the instructor as well as the learners. The performance report of the learner also helps to improve the result of the learner. This performance enhancement will also help the entire learner to get placement in various trades according to the norm. The educational institution gets benefited with the proposed system for their even and victorious running of the organization. The solution provided using the Discriminant Algorithm predicts the performance of a learner correctly filling the error gap to produce a sound enough result.

##### Figure 10:

Centroids (axes F1 and F2: 100.00%).

## References

1. AI-Radaideh, Q. A. , AI-Shawakfa, E. M. and AI-Najjar, M. I. 2006. “Mining student data using decision trees”, International Arab Conference on Information Technology Yarmouk University, Jordan, pp. 1–5.
2. Angel, B. , Candice, B. and Jennifer, H. 2011. “Evaluating the effectiveness of the North Carolina Virtual Public Schools System”. North Carolina Virtual Public School, pp. 1–12.
3. Bower, M. and Richards, D. 2005. The impact of virtual classroom laboratories in computer science education. Thirty-Sixth SIGCSE Technical Symposium of Computer Science Education, St. Louis, Missouri, pp. 292–296.
4. Brown, M. T. and Wicker, L. R. 2000. “Discriminant Analysis”, Handbook of Applied Multivariate Statistics and Mathematical Modeling, pp. 209–235.
5. Etchells, A. T. , Nebot, A. , Vellido, A. , Lisboa, P. J. and Mugica, F. 2006. “Learning what is important: feature selection and rule extraction in a virtual course”, The 14th European Symposium on Artificial Neural Networks ESANN, Bruges, pp. 401–406.
6. Ferm, S. R. and Naughton, N. l. 2002. Collaborative virtual environments to support communication and community in internet-based distance education. Journal of Information Technology Education 1(3): 201–211.
7. Gagne, R. 1985. “The Conditions of Learning and Theory of Instruction” 4th ed., Holt, Rinehart and Winston, New York, NY.
8. Galit, B. Z. , Hershkovitz, A. , Mintz, R. and Nachmias, R. 2007. “Examining online learning processes based on log files analysis: a case study”. Research, Reflection and Innovations in Integrating ICT in Education 1: 55–59.
9. Han, J. and Kamber, M. 2006. Data mining: concepts and techniques 2nd ed., The Morgan Kaufmann Series in Data Management Systems Jim Gray, Series Editor, Morgan Kaufmann Publishers, San Francisco.
10. Hiltz, S. R. 1994. The Virtual Classroom: Learning without Limits via Computer Networks Ablex Publishing Corporation, Norwood, MA.
11. Hwang, G. J. , Tsai, C. C. and Yang, S. J. H. 2008. “Criteria, strategies and research issues of context-aware ubiquitous learning”. Journal of Educational Technology and Society 11(2): 81–91.
12. Kankaanranta, M. and Makela, T. 2014. “ Valuation of emerging learning solutions. Proceedings World Conference on Educational Multimedia, Hypermedia and Telecommunications, Tampere Association for the Advancement of Computing in Education, pp. 168–172.
13. Karacapilidis, N. 2010. “Novel Developments in Web-based Learning Technologies: Tools for Modern Teaching” IGI Global, Hershey, PA.
14. Keller, J. and Knopp, T. 1987. “Instructional Theories in Action: Lessons Illustrating Theories and Models”, Associates Hillsdale, Erlbaum, NJ.
15. Kotsiantis, S. B. , Pierrakeas, C. J. and Pintelas, P. E. 2004. “Predicting students performance in distance learning using machine learning techniques”. Applied Artificial Intelligence 18(5): 411–426.
16. Magdalene Delighta Angeline, D. , Ramasubramanian, P. and Samuel Peter James, I. 2015. “Learner’s prognostic analysis using class association rule”. International Journal of Advanced Research in Computer Science and Software Engineering 5(9): 314–318.
17. Magdalene Delighta Angeline, D. , Ramasubramanian, P. and Samuel Peter James, I. 2017. “Increased success outcome of a learner with ensemble teaching and analysis with Naive Bayes Algorithm”. Ciencia E Tecnica Vitivinicola 32(6): 59–73.
18. Marks, R. B. , Sibley, S. D. and Arbaugh, J. B. 2005. “A structural equation model of predictors for effective online learning”. Journal of Management Education 29(4): 531–563.
19. McManus, D. A. 2001. “The two paradigms of education and the peer review of teaching”. Journal of Geosciences Education 49(5): 423–434.
20. Milheim, W. D. and Martin, B. L. 1991. “Theoretical bases for the use of learner control: three different perspectives”. Journal of Computer-based Instruction, Association for the Development of Computer-Based Instructional Systems, West Woodruff Columbus OH United States, 18(3): 99–105.
21. Minaei, B. and Punch, B. 2003. “Using genetic algorithms for data mining optimization in an educational web-based system”. Genetic and Evolutionary Computation 2: 2252–2263.
22. Nebot, A. , Castro, F. , Vellido, A. and Mugica, F. 2006. Identification of fuzzy models to predict students performance in an e-learning environment. In The Fifth IASTED International Conference on Web-Based Education, Puerto Vallarta, pp. 74–79 .
23. New York Smart Schools Commission Report 2014. New York Smart Schools, Available at: http://www.governor. ny.gov/sites/governor.ny.gov/files/archive/governor_files/SmartSchoolsReport.pdf (accessed November 2017).
24. Nunez, R. E. , Edward, L. D. and Matos, J. F. 1999. “Embodied cognition as grounding for situatedness and context in mathematics education”. Educational Studies in Mathematics 39: 45–65.
25. Oliver, K. , Osborne, J. , Patel, R. and Kleiman, G. 2009. “Issues surrounding the development of a New State wide Virtual Public Schools”. The Quarterly Review of Distance Education 10(1): 37–49.
26. Reid, J. M. 1987. “The learning style preferences of ESL students” TESOL Quarterly, 21(1): 87–8711.

### FIGURES & TABLES

Figure 1:

Cognitive skills and related behaviors.

Figure 2:

Figure 3:

Analysis of learning with various approaches.

Figure 4:

Overall percentage of cognitive skills when applied with various learning approaches in an instructor-led virtual classroom.

Figure 5:

Statistical distance of each observation to the mean vector.

Figure 6:

Comparison of the final outcome with periodical assessment.

Figure 7:

Chart of the eigenvalue.

Figure 8:

Outcome of predicted performance with Bartlett’s test.

Figure 9:

Observations (axes F1 and F2: 100.00%).

Figure 10:

Centroids (axes F1 and F2: 100.00%).

### REFERENCES

1. AI-Radaideh, Q. A. , AI-Shawakfa, E. M. and AI-Najjar, M. I. 2006. “Mining student data using decision trees”, International Arab Conference on Information Technology Yarmouk University, Jordan, pp. 1–5.
2. Angel, B. , Candice, B. and Jennifer, H. 2011. “Evaluating the effectiveness of the North Carolina Virtual Public Schools System”. North Carolina Virtual Public School, pp. 1–12.
3. Bower, M. and Richards, D. 2005. The impact of virtual classroom laboratories in computer science education. Thirty-Sixth SIGCSE Technical Symposium of Computer Science Education, St. Louis, Missouri, pp. 292–296.
4. Brown, M. T. and Wicker, L. R. 2000. “Discriminant Analysis”, Handbook of Applied Multivariate Statistics and Mathematical Modeling, pp. 209–235.
5. Etchells, A. T. , Nebot, A. , Vellido, A. , Lisboa, P. J. and Mugica, F. 2006. “Learning what is important: feature selection and rule extraction in a virtual course”, The 14th European Symposium on Artificial Neural Networks ESANN, Bruges, pp. 401–406.
6. Ferm, S. R. and Naughton, N. l. 2002. Collaborative virtual environments to support communication and community in internet-based distance education. Journal of Information Technology Education 1(3): 201–211.
7. Gagne, R. 1985. “The Conditions of Learning and Theory of Instruction” 4th ed., Holt, Rinehart and Winston, New York, NY.
8. Galit, B. Z. , Hershkovitz, A. , Mintz, R. and Nachmias, R. 2007. “Examining online learning processes based on log files analysis: a case study”. Research, Reflection and Innovations in Integrating ICT in Education 1: 55–59.
9. Han, J. and Kamber, M. 2006. Data mining: concepts and techniques 2nd ed., The Morgan Kaufmann Series in Data Management Systems Jim Gray, Series Editor, Morgan Kaufmann Publishers, San Francisco.
10. Hiltz, S. R. 1994. The Virtual Classroom: Learning without Limits via Computer Networks Ablex Publishing Corporation, Norwood, MA.
11. Hwang, G. J. , Tsai, C. C. and Yang, S. J. H. 2008. “Criteria, strategies and research issues of context-aware ubiquitous learning”. Journal of Educational Technology and Society 11(2): 81–91.
12. Kankaanranta, M. and Makela, T. 2014. “ Valuation of emerging learning solutions. Proceedings World Conference on Educational Multimedia, Hypermedia and Telecommunications, Tampere Association for the Advancement of Computing in Education, pp. 168–172.
13. Karacapilidis, N. 2010. “Novel Developments in Web-based Learning Technologies: Tools for Modern Teaching” IGI Global, Hershey, PA.
14. Keller, J. and Knopp, T. 1987. “Instructional Theories in Action: Lessons Illustrating Theories and Models”, Associates Hillsdale, Erlbaum, NJ.
15. Kotsiantis, S. B. , Pierrakeas, C. J. and Pintelas, P. E. 2004. “Predicting students performance in distance learning using machine learning techniques”. Applied Artificial Intelligence 18(5): 411–426.
16. Magdalene Delighta Angeline, D. , Ramasubramanian, P. and Samuel Peter James, I. 2015. “Learner’s prognostic analysis using class association rule”. International Journal of Advanced Research in Computer Science and Software Engineering 5(9): 314–318.
17. Magdalene Delighta Angeline, D. , Ramasubramanian, P. and Samuel Peter James, I. 2017. “Increased success outcome of a learner with ensemble teaching and analysis with Naive Bayes Algorithm”. Ciencia E Tecnica Vitivinicola 32(6): 59–73.
18. Marks, R. B. , Sibley, S. D. and Arbaugh, J. B. 2005. “A structural equation model of predictors for effective online learning”. Journal of Management Education 29(4): 531–563.
19. McManus, D. A. 2001. “The two paradigms of education and the peer review of teaching”. Journal of Geosciences Education 49(5): 423–434.
20. Milheim, W. D. and Martin, B. L. 1991. “Theoretical bases for the use of learner control: three different perspectives”. Journal of Computer-based Instruction, Association for the Development of Computer-Based Instructional Systems, West Woodruff Columbus OH United States, 18(3): 99–105.
21. Minaei, B. and Punch, B. 2003. “Using genetic algorithms for data mining optimization in an educational web-based system”. Genetic and Evolutionary Computation 2: 2252–2263.
22. Nebot, A. , Castro, F. , Vellido, A. and Mugica, F. 2006. Identification of fuzzy models to predict students performance in an e-learning environment. In The Fifth IASTED International Conference on Web-Based Education, Puerto Vallarta, pp. 74–79 .
23. New York Smart Schools Commission Report 2014. New York Smart Schools, Available at: http://www.governor. ny.gov/sites/governor.ny.gov/files/archive/governor_files/SmartSchoolsReport.pdf (accessed November 2017).
24. Nunez, R. E. , Edward, L. D. and Matos, J. F. 1999. “Embodied cognition as grounding for situatedness and context in mathematics education”. Educational Studies in Mathematics 39: 45–65.
25. Oliver, K. , Osborne, J. , Patel, R. and Kleiman, G. 2009. “Issues surrounding the development of a New State wide Virtual Public Schools”. The Quarterly Review of Distance Education 10(1): 37–49.
26. Reid, J. M. 1987. “The learning style preferences of ESL students” TESOL Quarterly, 21(1): 87–8711.