- Open Access
- Total Downloads : 385
- Authors : Miss. Ankita Rajesh Chordiya, Prof. S. B. Bagal
- Paper ID : IJERTV4IS010355
- Volume & Issue : Volume 04, Issue 01 (January 2015)
- Published (First Online): 17-01-2015
- ISSN (Online) : 2278-0181
- Publisher Name : IJERT
- License: This work is licensed under a Creative Commons Attribution 4.0 International License
Comparative Research of Clustering Algorithms for Prediction of Academic Performance of Students
Ankita R. Chordiya,
PG Student E &TC Department,
Lt. G. N. Sapkal College of Engineering Savitribai Phule Pune University
Prof. S. B. Bagal
Associate Professor & Head, E &TC Department, Lt. G. N. Sapkal College of Engineering, Savitribai Phule Pune University
Abstract- Clustering is a task of assigning a set of objects into groups called clusters. In general the clustering algorithms can be classified into two categories. One is hard clustering; another one is soft (fuzzy) clustering. Hard clustering, the datas are divided into distinct clusters, where each data element belongs to exactly one cluster. In soft clustering, data elements belong to more than one cluster, and associated with each element is a set of membership levels In order to monitor the progress of students efficiently, different clustering algorithms are applied to the academic results of students so as to categorize in appropriate class as per their performance. We proposed the use of FCM and KFCM clustering algorithms for prediction of students academic performance. Euclidean distance as a measure of similarity measurement is taken into consideration. These algorithms are applied and performance is evaluated on the basis of clustering output. FCM allows data points to belong to more than one cluster where each data point has a degree of membership of belonging to each cluster. The KFCM whereas uses a mapping function and gives better performance as compared to FCM. The summarized result shows that KFCM gives better performance than FCM
KeywordsFCM, KFCM, clustering, academic performance, membership function
-
INTRODUCTION
Pattern recognition is a branch of machine learning that focuses on the recognition of patterns and regularities in data clustering and classification are the major subdivisions of pattern recognition technique although clustering and classification are often used for purposes of segmenting data records, they have different objectives and achieve their segmentations through different ways. Classification is supervised learning technique used to assign predefined. tag to instance on the basis of features. So classification algorithm requires training data. Classification model is created from training data, then classification model is used to classify new instances. Clustering is unsupervised technique used to group similar instances on the basis of features. Clustering does not require training data. Clustering does not assign per-defined label to each and every group.
Fast and robust clustering algorithms play an important role in extracting useful information in large databases. The aim of cluster analysis is to partition a set of N object into C clusters such that objects within cluster should be similar to each other and objects in different clusters are should be dissimilar with each other. Clustering can be used to quantize the available data, to extract a set of cluster prototypes for the
compact representation of the dataset, into homogeneous subsets.
Clustering is a mathematical tool that attempts to discover structures or certain patterns in a dataset, where the objects inside each cluster show a certain degree of similarity. It can be achieved by various algorithms that differ significantly in their notion of what constitutes a cluster and how to efficiently find them. Cluster analysis is not an automatic task, but an iterative process of knowledge discovery or interactive multi-objective optimization. It will often necessary to modify preprocessing and parameter until the result achieves the desired properties. In Clustering, one of the most widely used algorithms is fuzzy clustering algorithms. Fuzzy set theory was first proposed by Zadeh in 1965 [1]& it gave an idea of uncertainty of belonging which was described by a membership function. The use of fuzzy set provides imprecise class membership function. Applications of fuzzy set theory in cluster analysis were early proposed in the work of Bellman, Zadeh, and Ruspini [8]. This paper opens door step of fuzzy clustering [2]. Integration of fuzzy logic with data mining techniques has become one of the key constituents of soft computing in handling challenges posed by massive collections of natural data. The central idea in fuzzy clustering is the non-unique partitioning of the data into a collection of clusters. The data points are assigned membership values for each of the clusters and fuzzy clustering algorithm allow the clusters to grow into their natural shapes [3].
The fuzzy clustering algorithms can be divided into two types . The FCM is the soft extension of the traditional hard c-means clustering[1]. Each cluster was considered as fuzzy set and the membership function measures the possibility that each training vector belongs to a cluster. so the vectors may be assigned to multiple clusters. Thus, it overcomes some drawbacks of hard clustering but it is effective only when the data is non-overlapping. By the contrast to the crisp c- partitions, in a fuzzy case elements can belong to several clusters and to different degrees [16]. This algorithm works by assigning membership to each data point corresponding to each cluster center on the basis of distance between the cluster center and the data point. More the data is near to the cluster center more is its membership towards the particular cluster center.
In KFCM, both the data and the cluster centers are mapped from the original space to a new space by [15] . An
important fact about kernel function is that it can be directly constructed in the original input space without knowing the form of .That is, a kernel function implicitly defines a
nonlinear mapping function [13].
N
u .x
m
ij i
N
c j i 1
ij
u m .
i 1
max u k 1 u k
(3)
The remainder of this paper is organized as follows:
When
ij ij
ij <
, where is a termination
Section II provides the basic algorithm of FCM. Section III provides kernel based FCM. Section IV proposes the application of FCM and KFCM in prediction of students academic performance and the results are compared for each of the algorithm in terms of cluster efficiency. Section V presents the conclusions drawn for the comparison of the FCM and KFCM on the basis of cluster efficiency.
-
FUZZY C-MEANS (FCM)
Fuzzy c-means (FCM) is a method of clustering which
criterion between 0 and 1, whereas k are the iteration steps, the iteration stops. This procedure converges to a local minimum
or a saddle point Jm.
The algorithm is composed of the following steps:
-
Randomly select cluster centre
-
Initialize U=[uij] matrix, U(0) Calculate the the uij using:
1
allows one piece of data to belong to two or more clusters. This method (developed by Dunn in 1973 and improved by
Bezdek in 1981) is frequently used in pattern recognition.
uij
c xi c j
2 m1
Straightly speaking, this algorithm works by assigning membership to each data point corresponding to each cluster
i1
x
i ck
center on the basis of distance between the cluster and the data point. More the data is near to the cluster center more is its membership towards the particular cluster center. Clearly, summation of membership of each data point should be equal to one [8]. The algorithm is based on minimization of the following objectve function:
-
At k-step: calculate the centres vectors C(k)=[cj] with U(k)
N
ij
i
u m .x
N
c j i 1
ij
u m .
N
J
um x c 2
(1)
i 1
m
C
i1
ij i j
j 1
-
Update U
(k)
, U(k+1)
where m (the Fuzziness Exponent) is any real number greater than 1, N is the number of data, C is the number of clusters,
uij
1
2
m1
uij is the degree of membership of xi in the cluster j, xi is the
c xi c j
ith of d-dimensional measured data, cj is the d-dimension
center of the cluster, and ||*|| is any norm expressing the
i1 xi ck
similarity between any measured data and the center. Fuzzy
partitioning is carried out through an iterative optimization of the objective function shown above, with the update of membership uij and the cluster centers cj by:
-
If || U(k+1) – U(k)||< or the minimum J is achieved, then STOP; otherwise return to step 2.
-
-
KERNEL FUZZY C-MEANS CLUSTERING
uij
1
c x c
2 m1
(2)
The KFCM algorithm adds kernel information to the traditional fuzzy c-means algorithm and it overcomes the disadvantage that FCM algorithm cant handle the small
i j
differences between clusters. The main idea of fuzzy kernel
i1
xi ck
c-means algorithm (KFCM) is described as follows. The kernel method maps nonlinearly the input data space into a
where ||xi – cj|| is the Distance from point i to current cluster
high dimensional feature space.
Given a dataset, X= {x ,…, xn} Rp , the original FCM
centre j, ||xi – ck|| is the Distance from point i to other cluster centers k.
algorithm partitions X into c fuzzy subsets by minimizing the following objective function
c n
ik
m
J (U ,V ) u m
i1 k 1
xk vi
2 (4)
where c is the number of clusters and selected as a specified value in this paper, n the number of data points, uik the
And it can be proven that d(x, y) defined in Eq. (7) is a metric in the original space in case that K(x,y) takes as the Gaussian
c kernel function. According to Eq. (6), the data point
xk is
membership of xk in class i, satisfying uik
i1
1, m the
endowed with an additional weight
K (xk
, vi ) , which
quantity controlling clustering fuzziness, and V the set of cluster centers or prototypes ( vi Rp). The function J m is minimized by a famous alternate iterative algorithm. Now consider the proposed kernel fuzzy c-means (KFCM) algorithm. Define a nonlinear map as 😡 (x) F, where x X . X denotes the data space, and F the transformed feature space with higher even infinite dimension. KFCM minimizes the following objective function :-
measures the similarity between xk and vi , and when xk is an outlier, i.e., xk is far from the other data points, then K (xk , vi ) will be very small, so the weighted sum of data points shall be more robust.
KFCM Algorithm
Step 1: Fix c ,tmax, m >1 and >0 for some positive constant;
ik
c n Step 2: Initialize the memberships u 0
J (U ,V ) u m (x ) (v ) 2 (5) Step 3: For t =1,2,,t , do:
max
m ik i i
i
i1 k 1
-
Update all prototypes
vt with Eqs. (9);
ik
u
ik
ik
where (xi ) (vi 2
-
Update all memberships u t with Eqs. (8);
(6)
Where K(x, y) = (x)T( y) is an inner product kernel
-
Compute t=t+1
Et max
i,k
t ut 1 ,if Et ,stop; else
function. If we adopt the Gaussian function as a kernel function, i.e.
x 2 2 2
according to Eq. (5), Eq. (6) can be rewritten as,
-
-
RESULTS
We applied the model on the data set (academic result of one semester) of a University of Pune. Table I shows the dimension of the data set (Students scores)in the form N by M matrices, where N is the rows (# of students) and M is the column (# of courses) offered by each student. In table II, Performance index is specified as per the average score of
c
n
ik
m
J m (U ,V ) 2u 1 K (x
i1 k 1
, vi )
k
(7)
every individual so as to categorize in different class
The result generated is shown in tables III and IV. The corresponding algorithm is applied individually to the dataset and the respective count i.e the number of samples in each
Minimizing Eqs. (4) under the constraint of uik , we have
cluster are evaluated. The count values obtained in each of
ik
1 1 K (xk
c
, v )1
i
1
m1
m1
the clustering algorithm are thus compared with the actual
values and the overall performance of each cluster is calculated. The summarized results shows that KFCM has
1 (1 K (xk ,vi ))
j 1
better performance as compared to FCM.
n
Students score
Number of students
Dimension(Total number of subjects)
Data
78
5
uik
K (xk
, vi )xk
TABLE I. STATISTICS OF DATA USED
v
k 1
i n
ik
k
u m K (x
k 1
, vi )
(9)
Here we just use the Gaussian kernel function for simplicity. If we use other kernel functions, there will be corresponding modifications in Eq. (5) and (6). In fact, Eq.(3) can be viewed as kernel-induced new metric in the data space, which is defined as the following :
70 & above
Excellent
60-69
Very Good
50-59
Good
45-49
Very Fair
40-44
Fair
TABLE II. PERFORMANCE INDEX
(x) ( y)
21 K(x, y)
TABLE III. INDIVIUAL CLUSTER EFFICIENCY FOR FCM
Sr.
No.
Cluster
Actual
Count
Cluster Efficiency (%)
1
1
14
18
94
2
2
9
25
84
3
3
18
5
87
4
4
25
9
84
5
5
12
43
69
TABLE IV. INDIVIDUAL CLUSTER EFFICIENCY FOR KFCM
Sr.
No.
Cluster
Actual
Count
Cluster Efficiency (%)
1
1
14
15
99
2
2
9
13
96
3
3
18
14
96
4
4
25
20
95
5
5
12
38
74
-
CONCLUSIONS
Thus we studied and applied different clustering algorithms for the purpose of result analysis of students academic performance. The results of the paper in terms of cluster accuracy confirme that KFCM has a better performance than FCM when applied to evaluate the academics result of students. Also in future,numerical interpretation of the results based on the clustering algorithms will be shown which will be helpful in making an effective decision by the academic Planners.
REFERENCES
-
James C. Bezdek, Robert Ehrlich and William Full, FCM: The Fuzzy c-means Clustering Algorithm , Computer and Geosciences Vol. 10 , No. 2-3, pp. 191-203, 1984.
-
Oyelade, Oladipupo and Obagbuwa, Application of k-means Clustering Algorithm for Prediction of Students Academic Performance, Internation Journal of Computer science and Information Security, Vol. 7, No. 1, 2010.
-
SoumiGhosh and Sanjay kumarDubey, Comparative Analysis of k- means and Fuzzy c-means Algorithms, International Journal of Advanced Computer Scienceand Applications, Vol. 4, No. 4, 2013.
-
MamtaBharadwaj, AnkitaWalia and HemantTulsani, Comparative Research on Fuzzy c-means and k-means Clustering Segmentation, International Journalof Computer Application and Information Technology, Vol. 3, Issue II, 2013, ISSN : 2278-7720.
-
G. Raghotham Reddy, K.Ramudu, Syed zaheeruddin and Dr. R.RameshwarRao, Image Segmentation Using Kernel Fuzzy c-means Clustering on Level Set Method on Noisy Images.
-
Dao-Qiang Zhang and Song-Can Chen, Kernel-based Fuzzy and Possibilities c-means Clustering.
-
Tara.Saikumar, B.K.Anoop and P.S.Murthy, Robust Adaptive Thresold Algorithm based on Kernel Fuzzy Clustering on Image Segmentation.
-
R.Suganya, R.Shanthi, Fuzzy C- Means Algorithm- AReviewInternational Journal of Scientific and ResearchPublications,
Volume 2, Issue 11, November 2012 1 ISSN2250-3153
-
Bezdek JC, Cluster validity with fuzzy sets, Journal ofCybernetics,
Vol.3, 1974, pp. 58-73
-
D.Q. Zhang, S.C. Chen, A novel kernelized fuzzy c-meansalgorithm with application in medical imagesegmentation,Art. Intell. Med, vol. 32, 2004, pp. 37-50
-
Bezdek JC, Pattern recognition with fuzzy objective functionalgorithm, New York: Plenum Press, 1981
-
D.L. Pham, J.L. Prince, An adaptive fuzzy c-meansalgorithm for Imagesegmentation in the presence of intensityinhomogeneities, PatternRecogn. Lett. vol. 20, 1999, pp.57-68.
-
Z. Wang, S. e. Chen, and T.K.Sun," Multiple kernel learning algorithm,"IEEE Transactions on patterns analysis and Machine Intelligence, Vol.30,no.2,pp.348-353,Feb2008
-
Hossein Yousefi-Banaem, Saeed Kermani,Davood Khodadad, An Improved Spatial FCM Algorithm forCardiac Image Segmentation, 13thIranian Conference on Fuzzy Systems (IFSC), 2013
-
Abdenour Mekhmoukh, Karim Mokrani and Mohamed Cheriet, A modified Kernelized Fuzzy C-Means algorithm for noisyimages segmentation: Application to MRI images,IJCSI International Journalof Computer Science Issues, Vol. 9, Issue 1, No 1, January 2012.
-
K. M. Bataineh, M. Naji, M. Saqer,A Comparison Study between Various Fuzzy Clustering Algorithms, Jordan Journal of Mechanicaland Industrial Engineering, Volume 5, Number 4, Aug. 2011 ISSN 1995-6665 Pages 335 343