An Adaptive FCM with Incremental Learning For Human Activity Recognition

DOI : 10.17577/IJERTV3IS030627

Download Full-Text PDF Cite this Publication

Text Only Version

An Adaptive FCM with Incremental Learning For Human Activity Recognition

K. Nanthini Student, Department of CSE Kongu Engineering College Perundurai, Erode-638052

Tamil Nadu, India

R. Manjula Devi

Assistant Professor, Department of CSE Kongu Engineering College Perundurai, Erode-638052

Tamil Nadu, India

Abstract: Fuzzy Logic is a multi-valued logic where the truth values lies between zero and one. In any system there are two phases namely the training (learning) phase and the testing phase. In the training or learning phase the data samples are given as input to the system for training the fuzzy system, to classify the inputs according to the characteristics of the problem. In the testing phase the data instances are given as input to check whether the system classifies correctly. To improve the training speed of the fuzzy system the number of instances introduced to the epochs must be adjusted. The fuzzy system, when introduced with a set of new input patterns, the fuzzy system must be able to easily learn additional information from new samples to improve its recognition accuracy and must be independent of the previously trained data. This work aims to improve the training speed of the network, where, Adaptive Skipping method is applied. Thus the number of instances introduced to the epochs for training the fuzzy system is reduced. To train the new input patterns, Incremental Adaptive Fuzzy C-Means Learning approach is used. This method does not require the whole dataset with old and new dataset to train the fuzzy system but trains only the new data with previously used conditions. So, the system can easily adapt to the new input patterns and can also reduce the training time. Thereby the overall training time is reduced and new input samples are trained easily.

Keywords: Clustering analysis, Fuzzy Clustering, Fuzzy C- Means (FCM), Human Activity Recognition, Incremental Learning.

I.INTRODUCTION

Artificial Neural Network (ANN) is a mimic of human brain. Every ANN is composed of collection of neurons arranged in layers namely input layer, hidden layer and output layer. GA is a search heuristic that mimics the process of natural selection. This heuristic is routinely used to generate useful solutions to optimization and search problems. Genetic algorithms are used in the applications like bioinformatics, phylogenetics, computational science, engineering, economics, chemistry, manufacturing, arithmetic, physics, pharmacometrics and other fields. FL is a multi-valued logic where the truth values lies between zero and one. Here FL is used for classification purpose. Fuzzy Classification is the process of grouping elements into a fuzzy set whose membership function is defined by the truth value of a fuzzy propositional function. Classification is the process of grouping individuals having

the same characteristics into a set. A classification corresponds to a membership function that indicates whether an individual is a member of a class.

Every system has two phases namely the training phase and the testing phase. The training phase is used to train the fuzzy system by making the system learn from the experience to perform a task. At the testing phase, the fuzzy system is given with a data instance and processes it to produce an output. If the fuzzy system has correctly learnt, it should be able to generalize and produce the actual output.

Compared to the testing phase, the training phase is important because the fuzzy system can acquire the knowledge from this phase to enhance its performance. The factors that are to be considered during the training phase include the time taken to train the fuzzy system and the ability to train the new incoming data. The time taken for training the fuzzy system with the large dataset may consume large amount of time. This can be resolved using adaptive skipping method. In the real-time scenario, the data samples may arrive periodically. So, the fuzzy system must be able to gain knowledge from the new data without accessing the new data and without losing its previously acquired knowledge. This can be done by using the incremental learning approach.

Adaptive skipping (AS) method is used to speed up the training time with maintaining the generalization process. This method is done by skipping the correctly classified samples for specific number of epochs, where the training time of the large dataset could be reduced.

An incremental learning (IL) algorithm is defined as the acquisition of new input samples periodically. By using this incremental algorithm, the fuzzy system has the ability to learn from new information whenever the information becomes available. A fuzzy system should be able to use any new training data to improve its performance, without using the previous data. This can also be used to accommodate new classes of data that are introduced with the new data.

In incremental learning, whenever a new input is given to the system, the new inputs are trained without affecting the previous knowledge. To train the new inputs the conditions used in training the previous data are used to improve the training speed of the fuzzy system. The

adaptive method is incorporated in the incremental method where the numbers of instances are reduced to speed up the training of the fuzzy system.

The rest of the paper is organized as follows. In Section 2, related works are reviewed. Section 3 describes the Problem Definition. Section 4 elaborates the proposed work. Section 5 deals with the results and the discussions employed on Human Activity Recognition dataset. Finally Section 6 concludes the paper and gives the future work that can be incorporated in the FCM for human recognition system.

  1. RELATED WORK

    A. Incremental Learning

    Polikar et al (2001) proposed an algorithm for incremental training of neural network (NN) pattern classifiers. The proposed algorithm enables supervised NN paradigms, such as the multilayer perceptron (MLP), to accommodate new data, including examples that correspond to previously unseen classes. Furthermore, the algorithm does not require access to previously used data during subsequent incremental learning sessions, yet at the same time, it does not forget previously acquired knowledge. The incremental learning algorithm as one that meets the following criteria as 1.It should be able to learn additional information from new data, 2.It should not require access to the original data, used to train the existing classifier, 3.It should preserve previously acquired knowledge, 4. It should be able to accommodate new classes that may be introduced with new data.

    Muhlbaier et al (2009) proposed an algorithm that describes an efficient incremental learning of multiple new classes with significantly few classifiers. To do so, the algorithm introduced dynamically weighted consult and vote(DW-CAV), a novel voting mechanism for combining classifiers: individual classifiers consult with each other to determine which ones are most qualified to classify a given instance, and decide how much weight, if any, each classifiers decision should carry.

    He et al (2011) proposed an adaptive incremental learning framework for the stream of data that appear over a specific amount of time. This algorithm is capable of learning from continuous raw data, accumulating experience over time, and using such knowledge to improve future learning and prediction performance.

    Wong et al (2011) proposed an algorithm on Radial Basis Function (RBF) with incremental learning method based on orthogonal least square method. It is designed to accommodate ew information without retraining the initial network. In our proposed method, the selection of the regressors for the new data is done locally, hence avoiding the expensive reselecting process. In addition, it accumulates previous experience and learns updated new knowledge of the existing groups to increase the robustness of the system.

    Pang et al (2011) proposed an incremental algorithm for incrementally learning the threshold-based classifiers. Threshold-based classifier is a simple and

    powerful pattern classification tool, which is frequently used in applications of object detection and recognition. A threshold-based classifier is associated with a unique one- dimensional feature. A properly selected threshold and a binary sign corresponding to the feature govern the classifier. Here, the learning process is generally done in a batch manner. The batch algorithms are not suitable for sequentially incoming data because of the limitation of storage and prohibitive computation cost. To deal with sequentially incoming data, this proposed incremental algorithm is used.

    Cavalin et al (2012) proposed the LoGID (Local and Global Incremental Learning for Dynamic Selection) framework, the main goal of which is to adapt hidden Markov model-based pattern recognition systems during both the generalization and learning phases. Given that the base line system is composed of a pool of base classifiers, adaptation during generalization is performed through the dynamic selection of the members of this pool that best recognize each test sample. This is achieved by the proposed K-nearest output profiles algorithm, while adaptation during learning consists of gradually updating the knowledge embedded in the base classifiers, by processing previously unobserved data. This phase employs two types of incremental learning: local and global. Local incremental learning involves updating the pool of base classifiers by adding new members to this set. The new members are created with the Learn++ algorithm. Global incremental learning, in contrast, consists of updating the set of output profiles used during generalization process.

    Lu et al (2012) proposed an Incremental Complete Linear Discriminant Analysis (ICLDA) method which can accurately update the discriminant vectors of Complete Linear Discriminant Analysis (CLDA) when new samples are inserted into the training set. The complete linear discriminant analysis algorithm has been proven to be an effective tool for face recognition. The CLDA method can make full use of the discriminant information of the training samples. However, the implementation of CLDA may not suitable for incremental learning problem. Thus, the ICLDA approach has been used for incremental learning ability.

    B) Clustering Algorithms

    Data clustering is the process of dividing data elements into classes or clusters so that items in the same class are as similar as possible, and items in different classes are as dissimilar as possible. Fuzzy clustering is a class of algorithms for cluster analysis in which the allocation of data points to clusters is fuzzy. In fuzzy clustering, every point in the cluster has a degree to indicate its importance rather than belonging to just one cluster.

    Bezdek proposed Fuzzy C-means that is applicable to a wide variety of geostatistical data analysis problems. This method generates fuzzy partitions and prototypes for any set of numerical data. These partitions are useful for corroborating known substructures or suggesting substructure in unexplored data. The clustering criterion used to aggregate subsets is a generalized least-

    squares objective function. Features of this program include a choice of three norms (Euclidean, Diagonal, or Mahalonobis), an adjustable weighting factor that essentially controls sensitivity to noise and acceptance of variable number of clusters.

    Pedrycz (1998) proposed a Conditional Fuzzy Clustering in the Design of RBF Neural Networks. This paper deals with the design of the neural networks, especially with the layer composed of RBFs, using the techniques of fuzzy clustering. The main idea of conditional clustering whose objective is to develop clusters preserving homogeneity of the clustered patterns with regard to their similarity in the input space as well as their respective values assumed in the output space.

    Pedrycz (2008) developed a concept of dynamic data granulation realized in presence of incoming data organized in the form of so-called data snapshots. For each of these snapshots a structure is reveals by running fuzzy clustering. The proposed algorithm of adjustable FCM exhibits a number of useful features which directly associate with the dynamic nature of the underlying data: (a) the number of clusters is adjusted from one data snapshot to another in order to capture the varying structure of patterns and its complexity, (b) continuity between the consecutively discovered structures is retained, viz the clusters formed for a certain data snapshot are constructed as a result of evolving the clusters discovered in the predeceasing snapshot. The cluster splitting is guided by conditional FCM while cluster merging involves two neighbouring prototypes. The criterion used to control the level of information granularity throughout the process is guided by a reconstruction criterion which quantifies an error resulting from pattern granulation and de-granulation. Numeric experiments provide a suitable illustration of the approach.

  2. PROBLEM DEFINITION

    The limitation of the existing methods is that the training phase consumes a large amount of time for large datasets which may affect the training rate. In order to improve the speed of the training in the training phase the adaptive skipping method is introduced. This adaptive skipping method skips the correctly clustered samples for specific number of iterations, which improves the training speed without affecting its generalization accuracy.

    When a new set of input patterns are introduced to the present network, all the input patterns are retrained instead of training only the new input patterns. In order to retrain the whole network, the Incremental training is employed. The Incremental Learning process does not require the previously trained dataset to train the newly incoming data but trains the new data with the previously used conditions for clustering and is capable of adding new activities.

    The training speed can be improved using the adaptive skipping method and the newly introduced dataset must be trained without retraining the whole dataset. The Adaptive FCM with Incremental Learning is used to improve the training time and to train the new dataset

    without using the previously trained dataset but with the previously used conditions to cluster the data instances.

  3. PROPOSED WORK

The paper proposes the fuzzy clustering with incremental learning capability for the daily human activities such as sitting, standing, walking, walking upstairs, walking downstairs and lying. The activities recorded for the dataset are proposed in [3]. In hard clustering the data samples can be assigned to only one cluster but in fuzzy clustering a single data sample may correspond to more than one cluster. In fuzzy clustering method each data point in the cluster is given with a membership degree to mark its importance in the cluster. The most widely used clustering algorithm is Fuzzy C- Means (FCM) algorithm which is used to cluster the data instances initially. The Adjustable Fuzzy Clustering (AFC) algorithm is employed for the human activity recognition for incorporating the incremental learning capability in the Human Activity Recognition system.

A) Adjustable Fuzzy Clustering

The Adjustable Fuzzy Clustering (AFC) is an improvement over the FCM algorithm. The Human Activity Recognition dataset is initially clustered using the FCM algorithm. The main limitation of the FCM algorithm is the lack of ability to learn incrementally. To make the algorithm learn incrementallythe AFC algorithm uses the previously used conditions to cluster the new data of the existing activities into the existing clusters or the new clusters are formed in the case of new activities.

In the Adjustable Fuzzy Clustering (AFC) the data samples are partitioned according to the activities to which they belong. Each activity containing a particular numbers of samples are clustered using the Fuzzy C Means algorithm and the corresponding cluster centres, degranulation samples and the reconstruction error are calculated as described as mentioned below.

The AFC uses the centres for calculating the degranulation samples to find the reconstruction error. The reconstruction error is used to find whether the clusters are to be merged or to be spitted.

deg=u (w) v / u (w)

H H

T T

h h h h=1 h=1

(1)

where uh(w) is the membership degree of the data samples in the cluster, T is the fuzzifier value, H is the total number of clusters and vh is the centre of the corresponding cluster.

The reconstruction criteria is calculated using difference between the degranulation samples and the data samples.

Vh =wdata |w-deg|

(2)

where w is the data samples and deg is the degranulation sample calculated for the corresponding data sample in the cluster.

The condition for the clusters to be merged or spitted is made according to the reconstruction criteria. If the reconstruction error is greater than the threshold value i.e., max(Vh)>threshold value the cluster are spitted otherwise the clusters are merged when the reconstruction criteria is less than the threshold value i.e., max(Vh)<threshold value. The new cluster centres and the membership degree for each data sample in the clusters are recalculated.

The new data samples can be of two types where the new data belong to the same activities or the data belong to the new activities. Whenever a new data sample of the existing activities are introduced to the Human Activity Recognition system the data samples are grouped to the corresponding cluster using the previous conditions such as centre and weight vectors. The second case corresponds to the data belonging to the new activity which is newly introduced. The new data for the new activity are clustered with the initial conditions as in FCM and the corresponding centres, membership functions, degranulation samples and the reconstruction criteria are calculated.

C) Probabilistic Neural Networks

The Probabilistic Neural Network (PNN) consists of four layers namely the input layer, pattern layer, summation layer and decision layer. The working of the layers are as followed.

Input Layer : In the input layer the input is acquired from the user and forwarded to the pattern layer.

Pattern Layer : The pattern layer contains the neurons equal to the number of centres calculated using the AFC. This layer dot products the input with the centre obtained from the AFC instead of introducing the whole training data instances as the weights to this layer.

m,k m.k

y =exp[(x*w -1)/2 ] (3)

where x is the input from the user wm,k is the centre obtained from the AFC, m is the number of classes, k is the kth centre of the class and is the smoothing factor calculated as

m m

K -1 K

M |wm,i -wm,j|

follows.

=1/M i=1 j=i+1

  1. Overall Algorithm for Adaptive Adjustable Fuzzy Clustering

    m=1

    (Km -1)(Km -2)

    (4)

    Step 1: Partition the dataset instances according to the activities to which they belong

    Step 2: Apply the Fuzzy C Means algorithm to cluster the dataset

    Step 3: During the iteration process for each data instance a skipping factor is calculated

    Step 4: If data sample is clustered correctly the skipping factor is incremented by one and

    where M is the total number of classes and Km is the total number of centre neurons belonging to one class, wm,i and wm,j are the centre neurons belonging to the particular class. Summation Layer : The Summation Layer consists of the neurons equal to the number of classes present in the dataset. The summation layer neuron summates the values from the pattern layer neurons that belong only to the class.

    1 1 Km

    respective data instance is skipped for that one

    zm = (2)L/2L K ym,k

    (5)

    iteration

    Step 5: Otherwise the incorrectly grouped data instance is introduced to the next iteration

    Step 6: Calculate the centre for the clusters and the membership degree of the each data sample in the cluster initially

    Step 7: Calculate the degranulation samples and reconstruction criteria

    Step 8: According to the degranulation sample the splitting mechanism and the merging mechanism are employed.

    Step 9: If a new data sample is introduced belonging to the existing activity that data sample is grouped to the cluster using the previous conditions such as centre and membership function

    Step 10: If the new data sample belonging to the new activity is introduced then the step 2 – 8 are followed

    Adjustable Fuzzy Clustering is used to find the centres by clustering the data samples. The centres found in the AFC are used in the pattern layer of the Probabilistic Neural Networks as explained below.

    m k=1

    Where L is the dimension of the data set ym,k is the output of the summation layer.

    Decision Layer : The decision layer consists of a single neuron which gets the inputs from the summation layer and decides to which the input neuron must be classified to.

    decision= arg max{pmzm|m=1,…,M} (6)

    m

    Where pm is the probability of the input belonging to one class where the value taken is 1/6.

    The classification of the input pattern to a particular class is done through the Probabilistic Neural Network with the use of Adjustable Fuzzy Clustering technique to reduce the memory usage of the system and to reduce the computation speed of the human activity recognition system. Thus the system is capable of training the samples in a reduced time to improve the training speed of the human activity recognition system.

    1. RESULT AND DISCUSSION

      The Adaptive Adjustable Fuzzy Clustering with Incremental Learning algorithm is used for the Human Activity Recognition dataset taken from the UCI machine repository. The data set consists of features using which the

      data samples are grouped into six clusters namely sitting, standing, walking, walking upstairs, walking downstairs and lying.

      A. Five-Fold cross validation

      unsupervised learning datasets to know the performance of the algorithm where the clustering can be used to cluster the data samples.

      VI.REFERENCES

      200

      Time in Seconds

      150

      100

      50

      0

      Training Time

      AFC Proposed

      Method

      1 2 3 4 5

      Training Step

      Fig.1

      1. Dong.G and Xie.M , Color Clustering and Learning for Image Segmentation Based on Neural Networks, IEEE Transactions on Neural Networks, vol. 16, no. 4, pp.925-937, 2005.

      2. Chatzis.P.S and Varvarigou.A.T , A Fuzzy Clustering Approach Toward Hidden Markov Random Field Models for Enhanced Spatially Constrained Image Segmentation, IEEE Transactions on Fuzzy Systems, vol. 16, no. 5, pp.1351-1362, 2008.

      3. Brezmes.T, Gorricho.J.L, Cotrina.J, Activity Recognition from Accelerometer Data on a Mobile Phone, IWANN 2009, Part II,

        LNCS 5518, pp. 796799, © Springer-Verlag Berlin Heidelberg 2009

      4. Sulaiman.N.S and Isa M.A.N , Denoising-based Clustering Algorithms for Segmentation of Low Level Salt-and-Pepper Noise- Corrupted Images, IEEE Transactions on Consumer Electronics, Vol. 56, No. 4,pp.2702-2711.,2010.

      5. Sulaiman.N.S and Isa M.A.N , Adaptive Fuzzy-K-means Clustering Algorithm for Image Segmentation, IEEE Transactions on Consumer Electronics, Vol. 56, No. 4,pp.2661-2669,2010.

      6. Eghbal G.M , FRBC: A Fuzzy Rule-Based Clustering Algorithm, IEEE ransactions on Fuzzy Systems, vol. 19, no. 5, pp.965-977, 2011.

        In this cross validation method the dataset is divided into five partitions say pi where i correspond from 1 to 5. During the first validation process partitions p1 to p4 are used for training the human activity recognition system and the p5 partition is used for testing the system. For the next validation process partitions p1 to p3 and p5 are taken for training and the partition p4 is used for testing. Likewise the iterations are continued for five times taking four partitions for training the system and the remaining one partition for testing the human activity recognition system.

        The Fig.1 and Fig.2 shows the training and the testing time of the human activity recognition system respectively with the existing Adjustable Fuzzy Clustering method and the proposed Adaptive Adjustable Fuzzy Clustering. The results show that the training time and the testing time of the Human Activity Recognition system has been improved by using the Adaptive Skipping method.

    2. CONCLUSION AND FUTURE WORK

    Thus the regular human activities have been classified into six classes using the Adjustable Fuzzy Clustering algorithm. Development of any system consists of two phases namely the training phase and testing phase as described in the introduction. The training phase of any system consumes considerably large amount of time compared to the testing phase. The Human Activity Recognition system is incorporated with the incremental learning ability with the use of AFC algorithm where the training is also reduced to improve the performance of the overall system. The training time is improved using the adaptive skipping method without degrading the generalization accuracy. The results obtained are shown in the results and discussions section which are satisfactory.

    The future work of this paper is to incorporate the Adaptive Adjustable Fuzzy Clustering for various datasets. This algorithm can also be tested using the

    1. Hung.C.C, Sameer.K and Kuo.B.C , A New Weighted Fuzzy C- Means Clustering Algorithm for Remotely Sensed Image Classification, IEEE Journal of Selected Topics in Signal Processing, vol. 5, no. 3, pp.543-554, 2012.

    2. Indrajit.S, Ujjwal.M,Sanghamitra.B and Dariusz.P , SVMeFC: SVM Ensemble Fuzzy Clustering for Satellite Image Segmentation, IEEE Geoscience and Remote Sensing Letters, vol. 9, no. 1, pp.52-56,2012.

    3. Karacali.B , Hierarchical Motif Vectors for Prediction of Functional Sites in Amino Acid Sequences Using Quasi-Supervised Learning, IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 9, no. 5, pp. 1432-1442, 2012.

    4. Zhang.Z, Yan.S and Zhao.M , Pairwise Sparsity Preserving Embedding for Unsupervised Subspace Learning and Classification, IEEE Transactions on Image Processing, vol. 22, no. 12, pp. 4640 – 4652, 2013.

    5. Bezdek C., Ehrlich R. and Full W.,Fuzzy C-Means Clustering Algorithm, Computers and Geosciences,Vol.10,No.2-3,pp.191- 203,1984.

    6. Chen S., Cowan C.F.N. and Grant P.M., Orthogonal Least Square Learning Algorithm For Radial Basis Function Networks, IEEE Transactions on Neural Networks, Vol.2, No.2, pp.302-310,1991.

    7. Hang H.G. and Qiao J.F., Adaptive Computation Algorithm for RBF Neural Network, IEEE Transactions on Neural Networks and Learning Systems, Vol.23, No.2, pp.342-348,2012.

    8. He H., Chen S., Li K. and Xu X., Incremental Learning From Stream Data, IEEE Transactions on Neural Networks, Vol.22,No.12,pp.1901-1915,2011.

    9. Lu G.F., Zou J. and Wang Y., Incremental LDA for Face Recognition, Pattern Recognition, pp.2510-2521,2012.

    10. Manjula Devi R., Kuppuswami S., and Suganthe R.C., Fast Linear Adaptive Skipping Training Algorithm For Training Artificial Neural Networks, Mathematical Problems in Engineering, Volume 2013, Article Id 346949,2013.

    11. Muhlbaier M., Topalis A., and Polikar R., Learn++.NC: Combining Ensemble of Classifiers With Dynamically Weighted Consult-and- Vote for Efficient Incremental Learning of New Classes, IEEE Transactions on Neural Networks, Vol.20,No.1, pp.152-168,2009.

    12. Pedrycz W., A dynamic data granulation through adjustable fuzzy clustering, Pattern Recognit. Lett., vol. 29, pp. 20592066,2008.

    13. Peñalver A. and Escolano F., Entropy-Based Incremental Variational Bayes Learning Of Gaussian Mixtures ,IEEE Transactions On Neural Networks and Learning Systems,Vol.23,No.3,pp.534- 542,2012.

    14. Polikar R., Udpa L., Udpa S.S. and Honavar V., Learn++: An Incremental Learning Algorithm for Supervised Neural Networks, IEEE Transactions On Systems, Man, And Cybernetics, Vol.31,No.4,pp.497-508,2001.

    15. Varnava T.M. and Meade A.J., An Initialization Method For Feed forward Artificial Neural Networks Using Polynomial Bases, Advances in Adaptive Data Analysis, Vol.3,No.4,pp.385-400,2011.

    16. Wang Z., Jiang M., Hu Y. and Li H., An Incremental Learning Method Based on Probabilistic Neural Networks and Adjustable Fuzzy Clustering For Human Activity Recognition by Using Wearable Sensors, IEEE Transactions on Information Technology in Biomedicine, Vol.16,No.4,pp.691-700,2012.

    17. Wong Y.W., Seng K.P., and Ang L., Radial Basis Function Neural Network with Incremental Learning for Face Recognition, IEEE Transactions on Systems, Man, and Cybernetics, Vol. 41, No. 4, pp.940-949,2011.

Leave a Reply