 Open Access
 Total Downloads : 662
 Authors : B.Rama Krishna, J.Ramesh
 Paper ID : IJERTV1IS7152
 Volume & Issue : Volume 01, Issue 07 (September 2012)
 Published (First Online): 25092012
 ISSN (Online) : 22780181
 Publisher Name : IJERT
 License: This work is licensed under a Creative Commons Attribution 4.0 International License
An Efficient Self Constructing Algorithm For Text Categorization
An Efficient Self Constructing Algorithm for Text Categorization
B.RAMA KRISHNA1, J.RAMESH2
1 M.tech Student, Dept of Computer Science and Engineering,
QIS COLLEGE OF ENGG. & TECH.,ONGOLE,Andhra Pradesh, India.
2 Asst.Professor, Dept. of Computer Science and Engineering
QIS COLLEGE OF ENGG. & TECH., ONGOLE, Andhra Pradesh, India
Abstract Text categorization is a predefined category to the natural language text. One of the major characteristic of text document classification problem is extremely reduce high dimensionality of text data in to low dimensionality. In this paper we introduce a naive algorithm for feature/word selection for the purpose of text classification. We use sequential forward selection methods based on improved mutual information criterion functions. The performance of the proposed evaluation functions compared to the information gain which evaluate features individually is discussed. We present experimental results using naive Bayes classifier based on multinomial model, linear support vector machine and knearest neighbour classifiers on the Reuters, Webkbs, 20 news subgroups data sets .
Feature clustering is a powerful method to reduce the dimensionality of feature vectors for text classification. In this paper, we propose a fuzzy similaritybased selfconstructing algorithm for feature clustering. The words in the feature vector of a document set are grouped into clusters, based on similarity test. Words that are similar to each other are grouped into the same cluster. Each cluster is characterized by a membership function with statistical mean and deviation. When all the words have been fed in, a desired number of clusters are formed automatically. We then have one extracted feature for each cluster. The extracted feature, corresponding to a cluster, is a weighted combination of the words contained in the cluster. By this algorithm, the derived membership functions match closely with and describe properly the real distribution of the training data. Besides, the user need not specify the number of extracted features in advance, and trialanderror for determining the appropriate number of extracted features can then be avoided.
Keywords Feature/word selection, Support vector machine, feature reduction, feature clustering, feature extraction, Tfidf and weight.

INTRODUCTION
The goal of text document classification is to assign automatically a new document into one or more predefined classes based on its contents. An increasing number of statistical classification methods and machine learning algorithms have been explored to build automatically a classifier by learning from previously labelled documents including naive Bayes, knearest neighbour, support vector
machines, neural network, decision trees and logistic regression. The overview of discusses the main approaches to text classification.
We propose a fuzzy similaritybased selfconstructing feature clustering algorithm, which is an incremental feature clustering approach to reduce the number of features for the text classification task. The words in the feature vector of a document set are represented as distributions and processed one after another. Words that are similar to each other are grouped into the same cluster. Each cluster is characterized by a membership function with statistical mean and deviation. If a word is not similar to any existing cluster, a new cluster is created for this word. Similarity between a word and a cluster is defined by considering both the mean and the variance of the cluster. When all the words have been fed in, a desired number of clusters are formed automatically. We then have one extracted feature for each cluster. The extracted feature corresponding to a cluster is a weighted combination of the words contained in the cluster. Three ways of weighting, hard, soft, and mixed, are introduced. By this algorithm, the derived membership functions match closely with and describe properly the real distribution of the training data. Besides, the user need not specify the number of extracted features in advance, and trialanderror for determining the appropriate number of extracted features can then be avoided. Experiments on realworld data sets show that our method can run faster and obtain better extracted features than other methods.

DEFINITIONS AND NOTATIONS
Let C = {c1, . . . , cC} be the set of C predefined classes and let D = {d1, . . . , dD} be the finite training document set. Let V = {w1, . . . , wV} be the vocabulary set containing V distinct words occurred in training documents. Given a set of document vectors {d1, . . . , dD} and their associated class labels c(dj ) = {c1, . . . , cC}, text classification is the problem of estimating the true class label of a new document. Text documents cannot be directly interpreted by a classifier. According to the bagofwords representation, the document d can be represented by a feature vector consisting of one feature variable for each word in the given vocabulary set V. A common characteristic of text data is its extremely high
dimensionality. The number of potential features (several tens of thousands) exceeds the number of training documents.
To process documents, the bagofwords model is commonly used. Let D = { d1; d2; . . . ; dn } be a document set of n documents, where d1, d2; . . . ; dn are individual documents, and each document belongs to one of the classes in the set {c1; c2; . . . . . . ; cp}. If a document belongs to two or more classes, then two or more copies of the document with different classes are included in D. Let the word set W = {w1; w2; . . . ; wm} be the feature vector of the document set. Each document di, 1 i n, is represented as di = <di1; di2; . . . ; dim>, where each dij denotes the number of occurrences of wj in the ith document. The feature reduction task is to find a new word set W' = {w'1; w'2; . . . ; w'k}, k << m, such that W and W' work equally well for all the desired properties with
D. After feature reduction, each document di is converted into
methods for word selection attempt to select from the original set V, the set V' of words with  V' V that, when used for document representation, yields the highest effectiveness. Different methods for feature subset selection have been developed in pattern recognition and machine learning using different evaluation functions and search procedures.
C. Feature Reduction
In general, there are two ways of doing feature reduction, feature selection, and feature extraction. By feature selection approaches, a new feature set W' = {w'1, w'2, . . . , w'k} is obtained, which is a subset of the original feature set W. Then W' is used as inputs for classification tasks. Information Gain (IG) is frequently employed in the feature selection approach [10]. It measures the reduced uncertainty by an information theoretic measure and gives each word a weight. The weight of a word wj is calculated as follows:
a new representation d' = <d' ; d' ; . . . ; d' > and the
i i1 i2 ik
converted document set is D' = {d' ; d' ; . . . ; d' }. If k is much IG(w ) = –
1 2 n j
smaller than m, computation cost with subsequent operations on D' can be drastically reduced.

Dimensionality Reduction
Dimensionality reduction is a very important step in text classification because irrelevant and redundant features often derade the performance of classification algorithms both in speed and classification accuracy. The number of features can be dramatically reduced by the domain dependent methods which include the elimination of stop words, stripping of special characters as well as stemming algorithms or morphological analysis. For a further dimensionality the domain independent methods can be used.

Feature Subset Selection
In text classification the dominant approach to dimensionality reduction is feature selection. Given a predefined integer  V' ,
+ P(wj)
+ P( ) (1)
where P (cl) denotes the prior probability for class cl, P(wj) denotes the prior probability for feature wj, P j) is identical to 1P(wj), and P(cl wj ) and P (cl j) denote the probability for class cl with the presence and absence, respectively, of wj. The words of top k weights in W are selected as the features in W'.
In feature extraction approaches, extracted features are obtained by a projecting process through algebraic transformations. An incremental orthogonal centroid (IOC) algorithm was proposed in [14]. Let a corpus of documents be represented as an m Ã— n matrix X RmÃ—n, where m is the number of features in the feature set and n is the number of documents in the document set. IOC tries to find an optimal transformation matrix F* RmÃ—k, where k is the desired number of extracted features, according to the following criterion:
F* arg max trace(FTSbF), (2) where F* RmÃ—k and FT F=I, and
Sb =
with P (cq ) being the prior probability for a pattern belonging to class cq, Mq being the mean vector of class cq, and Mall being the mean vector of all patterns.

Feature Extraction
Feature Extraction is a method of retrieving information from reduced data. That is which data you would like to retrieve then automatically extract from that area. Extracted data in the sense already pre processed data. This is also one of the method for text classification.

Text Classification
Text classification [10] (also known as text categorization or topic spotting) is the task of automatically sorting a set of
documents into categories from a predefined set. This task has several applications, including automated indexing of scientific articles according to predefined thesauri of technical terms, filing patents into patent directories, selective dissemination of information to information consumers, automated population of hierarchical catalogues of Web resources, spam filtering, identification of document genre, authorship attribution, survey coding, and even automated essay grading. Automated text classification is attractive because it frees organizations from the need of manually organizing document bases, which can be too expensive, or simply not feasible given the time constraints of the application or the number of documents involved. The accuracy of modern text classification systems rivals that of trained human professionals, thanks to a combination of information retrieval (IR) technology and machine learning (ML) technology.

Feature Clustering
i
Feature clustering is an efficient approach for feature reduction, which groups all features into some clusters, where features in a cluster are similar to each other. The feature clustering methods [are hard clustering methods, where each word of the original features belongs to exactly one word cluster. Therefore each word contributes to the synthesis of only one new feature. Each new feature is obtained by summing up the words belonging to one cluster. Let D be the matrix consisting of all the original documents with m features and D' be the matrix consisting of the converted documents with new k features. The new feature set W' = {w'1; w'2; . . . ; w'k}, corresponds to a partition ( , ,. ) of the original feature set W, i.e., Wt Wq = where 1 q; t the partition. Then, the tth feature value of the converted document d' is calculated as follows:
=
which is a linear sum of the feature values in Wt. The divisive informationtheoretic feature clustering (DC) algorithm, calculates the distributions of words over classes, P(Cjwj), 1 j m,
where C = {c1; c2; . . . . . . ; cp}, and uses KullbackLeibler divergence to measure the dissimilarity between two distributions. The distribution of a cluster Wt is calculated as follows:
P(CWt) = P(Cwj). (5)
The goal of DC is to minimize the following objective function:
which takes the sum over all the k clusters, where k is specified by the user in an Pixel Feature Clustering

Term Frequency and inverse document frequency:
Tfidf is a form of assigning frequent words in the documents. Basically, the term frequency describes how frequently the word appears in the document. At the same time the inverse document frequency describes how the term or word frequently appears in the remaining documents. We can add another feature for this classification, i.e, compactness. This compactness describes position of the word in particular document. Totally these features represents what is the first appearance of the word, what is the last appearance of the word. It is also describes middle of the word. Tfidf is most powerful method for classification.


PROPOSED PROCEDURE
There are some issues pertinent to most of the existing feature clustering methods. First, the parameter k, indicating the desired number of extracted features, has to be specified in advance. This gives a burden to the user, since trialanderror has to be done until the appropriate number of extracted features is found. Second, when calculating similarities, the variance of the underlying cluster is not considered. Intuitively, the distribution of the data in a cluster is an important factor in the calculation of similarity. Third, all words in a cluster have the same degree of contribution to the resulting extracted feature. Sometimes, it may be better if more similar words are allowed to have bigger degrees of contribution. Our feature clustering algorithm is proposed to deal with these issues.
Suppose, we are given a document set D of n documents d1, d2, . . . , dn, together with the feature vector W of m words w1,w2,. wm and p classes c1, c2, . . . , cp, as specified in Section 2. We construct one word pattern for each word in W. For word wi, its word pattern xi is defined, similarly as in [21], by
xi = <xi1, xi2, . . . , xip>,
= <P (c1wi), P (c2wi), . . . , P (cpwi)>
Where
(8)
mutual information between class ck and word w. We can consider three strategies for solving our feature selection problem.
The optimal strategy generates all the word subsets S and
for 1 j p. Note that dqi indicates the number of occurrences of wi in document dq, as described in Section 2.
Also, is defined as
(9)
Fig. 2 principal components can be now be clustered for SelfOrganizing Therefore, we have m word patterns in total. For example, suppose we have four documents d1, d2, d3, and d4 belonging to c1, c1, c2, and cn, respectively. Let the occurrences of w1 in these documents be 1, 2, 3, and 4, respectively. Then, the word pattern x1 of w1 is:
X1 = <0:3; 0:7>:
We consider the global filtering approach to feature selection in text document task. In this section novel algorithms for feature selection using mutual information are presented.
1. Feature Selection using Mutual Information
Our feature subset selection problem is formulated as follows: Given an initial set V with V features, find the subset Ss V with S features that maximizes the mutual information for text defined as mutual information for a set of words averaged over all classes given by the following formula:
MI(S) = (1)
The mutual information for a feature/word w (word w
occurred) averaged over all classes is defined as:
MI(w) = I( ,w) (2)
Here P(w) is the probability,that the word w occurred, P(ck) is the probability of the class ck, P(wck) is the conditional probability of the word w given the class ck, I(ck,w) is the
compares their MI(S). It is almost impossible for too many combinations. In the backward elimination strategy we remove the worst word from the complete set V one by one till the required number of words remain. This procedure has a lot of difficulties in computing I(ck, S).

CLUSTERING ALGORITHM
The whole clustering algorithm can be summarized below.
Initialization
of original word patterns: m
of classes: p
Threshold: Initial deviation:
Initial deviation: Initial of clusters: k = 0
Input:
xi = < xi1; xi2; . . . ; xip>, i m
Output:
Clusters G1, G2; . . . ; Gk
procedure SelfConstructingClusteringAlgorithm
for each word pattern xi, 1 i m temp_W = Gj Gj (xi) ; 1 j k}; if(temp_W== )
A new cluster Gh, h = k +1
else let Gt temp W be the cluster to which xi is closest by (19);
Incorporate xi into Gt by (20)(24);
endif; endfor;
return with the created k clusters;
endprocedure
Note that the word patterns in a cluster have a high degree of similarity to each other. Besides, when new training patterns are considered, the existing clusters can be adjusted or new clusters can be created, without the necessity of generating the whole set of clusters from the scratch. The order in which the word patterns are fed in influences the clusters obtained. We apply a heuristic to determine the order. We sort all the patterns, in decreasing order, by their largest components. Then the word patterns are fed in this order. In this way, more significant patterns will be fed in first and likely become the core of the underlying cluster. For example, let x1 = <0:1; 0:3;
Fig. 3 Feature Clustering Diagram
algorithm and feature Clustering can be applied to classify literatures in the field of electrical engineering. Future study direction will focus on the effect to performance when related parameters, such as crossingover rate, mutation rate, size of population, etc., have different values, and improve computational efficiency of the new algorithm further.
Similaritybased clustering is one of the techniques we have developed in our machine learning research. In this paper, we apply this clustering technique to text categorization problems. We are also applying it to other problems, such as image segmentation, data sampling, fuzzy modelling, web mining, etc. We found that when a document set is transformed to a collection of word patterns, the relevance among word patterns can be measured, and the word patterns can be grouped by applying our similaritybased clustering algorithm. Our method is good for text categorization problems due to the suitability of the distributional word clustering concept.
The following are the results for text classification. This classification basically depend on clustering after that each word in the input file must be compared to trained data set. Based on weight this algorithm classifies input data and represent individual classification results.
TABLE I
COMPARE OF RESULTS AMONG THREE ALGORITHMS IN IRIS DATASET
Accuracy Rate of
classification
Running time
Best
Worst
Avg.
Best
Worst
Avg.
SVM
1
0.833
0.9033
0.1406
0.2813
0.19064
GASVM
1
1
1
6.9063
7.8906
7.3172
EGASVM
1
1
1
6.6094
7.3594
6.93439
TABLE 2
CLASSIFICATION RESULT
List of Input Text
Documents
Classification(Based on
all algorithms)
Crude
C1
Wheat
C1
Reuters
C3
Webkb
C5

CONCLUSIONS
This paper mainly focused on the problem of feature selection for document classification. In this paper we presented methods based on novel improved mutual information measures. The proposed algorithms are new in the text classification field. It uses good optimization performance of support vector machines to improve classification performance of genetic algorithm with elite strategy. Iris dataset and a text dataset are chosen to validity performance of the combing algorithm. Its obviously that the hybrid
REFERENCES

K. Nigam, A. K. McCallum, S. Thrun, and T. Mitchell, Text classification from labeled and unlabeled documents using em, Machine Learning, vol. 39, pp. 103134, 2000.

A. McCallum and K. Nigam, A comparison of event models for naive

Bayes text classification, in Proceedings of the AAAI98 Workshop on Learning for Text Categorization, 1998.

S. M. Metev and V. P. Veiko, Laser Assisted Microtechnology, 2nd ed.,
R. M. Osgood, Jr., Ed. Berlin, Germany: SpringerVerlag, 1998.

J. Breckling, Ed., The Analysis of Directional Time Series: Applications to Wind Speed and Direction, ser. Lecture Notes in Statistics. Berlin, Germany: Springer, 1989, vol. 61.

S. Zhang, C. Zhu, J. K. O. Sin, and P. K. T. Mok, A novel ultrathin elevated channel lowtemperature polySi TFT, IEEE Electron Device Lett., vol. 20, pp. 569571, Nov. 1999.

M. Wegmuller, J. P. von der Weid, P. Oberson, and N. Gisin, High resolution fiber distributed measurements with coherent OFDR, in Proc. ECOC00, 2000, paper 11.3.4, p. 109.

R. E. Sorace, V. S. Reinhardt, and S. A. Vaughn, Highspeed digital toRF converter, U.S. Patent 5 668 842, Sept. 16, 1997.

(2002) The IEEE website. [Online]. Available: http://www.ieee.org/

M. Shell. (2002) IEEEtran homepage on CTAN. [Online]. Available: http://www.ctan.org/texarchive/macros/latex/contrib/supported/IEEEtr an/FLEXChip Signal Processor (MC68175/D), Motorola, 1996.

Support vector machine. http://en.wikipedia.org/wiki/Support_vector_machine [2009106]

Genetic algorithm. http://en.wikipedia.org/wiki/Genetic_algorithm. [2009106]

Text classification. http://en.wikipedia.org/wiki/Text_classification[2009106]

Vector space model. http://en.wikipedia.org/wiki/Vector_space_model [2009106]

PDCA1270 data sheet, Opto Speed SA, Mezzovico, Switzerland.

A. Karnik, Performance of TCP congestion control with rate feedback: TCP/ABR and rate adaptive TCP/IP, M. Eng. thesis, Indian Institute of Science, Bangalore, India, Jan. 1999.

J. Padhye, V. Firoiu, and D. Towsley, A stochastic model of TCP Reno congestion avoidance and control, Univ. of Massachusetts, Amherst, MA, CMPSCI Tech. Rep. 9902, 1999.

Wireless LAN Medium Access Control (MAC) and Physical Layer (PHY) Specification, IEEE Std. 802.11, 1997.

Li, S.T., Wu, X.X., and Hu, X.Y.: Gene selection using genetic algorithm and support vectors machines, Soft Computing, 2008, 12, (7), pp. 693698

FENG Helong and XIA Shengping. Web Page Classification Method Based On RSOMBayes. Computer Engineering, 2008,34(13),pp.61 63.

Liu Lizhen, HE Haijun, Lu Yuchang and Song Hantao. Application Research of Support Vector Machine in Web Information Classification. MINIMICRO SYSTEM, 2007,28(2), pp.337340.

Distributional features for text categorization, IEEE march 2009, XiaoBing Xue and ZhiHua Zhou, Senir Member, IEEE
AUTHORS PROFILE
Mr..Ramesh currently working as an Asst.Professor in Dept of CSE. He received his B.Tech from Chundi Ranganayakulu Engg. college, JNTU Hyderabad and M.Tech from QISCET,JNTU Hyderabad. He has 5 years of teaching experience and published many papers at various international journals and conferences. His Research areas includes Data
Mining and Data Warehousing .
B.Rama Krishna currently pursuing M.Tech (CSE) from QIS College of Engineering and Technology, Ongole, Affiliated to JNTU Kakinada. His Interested areas includes Data Mining and Data Warehousing.