ISSN ONLINE(2320-9801) PRINT (2320-9798)

All submissions of the EM system will be redirected to Online Manuscript Submission System. Authors are requested to submit articles directly to Online Manuscript Submission System of respective journal.

Improved Correlation Preserved Indexing For Text Mining

Vinnarasi Tharania. I1,M.Kanchana2,V.Kavitha3
  1. Assistant Professor, Department Of Information Technology, Karpaga Vinayaga College Of Engineering & Technology, China Kolambakkam, Madurantakam Taluk,Kanchipuram-603308, Tamilnadu, India
  2. Professor, Department Of Information Technology, Karpaga Vinayaga College Of Engineering & Technology, China Kolambakkam, Madurantakam Taluk,Kanchipuram-603308, Tamilnadu, India
  3. Assistant Professor, Department Of Information Technology, Karpaga Vinayaga College Of Engineering & Technology, China Kolambakkam, Madurantakam Taluk,Kanchipuram-603308, Tamilnadu, India
Related article at Pubmed, Scholar Google

Visit for more related articles at International Journal of Innovative Research in Computer and Communication Engineering

Abstract

Data mining is an excellent domain to work where new concepts are implemented. The knowledge extraction and knowledge discovery is a major task of engineering organization. Therefore a new field of study, Knowledge Discovery in Database (KDD), and data mining is explored. And many applications are been developed. In this paper a discussion about field of data mining is also placed. Document clustering is a field where we are implementing a concept of grouping of similar objects. Some groups are formed and the documents are placed under those groups. The main motive of this paper is comparison of various algorithms and giving the best result of the clustering. A new algorithm has been proposed in this paper is ICPI (Improved correlation preserving indexing) which is performed by the correlation similarity measure space. ICPI can successfully find out the essential structures rooted in high dimensional document space. The proposed work is to provide an efficient text mining algorithm to perform mining in the document.ICPI can successfully find out essential structures rooted in high dimensional document space.

Keywords

K-means, Latent Semantic Indexing (LSI), Locality Preserving Indexing (LPI), Correlation Preserving Indexing (CPI), Improved Correlation Preserving Indexing (ICPI), Knowledge Discovery in Database (KDD).

INTRODUCTION

More than hundred fifty years, the word "computer" started to appear in the vocabulary. Even twenty years back, a hard disk is transported only in airlines, due to huge size, for ex: a 4GB hard disk has 300Kg. Now a day, 1TB storage space is available in packet sized hard disk. The fields of application of the computers are very vast and various new applications are being searched out. The scientists applied it to prediction of weather earthquakes, storms, controlling of satellites, controlling of atomic reactions in reactors like this their application varies from small to large. The application is not limited, the use of computers are used also in designing cars, aeroplanes, ships, automobiles, bridges and tools etc. In the industry, the computers is playing a good job which applied for many tasks like word processing, record keeping, inventory controlling, patrol processing, account keeping and auditing, stock marketing and ticket reservations. Clustering is occurring as a main concept in the field of data mining. Where they are used in various applications like banking, finance etc.

II. RELATED WORK

From the year 1994 many researcher are researching in this clustering field. Till 2012 up-to-date the research is going in this field. Basically it is started with the k-means and many new researchers have found various new algorithms which improves the performance of mining. Text mining is a fundamental operation used in unverified document association, usual topic mining, and information retrieval. The well-known partitioning clustering algorithm is the K-means algorithm and its variants [2]. Latent Semantic Indexing (LSI)[4] is one among the most popular linear document indexing methods which constructs low dimensional representations. The main aim of LSI is to find the best subspace estimation to the original document space in the sense of minimizing the global reconstruction error. In other words, LSI[4] look forward to uncover the most representative features rather the most discriminative features for document representation.
K-MEANS
The K-mean algorithm takes the input as K and partitions a set of n Objects into K clusters so that the resulting intra-cluster similarity is high and the inter-cluster is low. The cluster similarity is measured with the help of the mean of each cluster. They are considered as the cluster’s centroid or Centre of Gravity. Firstly, they Randomly select K objects each of which initially represent a cluster mean or centre. Each of the remaining objects an object is assigned to the cluster to which it is most similar based on the cluster object and the mean value. Then they compute a new mean value. This process is repeated until there is no change.
To find the mean of the cluster

LATENT SEMANTIC INDEXING (LSI)

Latent semantic indexing is one of the most popular algorithm for document dimensionality reduction.It is fundamentally based on SVD(Singular value decomposition).suppose the rank of the term document A is r, then LSI decomposes X using SVD as follows:
are the singular value of A,U=[a1…….,ar] and ai is called the left singular vector, and the singular vector, and V=[v1………,vr],and vi is called the right singular vector.LSI uses the first k vectors in U as the transformation matrix to embed the original documents into k-dimensional subspace. It can be easily checked that the column vectors of U are the eigenvectors of AAT.

LOCALITY PRESERVING INDEXING (LPI)

The locality preserving indexing was proposed to discover the discriminant structure of the space of the document. It has been proved that it would have been more discriminant power than the LSI. However the computational complexity of the LPI is very costly due to the reason of that they involving Eigen-decomposition of two dense matrices. The LPI is not highly advised to use, since they use large data set.

The Algorithm used in Locality Preserving Indexing

 PCA Projection
We project the document set {Ai} into the PCA subspace by throwing away the smallest principal components. We denote the transformation matrix of PCA by Wpca
 Constructing the Adjacency Graph
Let G denote a graph with n nodes. The ith node corresponds to the document ai. We put an edge between nodes i and j if ai and aj are “close”, i.e. ai is among k nearest neighbors of ai is among k nearest neighbors of aj.
 Choosing the Weights
If node i and j are connected, put
Otherwise, put Sab = 0. The weight matrix S of graph G models the local structure of the document space.
 Eigenmap:
Compute the eigenvectors and eigenvalues for the generalized eigenvector problem:
Where D is a diagonal matrix whose entries are column (or row, since S is symmetric) sums of S,

CORRELATION PRESERVING INDEXING

Correlation preserving indexing (CPI), which clearly considers the multiple structure rooted in the similarities between the documents. It aims to find an optimal semantic subspace by simultaneously maximizing the correlations between the documents in the local patches and minimizing the correlations between the documents outside these patches. This is different from LSI and LPI, which are based on a dissimilarity measure (Euclidean distance), and are determined on detecting the intrinsic structure between widely separated documents rather than on detecting the intrinsic structure between nearby documents. Since the intrinsic semantic structure of the document space is often embedded in the similarities between the documents [7], CPI can effectively find the intrinsic semantic structure of the high-dimensional document space. CPI with euclidean distances performed competitively with CPI with correlation distance. Thus, similarity is an appropriate metric for measuring correlation between the documents.

III.PROPOSED METHOD

IMPROVED CORRELATION PRESERVING INDEXING

Compute Correlation between words i and j,
In maximum correlation the maximum similarity is found. And with the min correlation the max dissimilarity is found. Then those values of similarity are preserved. Correlation of documents to be preserved.Correlation Preserving Index to be maintained for all documents.The proposed work, improves existing CPI method for mining similarity pattern within the documents.
There are four modules in the proposed system. They are
 Document collection and pre-processing
 Implementing the correlation function
 Identification of optimal threshold
 Ranking, Index preserving and testing
Step1: Document Collection
Collecting the document from various data set. Data set Such as IEEE Explore, ICM Digital library, Scopus etc….IEEE Explore is a scholarly research database it contains indexes, abstract and provide a full text for the article and papers on various subjects. The IEEE Explore database contains the following collection such as journal, transactions, letters and magazines. Remove authors name and keywords. Convert uppercase letters to lower case letters. Remove stop list words.
Step2: Implementation of Correlation Function
Correlation is to find the similarity.Basically +1.00 positive correlation,-1.00 negative correlation and 0 no correlation.
Corr(I,J)=ITJ/√IT I√JTJ.Using this formula correlation is found for the documents pre-processed.
Step3: Identification of Optimal Threshold
From the value obtained from the correlation function a threshold value is found in this module. All the values obtained from them is stored in the form of a table. A intermediate value is obtained as the threshold value from the existing values. If the optimal threshold value is not up to the expectation then again the optimal threshold value can be changed and correlation can be done.
Step4: Ranking, Index Preserving and Testing
The values above the threshold values are ranked in a decrement order. Then those values are stored and an index is maintained. That index value is compared with the database were the dataset is also maintained. Then the output is displayed according to the input given.

IV.OUTPUT ANALYSIS

The fig 4.1 explains the document collection module initial step for this project the input is collected from the IEEE EXPLORE. So the above figure shows the home page of IEEE EXPLORE.
Here an abstract of an IEEE Paper is shown as the example with the sample abstract of Computer Science and Engineering.
The above fig.4 explains about the pre-processing of the documents in the sense it removes the stop-list words and the author names from the document collected. And convert them into the lower case letters and into the text document. This is done for a document of CSE.
Here in the above fig.7 the count of the literal is given along with the ranking. So in this figure they explain were all the words available in which file. And according to their count they are published in an order.

V.CONCLUSION

The proposed work is implemented using few set of documents which is collected from well known data sets such as IEEE EXPLORE, ICM DIGITAL LIBRARY and SCOPUS. The emerging researches from any well established research lab in this field of research can implement the proposed work in the own data sets. The result of this ICPI can be found with help any one single document given as the input. At the same instance many new solutions have been found in this ICPI. New solution is like solution to the problem of synonymy and polysemy. Since none of the algorithms give a better solution to these problem. But even though this algorithm has been proposed still improvement of paper is going on only the survey of this is done. So the future work will be with the result of the algorithm with other algorithm will be produce soon in the future.
 

Figures at a glance

Figure 1 Figure 2 Figure 3
Figure 1 Figure 2 Figure 3
Figure 1 Figure 2 Figure 3
Figure 4 Figure 5 Figure 6
 

References