Improving Scalability of R-Dictionary Using Keyword Based Approach | Open Access Journals

ISSN ONLINE(2320-9801) PRINT (2320-9798)

Yakışıklı erkek tatil için bir beldeye gidiyor burada kendisine türk Porno güzel bir seksi kadın ayarlıyor Onunla beraber otel odasına gidiyorlar Otel odasına rokettube giren kadın ilk önce erkekle sohbet ederek işi yavaş halletmeye çalışıyor sex hikayeleri Kocası fabrikatör olan sarışın Rus hatun şehirden biraz uzak olan bir türk porno kasabaya son derece lüks bir villa yaptırıp yerleşiyor Kocasını işe gönderip mobil porno istediği erkeği eve atan Rus hatun son olarak fotoğraf çekimi yapmak üzere türk porno evine gelen genç adamı bahçede azdırıyor Güzel hatun zengin bir iş adamının porno indir dostu olmayı kabul ediyor Adamın kendisine aldığı yazlık evde sikiş kalmaya başlayan hatun bir süre sonra kendi erkek arkadaşlarını bir bir çağırarak onlarla porno izle yapıyor Son olarak çağırdığı arkadaşını kapıda üzerinde beyaz gömleğin açık sikiş düğmelerinden fışkıran dik memeleri ile karşılayıp içeri girer girmez sikiş dudaklarına yapışarak sevişiyor Evin her köşesine yayılan inleme seslerinin eşliğinde yorgun düşerek orgazm oluyor

Improving Scalability of R-Dictionary Using Keyword Based Approach

G.Kokila1, Dr.R.Rajasekaran2
P.G Student CSE Dept, V.P.M.M Engineering College for Women, Krishnankovil, Tamilnadu, India
Related article at Pubmed, Scholar Google

Visit for more related articles at International Journal of Innovative Research in Computer and Communication Engineering


Translators and language professionals in general, have long claimed that dictionaries are deficient, especially for regarding access and updating of content. Some authors have also noted that these deficiencies are compounded by the fact that language professionals do not receive (proper) training in dictionary use, and therefore do not fully benefit from them. Electronic dictionaries include new search capabilities, not found in traditional dictionaries that could meet user’s needs. Most of the recent work in semantic area suggests the use of Domain specific Ontology to achieve meaningful classification. Hence the design and implementation of a reverse dictionary is described based on the semantic sentence similarity. The objective of this project is to make research in the area of semantic similarity. Semantic similarity is a widely adopted approach to language understanding in which the meaning of a text A is inferred based on how similar it is to another text B. Its scope is typically used in information retrieval with more efficiency and to build a machine that could communicate in natural language.



Reverse Dictionary (RD), Text mining, Information Retrieval, Semantic Similarity, Glosses.


The use of electronic dictionaries has many advantages over the traditional paper dictionary. However, access to the lexicon and terminology of a dictionary presents certain difficulties, partly due to the lack of user knowledge (even among language experts such as translators) about how a dictionary can be queried to access this kind of information, and partly due to the diversity of ways a dictionary can be consulted (in different areas of the dictionary, with different operators, in widely varied interfaces) which vary from one dictionary to another. Electronic dictionaries are available in both online and offline modes.
Electronic dictionaries and search possibilities: The development of new technologies and the Internet have progressively modified the concept of the dictionary. Many paper dictionaries have been converted to electronic formats, such as CD-ROM, while others are available online. Electronic dictionaries can be classified in various types according to different criteria. In this classification, the author distinguishes between newly developed electronic dictionaries (new development) and electronic versions of paper dictionaries (based on paper) (Gross 1997, Jacquet-Pfau 2002: 90). Nesi (2000a: 140) states that fully electronic dictionaries are more effective than electronic dictionaries adapted from paper versions: “electronic dictionaries would be most effective if they were designed from scratch with computer capabilities and computer search mechanisms in mind”. Electronic dictionaries can be easily updated (Kay 1984: 461, Carr 1997: 214, Harley 2000) and allow a quicker, more precise and exhaustive search, in which a variety of search criteria can be combined.
Types of searches includes
In the list of entries: The dictionary shows an alphabetical list of all the dictionary entries starting with the word or sequence of characters introduced by the user. In the entries: The dictionary retrieves a list of words that are alphabetically similar to that introduced by the user when the searched word is not found in the dictionary. In the definitions: The dictionary generates a list of words whose definitions contain all the words introduced in the dictionary with the operator ‘&’ or some of the words introduced with the operator ‘|’. (REVERSE DICTIONARY)Anagrams: The dictionary retrieves words that result from a combination of all the letters introduced.
Reverse Dictionary is an electronic dictionary where a user submits a sentence or phrase or series of words to a search engine and that search engine then produces that word that the person was defining. Also known as a “rictionary”. Follows search type based on in the definition. Hence it is concept based dictionary. The input is set of keywords or terms in the documents. Each keyword has both the property of syntactic similar and semantic similar. Semantic similarity or semantic relatedness is a concept whereby a set of documents or terms within term lists are assigned a metric based on the likeness of their meaning / semantic content.
Concretely, this can be achieved for instance by defining a topological similarity, by using ontology’s to define a distance between words (a naive metric for terms arranged as nodes in a directed acyclic graph, like a hierarchy, would be the minimal distance in separating edges between the two term nodes), or using statistical means such as a vector space model to correlate words and textual contexts from a suitable text corpus (co-occurrence).
Semantic Sentence Similarity is to find similarity between two sentences. Both semantic and syntactic information make contributions to the meaning of a sentence. Sentence compression may involve recognizing the relationship between parts of a sentence. The relationship may be implicit or explicit. The range value is 0 to 1.
The challenges are databases are rapidly growing due to the increasing amount of information available in electronic form is one the challenge; computing the semantic similarity is another challenge this is referred to Concept Similarity Problem (CSP).
Some other challenges are information is unstructured/semi structured; large textual data base and Noisy data for (e.g.) spelling mistakes; very high number of possible “dimensions” (but sparse); all possible word and phrase types in the language; word sense ambiguity for e.g. Apple (the company) or apple (the fruit).


Reverse Dictionary problem description is very simple. There are two main problems.
First, the user input is unlikely to exactly match the definition of a word in the dictionary database. For example, a user may enter the concept or sense phrase (query) “big building” when looking for word phrase (result) such as “castle” whose dictionary definition be “large building” which is conceptually similar but does not contain any of the same words in user input.
Second, the response efficiency needs to be similar to that of forward dictionary lookups.
Hence the problems are as follows upon receiving a search concept; the reverse dictionary consults the forward dictionary at its disposal and selects those words whose definitions are similar to this concept. Those words then form the output of reverse dictionary lookup. This problem is referred as concept similarity problem (CSP).
Estimating the semantic similarity of concepts is an important problem. Results of such studies are reported in a variety of fields, including psychology, natural language processing, information retrieval, language modeling and database systems
Some related works are helpful in understanding the problem of multiword similarity. One area of such work addresses the problem of finding the similarity of multiword phrases across a set of documents in Wikipedia. The documents contain sufficient contextual information (at least 50-100 words) for similarity matching, which fits within the traditional notion of “short documents” in IR research. In another area of related work, i.e., passage retrieval [6], [7], the concept of interest is modeled as a multiword input query. Work in this area attempts to identify relevant passages in largetext documents.
For a RD, semantic similarities must be computed between multiword phrases. In contrast, in the reverse dictionary scenario, the “documents” considered for similarity are very short dictionary definitions (often consisting of fewer than twenty words), which contain very little contextual information. The lack of contextual information in the RD case adds to the difficulty of addressing this problem.
In the system [1], this system use five pre-create database for dictionary.
One of the databases is RMS database, which contains a table of reverse mapping for each word. In case, new word is add to database the possible RMS is not updated automatically this is drawback of this system.


In RD, semantic similarity is calculated based on sentence phrase which is effective as compared with similarity based on each word. And finding passage similarity is unnecessary since RD contains only fewer words. It reduces the CSP effectively. The drawback of [1] is overcome from this proposed system. This paper is in proceeding.

Architectural design and implementation methodology

The overall architectural design of r-dictionary is somewhat similar to the text mining process because reverse dictionary is process based on the user input (consider as keyword) which is a text only.
There are three elements for processing the reverse dictionary: a query, a resource and a result. The query is the word or phrase introduced by the user in the interface of a resource.
The resource is the resource or part of the resource in which the word or phrase is searched.
The result is the element obtained when a query is searched in a resource.
Solution overview has four modules which are text preprocessing; finding sentence similarity; ranking; analysis of scalability.
Text preprocessing includes stop word removal and stemming. The input is user sentence which is denoted as sense phrase (S).S is query element.
First remove all stop words. Since these stop words do not add specificity to the input, removing them does no harm.
Apply stemming for converting the each word in the sentence to their base form. For that porter stemmer is used. From [2], the porter stemmer is effective when compared with other stemming algorithm.
The output of the text preprocessing is referred as the sense phrase 1(S1) which is input for finding sentence similarity. The S1 is compared with the database sentences (i.e. meaning of all words) for finding sentence similarity there are many algorithms some of them are listed in the table below. The Flowchart will represent the how semantic similarity is measuring.
For finding semantic similarity SEMILAR API is used. SEMILAR API offers a variety of similarity methods based on WordNet (Fellbaum, 1998), Latent Semantic Analysis (LSA; Landauer et al., 1998), Latent Dirichlet Allocation (LDA; Blei et al., 2003), BLEU (Papieni et al., 2002), Meteor (Denkowski and Lavie, 2011), Pointwise Mutual Information (PMI) (Church et al., 1990), methods that use syntactic dependencies, optimized lexico-syntactic matching methods based on Quadratic Assignment, methods that incorporate negation handling, etc. Some methods have their own variations which, coupled with parameter settings and user’s selection of preprocessing steps, could result in a huge methods space. SEMILAR provides a framework for the systematic comparison of various semantic similarity methods.
I choose similarity measure based on sentence.

Approaches for finding semantic sentence similarity

Semantic sentence similarity can be finding in two ways.
One is to expand word-to-word similarity (i.e. use similarity of word in one sentence to a word in another sentence and by some means calculate sentence level similarity score) and another approach is to use the semantic representations of sentences to calculate the sentence similarity directly.


Measure based on sentence directly
o Corley Mihalcea Comparer
o BLEU Comparer
o METEOR Comparer


The Table 1 measures come under this category.
Then the semantic measures is calculated based on word-to-word is more efficient. Then the measures are sorted and ranked with database sentences, here ranking is done based on relevance and then the corresponding words are considered as candidate words (W). W are listed which is result element of the RD.


By using the semantic measures we can improve the relevant measure score. We have to measure the scalability property and compared with existing RD’s. For RD, the performance measure execution time of based on word-to-word semantic algorithms are calculated.
In each algorithm, the execution time is large for sample of 200 glosses. The original dataset contains one lakh words the execution time is upto three hours this time.
When user sentence (S) is compared with considered all database sentences purpose we can introduce the clustering concept. We create a cluster based on semantic sentence similarities in the dictionary database. This leads to reduction of database size and computation of similarity measures and time is minimized by introducing the classification.

Measure of RD

Precision: the percentage of retrieved documents that is in fact relevant to the query (i.e., “correct” responses)
Recall: the percentage of documents that are relevant to the query and were, in fact, retrieved.
The following graph shows the precision and recall score.


Thus RD is effectively created with the idea of semantic similarity. Sentence similarity is achieved by finding word-to-word algorithms.


We have an idea of introducing the wild card characters in user query and eager to develop for bilingual languages.

Tables at a glance

Table icon Table icon
Table 1 Table 2

Figures at a glance

Figure 1
Figure 1