ISSN ONLINE(2320-9801) PRINT (2320-9798)
K.Deepika1, N.Naveen Prasad2, Prof.S.Balamurugan3, S.Charanyaa4
|
Related article at Pubmed, Scholar Google |
Visit for more related articles at International Journal of Innovative Research in Computer and Communication Engineering
This paper reviews methods developed for anonymizing data from 2009 to 2010. Publishing microdata such as census or patient data for extensive research and other purposes is an important problem area being focused by government agencies and other social associations. The traditional approach identified through literature survey reveals that the approach of eliminating uniquely identifying fields such as social security number from microdata, still results in disclosure of sensitive data, k-anonymization optimization algorithm ,seems to be promising and powerful in certain cases ,still carrying the restrictions that optimized k-anonymity are NP-hard, thereby leading to severe computational challenges. k-anonimity faces the problem of homogeneity attack and background knowledge attack . The notion of ldiversity proposed in the literature to address this issue also poses a number of constraints , as it proved to be inefficient to prevent attribute disclosure (skewness attack and similarity attack), l-diversity is difficult to achieve and may not provide sufficient privacy protection against sensitive attribute across equivalence class can substantially improve the privacy as against information disclosure limitation techniques such as sampling cell suppression rounding and data swapping and pertubertation. This paper aims to discuss efficient anonymization approach that requires partitioning of microdata equivalence classes and by minimizing closeness by kernel smoothing and determining ether move distances by controlling the distribution pattern of sensitive attribute in a microdata and also maintaining diversity.
Keywords |
Data Anonymization, Microdata, k-anonymity, Identity Disclosure, Attribute Disclosure, Diversity |
INTRODUCTION |
Need for publishing sensitive data to public has grown extravagantly during recent years. Though publishing demands its need there is a restriction that published social network data should not disclose private information of individuals. Hence protecting privacy of individuals and ensuring utility of social networ data as well becomes a challenging and interesting research topic. Considering a graphical model [35] where the vertex indicates a sensitive label algorithms could be developed to publish the non-tabular data without compromising privacy of individuals. Though the data is represented in graphical model after KDLD sequence generation [35] the data is susceptible to several attacks such as homogeneity attack, background knowledge attack, similarity attacks and many more. In this paper we have made an investigation on the attacks and possible solutions proposed in literature and efficiency of the same. |
THE EUCALYPTUS OPEN SOURCE CLOUD COMPUTING SYSTEM[2009] |
Daniel nurmi, Rich wolski, Chris grzegorczyk, Graziano obertelli, sunil soman, lamia youseff, Dmitrii zagorodnov Generally ,cloud computing systems fundamentally provide access to large pool of data and computational resources through a variety of interfaces.Those interfaces are similar to existing grid and HPC resources management and programming systems. Today,the most cloud computing systems are perfectly rely upon infrastructure. That infrastructure is invisible to the research community.In this paper,the authors presented the EUCALYPTUS which is an open source software framework for cloud computing. This open source software framework implements infrastructure as a service(IaaS). The architecture of the EUCALYPTUS system is simple,flexible and modular with a hierarchical design reflecting common resource environments found in many academic settings. The authors described the four high level components,each with its own web service interface,that comprise a EUCALYPTUS installation . It includes node controller, cluster controller, storage controller and cloud controller. Node controller executes on every node that is designated for hosting VM instances. Each and every node controller makes queries to discover the node’s physical resources like the number of cores,the size of the memory,the available disk space and also to learn about the state of virtual machine instances on the node. And next the author says that clster node generally executes on a cluster front end machine, or any machine that has network connectivity to both the nodes running NC’s and to the machine running the cloud controller. Many of the cloud controller operations are similar to the NC’s operations but are generally plural instead of singular(eg: run instances,terminate instances). Basically the cloud controller calculates how many simultaneous instances of the specific “type” can execute on its collection of NC’s and reports that number back to the CLC. VM instance interconnectivity is one of the most interesting challenges in the design of cloud computing infrastructure. While designing EUCALYPTUS, the authors recognized that the VM instance network solution must address connectivity isolation ad performance. This EUCALYPTUS design attempts to maintain inter-VM network performance as close to native as possible. |
Within EUCALYPTUS, the CC currently handles the three modes . the first configuration instructs the system to attach the VM’s interface directly to a software ethernet bridge connected to a real physical machines network. The second configuration allows an admin to define static media access control(MAC) and IP address tuples. In this mode each new instance created by the system is assigned a free MAC/IP tuple, which is released when the instance is terminated. In these modes, the performance of inter-VM communication is near native, when VM’s are running on the same cluster but there is not inter –VM network isolation. Finally ,this work aims to illustrate the fact that the EUCALYPTUS system has filled on important niche in the cloud computing design space by providing a system that is easy to deploy a top existing resources, that lands itself to experimentation by being modular and open source, and that provides powerful feature out of the box through an interface compatible. The authors provided that they successfully deployed the complete system on resources ranging from a single laptop to small linux clusters. In addition, they had made a installation available to all who wish to tryout the system without installing any software. Hereby, they concluded that their experiences so far has been extremely positive, leading us to the conclusion that EUCALYPTUS is helping to provide the research community with a much needed, open source software framework around which a user base of cloud computing researchers can be developed. |
THE SECURITY OF CLOUD COMPUTING SYSTEM ENABLED BY TRUSTED COMPUTING TECHNOLOGY (2010) |
Generally,the basic networks need security to transmit information authenticately. Cloud computing provides people the way to share distributed resources and services that belong to different organizations or sites.Since distributed systems and network computing were used wildly, security has become an urgent problem and will be more important in the future. In order to improve the work efficiency, the different services are distributed in different servers that are distributed in different places.Users from multiple environment hope use the distributed computing more efficient, just like using the electric power. Then, cloud computing has become a new information for this demand.Cloud computing provides a facility that enable large-scale controlled sharing and interoperation among resources that are dispersedly owned and managed. The authors noted that the security is therefore a major element in any cloud computing infrastructure, because it is necessary to ensure that only authorized access is permitted and secure behavior is accepted.Because the cloud computing is composed of different local systems and includes the members from multiple environments, therefore the security in cloud is complicate. In one side, the security mechanism should provide guarantees secure enough to the user, on the other side, the security mechanism should not be too complex to put the users into an inconvenient situation.The authors proposed a new way that is conducive to improve the secure and dependable computing in cloud. In their design, the authors integrated the Trusted Computing Platform (TCP), which is based on Trusted Platform Module (TPM), into the cloud computing system. The TCP will be used in authentication, confidentiality and integrity in cloud computing environment.The TCP can improve the cloud computing security and will not bring much complexity to users. Because the TCP is based on relatively independent hardware modules, it does not cost too much resource of CPU, and can improve the performance of processing cryptographic computation. The authors also designed a software middleware, the Trusted Platform Support Service (TSS), on which the cloud computing application can use easily the security function of TPM.The authors then discussed about the security model of the cloud computing. In order to achieve security in cloud computing system, some technologies have been used to build the security mechanism for cloud computing. The cloud computing security can be provided as security services. Security messages and secured messages can be transported, understood, and manipulated by standard Web services tools and software. The authors noted that this mechanism is a good choice because the web service technology has been well established in the network-computing environment.The CLOUD includes distributed users and resource |
from distributed local systems or organizes, which have different security policies. According to this reason, how to build a suitable relationship among them is a challenge. In fact, the requirements for the security in cloud computing environment have some aspects, including confidentiality. multiple security policy, dynamic of the services., the trust among the entities, dynamically building trust domains.The authors proposed the mechanism of trusted computing platform and other related functions that aid to achieve the trusted cloud computing, which has a trusted computing environment.The word trust is defined as “A trusted component, operation, or process is one whose behavior is predictable under almost any operating condition and which is highly resistant to subversion by application software, viruses, and a given level of physical interference.” Then the authors concentrated on the trusted computing platform.TCP operates through a combination of software and hardware: manufacturers add some new hardware to each computer to support TC functions, and then a special TC(trusted computing) operating system mediates betweenthe hardware and any TC-enabled applications. TCP provides two basic services, authenticated boot and encryption, which are designed to work together. An authenticated boot service monitors what operating system software is booted on the computer and gives applications a sure way to tell which operating system is running. It does this by adding hardware that keeps a kind of audit log of the boot process.The authors keynoted that the build trusted cloud computing system using TCP.The trusted computing mechanism can provide a way that can help to establish a security environment. The model of trusted computing is originally designed to provide the privacy and trust in the personal platform and the trusted computing platform is the base of the trusted computing. Since the internet computing or network computing has been the main computing from the end of the last century, the model of trusted computing is being developed to the network computing, especially the distributed systems environment. The cloud computing is a promising distributed system model and will act as an important role in the e-business or research environments.The authors specified the Authentication cloud computing environment in TCP.In cloud computing environment, different entities can appeal to join the CLOUD. Then the first step is to prove their identities to the cloud computing system administration. Because cloud computing should involve a large amount of entities, such as users and resources from different sources, the authentication is important and complicated. Considering these, we use the TCP to aid to process the authentication in cloud computing. Then the authors were dealt about the Role based access control model in cloud computing environment. In order to reach the goal of trusted computing, the users should come from the trusted computing platform, and take the security mechanism on this platform to achieve the privacy and security for themselves. The user has his personal ID and secret key, such as the USB Key, to get the right to use the TCP. They can use the decryption function to protect their data and other information. |
The authors described that by using the remote attest function, the user in the TCP could to notify their identities and relevant information to the remote machine that they want to make access to. The authors concluded that they have analyzed the trusted computing in the cloud computing environment and the function of trusted computing platform in cloud computing. The advantages of their proposed approach are to extend the trusted computing technology into the cloud computing environment to achieve the trusted computing requirements for the cloud computing and then fulfill the trusted cloud computing. TCP is used as the hardware base for the cloud computing system. In their design, TCP provides cloud computing system some important security functions, such authentication, communication security and data protection. Related methods for these implementations are also proposed by the authors in this paper. |
CONCLUSION AND FUTURE WORK |
Various methods developed for anonymizing data from 2009 to 2010 is discussed. Publishing microdata such as census or patient data for extensive research and other purposes is an important problem area being focused by government agencies and other social associations. The traditional approach identified through literature survey reveals that the approach of eliminating uniquely identifying fields such as social security number from microdata, still results in disclosure of sensitive data, k-anonymization optimization algorithm ,seems to be promising and powerful in certain cases ,still carrying the restrictions that optimized k-anonymity are NP-hard, thereby leading to severe computational challenges. k-anonimity faces the problem of homogeneity attack and background knowledge attack . The notion of ldiversity proposed in the literature to address this issue also poses a number of constraints , as it proved to be inefficient to prevent attribute disclosure (skewness attack and similarity attack), l-diversity is difficult to achieve and may not provide sufficient privacy protection against sensitive attribute across equivalence class can substantially improve the privacy as against information disclosure limitation techniques such as sampling cell suppression rounding and data swapping and pertubertation. Evolution of Data Anonymization Techniques and Data Disclosure Prevention Techniques are discussed in detail. The application of Data Anonymization Techniques for several spectrum of data such as trajectory data are depicted. This survey would promote a lot of research directions in the area of database anonymization. |
References |
|