A framework on Group Data Sharing Based Agreement Implementation in Cloud Computing

Cloud computing and cloud storage have become hot topics in recent decades. Both are changing the way we live and greatly improving production efficiency in some areas. At present, due to limited storage resources and the requirement for convenient access, we prefer to store all types of data in cloud servers, which is also a good option for companies and organizations to avoid the overhead of deploying and maintaining equipment when data are stored locally. The cloud server provides an open and convenient storage platform for individuals and organizations, but it also introduces security problems. Due to the particularity of some industries, some enterprises have extremely high requirements on the security of data. If the data is lost or destroyed, the enterprise will bear serious consequences. The second problem is reliability. Only by ensuring that the data records are accurate and reliable can the data transmission be more secure and the enterprise win a higher reputation. In addition, the issue of cloud computing regulation is also very important. Only through efficient and orderly regulation can enterprises operate healthily and normally. At present, many small and medium sized enterprises need to spend huge costs on data storage, which will continue to increase with the expansion of data volume. However, the emergence of cloud storage services has brought the dawn for these enterprises, they will need to store the data transmission to the cloud, through the cloud storage service provider to solve this problem, so that they can save the cost of their own data center, use low price to achieve the purpose of data storage.


INTRODUCTION
The main process includes inputting the dongle after the computer is running and then verifying the media.If the conditions are met, the media will be further loaded, and if the media verification is not passed, the program will be forbidden.At this time, the entire system cannot continue to run, and the judgment efficiency is high.After the loading medium is passed, the next step is to verify the certificate file and then verify the user' name and password.After these subroutines are verified, the program can be successfully opened, and if any intermediate link fails to pass the verification before starting the program, then the safety protection system will all recognize it as a failure and return to the prohibition of loading the program again.It can be seen that the authentication of login authority plays a vital role in the security protection of the entire computer system, and the security protection level and overall operation efficiency are relatively high.

A. Process Security Protection
Process security protection belongs to the application layer of the computer multimedia security protection system.Its existence can control the process of the computer multimedia security protection system to start, terminate, verify external information, exclude untrusted processes, and add trusted process.In the actual running process, when the process security protection module is opened, if it is judged that it is a safe condition at this time, the process will be further created, and if the process security protection module fails to open or the process creation fails, the program will be directly terminated.After the program is created successfully, the computer will further execute the underlying driver and verify the underlying process information after success, open the process after passing it, send the program to the application layer after failure, and use the query open program mode to operate the entire running process.Only the program that trusts the current operation is executed in the open program, and the untrusted operation is added to the blacklist.After the trusted program is successfully run, it can be further added to the trusted program to speed up the efficiency of the next operation.During the entire running process of the security protection module, the entire running program will end whether it is the failure of the underlying driver or the failure of the program to be sent to the program.

B. Document Security Protection
Document security protection is also subordinate to the application layer of the computer multimedia security protection system.Its existence can ensure the integrity and confidentiality of data files.During the operation process, it plays a role in starting the program, loading the driver, reading the document, and controlling the document application effect.Since the current computer system generally uses the drive system provided by Windows officially, although this drive system is stronger than the application programming interface in protection and safety, it has greater difficulties in specific technical implementation.

C. Mobile Media Security Protection
The mobile media security protection also belongs to the application layer of the computer multimedia security protection system, and its existence can realize the functions of computer reading and writing control and information transmission.In the process of daily work, life, and study, mobile media are ubiquitous, and the security protection of such mobile media with communication and connection attributes is considered to be a module that needs to be strengthened.From the perspective of past mobile media security protection strategies, the overall structure should be mainly carried out from the steps of monitoring the system before use, preventing during use, and eliminating hidden dangers after use D. Objectives

•
To share data securely and efficiently between many participants in cloud environment, with help of key generation.To achieve the same key agreement protocols is used along with the encryption and decryption technique.Same key is shared with the participants and to avoid malicious attack the fault tolerance property is used to deliver secure data.

•
To provide security to all these components and interaction of these components with each other needs to be addressed.

•
To improve the quality of service delivered on the network is another concern.

•
To control loss over physical, Logical of system, and alternative control to client's assets, mismanagement of assets Are some additional concerns.Setting up the data center for cloud is expensive and running the infrastructure needs expertise as well as a lot of resources such as high power for cooling, redundant power for assured availability, etc.For example, 45% of the data center cost goes to the acquisition of servers, 25% goes to specialized infrastructure for fault tolerance, redundant power, cooling systems, and backup batteries, while electrical cost consumed by the machines accounts for 15% of the amortized total cost.In addition to the vast number of servers used in data centers, there are billions of Personal Computers (PCs) owned by individuals and organizations worldwide.[3] Zhao Tianhai et al. proposed that Cloud computing is the new research and cooperation pattern for science computing.Keahey, et al. proposed one of the first cloud-based infrastructures for computational science, Science Cloud [1].Software-as-a-service [2] is at the top end of the cloud computing stack, which is seen as a replacement to traditional software.With SAAS, the cloud operator provides end users with an integrated service which comprising hardware, software and development platform.The resource management is the key of application software service in science cloud computing.The optimization decomposition approach to solve cloud resource allocation for satisfying the cloud user's needs and the profits of the cloud providers [3].A new resource management framework presented provides efficient green enhancements within a scalable cloud computing architecture [4].Wang Xiaoyu et al. proposed that Cloud computing has many definitions, but they are all pretty much the same.So called cloud computing, is composed of distributed processing, parallel processing and grid computing, development of a dynamic, good extension, virtualization technology is used to establish a unified infrastructure, services, applications and information resources pool, with distributed technology to the resource pool of infrastructure to effectively organize and run a computing model.Cloud computing as a new IT resources service mode, is composed of a large number of computer resources Shared IT resources pool, users have to use the data will be uploaded to the data resources pool, we can see IT as a huge data centres, again after the data integration to provide users with computing and storage services, to a great extent, improve the efficiency of resource use.After the cloud users upload their data to the cloud, the management of the data will no longer be controlled by the cloud users.At this time, the user data is in an uncontrollable domain, which makes the users lack sufficient trust in the cloud service provider, and of course, they do not want their data to be seen by others.[5] Yanhong Shang1 et al.states that the outbreak of various information data leakage incidents have made computer network users increasingly demanding personal information security.The construction of a computer network information security environment can not only effectively ensure information security but also achieve greater effective protection of information security.This requires that the traditional information security protection measures be used as the basis to continuously increase the construction of information security.[6]

MULTIDIMENSIONAL DATA SECURITY PROTECTION SYSTEM A. Cloud Computing Services
Cloud services are infrastructure, platforms, or software that are hosted by third-party providers and made available to users through the internet.Cloud services facilitate the flow of user data from frontend clients (e.g., users' servers, tablets, desktops, laptops-anything on the users' ends), through the internet, to the provider's systems, and back.Cloud services promote the building of cloud-native applications and the flexibility of working in the cloud.Users can access cloud services with nothing more than a computer, operating system, and internet connectivity.Cloud computing has risen massively in terms of popularity in recent times.This is due to the way it reduces on-premise infrastructure cost and improves efficiency.Primarily, the cloud model has been divided into three major service categories: • Platform as a Service (PaaS) Platform as a service refers to the cloud computing platform as a service, which provides the application running environment for users to meet their needs for the development environment.PaaS can integrate various resources, calculate the resource requirements according to the business requirements of customers, invoke the corresponding hardware resources through the interface provided by IasS, monitor the utilization of resources, and provide services to SaaS through the relevant interface.When using the service, users only need to use the API provided by the platform to easily develop and deploy applications without the need for cloud infrastructure management and control.
• Software as a Service (SaaS) SaaS is a mode in which application software is uniformly installed on various cloud platforms and users order application software services through the Internet.Users can use the whole software or only some functions of a certain software.The cloud service provider is responsible for the hardware facilities, management and maintenance of the software, so that users can use the software conveniently and quickly at any time and place through the Internet.Under the SaaS service mode, customers only need to pay for the corresponding services, which greatly reduces the operating cost of the enterprise, which is also the most efficient operation mode of network applications.
• Infrastructure as a Service (IaaS) Infrastructure as a Service often provides the infrastructure such as servers, virtual machines, networks, operating system, storage, and much more on a pay-as-you-use basis.IaaS providers offer VM from small to extra-large machines.Infrastructure-as-a-Service, commonly referred to as simply "IaaS," is a form of cloud computing that delivers fundamental compute, network, and storage resources to consumers on-demand, over the internet, and on a pay-as-you-go basis.IaaS enables end users to scale and shrink resources on an as-needed basis, reducing the need for high, up-front capital expenditures or unnecessary "owned" infrastructure, especially in the case of "spiky" workloads.

B. Hash Tree Technology and HDFS Encryption Algorithm
The system gives a binary hash tree (tiger hash tree commonly used two binary hash tree, but also the form).It is often used in some distributed hash tree in distributed storage system or the anti-entropy mechanism (Antientropy), also called to entropy [11][12][13].These applications include Amazon Dynamo and Apache Cassandra database, through to the entropy can do each synchronization of different nodes, each node keeps abreast of the latest information is.The characteristics of the hash tree is very clear: the leaf node is stored in the data file, and the nonleaf node is stored in the hash value of its child nodes (called Message Digest) of the non-leaf nodes of the Hash is called the path hash value, Hash value of leaf nodes is the real data.

PROPOSED SYSTEM
• Through block design-based key agreement protocol that supports multiple participants, which could flexibly extend the quantity of participants in associate extremely cloud setting in step with the structure of the block vogue.• Supported the projected cluster info sharing model, we've got an inclination to gift general formulas for generating the common conference key K for multiple participants.Note that by creating the foremost of the (v; k + 1; 1)-block vogue, the procedure quality of the projected protocol linearly can increase with the quantity of participants and thus the communication quality is greatly reduced.In addition, the fault tolerance property of our protocol permits the cluster info sharing in cloud computing to set about to all totally different key attacks.• A key agreement protocol is used to return up with a customary conference key for multiple participants to create positive the security of their later communications, and this protocol is applied in cloud computing to support secure and economical info sharing.

A. Architecture including HDFS Client
Fig. 1 Architecture of HDFS with algorithm process HDFS is an open source, distributed and extensible file system.It is one of the core subprojects of Hadoop project and is written and implemented in pure java language.Because of its lack of POSIX compatibility and its inability to mount, HDFS provides shell commands and Java API interfaces similar to other file systems.The Hadoop cluster nominally has a single data node and a data node cluster, although due to its criticality, redundancy options are available for name node.Each data node can provide data blocks over the Internet and using HDFS specific block protocols.The HDFS file system uses TCP / IP sockets for communication clients use remote procedure calls (RPCs) to communicate with each other.Figure 4.1 shows the overall architecture of HDFS.It can be seen that HDFS is an extensible distributed file system with master-slave structure.Many computers or servers that make up HDFS are called clusters.In the HDFS cluster, there is a name node node, which is the main server responsible for maintaining the file system namespace and regulating and controlling the client's access to the file system.In addition, there are many data node nodes, which are usually each node in the HDFS cluster, responsible for managing the data storage attached to the nodes they run.The architecture provides a namespace without reservation.Users can use the namespace to store their own private or non-confidential information in the file.All the attribute information related to the file is maintained by name node.The data stored in HDFS is not the original complete single file, but needs to take some measures to cut these large data into one block, or many blocks; then the small block with certain original information is stored in the specified data node or multiple slave nodes.The main function of name node is to execute a series of namespace related instructions, such as file open, file close and file rename.In addition, it can determine the mapping relationship between each small data block and the corresponding data node.The main function of the corresponding large number of data nodes is to receive the requests from the client to read and write data.At the same time, it can also perform the operations related to the data block according to the instructions given by the name node.

CONCLUSION AND FUTURE SCOPE A. Conclusion
In this way the data sharing in cloud computing enables multiple participants to freely share the group data, which improves the efficiency of work in cooperative environments and has widespread potential applications.It also tries to ensure the security of data sharing within a group and how to efficiently share the outsourced data in a group manner are formidable challenges.Note that key agreement protocols have played a very important role in secure and efficient group data sharing in cloud computing.The main advantage of the symmetric balanced incomplete block design (SBIBD), by presenting a novel block design-based key agreement protocol that supports multiple participants, which can flexibly extend the number of participants in a cloud environment according to the structure of the block design was apparently design and help to provide security more than the previous system.The proposed group data sharing model, we present general formulas for generating the common conference key K for multiple participants.And the protocol linearly increases with the number of participants and the communication complexity is greatly reduced.In addition, the fault tolerance property of our protocol enables the group data sharing in cloud computing to withstand different key attacks.

B. Future work
Aiming at the security problems of the internal personnel of cloud service providers that are easy to be ignored, this paper puts forward the identity authentication and role-based access control strategies based on account and certificate, analyzes and studies the cloud security standards and legal maintenance, and puts forward the cloud security assessment system and relevant legal suggestions and measures to provide a strong guarantee for data security protection.

Jian
Shen et al.  proposed an efficient and secure block design-based key agreement protocol by extending the structure of the SBIBD to support multiple participants, which enables multiple data owners to freely share the outsourced data with high security and efficiency.Note that the SBIBD is constructed as the group data sharing model to support group data sharing in cloud computing.Moreover, the protocol can provide authentication services and a fault tolerance property.[1] Mehdi Bahrami et al. proposed fields to establish a virtual IT department via the Internet.The cloud computing offers different virtual services like traditional IT department, such as storage, stream server and database server.The cloud provides a cost effective model through pay-per-use that allows each individual or businesses in healthcare start a cloud based service with minimum investment.[2] Tessema Mengistu at al. proposed that The current Cloud Computing services are based on the "data center" approach, where hundreds of thousands of dedicated servers are setup to give the services.

Fig. 4 FigFig. 9
Fig. 4 Data Selection and generation of key