A New Framework to Integrate WSN with Cloud Computingand Security Using Dynamo DB

DOI : 10.17577/IJERTCONV2IS13073

Download Full-Text PDF Cite this Publication

Text Only Version

A New Framework to Integrate WSN with Cloud Computingand Security Using Dynamo DB

Lalleshwari. B. P

Dept. of CSE

SJBIT, BANGALORE-560060 KARNATAKA, INDIA

Lalleswari2812@gmail.com

Sunitha. K , asst. prof

Dept. of CSE

SJBIT, BANGALORE-560060 KARNATAKA, INDIA

ksunitha@sjbit.edu.in

Abstract-Wireless sensors networks have severalapplications of their own. These applications can furtherenhanced by integrating a local wireless sensor network tointernet, which can be used in real time applications wherethe results of sensors are stored on the cloud. We proposean architecture that integrates a wireless sensor networkto the internet using cloud technology. The resultantsystem is proved to be reliable, available and extensible. Inthis paper a new framework is proposed for WSNintegration with Cloud computing model, existing WSNwill be connected to the proposed framework. Threedeployment layer are used to serve user request ( IaaS,PaaS, SaaS) either from the library which is made fromdata collected from data centric DC by WSN periodically.The integration controller unit of the proposed frameworkintegrates the sensor network and cloud computingtechnology which offers reliability, availability andextensibility.

Keywords-Saas,Paas,Iaas,Dynamo DB

  1. INTRODUCTION

    The idea of wireless networks was developed by sendingmessages through fire and smoke. It was a sign of presence ofpeople in specific territory. In 80's researches came up withthe idea of TNC (Terminal Node Controller). The aim of thisinvention was to communicate between a computer and radio,but desired performance was not achieved. Later on in 1985FCC (Federal Communication Commission) allocated ISMbands for different applications. The assigned ISM bands were902 MHz and 5.85 GHz. and access was limitedto licensed users. Wireless Sensor Network (WSN) consists ofand wireless communications. The nodes are connected in adhocfashion, operating independently of other nodes. There arevarious factors associated with the performance of WSN suchpower management, data dissemination and routing ofinformation. A lot of work has been undergoing in these areaswhere energy awareness of essential design issue; routing anddata dissemination is application dependent. WSN architecturecould be either centralized or distributed. In centralizedarchitecture the central node is the weak point of the network.If it fails, whole network collapse. However, distributedarchitecture provides failure resistant sensor network [1].Comprehensive architecture and various wireless sensorapplications emergence is represented and discussed as we can

    see in the Figure 1.distributed nodes with the capability of sensing, computation

    Figure 1: Emergence of Wireless Sensor Network in

    Various Applications

    1. Issues in Wireless Sensor Network

      Ad hoc network inherited some of the tribulations of wireless communication networks like lossy links, unreliable time in varying and asymmetric channel, hidden node and exposed node and improperly defmed coverage boundary. Ad hoc network puts a messy contribution in this knotted portfolio in the form of multi hop environment, location awareness, and node mobility, dynamically changing topology, vulnerability of channel and nodes and different aspects in QoS. Specifically in WSN, in addition to most of the aforementioned issues, network partitioning, localization, calibration, data fusion, aggregation and dissemination, coverage issues, self- organizing and self-administration, scalability, load balancing, node clustering, topology management, end-to-end delay constraint routing, security and privacy, heterogeneity, and other energy, memory, power and bandwidth constraints are the active challenges. In the closer view, node scheduling, whole problem, avoiding and coping with void node areas, node failure and QoS relating factors are under great concentration of researchers where QoS is a level of service in

      achieving the target with sufficient resources byfulfilling the requested QoS parameters. End-to-Endperformance, delay, bandwidth, energy consumption,transmission power, memory usage, probability of packet loss,jitter, bit error rate, miss ratio, packet over-head and packetsuccess delivery are considered under the umbrella of QoSparameters. Real-time routing has the core importance in QoSawarenetwork layer issues which ensures catching thedestination within a limited required time and resources.

      Table 1: Wireless Sensor Network Applications

  2. CLOUD COMPUTING

    Cloud Computing is internet based computing, where resources allocations are shared, software and information areprovided to computer and all other devices on demand as requested. It is considered as an alternative to traditional server or web hosting servers. Cloud provides much more services then that, having different layer to provide application based services. [2].

    Cloud Computing

    Figure 2: Overview of Cloud Computing

    IBM report stated that "Cloud is a new consumption and deliver model for many IT-based services, in which the user

    sees only the service, and has no need to know anything about the technology or implementation" [10]. NIST classify the cloud computing as "Cloud computing is a model for enabling convenient, on demand network access to shared pool of conjigurable computing resources" [11].

    Cloud Computing Architecture Layers

      1. SaaS (Software as a Service)

      2. PaaS (platfonn as a Service)

      3. IaaS (Infrastructure as a Service ).

        General overview of cloud computing architecture layers as we can see in the Figure 3.

        Figure 3: Cloud Computing Architecture

        SaaS provides board market solutions where the vendor provides access to hardware and software products through portal interface [12].

        PaaS allows the consumer to run the specified application on the platform. In these types of services, consumer has no control over the infrastructure as well as the installed applications [13], [14]. IaaS provides consumers with the benefit to consume the infrastructure that includes processing power, data storage, and network etc.

  3. LITERATURE SURVEY

    The Cloud computing, coined in late of 2007, currently emerges as a hot topic due to its abilities to offer flexible dynamic IT infrastructures, QoS guaranteed computing environments and configurable software services. Cloud computing can be defined as follows: A Cloud is a type of parallel and distributed system consisting of a collection of interconnected and virtualized computers that are dynamically provisioned and presented as one or more unified computing resources based on service-level agreements established through negotiation between the service provider and customers and can be ubiquitously accessed from any connected devices over the internet Cloud computing started quietly from several seeding technologies such as grid computing, virtualization, SalesForce.com innovative

    subscription-based business model or Amazons effort to scale their e-commerce platform. However, it differs from traditional ones in that:

        1. it is massively scalable,

        2. can be encapsulated as an abstract entity that delivers different levels of services to customers anywhere, anytime, and

        3. it is driven by economies of scale that is the services can be dynamically.

    Configured (via virtualization or other approaches) and delivered on-demand. The Web search popularity, as measured by the Google search trens during the last 12 months, for terms Cluster computing, Grid computing, and Cloud computing is shown in Figure 1. From the Google trends, it can be observed that cluster computing was a popular term during 1990s, from early 2000 Grid computing become popular, and recently Cloud computing started gaining popularity. Meanwhile, market-research firm IDC expects IT Cloud-services spending to grow from about $16 billion in 2008 to about $42 billion by 2012 as Figure 2 shows. IDC also predicts Cloud computing spending will account for 25 percent of annual IT expendituregrowth by 2012 and nearly a third of the growth the following year.

    Cloud Computing has many benefits that the public sector and government IT organizations are certain to want to take advantage of. In very brief summary form they are as follows:

    1. Reduced cost, higher gains:

      Cloud technology is paid incrementally, saving organizations money.

    2. Increased storage:

      Organizations can store more data than on private computer systems.

    3. Highly automated:

    No longer do IT personnel need to worry about keeping software up to date.

    D .Flexibility:

    Cloud computing offers much more flexibility than past computing methods.

    1. More mobility:

      Employees can access information wherever they are, rather than having to remain at their desks. In Cloud computing, customers do not own the infrastructure they are using; they basically rent it, or pay as they use it. One of the major selling points of cloud computing is

    2. Lower costs:

      Companies will have lower technology-based capital expenditures, which should enable companies to focus their money on delivering the goods and services that they specialize in. There will be more device and location independence, enabling users to access systems no matter where they are located or what kind of device they are using. The sharing of

      costs and resources amongst so many users will also allow for efficiencies and cost savings around things like performance, load balancing, and even locations (locating data centers and infrastructure in areas with lower real estate costs, for example). The general architecture of Cloud computing is shown below.

    3. No Existing Infrastructure for Integration of WSN to Cloud:

    In the past few years, wireless sensor networks (WSNs) have been gaining increasing attention to create decision making capabilities and alert mechanisms, in many Life care application areas including Life care monitoring for patients, environmental monitoring, pollution control, disaster recovery, military surveillance etc. For example, MIT wireless sensor ring can measure heart rate, heart rate variability, Oxygen saturation and blood pressure for the person wearing the ring.

    Our proposed Dynamo DB can be deployed for various u-Life care services and WSN services.

  4. PROPOSED FRAME WORK OVERVIEW Figure 5 shows the proposed integration framework of

    cloud computing and WSN. The major components of the Publisher/Subscriber Broker, Request Subscriber, Identity and Access Management Unit (lAMU) and Data Repository (DR).The data gathered from WSN is passed through gateway to DPU, which process data and add it to DR. In order to access the stored data from cloud services, user connects through secured IAMU; on successful connection establislunent user will be given the access according to the account policies. User data request is forwarded to RS which creates a request subscription and forward the subscription to the Pub/Sub Broker. When DPU receives the data from gateway, it forwards the data to Pub/Sub Broker. When the event matches the subscription, data is made available to the respective user. There has to be effective way for user to access the data sensed by the sensors. The idea of connecting a standalone wireless network to internet is comparatively new and still is an open research area. The sensor networks are known to be pervasive and ubiquitous. It is impossible to connect each and every node to internet, as we are constrained by the IP addresses. This article proposes architecture as a solution where industrial sensor networks can be integrated with internet through cloud technology and Service Oriented Architecture (SOA). The system proposed main modules as application server, integrated controller, and a register agent. The role of integrated controller is to provide the storage as well as the recovery mechanism to the sensed data. This has been achieved by uploading the sensed data to internet using cloud services. The user can access the data from any location in the world. The sensor networks are ideally considered to be energy efficient and it's the major criticality of the network that must be answered. In addition to that, short range hop communications is preferred in order to communicate with a long range destination. Therefore, the information from source is distributed across intermediate nodes in the path towards destination node. The intermediate nodes must process the received data, aggregate it to remove duplication and to minimize the network traffic.

    1. Access Control Enforcement Unit:

      ACEU is used to authenticate the user and it is consists of EN and three servers i.e. AS, TGS and SS. The request received by EN is sent to AS. EN implements Kerberos in order to authenticate the user with AS.

    2. Access Control Decision Unit :

      ACDU is used to enforce the policy rules. It consists ofRBAC processor and policy storage. It communicates with ACEU through SS. After successful authentication; user is given the access to the resources as constrained by the access policies.

    3. Communication flow between User andIAMU :

    The description of different messages those have been exchanged among different servers and edge node (EN) are left out of the scope of this poster due to the space constraint.

    .

    Figure 5: Proposed Architecture General Overview

  5. DYNAMO DB

    Dynamo DB is a fast, fully managed Nosily database service that makes it simple and cost-effective to store and retrieve any amount of data, and serve any level of request traffic. Its guaranteed throughput and single-digit millisecond latency make it a great fit for gaming, ad tech, mobile and many other applications.

    1. Features and benefits:

      Dynamo DB delivers seamless throughput and storage scaling via API and easy-to-use management console, so you can easily scale up or down to meet your needs. Many of our customers have, with the click of a button, created Dynamo DB

      deployments in a matter of minutes that are able to serve trillions of database requests per year.

      Dynamo DB tables do not have fixed schemas, and each item may have a different number of attributes. Multiple data types add richness to the data model. Secondary indexes add flexibility to the queries you can perform, without impacting performance.

      Performance, reliability and security are built-in, with SSD-storage and automatic 3-way replication. Amazon Dynamo DB uses proven cryptographic methods to securely authenticate users and prevent unauthorized data access.

    2. Scalable:

      Amazon Dynamo DB is designed for seamless throughput and storage scaling.

    3. Automated Storage Scaling:

      There is no limit to the amount of data you can store in a Dynamo DB table, and the service automatically allocates more storage, as you store more data using the Dynamo DB write APIs.

    4. Provisioned Throughput:

      When creating a table, simply specify how much request capacity you require. Dynamo DB allocates dedicated resources to your table to meet your performance requirements, and automatically partitions data over a sufficient number of servers to meet your request capacity. If your throughput requirements change, simply update your table's request capacity using the AWS Management Console or the Amazon Dynamo DB APIs. You are still able t achieve your prior throughput levels while scaling is underway.

    5. Distributed, Shared Nothing Architecture:

      Amazon Dynamo DB scales horizontally and can seamlessly scale a single table over hundreds of servers.

    6. Easy Administration:

      Amazon Dynamo DB is a fully managed service you simply create a database table and let the service handle the rest. You dont need to worry about hardware or software provisioning, setup and configuration, software patching, operating a reliable, distributed database cluster, or partitioning data over multiple instances as you scale.

    7. Flexible:

    Amazon Dynamo DB gives you the flexibility to query on any attribute using secondary indexes which have the same performance, scalability, and fault-tolerance characteristics as that of a Dynamo DB table.

  6. SECURITY

The data owner and all of the cloud servers in the cloud Share a synchronized clock, and there is no transmission Andqueuing delays when executing read and write Commands.

  1. Intuition

    The data owner will first generate a shared secret key to theCSP. Then, after the data owner encrypts each file with theappropriate attribute structure and time slice, the data owneruploads the file in the cloud. The CSP will replicate the fileto various cloud servers. Each cloud server will have a copyof the shared secret key.Let us assume that a cloud server stores an encrypted file

    F with A and TSi. When a user queries that cloud server, thecloud server first uses its own clock to determine the currenttime slice. Assuming that the current time slice is TSi+k,the cloud server will automatically re-encrypt F with TSi+kwithout receiving any command from the data owner. Duringthe process, the cloud server cannot gain the contents of thecipertext and the new decryption keys. Only users with keyssatisfying A and TSi+k will be able to decrypt F.

  2. Protocol Description

    We divide the description of the basic R3 scheme into threecomponents: data owner initialization, data user read data anddata owner write data. We will rely on the following functions.Table II shows the notations used in the description.

    1. Setup() (PK,MK, s) : At TS0, the data ownerpublishes the system public key PK, keeps the system

      Algorithm 1 Basic R3 (synchronized clock with no delays) while Receive a write command W(F, seqnum) at TSi does Commit the write command in order at the end of TSi while Receive a read command R(F) at TSi do

      Re-encrypt file with TSi

      master key MK secret, and sends the shared secret key s to the cloud.

    2. GenKey(PK,MK, s,PKAlice,A, T ) (SKAlice, {SKT Alice,A}) : When the data owner wants to grant data user Alice attributes A with valid time period T , the data owner generates SKAlice and {SKT Alice,A} using the system public key, the system master key, the shared secret key, Alices public key, Alices attributes and eligible time.

    3. Encrypt(PK,A, s, TSt, F) (CtA ) : At TSt, the data owner encrypts file F with access structure A, and produces ciphertext CtA using the system public key, access structure, the system secret key, time slice, and plaintext file

    4. Decrypt(PK,CtA , SKAlice, {SKt Alice,aij }1jni ) F : At TSt, user U, who possesses version t attribute secret keys on all attributes in CCi, recovers F using the system public key, the user identity secret key, and the user attribute secret keys.

    5. REncrypt(CtA , s, TSt+k) Ct+k A : When the cloud server wants to return a data user with the file at TSt+k, it updates the ciphertext from CtA to Ct+k A using the shared secret key

      CONCLUSION

      Both wireless sensor network and Cloud Computing technologies along with their applications are discussed in this paper. We gave an overview of architectural extension to wireless sensor network. Application of cloud computing to enhance the reliability and availability of wireless sensor networks is discussed with special emphasis on its application in distributed manufacturing engineering. The proposed system is based on ideas taken from an in depth study and support of various technologies. The proposed system has its useful applications and important role in medical sciences field. It is supposed to help in efficient cure of Strokes and Parkinson. However the security issues involved in the integration process are of key importance and need critical focus. Further efforts will help extend the applications to military and manufacturing services.

      REFERENCES

      1. Wendi B. Heinzelman, Amy 1. Murphy, Hervaldo S. Carvalho, Mark A.Perillo; "Middleware to Support Sensor Network Applications"; IEEE Network, pp. 6-14, January/February 2004.

      2. F. Schepers. (2010) Security in Cloud Computing, IBM Tivoli InternetSecurity Systems.[Online]. Available: http://www.cpdpconferences.orgiResources/Schepers.pdf. Last accessed:10/1112010.

      3. P. McDaniel, and S. W. Smith, "Outlook: Cloud Computing with aChance of Security Challenges and Improvements," IEEE Computer andReliability Socities 2010, pp. 77-80, Jan. 2010.

      4. P. Barham, B. Dragovic, K. Fraser, S. Hand, T. Harris, A. Ho, R.Neugebauer, I. Pratt, A. Warfield, "Xen and the Art of Virtualization,"in Proc. of 19th ACM symposium on Operating Systems Principles,Bolton Landing, NY, USA, October 2003, pp. 164-I 77.

      5. (2010) Google App Engine. [Online]. Available: http://code.google.com/appengine/. Last Accessed: 15/07/2011

      6. (2007) Sales Force. [Online]. Available: http://www.salesforce.com/platform/. Last Accessed: 10/1112010

      7. A. Dubey, and D. Wagle. (2007) Delivering software as a service TheMcKinsey Quarterly.[Online]. Available:

        http://www.mckinsey.de/downloads/publ ikation/mck on btl2007/mobt_Delivering_Software_as_a_Service.pdf Last Accessed: 15/08/2011.

      8. D. Harkins and D. Carrel. (1998) The Internet Key Exchange (IKE), RFC 2409, IETF Network Working Group. [Online]. Available:http://www.ietf.orglrfc/rfc2409.txt. Last Access: 12/01/20 II. . Heidemann, et aI., Building efficient wireless sensor networks with low- level naming, in: Proc. Symposium on Operating Systems Principles, Chateau Lake Louise, Banff, Alberta, Canada, ACM. (Oct.,2001) pp. 146–1 59.Availablein: http://www.isi.edu/johnh/PAPERS/HeidemannOI.html.

      9. Nils Hoeller, ChristophReinke, Jana Neumann, Sven Groppe, DanielBoeckmann,VolkerLinnemann,"Efficient XML Usage within WirelessSensor Networks" WICON '08, November 17-I 9, 2008, Maui,Hawaii,USA.

      10. LimHocBeng, "Sensor cloud:towards sensor-enabled cloud services"Intelligent systems center, Nanyang Technological University, 13 April2009.

[II] P. McDaniel, and S. W. Smith, "Outlook: Cloud Computing with aChance of Security Challenges and Improvements," IEEE Computer and Reliability Socities 2010, pp. 77-80, Jan. 2010.

  1. P. Barham, B. Dragovic, K. Fraser, S. Hand, T. Harris, A. Ho, R.Neugebauer, I. Pratt, A. Warfield, "Xen and the Art of Virtualization,"in Proc. of 19th ACM symposium on Operating Systems PrinciplesBolton Landing, NY, USA, October 2003, pp. 164-I 77.

  2. (2010) Google App Engine. [Online]. Available: http://code.google.com/appengine/. Last Accessed: 15/07/2011

  3. (2007) Sales Force. [Online]. Available: http://www.salesforce.com/platform/. Last Accessed: 10/1112010

Leave a Reply