Cost Effective Selection of Data Center in Cloud Environment Manoranjan Dash1, Amitav Mahapatra2 & Narayan Ranjan Chakraborty3 1 Institute of Business & Computer Studies, Siksha O Anusandhan University, Bhubaneswar, Odisha 2 College of Engineering and Technology, Bhubaneswar, Odisha 3 Department of CSE, Daffodil International University, Bangaldesh E-mail: manoranjanibcs@gmail.com1 Abstract – Cloud computing today has now been rising as new technologies and new business models. The increasing cloud computing services offer great opportunities for clients to find the maximum service and finest pricing, which however raises new challenges on how to select the best service out of the huge group. Cloud computing employs a variety of computing resources to facilitate the execution of large-scale tasks. Therefore, to select appropriate node for executing a task is able to enhance the performance of largescale cloud computing environment. It is time-consuming for consumers to collect the necessary information and analyze all service providers to make the decision. This is also a highly demanding task from a computational perspective, because the same computations may be conducted repeatedly by multiple consumers who have similar requirements. Load balancing is the process of distributing the load among various nodes of a distributed system to improve both resource utilization and job response time .Load balancing ensures that all the processor in the system. or every node in the network does approximately an equal amount of work at any instant of time. CloudAnalyst is a tool that helps developers to simulate large-scale Cloud applications with the purpose of understanding performance of such applications under various deployment configurations. The simulated results provided in this paper based on the scheduling algorithm Throttled load balancing policy across VM’s in a single data center and is being compared with round robin scheduling algorithm to estimate response time , processing time . cloud computing in determining whether to adopt it as an IT strategy. The availability of advance processors and communication technology has resulted the use of interconnected, multiple hosts instead of single highspeed processor which incurs cloud computing. In case of Cloud computing services can be used from diverse and widespread resources, rather than remote servers or local machines. There is no standard definition of Cloud computing. Generally it consists of a bunch of distributed servers known as masters, providing demanded services and resources to different clients known as clients in a network with scalability and reliability of datacenter. Cloud computing is revolutionizing the way IT resources are managed and provisioned. Cloud computing is on demand service in which shared resources, information, software and other devices are provided according to the client requirement at specific time. Cloud computing is an evolving paradigm with changing definitions, it is defined as a virtual infrastructure which provides shared information and communication technology services, via an internet i.e. cloud. Cloud computing provides a computer user access to Information Technology (IT) services (i.e., applications, servers, data storage) without requiring an understanding of the technology or even ownership of the infrastructure. Cloud Computing is getting advanced day by day. Cloud service providers are willing to provide services using large scale cloud environment with cost effectiveness. Also, there are some popular large scaled applications like social-networking and ecommerce. These applications can benefit to minimize the costs using cloud computing. Cloud computing is modeled to provide service rather than a product. Services like computation, software, data access and storage are provided to its user without its knowledge about physical location and configuration of the server which is providing the services. Cloud works on the principle of virtualization of resources with on-demand and pay-as–you go model policy. . Keywords – Cloud computing, round robin and throttled algorithms, load balancing. I. INTRODUCTION Cloud computing is a distributed computing paradigm that focuses on providing a wide range of users with distributed access to scalable, virtualized hardware and/or software infrastructure over the internet. Despite this technical definition cloud computing is in essence an economic model for a different way to acquire and manage IT resources. An organization needs to weigh cost, benefits and risks of ISSN (Print) : 2319 – 2526, Volume-2, Issue-1, 2013 1 Special Issue of International Journal on Advanced Computer Theory and Engineering (IJACTE) memories for their resources to scale with increased demands. This service is optional and depends on the clients business needs. So load balancing serves two important needs, primarily to promote availability of Cloud resources and secondarily to promote performance. In order to balance the requests of the resources it is important to recognize a few major goals of load balancing algorithms: a) Cost effectiveness: primary aim is to achieve an overall improvement in system performance at a reasonable cost. Fig.1: View of the Cloud Computing Environment b) Scalability and flexibility: the distributed system in which the algorithm is implemented may change in size or topology. So the algorithm must be scalable and flexible enough to allow such changes to be handled easily. II. c) LOAD BALANCING IN CLOUD COMPUTING: Load balancing in clouds is a mechanism that distributes the excess dynamic local workload evenly across all the nodes. It is used to achieve a high user satisfaction and resource utilization ratio, making sure that no single node is overwhelmed, hence improving the overall performance of the system. Proper load balancing can help in utilizing the available resources optimally, thereby minimizing the resource consumption. It also helps in implementing fail-over, enabling scalability, avoiding bottlenecks and overprovisioning, reducing response time etc. III. Priority: prioritization of the resources or jobs need to be done on before hand through the algorithm itself for better service to the important or high prioritized jobs in spite of equal service provision for all the jobs regardless of their origin. DISTRIBUTED LOAD BALANCING ALGORITHM FOR CLOUD A. Round Robin Algorithm Round robin algorithm is random sampling based. It means it selects the load randomly in case that some server is heavily loaded or some are lightly loaded. B. Throttled Load Balancing Algorithm Throttled algorithm is completely based on virtual machine. In this client first requesting the load balancer to check the right virtual machine which access that load easily and perform the operations which is give by the client or user. In this algorithm the client first requests the load balancer to find a suitable Virtual Machine to perform the required operation C. VectorDot Singh et al. proposed a novel load balancing algorithm called VectorDot. It handles the hierarchical com-plexity of the data-center and multidimensionality of resource loads across servers, network switches, and storage in an agile data center that has integrated server and storage virtualization technologies. VectorDot uses dot product to distinguish nodes based on the item requirements and helps in removing overloads on servers, switches and storage nodes. Fig. 2 : Load Balancing in Cloud Computing Load balancing is the process of distributing the load among various resources in any system. Thus load need to be distributed over the resources in cloud-based architecture, so that each resources does approximately the equal amount of task at any point of time. Basic need is to provide some techniques to balance requests to provide the solution of the application faster. Cloud vendors are based on automatic load balancing services, which allow clients to increase the number of CPUs or D. Compare and Balance Y. Zhao et al. addressed the problem of intra-cloud load balancing amongst physical hosts by adaptive live ISSN (Print) : 2319 – 2526, Volume-2, Issue-1, 2013 2 Special Issue of International Journal on Advanced Computer Theory and Engineering (IJACTE) migration of virtual machines. A load balancing model is designed and implemented to reduce virtual machines’ migration time by shared storage, to balance load amongst servers according to their processor or IO usage, etc. and to keep virtual machines’ zero-downtime in the process. A distributed load balancing algorithm COMPARE AND BAL-ANCE is also proposed that is based on sampling and reaches equilibrium very fast. This algorithm assures that the migration of VMs is always from high-cost physical hosts to low-cost host but assumes that each physical host has enough memory which is a weak assumption. Performance is used to check the efficiency of the system. This has to be improved at a reasonable cost, e.g., reduce task response time while keeping acceptable delays. IV. INTRODUCTION TO CLOUD ANALYST GUI Package - It is responsible for the graphical user interface, and acts as the front end controller for the application, managing screen transitions and other UI activities. Simulation - This component is responsible for holding the simulation parameters, creating and executing the simulation. E. Event-driven V. Nae et al. presented an event-driven load balancing algorithm for real-time Massively Multiplayer Online Games (MMOG). This algorithm after receiving capacity events as input, analyzes its components in context of the resources and the global state of the game session, thereby generating the game session load balancing actions. It is capable of scaling up and down a game session on multiple resources according to the variable user load but has occasional QoS breaches. UserBase - This component models a user base and generates traffic representing the users. DataCenterController - This component controls the data center activities. Internet - This component models the Internet and implements the traffic routing behavior. InternetCharacteristics - This component maintains the characteristics of the Internet during the simulation, including the latencies and available bandwidths between regions, the current traffic levels, and current performance level information for the data centers. Metrics for Load Balancing In Clouds Various metrics considered in existing load balancing techniques in cloud computing are discussed below VmLoadBalancer - This component models the load balance policy used by data centers when serving allocation requests. Default load balancing policy uses a round robin algorithm, which allocates all incoming requests to the available virtual machines in round robin fashion without considering the current load on each virtual machine. Additionally, it is possible application of a throttled load balancing policy that limits the number of requests being processed in each virtual machine to a throttling threshold. If requests are received causing this threshold to be exceeded in all available virtual machines, then the requests are queued until a virtual machine becomes available. CloudAppServiceBroker This component models the service brokers that handle traffic routing between user bases and data centers. The default traffic routing policy is routing traffic to the closest data center in terms of network latency from the source user base. Throughput is used to calculate the no. of tasks whose execution has been completed. It should be high to improve the performance of the system. Overhead Associated determines the amount of overhead involved while implementing a load-balancing algorithm. It is composed of overhead due to movement of tasks, inter-processor and inter-process communication. This should be minimized so that a load balancing technique can work efficiently. Fault Tolerance is the ability of an algorithm to perform uniform load balancing in spite of arbitrary node or link failure. The load balancing should be a good faulttolerant technique. Migration time is the time to migrate the jobs or resources from one node to other. It should be minimized in order to enhance the performance of the system. V. Response Time is the amount of time taken to respond by a particular load balancing algorithm in a distributed system. This parameter should be minimized. PERFORMANCE ANALYSIS We had used the cloud analyst tool to evaluate the algorithms round robin, and throttled algorithm for the case closest data center by using g user base (1-3) with different regions & data centers (1-4) with different virtual machine monitors. Resource Utilization is used to check the utilization of re-sources. It should be optimized for an efficient load balancing. ISSN (Print) : 2319 – 2526, Volume-2, Issue-1, 2013 3 Special Issue of International Journal on Advanced Computer Theory and Engineering (IJACTE) A. User Base VI. RESULT The design model use the user base to represent the single user but ideally a user base should be used to represent a large numbers of users for efficiency of simulation. Fig. 3 : Simulation Configuration B. Datacenter Datacenter manages the data management activities virtual machines creation and destruction and does the routing of user requests received from user base via the internet to virtual machines. After performing the simulation the result computed by cloud analyst is shown in following below figures. We have used the above defined configuration for each load balancing policy one by one and depending (Response Time and Processing Cost Using Round Robin) ISSN (Print) : 2319 – 2526, Volume-2, Issue-1, 2013 4 Special Issue of International Journal on Advanced Computer Theory and Engineering (IJACTE) VIII. (Response Time and Processing Cost Using Throttled ) [1] Anthony T.Velte, Toby J.Velte, Robert Elsenpeter, Cloud Computing A Practical Approach, TATA McGRAW-HILL Edition 2010. [2] Bhathiya, Wickremasinghe. Cloud Analyst: A Cloud Sim-based Visual Modeller for Analysing Cloud Computing Environments and Applications [3] Chhabra, G. Singh, Qualitative Parametric Comparison of Load Balancing Algorithms in Distributed Computing Environment,14th International Conference on Advanced Computing and Communication, July 2006 IEEE, pp 58 – 61. [4] Fang Y., Wang F. and Ge J. (2010) Lecture Notes in Comput-er Science, 6318, 271-277. [5] Hu J., Gu J., Sun G. and Zhao T. (2010) 3rd International Symposium on Parallel Architectures, Algorithms and Pro-gramming, 89-96. [6] Mehta H., Kanungo P. and Chandwani M. (2011) International Conference Workshop on Emerging Trends in Technology, 370-375 [7] Nae V., Prodan R. and Fahringer T. (2010) 11th IEEE/ACM International Conference on Grid Computing (Grid), 9-17. [8] Ram Prasad Padhy (107CS046), P Goutam Prasad Rao (107CS039).Load balancing in cloud computing system Department of Computer Science and Engineering National Institute of Technology, Rourkela Rourkela-769 008, Orissa, India May, 2011. [9] Randles M., Lamb D. and Taleb-Bendiab A. (2010) 24th Inter-national Conference on Advanced Information Networking and Applications Workshops, 551-556 [10] T.R.Gopalakrishnan Nair Vaidehi .M2 Suma. V Improved Strategies for Enhanced Business Performance in Cloud based IT Industries, Research & Industry Incubation Centre, DSI, Bangalore, India. [11] Wenhong Tian, Yong Zhao, Yuanliang Zhong, Minxian Xu, Chen Jing(2011) , A Dynamic And Integrated Load balancing Scheduling Algorithm For Cloud Datacenters, University of Electronic Science and Technology [12] Zhao Y. and Huang W. (2009) 5th International Joint Confer-ence on INC, IMS and IDC, 170-175. Fig. 4 : Comparison of Load Balancing Algorithm) Comparing with the table and graph, overall response time and data centre processing time is improved. It is also seen that the virtual machine cost and data transfer time in the round robin algorithm is much better when compared to throttled algorithms. VII. CONCLUSIONS The response time and data transfer cost is a challenge of every cloud engineer to build up the products that can increase the business performance in the cloud industry adopted sector. Several strategies lack efficient scheduling and load balancing resource allocation techniques leading to increased operational cost and give customer dissatisfaction. The paper tries to give a bird’s eye view on enhanced strategies through improved job and load balancing resource allocation techniques. REFERENCES ISSN (Print) : 2319 – 2526, Volume-2, Issue-1, 2013 5