International Journal On Engineering Technology and Sciences – IJETS™ ISSN (P): 2349-3968, ISSN (O): 2349-3976 Volume 1 Issue 7, November 2014
Deadline Based Task Scheduling in Cloud with Effective Provisioning Cost using LBMMC Algorithm Ms.K.Sathya, M.E., (CSE), Jay Shriram Group of Institutions, Tirupur
[email protected]
Dr.S.Rajalakshmi, Associate Professor/CSE, Jay Shriram Group of Institutions, Tirupur
[email protected]
Abstract- In cloud computing environment, the major issue is cost provisioning problem. To reduce the impact of performance variation of cloud resources in the deadlines of workflows, proposed a algorithm called EIPR, which takes into consideration the behavior of cloud resources during the scheduling process and also applies replication of tasks to increase the chance of meeting application deadlines. As the execution of the workflow in the Cloud incurs financial cost, the workflow may also be subject to a budget constraint. Although the budget constraint of a workflow may be a limiting factor to its capability of being scaled across multiple resources in Clouds, its structure itself also imposes significant limitation. This new method called Location based Minimum Migration in Cloud (LBMMC).This method monitors all the virtual machines in all cloud locations centre. It identifies the state virtual machines whether it is sleep, idle, or running (less or over) state. And also checks number of tasks running and number of tasks can able to run. When a virtual machine is considered to be overloaded requiring live migration to one or more VMs from the nearby location cloud centers. When a virtual machine is considered to be under loaded (minimum task running) requiring live migration to one or more VMs (which are already running to perform some jobs) from the nearby location cloud centers. Location based VM selection policy (algorithm) has to be applied to carry out the selection process. Index Terms—Cloud computing, EIPR, LBMMC , Migration
Fig.1 Architecture of Cloud Computing I.INTRODUCTION
Cloud
computing has been envisioned as the next generation information technology (IT) architecture for enterprises, due to its long list of unprecedented advantages in the IT history: on-demand self-service, ubiquitous network access, location independent resource pooling, rapid resource elasticity, usage-based pricing and transference of risk. As a disruptive technology with profound implications, cloud computing is transforming the very nature of how businesses use information technology. One fundamental aspect of this paradigm shifting is that data are being centralized or outsourced to the cloud. From users’ perspective, including both individuals and IT enterprises, storing data remotely to the cloud in a flexible on-demand manner brings appealing benefits: relief of the burden for storage management, universal data access with location independence, and avoidance of capital expenditure on hardware, software, and personnel maintenances, etc.,
A. Scheduling The primary benefit of moving to Clouds is application scalability. Unlike Grids, scalability of Cloud resources allows real-time provisioning of resources to meet application requirements. Cloud services like compute, storage and bandwidth resources are available at substantially lower costs. Usually tasks are scheduled by user requirements. New scheduling strategies need to be proposed to overcome the problems posed by network properties between user and resources. New scheduling strategies may use some of the conventional scheduling concepts to merge them together with some network aware strategies to provide solutions for better and more efficient job scheduling. Initially, scheduling algorithms were being implemented in grids. Due to the reduced performance faced in grids, now there is a need to implement scheduling in cloud. The primary benefit of moving to Clouds is application scalability. Unlike Grids, scalability of Cloud resources allows real-time provisioning of resources to meet application requirements. This enables workflow management systems to readily meet Quality ofService (QoS) requirements of applications, as opposed to the traditional approach that required advance reservation of resources in global multi -user Grid environments. Cloud services like compute, storage and bandwidth resources are available at substantially lower costs. Cloud applications often require very complex execution environments. These environments are difficult to create on grid resources. In
167
International Journal On Engineering Technology and Sciences – IJETS™ ISSN (P): 2349-3968, ISSN (O): 2349-3976 Volume 1 Issue 7, November 2014 addition, each grid site has a different configuration, which results in extra effort each time an application needs to be ported to a new site. Virtual machines allow the application developer to create a fully customized, portable execution environment configured specifically for their application. Traditional way for scheduling in cloud computing tended to use the direct tasks of users as the overhead application base. The problem is that there may be no relationship between the overhead application base and the way that different tasks cause overhead costs of resources in cloud systems. For large number of simple tasks this increases the cost and the cost is decreased if we have small number of complex tasks.
factors and can cause performance variation of up to 30 percent for execution times and 65 percent for data transfer time. Fluctuations in performance of Cloud resources delay tasks execution, what also delays such tasks’ successors. If the delayed tasks were part of the critical path of the workflow, it will delay its completion time, and may cause its deadline to be missed. Workflow execution is also subject to delays if one or more of the virtual machines fail during task execution. However, typical Cloud infrastructures offer availability above 99.9 percent, therefore performance degradation is a more serious concern than resource failures in such environments. Here, this algorithm that uses idle time of provisioned resources to replicate workflow tasks to mitigate effects of performance variation of resources so that soft deadlines can be met. The amount of idle time available for replication can be increased if there is budget available for replica execution, which allows provisioning of more resources than the minimal necessary for meeting the deadline. Therefore, this algorithm applies a deadline constraint and variable budget for replication to achieve its goals. To reduce the impact of performance variation of public Cloud resources in the deadlines of workflows, we proposed a new algorithm, called EIPR, which takes into consideration the behavior of Cloud resources during the scheduling process and applies replication of tasks to increase the chance of meeting application deadlines.
B. Scientific Workflows Research in execution of scientific workflows in Clouds either try to minimize the workflow execution time ignoring deadlines and budgets or focus on the minimization of cost while trying to meet the application deadline. It implement limited contingency strategies to correct delays caused by underestimation of tasks execution time or fluctuations in the delivered performance of leased public Cloud resources. To mitigate effects of performance variation of resources on soft deadlines of workflow applications, an algorithm is proposed that uses idle time of provisioned resources and budget surplus to replicate tasks. Typical Cloud environments do not present regular performance in terms of execution and data transfer times. This is caused by technological and strategic factors and can cause performance variation of up to 30 percent for execution times and 65 percent for data transfer time. Fluctuations in performance of Cloud resources delay tasks execution, what also delays such tasks’ successors. If the delayed tasks were part of the critical path of the workflow, it will delay its completion time, and may cause its deadline to be missed. Workflow execution is also subject to delays if one or more of the virtual machines fail during task execution. However, typical Cloud infrastructures offer availability above 99.9 percent; therefore, performance degradation is a more serious concern than resource failures in such environments.
A.EIPR Algorithm The existing algorithm performs following steps 1. Combined Provisioning and Scheduling 2. Data-Transfer Aware Provisioning Adjust 3. Task Replication Combined Provisioning and Scheduling The first step of the EIPR algorithm consists in the determination of the number and type of VMs to be used for workflow execution as well as start and finish time of each VM (provisioning) and the determination of ordering and placement of tasks on such allocated resources (scheduling). The provisioning and scheduling problems are closely related, because the availability of VMs affects the scheduling, and the scheduling affects finish time of virtual VMs. Therefore, a more efficient scheduling and provisioning can be achieved if both problems are solved as one rather than independently.
II.OBJECTIVE The objective of the project is to minimum number of virtual machine is allocated to complete the user tasks within the deadline based on the migration of tasks. Less number of should be selected, to avoid more processing cost for migration process. Migration of Virtual machine tasks to reduce the user provisioning cost, by finishing the user’s tasks in their deadline.
Data-Transfer Aware Provisioning Adjust In the second step of the EIPR algorithm, the provisioning decision performed in Step 1 is adjusted to account for the aforementioned factors. For each nonentry task scheduled as first task of a virtual machine, and for each non-exit task scheduled as the last task of a
III. EXISTING SYSTEM Typical Cloud environments do not present regular performance in terms of execution and data transfer times. This is caused by technological and strategic
168
International Journal On Engineering Technology and Sciences – IJETS™ ISSN (P): 2349-3968, ISSN (O): 2349-3976 Volume 1 Issue 7, November 2014 virtual machine, the algorithm meets the required communication time by setting the start time of the machine earlier than st of the first task, and/or setting the end time of the machine later than the finish time of the last task, depending on where the extra time is required. Finally, the beginning of the first allocation slot of each virtual machine is anticipated by the estimated deployment and boot time for virtual machines, which was accounted for during the scheduling process.
3. More Energy consumed and high cost for the providers side, when more virtual machines are allocated. IV.PROPOSED SYSTEM The proposed method called Location based Minimum Migration in Cloud (LBMMC).This method monitors all the virtual machines in all cloud locations centre. It identifies the state virtual machines whether it is sleep, idle, or running (less or over) state. And also checks number of tasks running and number of tasks can able to run. When a virtual machine is considered to be overloaded requiring live migration to one or more VMs from the nearby location cloud centers. When a virtual machine is considered to be under loaded (minimum task running) requiring live migration to one or more VMs (which are already running to perform some jobs) from the nearby location cloud centers. Location based VM selection policy (algorithm) has to be applied to carry out the selection process. The main contributions are shown below, Migration of Virtual machine tasks where less number of tasks is running, which occupies extra energy and power consumption. Migration of Virtual machine tasks where more number tasks are running, which gives overhead to the system performance. Less number of migrations should be selected, to avoid more processing cost for migration process. Migration of Virtual machine tasks to reduce the user provisioning cost, by finishing the user’s tasks in their deadline.
Task Replication The task replication process works as a semi-active replication technique for fault tolerance, with the difference that here tasks are replicated across performance- independent hosts rather than failureindependent locations. As in the case of replication for fault tolerance, in our approach replication is also managed by a single entity. The process operates as follows. Initially, the algorithm generates new VMs based on the available replication budget. For selection of the type of VM to be used for the new VM, the list of already provisioned VMs is sorted in descending order of number of scheduled tasks. Starting from the beginning of the sorted list, the first VM type whose cost is smaller than the available budget is chosen. The chosen VM is replicated, provisioning information is updated , the VM is moved to the end of the list , the available budget is updated , and the search is restarted. The process is repeated while the budget admits new instantiations. The new provisioned VM and the correspondent time slot reproduce the same start and end times than the original VM. Next, a list of all possible idle slots is created. This list includes both paid slots, which are slots were the VM is actually provisioned and no task is scheduled on it; and unpaid slots, which are the times between the start time and the deadline were the VM is not provisioned Once paid and unpaid slots are identified, the slots are sorted in increasing order of size . The first unpaid slot on the list is inserted after the last paid slot , so paid time slots have precedence over unpaid in the slots selection
A. Advantages The advantages of proposed system are 1. Minimum number of Virtual machines is allocated to complete the user tasks in the deadline. 2. Migrations of tasks in Virtual machines take minimum cost and energy. 3. The underload and overload of tasks are considered while allocation of tasks in Virtual machines. B. Data Flow Diagram
Fig.2 Paid And Unpaid Idle Time Slots B. Disadvantages The disadvantages of existing system are 1. Workflow execution is also subject to delays if one or more of the virtual machines fail during task execution. 2. Takes more number of Virtual Machines to complete the user tasks in their deadlines.
169
International Journal On Engineering Technology and Sciences – IJETS™ ISSN (P): 2349-3968, ISSN (O): 2349-3976 Volume 1 Issue 7, November 2014 VI.SCREENSHOTS A .Home Screen to Get Input
Fig.3 Data Flow Diagram of LBMMC Algorithm Fig.5 Home Screen to Get Input
C. System Architecture
B. Users Resource Assumption
Fig.6 Users Resource Allocation Fig.4 System Architecture of LBMMC Algorithm
C. Users Task Input
V. MODULE DESCRIPTION A. Resource Assumption and Task Analyze For developing one cloud environment, have to analyze the requirements for the resources and the number of resources. In this module, the resource assumption takes place which means need to mention the number of resources and the resources configuration (Processor, RAM, Storage, Capacity, etc). To analyze the task, the task from the user is analyzed and forward for the allocation process.
170
International Journal On Engineering Technology and Sciences – IJETS™ ISSN (P): 2349-3968, ISSN (O): 2349-3976 Volume 1 Issue 7, November 2014 Energy consumed and high cost for the providers side, when more virtual machines are allocated.
Fig.7 Users Task Input D. Calculating Security Demand and Resource Display
To overcome this, LBMMC algorithm was proposed. This algorithm involves A.)Resource Assumption and Task Analyze, B.)Allocate resource for jobs, C).Model Analyze for Migration, D.)Migration of tasks. In this paper , presented a resource assumption and task analyze. The resource assumption takes place which means need to mention the number of resources and the resources configuration. To analyze the task, the task from the user is analyzed and forward for the allocation process. In the future work, the resources are allocated for the user given jobs. The provision of these computational resources is controlled by a provider, and resources are allocated in an elastic way, according to consumers’ needs. To accommodate unforeseen demands on the infrastructure in a scalable and elastic way, the process of allocation and reallocation in Cloud Computing must be dynamic. The trust levels are given for each all resources and the user will give the task with the security demands to run the task in resources. Based on the user demanded security, the resources are allocated for each user jobs. Model Analyze takes place for migration process, where what all the tasks to be migrated. Location based VM selection policy (algorithm) has to be applied to carry out the selection process. The tasks are migrated based on the analyze of the last process.
Fig.8 Calculating Security Demand and Resource Display E. Users Resource Allocation
VIII.REFERENCES [1].Rodrigo N.Calheiros, Rajkumar Buyya,”Meeting Deadlines of Scientific Workflows in Clouds with Tasks Replication”, IEEE Transactions on Parallel And Distributed Systems,VOL.25, NO.7,JULY 2014. [2]. R. Buyya, C.S. Yeo, S. Venugopal, J. Broberg, and I. Brandic, ‘‘Cloud Computing and Emerging IT Platforms: Vision, Hype, and Reality for Delivering Computing as the 5th Utility,’’ Future Gener. Comput. Syst., vol. 25, no. 6, pp. 599-616, June 2009. [3]. Z. Shi and J.J. Dongarra, ‘‘Scheduling Workflow Applications on Processors with Different Capabilities,’’ Future Gener. Comput.Syst., vol. 22, no. 6, pp. 665-675, May 2006.
Fig.9 Users Resource Allocation
[4]. M. Mao and M. Humphrey, ‘‘Auto-Scaling to Minimize Cost and Meet Application Deadlines in Cloud Workflows,’’ in Proc. Int’l Conf. High Perform. Compute., Netw., Storage Anal. (SC), 2011, p. 49.
VII. CONCLUSION AND FUTURE WORK To reduce the impact of performance variation of public Cloud resources in the deadlines of workflows, proposed a existing algorithm called EIPR, which takes into consideration the behavior of Cloud resources during the scheduling process and also applies replication of tasks to increase the chance of meeting application deadlines. In this algorithm, Workflow execution is also subject to delays if one or more of the virtual machines fail during task execution. Takes more number of Virtual Machines to complete the user tasks in their deadlines. More
[5]. S. Abrishami, M. Naghibzadeh, and D. Epema, ‘‘Deadline-Constrained Workflow Scheduling Algorithms for IaaS Clouds,’’Future Gener. Comput. Syst., vol. 29, no. 1, pp. 158-169, Jan. 2013.
171
International Journal On Engineering Technology and Sciences – IJETS™ ISSN (P): 2349-3968, ISSN (O): 2349-3976 Volume 1 Issue 7, November 2014 [6].R. Abbott and H. Garcia-Molina, ‘‘Scheduling RealTime Transactions,’’ACM SIGMOD Rec., vol. 17, no. 1, pp. 71-81, Mar. 1988.
and Simulation of Cloud Computing Environments and Evaluation of Resource Provisioning Algorithms,’’ Softw., Pract. Exper., vol. 41, no. 1,pp. 23-50, Jan. 2011.
[7]. K.R. Jackson, L. Ramakrishnan,K.Muriki, S. Canon, S. Cholia,J. Shalf, H.J.Wasserman, and N.J.Wright, ‘‘Performance Analysis of High Performance Computing Applications on the Amazon Web Services Cloud,’’ in Proc. 2nd Int’l Conf. CloudCom, 2010,pp. 159-168.
AUTHORS BIOGRAPHY K.Sathya received her B.E degree in Avinashilingam Institute for Home Science and Higher Education for Women, Coimbatore, India and currently pursuing M.E degree in Jay Shriram Group of Institutions, Tiruppur, India. Her research interests include Cloud Computing, Advanced Database, Computer networks Operating System .System
[8]. J. Yu, R. Buyya, and K. Ramamohanarao, ‘‘Workflow Scheduling Algorithms for Grid Computing,’’ in Metaheuristics for Scheduling in Distributed Computing Environments, F. Xhafa and A.Abraham, Eds. New York, NY, USA: Springer-Verlag, 2008.
Dr.S. Rajalakshmi received her B.E. degree in Periyar University, Salem, India and M.E. degree in Anna University, Chennai, India and Ph.D. in Data mining in Anna University, India. Currently she is working as an Associate Professor in Jay Shriram Group of Institutions, Tirupur, India. Her research interests include Data mining, Cloud Computing and Big Data.
[9]. A.Hirales-Carbajal, A. Tchernykh, R. Yahyapour, J.L. Gonza´lez-Garcı´a,T. Ro¨ blitz, and J.M. Ramı´rezAlcaraz, ‘‘Multiple Workflow Scheduling Strategies with User Run Time Estimates on a Grid,’’ J. Grid Comput., vol. 10, no. 2, pp. 325-346, June 2012. [10]. C. Lin and S. Lu, ‘‘SCPOR: An Elastic Workflow Scheduling Algorithm for Services Computing,’’ in Proc. Int’l Conf. SOCA, 2011, pp. 1-8. [11]. C.J.Reynolds, S.Winter, G.Z.Terstyanszky, T. Kiss,P.Greenwell, S. Acs, and P. Kacsuk, ‘‘Scientific Workflow Makespan Reduction through Cloud Augmented Desktop Grids,’’ in Proc. 3rd Int’l Conf. CloudCom, 2011, pp. 18-23. [12]. M. Xu, L. Cui, H. Wang, and Y. Bi, ‘‘A Multiple QoS Constrained Scheduling Strategy of Multiple Workflows for Cloud Computing,’’ in Proc. Int’l Symp. ISPA, 2009, pp. 629-634. [13]. K. Plankensteiner and R. Prodan, ‘‘Meeting Soft Deadlines in Scientific Workflows Using Resubmission Impact,’’ IEEE Trans. Parallel Distrib. Syst., vol. 23, no. 5, pp. 890-901, May 2012. [14]. R. Prodan and T. Fahringer, ‘‘Overhead Analysis of Scientific Workflows in Grid Environments,’’ IEEE Trans. Parallel Distrib. Syst., vol. 19, no. 3, pp. 378-393, Mar. 2008. [15]. W. Chen and E. Deelman, ‘‘WorkflowSim: A Toolkit for Simulating Scientific Workflows in Distributed Environments,’’ in Proc. 8th Int’l Conf. EScience, 2012, pp. 1-8. [16]. P. Chevochot and I. Puaut, ‘‘Scheduling FaultTolerant Distributed Hard Real-Time Tasks Independently of the Replication Strategies,’’ in Proc. 6th Int’l Conf. RTCSA, 1999, p. 356. [17]. R.N. Calheiros, R. Ranjan, A. Beloglazov, C.A.F.D. Rose, and R. Buyya, ‘‘CloudSim: A Toolkit for Modeling
172