共查询到20条相似文献,搜索用时 0 毫秒
1.
Energy efficiency is the predominant issue which troubles the modern ICT industry. The ever-increasing ICT innovations and services have exponentially added to the energy demands and this proliferated the urgency of fostering the awareness for development of energy efficiency mechanisms. But for a successful and effective accomplishment of such mechanisms, the support of underlying ICT platform is significant. Eventually, Cloud computing has gained attention and has emerged as a panacea to beat the energy consumption issues. This paper scrutinizes the importance of multicore processors, virtualization and consolidation techniques for achieving energy efficiency in Cloud computing. It proposes Green Cloud Scheduling Model (GCSM) that exploits the heterogeneity of tasks and resources with the help of a scheduler unit which allocates and schedules deadline-constrained tasks delimited to only energy conscious nodes. GCSM makes energy-aware task allocation decisions dynamically and aims to prevent performance degradation and achieves desired QoS. The evaluation and comparative analysis of the proposed model with two other techniques is done by setting up a Cloud environment. The results indicate that GCSM achieves 71 % of energy savings and high performance in terms of deadline fulfillment. 相似文献
2.
The problem of constrained workflow scheduling on heterogeneous computing systems has been of major interest in the recent years. The user requirements are described by defining constraints on the workflow makespan and/or its execution cost. The uncertainty in the activity execution path and the dynamicity in the resource workload may cause some run-time changes of the makespan or cost. To prohibit run-time constraint violation, the system needs robust schedules. In this paper, probability of violation (POV) of constraints is proposed as a criterion for the schedule robustness. An ant colony system is then used to minimize an aggregation of violation of constraints and the POV. Simulation results on real world workflows show the effectiveness of the proposed method in finding feasible schedules. The results also indicate that the proposed method decreases the POV, as well as the expected penalty at run-time. 相似文献
3.
Chakravarthi K. Kalyana Neelakantan P. Shyamala L. Vaidehi V. 《Cluster computing》2022,25(2):1189-1205
Cluster Computing - The resource provisioning and workflow execution in a multi-cloud environment using a pay-as-you-use framework have recently gained the attention of the cloud computing research... 相似文献
4.
Energy preservation is very important nowadays. A large number of applications in science, engineering, astronomy and business analytics are classified as Bag-of-Tasks (BoT) applications. A BoT is a collection of independent tasks that do not communicate with each other during execution. BoT scheduling has been severely studied from a performance point of view. In this paper, we address the problem of energy-efficient BoT scheduling in a heterogeneous environment with the twin objectives of minimizing finish time and energy consumption. Specifically, we extend two performance-oriented scheduling policies, Min–Min and Max–Min, and propose power-aware centralized scheduling policies that incorporate a dynamic voltage/frequency scaling mechanism and can power on and off unneeded computing nodes of a heterogeneous cluster environment using dynamic power management. Additionally, to evaluate the system using a more realistic workload, high-priority tasks with and without time-constraints are also submitted. A series of simulation experiments conducted, show that we can achieve significant energy savings without affecting significantly the execution of BoTs and high-priority tasks. Additional experiments on a real system also confirmed the effectiveness of our policies. 相似文献
5.
Previously, DAG scheduling schemes used the mean (average) of computation or communication time in dealing with temporal heterogeneity. However, it is not optimal to consider only the means of computation and communication times in DAG scheduling on a temporally (and spatially) heterogeneous distributed computing system. In this paper, it is proposed that the second order moments of computation and communication times, such as the standard deviations, be taken into account in addition to their means, in scheduling “stochastic” DAGs. An effective scheduling approach which accurately estimates the earliest start time of each node and derives a schedule leading to a shorter average parallel execution time has been developed. Through an extensive computer simulation, it has been shown that a significant improvement (reduction) in the average parallel execution times of stochastic DAGs can be achieved by the proposed approach. 相似文献
6.
Efficient application scheduling is critical for achieving high performance in heterogeneous computing (HC) environments. Because of such importance, there are many researches on this problem and various algorithms have been proposed. Duplication-based algorithms are one kind of well known algorithms to solve scheduling problems, which achieve high performance on minimizing the overall completion time (makespan) of applications. However, they pursuit of the shortest makespan overly by duplicating some tasks redundantly, which leads to a large amount of energy consumption and resource waste. With the growing advocacy for green computing systems, energy conservation has been an important issue and gained a particular interest. An existing technique to reduce energy consumption of an application is dynamic voltage/frequency scaling (DVFS), whose efficiency is affected by the overhead of time and energy caused by voltage scaling. In this paper, we propose a new energy-aware scheduling algorithm with reduced task duplication called Energy-Aware Scheduling by Minimizing Duplication (EAMD), which takes the energy consumption as well as the makespan of an application into consideration. It adopts a subtle energy-aware method to search and delete redundant task copies in the schedules generated by duplication-based algorithms, and it is easier to operate than DVFS, and produces no extra time and energy consumption. This algorithm not only consumes less energy but also maintains good performance in terms of makespan compared with duplication-based algorithms. Two kinds of DAGs, i.e., randomly generated graphs and two real-world application graphs, are tested in our experiments. Experimental results show that EAMD can save up to 15.59 % energy consumption for HLD and HCPFD, two classic duplication-based algorithms. Several factors affecting the performance are also analyzed in the paper. 相似文献
7.
In this paper, the task scheduling in MapReduce is considered for geo-distributed data centers on heterogeneous networks. Adaptive heartbeats, job deadlines and data locality are concerned. Job deadlines are divided according to the maximum data volume of tasks. With the considered constraints, the task scheduling is formulated as an assignment problem in each heartbeat, in which adaptive heartbeats are calculated by the processing times of tasks, jobs are sequencing in terms of the divided deadlines and tasks are scheduled by the Hungarian algorithm. Taking into account both the data transfer and processing times, the most suitable data center for all mapped jobs are determined in the reduce phase. Experimental results show that the proposed algorithms outperform the current existing ones. The proposals with sorted task-sequences have better performance than those with random task-sequences. 相似文献
8.
Cluster Computing - Applications are evolving in ways that demand geographically distributed resources to co-operate in order to give users better Quality of Service (QoS). There is a plethora of... 相似文献
9.
Energy consumption in high performance computing data centers has become a long standing issue. With rising costs of operating the data center, various techniques need to be employed to reduce the overall energy consumption. Currently, among others there are techniques that guarantee reduced energy consumption by powering on/off the idle nodes. However, most of them do not consider the energy consumed by other components in a rack. Our study addresses this aspect of the data center. We show that we can gain considerable energy savings by reducing the energy consumed by these rack components. In this regard, we propose a scheduling technique that will help schedule jobs with the above mentioned goal. We claim that by our scheduling technique we can reduce the energy consumption considerably without affecting other performance metrics of a job. We implement this technique as an enhancement to the well-known Maui scheduler and present our results. We propose three different algorithms as part of this technique. The algorithms evaluate the various trade-offs that could be possibly made with respect to overall cluster performance. We compare our technique with various currently available Maui scheduler configurations. We simulate a wide variety of workloads from real cluster deployments using the simulation mode of Maui. Our results consistently show about 7 to 14 % savings over the currently available Maui scheduler configurations. We shall also see that our technique can be applied in tandem with most of the existing energy aware scheduling techniques to achieve enhanced energy savings. We also consider the side effects of power losses due to the network switches as a result of deploying our technique. We compare our technique with the existing techniques in terms of the power losses due to these switches based on the results in Sharma and Ranganathan, Lecture Notes in Computer Science, vol. 5550, 2009 and account for the power losses. We there on provide a best fit scheme with the rack considerations. We then propose an enhanced technique that merges the two extremes of node allocation based on rack information. We see that we can provide a way to configure the scheduler based on the kind of workload that it schedules and reduce the effect of job splitting across multiple racks. We further discuss how the enhancement can be utilized to build a learning model which can be used to adaptively adjust the scheduling parameters based on the workload experienced. 相似文献
10.
Cluster Computing - The problem of minimizing the execution monetary cost of applications on cloud computing platforms has been studied recently, and satisfying the deadline constraint of an... 相似文献
11.
Live virtual machine migration can have a major impact on how a cloud system performs, as it consumes significant amounts of network resources such as bandwidth. Migration contributes to an increase in consumption of network resources which leads to longer migration times and ultimately has a detrimental effect on the performance of a cloud computing system. Most industrial approaches use ad-hoc manual policies to migrate virtual machines. In this paper, we propose an autonomous network aware live migration strategy that observes the current demand level of a network and performs appropriate actions based on what it is experiencing. The Artificial Intelligence technique known as Reinforcement Learning acts as a decision support system, enabling an agent to learn optimal scheduling times for live migration while analysing current network traffic demand. We demonstrate that an autonomous agent can learn to utilise available resources when peak loads saturate the cloud network. 相似文献
12.
We describe DAGGER, an ab initio gene recognition program which combines the output of high dimensional signal sensors in an intuitive gene model based on directed acyclic graphs. In the first stage, candidate start, donor, acceptor, and stop sites are scored using the SNoW learning architecture. These sites are then used to generate a directed acyclic graph in which each source-sink path represents a possible gene structure. Training sequences are used to optimize an edge weighting function so that the shortest source-sink path maximizes exon-level prediction accuracy. Experimental evaluation of prediction accuracy on two benchmark data sets demonstrates that DAGGERis competitive with ab initio gene finding programs based on Hidden Markov Models. 相似文献
13.
High availability plays an important role in heterogeneous clusters, where processors operate at different speeds and are
not continuously available for processing. Existing scheduling algorithms designed for heterogeneous clusters do not factor
in availability. We address in this paper the stochastic scheduling problem for heterogeneous clusters with availability constraints.
Each node in a heterogeneous cluster is modeled by its speed and availability, and different classes of tasks submitted to
the cluster are characterized by their execution times and availability requirements. To incorporate availability and heterogeneity
into stochastic scheduling, we introduce metrics to quantify availability and heterogeneity in the context of multiclass tasks.
A stochastic scheduling algorithm SSAC (stochastic scheduling with availability constraints) is then proposed to improve availability of heterogeneous clusters while reducing average response time of tasks.
Experimental results show that our algorithm achieves a good trade-off between availability and responsiveness.
相似文献
Tao XieEmail: |
14.
15.
Data centers are the backbone of cloud infrastructure platform to support large-scale data processing and storage. More and more business-to-consumer and enterprise applications are based on cloud data center. However, the amount of data center energy consumption is inevitably lead to high operation costs. The aim of this paper is to comprehensive reduce energy consumption of cloud data center servers, network, and cooling systems. We first build an energy efficient cloud data center system including its architecture, job and power consumption model. Then, we combine the linear regression and wavelet neural network techniques into a prediction method, which we call MLWNN, to forecast the cloud data center short-term workload. Third, we propose a heuristic energy efficient job scheduling with workload prediction solution, which is divided into resource management strategy and online energy efficient job scheduling algorithm. Our extensive simulation performance evaluation results clearly demonstrate that our proposed solution has good performance and is very suitable for low workload cloud data center. 相似文献
16.
The demand for cloud computing is increasing dramatically due to the high computational requirements of business, social, web and scientific applications. Nowadays, applications and services are hosted on the cloud in order to reduce the costs of hardware, software and maintenance. To satisfy this high demand, the number of large-scale data centers has increased, which consumes a high volume of electrical power, has a negative impact on the environment, and comes with high operational costs. In this paper, we discuss many ongoing or implemented energy aware resource allocation techniques for cloud environments. We also present a comprehensive review on the different energy aware resource allocation and selection algorithms for virtual machines in the cloud. Finally, we come up with further research issues and challenges for future cloud environments. 相似文献
17.
18.
Sedimentation velocity analysis of highly heterogeneous systems 总被引:3,自引:0,他引:3
This article discusses several improvements to the van Holde-Weischet (vHW) method [Biopolymers 17 (1978) 1387] that address its capability to deal with sedimentation coefficient distributions spanning a large range of s values. The method presented here allows the inclusion of scans early and late in the experiment that ordinarily would need to be excluded from the analysis due to ultracentrifuge cell end effects. Scans late in the experiment are compromised by the loss of a defined plateau region and by back-diffusion from the bottom of the cell. Early scans involve partial boundaries that have not fully cleared the meniscus. In addition, a major refinement of the algorithm for determining the boundary fractions is introduced, taking into account different degrees of radial dilution for different species in the system. The method retains its desirable model-independent properties (the analysis of sedimentation data does not require prior knowledge of a user-imposed model or range of sedimentation coefficients) and reports diffusion-corrected s value distributions, which can be presented either in a histogram format or the traditional integral distribution format. Data analyzed with the traditional vHW method are compared with those of the improved method to demonstrate the benefit from the added information in the analysis. 相似文献
19.