SciresolSciresolhttps://indjst.org/author-guidelinesIndian Journal of Science and Technology0974-564510.17485/IJST/v15i2.1806research articleEvolutionary Computing based Web Service Composition Technique for Scheduling of Workload under Cloud EnvironmentHanabarattiKavita Dkdhanabaratti@git.edu1RoddSunil F2Assistant Professor, Department of Computer Science and Engineering, KLS Gogte Institute of TechnologyBelagaviProfessor, Department of Computer Science and Engineering, KLS Gogte Institute of TechnologyBelagavi152692022Abstract
Objectives:The objective of this model is to design a technique to schedule the workload for the computation of web services in the heterogeneous cloud environment. As the Web Service Composition (WSC) for the execution of scientific workload in a heterogeneous cloud computing environment is a challenging task. Modern workload requires dynamic resource provisioning technique as there exist parallelization among sub-tasks and different task demands different Quality of Service (QoS) requirement. Methods: This study presents an Evolutionary Computing based Web Service Composition (EC-WSC) technique to execute a large-scale scientific workload in a heterogeneous cloud environment. A multi-objective metric for improving energy efficiency and resource utilization is modelled. Then, an improved searching mechanism for the dragonfly evolutionary computing algorithm is modelled. Findings: Experiment outcomes show EC-WSC model attains superior performance in execution time performance analysis and energy efficiency performance analysis when compared with existing resource provisioning models of workload service composition such as Deadline and Budget-Aware Workflow Scheduling (DBAWS) 1, Evolutionary Computing Multi-objective optimization for Hybrid Clouds (EC-MOH) 2, Web Service Composition (WSC) 3, and Evolutionary Multi-Objective Optimization for clouds (EMOC)4 in terms of heterogeneous computing, workload size, multi-objective optimization, QoS metric, and optimization strategy. Our model EC-WSC has proved to be more efficient in terms of energy efficiency by a reduction of 52.13% and also reduction in execution time by 71% when compared with the WSC3 existing Web Service Composition model. Novelty: Existing resource provisioning predominantly focused on reducing computation cost and time; however, induces task execution latency and energy overhead. However, EC-WSC is modelled to utilize resources more efficiently and meet task QoS requirements by assuring energy minimization constraints.
Cloud computing environment offers virtual computing nodes (i.e., instance) for executing workflow tasks submitted by the user. The user can use several cloud instances in a scalable manner for executing tasks. Generally, these instances/services are given to users based on predefined SLA’s. The SLA defines a set of agreements between the service provider and service consumers for meeting desired QoS. The users are charged based on QoS and SLA defined. Considering these benefits, grid-based workflow providers such as Pegasus and ASKALON have started supporting cloud-based workflow execution. Thus, it is a challenging task for finding suitable resources (i.e., web service composition mechanism) for executing workload in a multi-processing framework. Generally, the standard web service composition technique hypothesis is that users submit the workload with constraints to the cloud platform. Cloud platform checks and informs its scheduler to find web service available and its configuration for executing workload as per QoS defined. Here the scheduler can allocate resources to workload statically or dynamically. From extensive survey varied, it can be seen an effective web service composition technique generally reduce overall make span with improved resource utilization. However, the modern workload with green computing norms demands the task must be completed within a certain deadline and to meet energy constraints. Existing web composition technique for executing workload focused just for minimizing energy; thus, leads to higher service violation.
Recently, extensive work has been completed on the workload resource provisioning issue in homogeneous processing conditions. In1 because of the workload issues and resource provision in cloud computing and for the storage requirements, an optimized method using the multi-objective has been proposed. This method mainly focuses to reduce the cost and time required for the execution of the scientific cloud problems. In5 a survey on various researches has been done to show the improvement in the allocation of virtual machines considering all the researches from 2015 to 2021. This survey comprises the data about the various attributes and parameters required for the computation of the scientific workloads. In the conclusion of this survey, they have presented the major role of machine learning algorithms in cloud computing to balance the workload in virtual machines and the advantages of virtual machines in the cloud network. In6 to improve the QoS requirement and to perform the execution of the workload with less cost and time, an optimized method to allocate the resources using the genetic algorithm has been proposed. The results have been tested using the different workloads deployed in the Amazon Web Services (AWS). In any case,56 didn't consider limiting energy utilization for the execution of DAG application tasks. In this manner, incites greater expense of execution. For tending to energy effectiveness, proficient execution, and dependable handling prerequisite of present-day Bigdata computational model/systems 7 thinks about that parallel computational platform structure can self-sufficiently enhance the exhibition boundary for uniquely allowing calculation measure as per client QoS or SLA necessities, asset openness, energy limitation, and execution prerequisite. This adaptive crosses directly from the application, for example, planning and mapping of the task to respective computing platforms through power gating and frequencies optimization. Cloud computing platform offers computation of workload in a distributed manner. These platforms are heterogeneous as they offer computation for the large-scale storage and computation process for the workload execution 8. These techniques include enormous computational costs and effects climate. This is because of high energy dispersal at various levels of task execution and storage 9. In10 for the scheduling of the tasks and also to estimate the workload of the cloud, an evolutionary method has been proposed based on the hybrid task scheduling. This method uses the PSO scheme and PTP scheme to schedule the resources required by the task and decrease the execution time. This method has also used the machine learning algorithm which has been employed using the GDM scheme. In11 to increase the usage of the resources with less make span and also to reduce the SLA violation, a load balancing scheduling method, ALTS, has been proposed. In this method, the algorithm is mapped to the incoming tasks and these tasks are then mapped to the respective free virtual machine for the execution of the scientific workflow. In any case, these assessments are kept to self-confined tasks. As heterogeneous systems continue being upgraded, DAG applications tasks with selectivity are also continuing to rise. The task scheduling on heterogeneous computation platforms is generally NP-hard.
Different meta-heuristic strategies, for instance, annealing, Ant Colony Optimization (ACO), GA are extensively used in DAG applications task scheduling 12. In 13, examined that by upgrading execution as far as decreasing calculation cost for the execution of image-based task; they zeroed in on blending prefetching and task provisioning. The response makespan minimization, execution cost, and enhancement issues are demonstrated through heuristic strategy. In14 exhibited the critical difficulties that occur during heterogeneous task resource provisioning under heterogeneous computation environment. Here tasks are scheduled to private and hybrid cloud-based cost minimization model. To guarantee prevalent asset usage, they introduced a heterogeneous resource provisioning model under a private Cloud platform. Additionally, for guaranteeing the responsibility execution is finished inside cutoff time imperative a resource provisioning technique is introduced utilizing Back Propagation Neural Network (BPNN) under heterogeneous computing platform. Similarly, in2 addressed DAG task is scheduled within the monetary financial plan and processing efficiency requirement considering heterogeneous computation environment. The main resource provisioning scheme is planned to utilize the Single Objective (SO) metric to be specific Deadline Constrained Cost Optimization (DCOH) model. The DCOH is planned by lessening the financial spending plan of task execution with the cutoff time essential. Second, they introduced a Multi-Objective (MO) resource provision technique specifically Multi-Objective Optimization Method (MOH). The MOH strategy is intended in balancing the financial spending plan and execution time for DAG task computation. The drawback of DCOH is it assumes the makespan of DAG tasks are fixed in nature, and resource provisioning models are designed considering such fix makespan assumption leading to poor performance. Further, it is important to reduce energy consumption and meet the application QoS prerequisites. This paper presents an improved QoS metric to improve resource utilization and reduce energy consumption for DAG task execution. The optimization process is through the development of an improved evolution computing model which is discussed in the next section.
For addressing the tradeoffs problem here is an efficient web service composition technique that considers minimizing execution time and energy consumption by employing multi-objective improved dragonfly algorithm by effectively utilizing resources in multi-core cloud framework.
The contributions of the work are as follows:
Presented an evolutionary computing-based WS provisioning technique to execute data-intensive workload in the heterogeneous cloud platform.
The EC-WSC technique executes workload by minimizing energy consumption with high resource utilization.
The EC-WSC reduces execution time and energy consumption for executing workload of varied size and complexity when compared with the existing WSC technique.
The manuscript is arranged as described. In section II, an evolutionary computing-based web service composition technique is presented. In section III, the EC-WSC model performance is compared with the standard WSC technique. In section IV the research is concluded and also future work in improving the WSC model is discussed.
Evolutionary Computing based Web Service Composition Technique for Scheduling Workloadin Heterogeneous Cloud Environment
This section presents an Evolutionary Computing-based EC-WSC technique for scheduling workload in the heterogeneous cloud computing environment. First, the workload execution model is presented. Second, presents a system model used for executing the scientific workload. Third, energy consumption for executing scientific workload is presented. The resource provisioning constraint and scheduling metrics to reduce energy dissipation and utilize resources more efficiently for workload execution has been presented. Finally, the dragonfly evolutionary computation algorithm is used for obtaining optimal resource provisioning performance.
2.1 Scientific Workload Model
The data-intensive workload can be described using the following equation
X=x1,x2,x3,…,xn,
and xjj=1,2,…n represent jth workload. The workload xj can be expressed using the following equation
xj=bj,ej,Hj,
where bj depicts workload arrival time, ej represent workload deadline time, and Hj depicts workload structure. Different workload has a different structure which can be defined as a directed acyclic graph as
Hj=Uj,Fj
where Uj defines the set of tasks in workload xj and it is expressed as follows
Uj=u1j,u2j,…,uUjj
and Fj defines data dependencies between the workload xj and it is expressed as follows
Fj⊆Uj×Uj
An edge fqkj∈Fjdescribes the beginning of the task ukj that relies on the output information of the ongoing taskuqj; thusukj is called a forthcoming task and uqj depict the preceding task.
2.2 System model for workload execution on heterogeneous cloud environment
This system model used in this work is shown in Fig. 1. The modern heterogeneous cloud computing environment is generally composed of a huge number of dynamic physical machineI and is described as follows
I=I1,I2,I3,…,Io,
where o defines the number of physical machines present in a heterogeneous cloud environment. For respective physical machineIl∈I, can be exhibited as follows
Il=stl,nl,ql↑,ol,gl,wl,Vl,
where stl depicts storage size,nlrepresent memory capacity, ql↑defines peak energy level, ol bandwidth, capacity, gl,wl depicts corresponding voltage and frequency level and Vl represent the virtual machine presented in the physical machine. The discrete voltage and frequencies of the physical machineIl can be expressed as follows
gl,wl=gl1,wl1,gl2,wl2,…,gl↑,wl↑
Similarly, the virtual machine in the physical machineIl is exhibited as follows
Vl=vl,1,vl,2,…,vl,Vk
Further, each virtual machine can be represented as follows
vl,m=gl,m,nl,m,stl,m
where gl,m depicts virtual machine frequencies level, nl,m represent the virtual machine memory capacity, and stl,m defines storage size. Here the hardware resource can be shared among virtual machines and can migrate between different physical machines.
Proposed system model 3
2.3 Energy consumption model of physical machines
Let consider tl as the static energy consumption and ql↑ represent maximum energy of the physical machineil. Using the model described in15 the energy consumption of physical machineil can be expressed as follows
Il=tl*ql↑*zlu+1-tl*ql↑gl↑3*gl3,
where zlu∈1,0 defines whether a physical machineil is active or not at timeu, glrepresent actual CPU frequencies at time instanceu, andgl↑ represent maximal CPU frequencies at time instanceu. From this parameter, the energy consumption of o physical machines considering different time instances (i.e., xt to yt) can be estimated using the following equation
E=∑j=1o∫xtyttl*ql↑*zlu+1-tl*ql↑gl↑3*gl3dt,
wherezlu andgl varies with respect to time; thus, are time-dependent; and the other parameter doesn't vary with respect to time; thus, they are static in nature.
Let yk,lmj represent the mapping association among task ukj and virtual machine vl,m on physical machineil. The value of yk,lmj will be equal to 0 if ukj is mapped to the virtual machinevl,m, otherwise, it will be 1; which can be mathematically described using the following equation
y_{k, l m}^{j}= \begin{cases}0, & \text { if } u_{k}^{j} \text { is not mapped to } v_{l, m} \\ 1, & \text { otherwise. }\end{cases}
The data dependencies between the tasks in different scientific workloads should overcome the following resource provisioning constraints
where {gu}_{q,lm}^j represent task u_q^j’s completion time, and {uu}_{qk}^j depicts time taken for transmission among task u_q and u_j. After completing the execution of all the tasks within the x_jto respective virtual machines, the maximum time taken for completing it is described using the following equation
Further, the resource provisioning model should meet task deadline requirement as described in the below equation
{gu}_j\leq e_j,\;\;\;\forall x_j\in X.
Along with, because the restricted resources available with physical machines restrict the number of virtual machines to be used; thus, makes resource provisioning even more difficult considering constraints defined below
g_l^\uparrow-{\sum_{m=1}^{\left(V_l\right|}{g_{l,m}\geq0,}}\;\;\;\forall i_l\in I;n_l-{\sum_{m=1}^{\left(V_l\right|}{n_{l,m}\geq0,}}\;\;\;\forall i_l\in I.
This work aimed at satisfying the constraint defined in Eq. (13), (15) (16), and (17) for minimizing energy dissipation in resource provisioning for minimizing for executing scientific workload as described in below equation
where o defines physical machine size, xt defines start time for workload execution and yt represents end time for workload execution. Further, this work aimed at using resources more efficiently; thus, the following objective functions are defined
where n defines the workload size of X and \left(U_j\right| represent the task size of the workload{\;x}_j, {cpu}_k^j represent task u_k^j’s CPU frequencies prerequisite,{\;T}_k^j represent time take to complete workload execution, defines physical machines size in a heterogeneous cloud environment, and A_l depicts an active time of the physical machinei_l’s. Scheduling these tasks on multiple computing machines is an NP-Hard problem. An extensive survey carried out in3 shows dragon fly evolutionary computing algorithm is efficient in solving such problems. The evolutionary computing-based WSC technique for scheduling scientific workload in a heterogeneous cloud computing environment is shown in Figure 2.
This section presents an evolutionary computing-based WSC technique for workload execution in the heterogeneous cloud computing environment. The objective of this work is to assign resource i.e., virtual machine using an improved dragonfly algorithm (IDA)that maximize resource utilization with minimal energy dissipation. Reducing energy dissipation plays a very important role in reducing the overall cost of provisioning service. The proposed work aimed in bringing good tradeoffs between energy minimization and resource utilization maximization by employing multi-objective-based workload scheduling using IDA. Dragonfly algorithm is a metaheuristic model designed using static and dynamic swarming behaviors of dragonflies in nature 3.
Improved dragonfly algorithm
Dragonfly has two main objectives, Hunting (static) and migration (dynamic). In the migration phase, a huge number of dragonflies move long distances toward diverse territorial regions; this results in the exploration phase. In the static hunting phase, the dragonfly moves with a huge herd and along one bearing with nearby developments and abrupt changes in the flying way, which is reasonable in the exploitation stage. To improve the searching ability and avoid the local optimum, here, improved DA is modeled. The entire process of the improved dragonfly algorithm is shown in Figure 3. The step-by-step process of the proposed multi-objective workload scheduling is given below:
Initialization phase and fitness computation: The population of the dragonflies is randomly initialized in this step. The position of i^{th} dragonfly is established using the following equation
X_i=\left(x_i^1,x_i^2,x_i^3,\dots,x_i^d\right)
where i=1,2,\dots,\;N, x_i^d defines the i^{th} dragonfly position within the d^{th} dimension of searching space and N represent the searching agent’s size. The population is consisting of a set of solutions. The solution is created based on several user tasks and virtual machines (VM). Initially, the tasks are assigned to VM randomly. Then, based on the fitness function the solutions are updated. Here the weights are initialized randomly for the different operators. The position and velocity of dragonflies are updated using Eq. (21) to Eq. (24).
Updating solution: The dragonfly algorithm is composed of five elements such as separation alignment, cohesion, attraction toward food, and distraction from the enemy. The separation is computed using the below equation
{SP}_i=-{\sum_{j=1}^N{X-X_j}}
where X defines the current individual position, X_j defines the j^{th} neighborhood individual position and N represent the size of the adjacent individual. The alignment is computed using the following equation
A_i=\frac{{\sum_{j=1}^N{V_j}}}N
where V_jdepicts the speed of the j^{th} adjacent individual. The cohesion is computed using the following equation
C_i=\frac{{\sum_{j=1}^N{X_j}}}N-X
Attraction toward food is computed using the following equation
F_j=X^+-X
where X^+ defines the food source position. The distraction from the enemy is computed using the following equation
E_i=X^--X
where X^- defines the enemy position. Then, the speed vector is computed using the following equation
wheres define distinct weights, S_i defines the distinct weight of i^{th} individual, cand C_idepicts the cohesion weight and distinct weight of i^{th} individual, respectively. The position vector is computed using the following equation
X_{k+1}=X_k+\triangle X_{k+1}
where k represents the present iteration. If it doesn’t find a dragonfly within its neighborhood the position of the dragonfly is optimized using the levy flight equation defined in Eq. (28). This aid in enhancing randomness behavior and global searching mechanism for achieving obtaining optimal solution as described in below equation
where d depicts the number of position vectors considered, r_1 and r_2 defines the random number ranging between 0 and 1, \alpha and \beta is a small constant value equal to 0.01. In this work, the fitness is computed using position and updating the velocity of the objective. The position is updated till it obtains optimal solution or it reaches the termination conditions. The EC-WSC technique is very efficient in minimizing task execution time with minimal energy dissipation which is proved through experimental study.
Results and Discussion
Here experiment is carryout to evaluate the performance of the evolutionary computing-based Web Service Composition Technique and existing web service composition technique 2, 3 for executing data-intensive workload on the cloud computing environment. Both EC-WSC and existing WSC techniques are implemented using the CloudSim simulator. Energy efficiency and execution time are metrics used for validating the performance of different web service composition models. The Inspiral workload is used for validating the performance of EC-WSC and the existing WSC model. A sample representation of Inspiral workload is shown in Figure 4. The Inspiral workload is memory and CPU-intensive in nature and is used for analyzing binary neutron stars and black holes. More details of workload can be obtained from 16. Here the Inspiral workload size is varied to\;30,\;50,\;100,\;\&\;1000 and experiments have been conducted.
Inspiral workflow
3.1 Execution time performance analysis
Here experiment is conducted by varying Inspiral workload size to 30,\;50,\;100,\;\&\;1000 and execution time achieved using EC-WSC and existing WSC 2, 3 techniques are graphically shown in Fig. 5. From the figure, it can be seen the EC-WSC technique reduces execution time in comparison with the existing WSC technique for both smaller workload sizes and also for larger workload sizes. An average execution time reduction of 83.39% is achieved using EC-WSC in comparison with standard WSC techniques. An important to think from the result we can note that EC-WSC techniques are very efficient for executing larger workload sizes. Further, we evaluate the average execution time efficiency per job using EC-WSC and existing WSC3 as shown from Fig 6. An average execution time reduction of 30.257% is achieved using EC-WSC in comparison with the standard WSC technique. Similarly, we evaluate the average execution time efficiency per job using EC-WSC and existing WSC 2, 3 as shown from Figure 7. An average execution time reduction of 88.28% is achieved using EC-WSC in comparison with the standard WSC model.
Inspiral Workload ExecutionTime using EC-WSC and standard WSC technique
Inspiral Workload Average Execution Time Efficiency per Job using EC-WSC and standard WSC technique 3
Inspiral Workload Average Execution Time Efficiency per Job using EC-WSC and existing WSC technique 2, 16
3.2 Energy efficiency performance analysis:
Here, the experiment is conducted by varying Inspiral workload size to 30,\;50,\;100,\;\&\;1000 and energy efficiency achieved using EC-WSC and existing WSC3 techniques is graphically shown in Fig. 8. From the figure, it can be seen the EC-WSC technique reduces energy consumption in comparison with the existing WSC technique. An average energy consumption reduction of 41.35% is achieved using EC-WSC in comparison with standard WSC techniques. An important to think from the result we can note that the EC-WSC technique is very efficient for executing both smaller and as well as larger workload sizes.
Inspiral Workload Average Energy Efficiency using EC-WSC and existing WSC technique
The Table 1 describes the comparatives analysis of EC-WSC concerning other WSC models such as EC-MOH 2, WSC 3, DBAWS 1, and EMOC 4. The following work has been considered for comparative analysis because of the multi-objective optimization strategy adopted. From Table 1, we can say that all WSC models except WSC can allow execution of the complex workload. Similarly, the existing model can support the execution of the complex workload of Small to medium-large. However, the proposed EC-WSC allows execution of complex workload of small-large. All existing work minimizes cost-makespan and no prior work has considered minimizing energy which plays a very important role in defining cost when resources are heterogeneous. However, the proposed EC-WSC is designed to minimize makespan and maximize resource utilization by optimizing energy constraints and meeting application deadlines. Most of the existing strategies have adopted genetic algorithms for solving multi-objective optimization problems; however, WSC 2 showed dragonfly algorithms are much efficient in comparison with genetic algorithms for optimizing NP-Hard problems; thus, in our work, an improved version of the dragonfly is modeled for solving a multi-objective optimization problem. The result achieved proves the efficiency of the EC-WSC model. The EC-WSC model reduces energy consumption and at the same time improves execution efficiency for workload execution. Thus, proves that the new improved evolution computing algorithm of the dragonfly algorithm can search better (i.e., optimal) resources in executing the workload. As a result, achieves much better performance than the recently modeled WSC technique1, 2, 3, 4.
Comparative analysis of proposed EC-WSC with various other existing WSC models
EC-MOH2, 2019
WSC 3, 2020
DBAWS 1, 2021
EMOC 4, 2021
EC-WSC
Heterogeneous computing
Yes
No
No
No
Yes
Workload type
Complex
Simple
Complex
Complex
Complex
Workload size
Small to medium
small
Small to medium-large
Small to medium-large
Small to large
Multi-objective optimization
Yes
Yes
Yes
Yes
Yes
QoS Metrics
Makespan & cost
Makespan, cost, &load balancing
Deadline & budget
Makespan & cost
Energy, Processing time, resource utilization, & deadlines
Optimization strategy
Genetic Algorithm
Dragonfly algorithm
Monte Carlo & Genetic Algorithm
Chaotic system & Genetic Algorithm
Improved dragonfly algorithm
Conclusion
In this paper, various existing WSC techniques for executing workload in the cloud environment have been discussed. This paper also presents a solution for the trade-off problems (NP-Hard problems) using QoS metrics and a new improved searching mechanism through an evolutionary computing algorithm which is introduced to obtain the optimal solution in quick time. In this model of WCS, the multi-objective dragonfly evolutionary computing algorithm has been used. Experiment outcomes show that the EC-WCS is very efficient in reducing the energy consumption by 23% when compared with the exiting method EC-MOH-WSC and also meets the QoS requirement of workload application. Experiment outcomes also show EC-WSC model attains superior performance in execution time performance analysis and energy efficiency performance analysis when compared with existing resource provisioning models of workload service composition such as EC-MOH 2, WSC 3, DBAWS1, and EMOC 4 in terms of heterogeneous computing, workload size, multi-objective optimization, QoS metric, and optimization strategy. Our model EC-WSC has proved to be more efficient in terms of energy efficiency by a reduction of 52.13% and also a reduction in execution time by 71% when compared with the WSC 3 existing Web Service Composition model. As the size of the workload scheduling increases our model performs efficiently when compared with other existing models.
Future work would be to improvise the utilization of physical machines. Further, would consider the minimization of computation time and I/O overhead by effectively utilizing system resources (idle server) of the cloud.
ReferencesCalzarossaMaria CarlaVedovaMarco L. DellaMassariLuisaNebbioneGiuseppeTesseraDanieleMulti-Objective Optimization of Deadline and Budget-Aware Workflow Scheduling in Uncertain Clouds2021989891899052169-3536Institute of Electrical and Electronics Engineers (IEEE)https://dx.doi.org/10.1109/access.2021.3091310ZhouJunlongWangTianCongPeijinLuPingpingWeiTongquanChenMingsongCost and makespan-aware workflow scheduling in hybrid clouds20191001016311016311383-7621Elsevier BVhttps://dx.doi.org/10.1016/j.sysarc.2019.08.004NeelimaP.ReddyA. Rama MohanAn efficient load balancing system using adaptive dragonfly algorithm in cloud computing2020234289128991386-7857, 1573-7543Springer Science and Business Media LLChttps://dx.doi.org/10.1007/s10586-020-03054-wPaknejadPeymanKhorsandReihanehRamezanpourMohammadrezaChaotic improved PICEA-g-based multi-objective optimization for workflow scheduling in cloud environment202111712280167-739XElsevier BVhttps://dx.doi.org/10.1016/j.future.2020.11.002UllahArifNawiNazri MohdOuhameSoukainaRecent advancement in VM task allocation system for cloud computing: review from 2015 to202120211450269-2821, 1573-7462Springer Science and Business Media LLChttps://dx.doi.org/10.1007/s10462-021-10071-7DevarasettyPrasadReddySatyanandaGenetic algorithm for quality of service based resource allocation in cloud computing20211423813871864-5909Springer Science and Business Media LLChttps://link.springer.com/article/10.1007/s12065-019-00233-6DoppaJanardhan RaoKimRyan GaryIsakovMihailoKinsyMichel AKwonHyouk JunKrishnaTusharAdaptive Manycore Architectures for Big Data Computing2017918ACMhttps://doi.org/10.3390/pr9091514XieGuoqiZengGangLiRenfaLiKeqinEnergy-Aware Processor Merging Algorithms for Deadline Constrained Parallel Applications in Heterogeneous Cloud Computing20172262752377-3782, 2377-3790Institute of Electrical and Electronics Engineers (IEEE)https://dx.doi.org/10.1109/tsusc.2017.2705183LiZhongjinGeJidongHuHaiyangSongWeiHuHaoLuoBinCost and Energy Aware Scheduling Algorithm for Scientific Workflows with Deadline Constraint in Clouds20181147137261939-1374, 2372-0204Institute of Electrical and Electronics Engineers (IEEE)https://dx.doi.org/10.1109/tsc.2015.2466545LakhaniB.AgrawalA.A Task Scheduling Approach for Cloud Environments Employing Evolutionary Algorithms20211324234382070-0245, 2070-0237Bangladesh Journals Online (JOL)https://dx.doi.org/10.3329/jsr.v13i2.49944MubeenAroosaIbrahimMuhammadBibiNargisBazMohammadHamamHabibCheikhrouhouOmarAlts: An Adaptive Load Balanced Task Scheduling Approach for Cloud Computing202199151415142227-9717MDPI AGhttps://dx.doi.org/10.3390/pr9091514ZhuZhaomengZhangGongxuanLiMiqingLiuXiaohuiEvolutionary Multi-Objective Workflow Scheduling in Cloud2016275134413571045-9219Institute of Electrical and Electronics Engineers (IEEE)https://dx.doi.org/10.1109/tpds.2015.2446459KhorramnejadKavehFerdouseLilatulGuanLingAnpalaganAlaganPerformance of integrated workload scheduling and pre-fetching in multimedia mobile cloud computing2018711142192-113XSpringer Science and Business Media LLChttps://dx.doi.org/10.1186/s13677-018-0115-6ChunlinLiJianhangTangYoulongLuoHybrid Cloud Adaptive Scheduling Strategy for Heterogeneous Workloads20191734194461570-7873, 1572-9184Springer Science and Business Media LLChttps://dx.doi.org/10.1007/s10723-019-09481-3ZhuXiaominYangLaurence T.ChenHuangkeWangJiYinShuLiuXiaochengReal-Time Tasks Oriented Energy-Aware Scheduling in Virtualized Clouds2014221681802168-7161Institute of Electrical and Electronics Engineers (IEEE)https://dx.doi.org/10.1109/tcc.2014.2310452BharathiShishirChervenakAnnDeelmanEwaMehtaGaurangSuMei-HuiVahiKaranCharacterization of scientific workflows2008IEEEAustin, TX.10.1109/WORKS.2008.4723958