CN109684070A - A kind of dispatching method in cloud computing parallel work-flow - Google Patents

A kind of dispatching method in cloud computing parallel work-flow Download PDF

Info

Publication number
CN109684070A
CN109684070A CN201810997296.7A CN201810997296A CN109684070A CN 109684070 A CN109684070 A CN 109684070A CN 201810997296 A CN201810997296 A CN 201810997296A CN 109684070 A CN109684070 A CN 109684070A
Authority
CN
China
Prior art keywords
task
virtual machine
time
starting
platform
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810997296.7A
Other languages
Chinese (zh)
Other versions
CN109684070B (en
Inventor
马建峰
张兆一
李辉
张世哲
李金库
姚青松
宁建斌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xidian University
Original Assignee
Xidian University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xidian University filed Critical Xidian University
Priority to CN201810997296.7A priority Critical patent/CN109684070B/en
Publication of CN109684070A publication Critical patent/CN109684070A/en
Application granted granted Critical
Publication of CN109684070B publication Critical patent/CN109684070B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • G06F9/5077Logical partitioning of resources; Management or configuration of virtualized resources
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/50Indexing scheme relating to G06F9/50
    • G06F2209/5011Pool
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/50Indexing scheme relating to G06F9/50
    • G06F2209/5017Task decomposition
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Debugging And Monitoring (AREA)

Abstract

The invention discloses the dispatching methods in a kind of cloud computing parallel work-flow, after determining virtual machine configuration, while opening 1,2,3 ..., n platform virtual machine, and count the average used time t that every virtual machine is opened in test every time1,t2,t3,…,tn, the maximum likelihood value of a and b is obtained according to formula ti=ai+b;Then calculating task used time τ, then starting m platform virtual machine executing tasks parallelly is optimal case;Secondly task approximation is averagely divided n parts (n > > m) by task window, constitute set of tasks, it is placed in task pool, task dispatcher successively creates m platform virtual machine, start parallel processing task process simultaneously, whenever having, when task is completed on new virtual machine starting or virtual machine, task dispatcher is that the virtual machine distributes task, and last whole virtual machines complete whole tasks and simultaneously return to task result.Optimal scheduling strategy in the present invention is designed the characteristics of starting the time by virtual machine, the speed-up ratio of near-optimization may be implemented, guarantee that each virtual machine task is almost completed at the same time, the waste of system resource is reduced while greatly reducing task execution time and improving task efficiency.

Description

A kind of dispatching method in cloud computing parallel work-flow
Technical field
The invention belongs to field of computer technology, grasp parallel further to the cloud computing of one of field of cloud computer technology Dispatching method in work, the present invention can be used for carrying out computation-intensive task under cloud computing environment.
Background technique
Cloud computing is due to its convenient and fast storage service, the advantages that flexible charge method so that more and more enterprises and Individual by local datastore on Cloud Server, to reduce the storage burden and administration overhead of local data.In recent years in phase When obtaining great reputation in the enterprises and individuals of quantity.User does not need the hardware and software resource for retaining oneself, he Can send request to cloud at any time and calculate operation.When executing large-scale calculations task or transformation task, cloud can star Virtual machine (VM) on platform executes parallel computation or parallel transmission.For example, privacy in order to protect data, we will be big The plaintext of amount is encrypted as ciphertext, and by a large amount of ciphertext from server transport to client, this needs to take a lot of time ability It completes.In general, Task-decomposing can be multiple subtasks by we, and started a large amount of virtual machines and carried out subtasking, with Shorten the deadline.
But it needs to solve there are also many problems.Since the starting of virtual machine is also required to the time, so being not virtual machine More, task total deadline is shorter.How many a virtual machines should so be started, how to give virtual machine distribution task? it is thus determined that A kind of optimal scheduling scheme, allows cloud platform in an optimal manner by duty mapping into virtual machine, completes task always Time further shortens the hot issue for becoming field of cloud calculation instantly.
Neusoft Group Co., Ltd application a kind of parallel calculating method of patent and system (application number: 201310591160.3, publication number: 103617086B), a kind of parallel calculating method is disclosed, method includes: to each calculating Node is monitored to obtain monitoring nodes data;The load energy of the calculate node is calculated based on the monitoring nodes data Power;Task to be allocated is allocated according to the load capacity of the calculate node.The deficiency of this method is: only for Platform resource is dispatched, and does not make optimization to the dispatching method of task.
Shandong University application patent based on Linux parallel computing platform dynamic load balancing method (application number: 201310341592.9, publication number: 103399800B), the dynamic load for disclosing kind of the parallel computing platform based on Linux is equal Weighing apparatus method, method content are as follows: during parallel computation, total calculating task is divided into execution time equal multiple ranks Duan Zhihang.Using the regular works dispatching technique in system, before the parallel computation of each time phase starts, first read each The current resource utilization of a node, and the calculated performance and computation complexity of each node are combined, to the calculating task of node It is dynamically distributed, guarantees that the calculating time of each node of each stage is of substantially equal, reduce the synchronous delay waited of system.It is logical This dynamic adjustable strategies are crossed, it can be achieved that completing total calculating task with higher resource utilization.The deficiency of this method is: It is only limitted to Linux platform, and the not influence in view of the time of system starting to execution efficiency.
The main policies that are widely studied and practices at present, such as to cpu frequency progress dynamic adjustment, closing idle machine or The migration merging etc. of sleep state, virtual machine is made it into, essence of these strategies are all by data center resource Dynamic dispatching and integration, maximize energy utilization rate, however there is no a kind of in one timing of task amount and system resource, Make task total used time shortest optimal acceleration strategy.
Therefore, when system resource is fixed, when how how to shorten general assignment completion as far as possible in executing tasks parallelly Between, seek suitable virtual machine quantity, maximizing optimal speed-up ratio becomes urgent problem to be solved.
Summary of the invention
In view of the deficienciess of the prior art, the object of the present invention is to provide the dispatching parties in a kind of cloud computing parallel work-flow Method determines the virtual machine quantity that optimal acceleration can be achieved by system resource and task amount before task execution, then right Task is split, and successively starts virtual machines performing tasks.
The present invention is to be achieved through the following technical solutions:
A kind of dispatching method in cloud computing parallel work-flow, includes the following steps:
S1, cloud platform parameter is calculated:
After determining virtual machine configuration, while opening 1,2,3 ..., n platform virtual machine, and it is every to count unlatching in test every time The average used time t of platform virtual machine1,t2,t3,…,tn, two groups of data are substituted into tiIn=ai+b, the likelihood of a and b is calculated Value, and record;
Wherein, tiWhen to open i virtual machine simultaneously, every virtual machine is opened average time used, and a and b system are special Levy parameter;
S2, optimal acceleration strategy is determined:
Task is executed in a virtual machine, calculating task used time τ, then starting m platform virtual machine executing tasks parallelly is most Excellent scheme;
S3, segmentation task:
When task reaches task window, task window is averagely divided task approximation n parts (n > > m), constitutes task-set It closes, is placed in task pool;
S4, creation virtual machine:
Task dispatcher successively creates m platform virtual machine, while starting parallel processing task process;
S5, task distribution:
Whenever having, when task is completed on new virtual machine starting or virtual machine, task dispatcher is that virtual machine distribution is appointed Business;
S6, task is completed:
It completes whole tasks and returns to task result.
Preferably, the specific method of task dispatcher distribution task is that task dispatcher takes from task pool in step S5 1 part of task out, by duty mapping to the virtual machine, virtual machine starts execution task, after task is finished on virtual machine It checks task pool, if task pool is not sky, continues to assign tasks to virtual machine, if task pool is sky, which is restored For original state.
Compared with prior art, the invention has the following beneficial technical effects:
1) optimal scheduling strategy in the present invention is designed the characteristics of starting the time by virtual machine, and approximation may be implemented Optimal speed-up ratio guarantees that each virtual machine task is almost completed at the same time, and improves task effect greatly reducing task execution time Reduce the waste of system resource while rate.
2) the optimal acceleration strategy in the present invention is realized by control virtual machine quantity and plan target.Once providing Calculating task, near-optimization speed-up ratio system can calculate optimal virtual machine quantity, and automatic plan target.It is applicable in Range is more extensive, and can be used in conjunction with existing resource dynamic scheduling scheme, greatly improves system effectiveness.
Detailed description of the invention
Fig. 1 is the flow diagram of dispatching method;
Fig. 2 is the curve graph of more virtual machine starting total times;
Fig. 3 is the deterministic process schematic diagram of task used time τ;
Fig. 4 is task distribution schematic diagram.
Specific embodiment
The present invention will be further described with reference to the accompanying drawings and detailed description.
As shown in Figure 1, for the optimal case step schematic diagram in a kind of cloud computing parallel work-flow of the present invention, system it is specific Process is described as follows:
Step 1: it calculates cloud platform parameter: after determining virtual machine configuration, while opening 1,2,3 ..., n platform virtual machine, And count the average used time t that every virtual machine is opened in test every time1,t2,t3,…,tn, two groups of data are substituted into ti=ai+ In (i=1,2 ..., n) likelihood value of a and b is calculated, and record in b;
Step 2: it determines optimal acceleration strategy: executing task in a virtual machine, calculating task used time τ then startsPlatform virtual machine executing tasks parallelly is optimal case;
Step 3: segmentation task: when task reaches task window, task approximation is averagely divided into n parts by task window (n > > m) constitutes set of tasks, is placed in task pool;
Step 4: creation virtual machine: task dispatcher successively creates m platform virtual machine, while starting parallel processing task mistake Journey;
Step 5: task distribution: whenever having, when task is completed on new virtual machine starting or virtual machine, task dispatcher is The virtual machine distributes task;
Step 6: it completes task: completing whole tasks and return to task result.
In step 1, at the same when starting more virtual machines every virtual machine average start-up time calculation method T=AN+B (N is virtual machine quantity, and T is average start-up time), which is tested by many experiments and is obtained.Experiment content are as follows: Start m platform virtual machine simultaneously on OpenStack platform, records first virtual machine starting time Tmin, start total used time Tmax, and Calculate the average used time T of every virtual machine of startingavg, data are obtained after many experiments as shown in Fig. 2, it can be found that more virtual Machine starting total time approximation meets T=AN+B curve.
In step 2, the deterministic process schematic diagram of task used time τ is as shown in figure 3, when the virtual machine of starting is constantly from task When pond takes task execution, it is believed that all virtual machines are approximate while ending task, and are equal to so general assignment completes the moment The task finish time of every virtual machine, it is also equal to the average start-up time of every virtual machine and the average time of task execution Sum, i.e. Tall=Tavg+ τ/m=am+b+ τ/m is observed it is found that when m takesWhen, total time obtains minimum value.
In step 5, Fig. 4 be for task distribute schematic diagram, specifically includes the following steps:
Step 1: taking out 1 part of task from task pool when virtual machine is opened, by duty mapping to virtual machine, start Execution task;
Step 2: when on virtual machine task be finished after check task pool, if task pool be not it is empty, continue the first step, If task pool is sky, which is reverted into original state.
Above description is only example of the present invention, does not constitute any limitation of the invention.Obviously for this It, all may be before without departing substantially from the principle of the invention, structure after understanding summary of the invention and principle for the professional in field It puts, carries out the amendment and improvement of algorithm, but these amendments and improvement based on inventive algorithm are wanted in right of the invention It asks within protection scope.

Claims (3)

1. the dispatching method in a kind of cloud computing parallel work-flow, which comprises the steps of:
S1, cloud platform parameter is calculated:
After determining virtual machine configuration, while opening 1,2,3 ..., n platform virtual machine, and count in test every time and open every void The average used time t of quasi- machine1,t2,t3,…,tn, the n data that statistics obtains are substituted into t respectivelyiIn=ai+b, a and b is calculated Maximum likelihood value;
Wherein, tiWhen to open i virtual machine simultaneously, every virtual machine opens average time used, a and b as system features ginseng Number;
S2, optimal acceleration strategy is determined:
Task is executed in a virtual machine, calculating task used time τ, then starting m platform virtual machine executing tasks parallelly is optimal side Case;
S3, segmentation task:
When task reaches task window, task window is averagely divided task approximation n parts (n > > m), constitutes set of tasks, It is placed in task pool;
S4, creation virtual machine:
Task dispatcher successively creates m platform virtual machine, while starting parallel processing task process;
S5, task distribution:
Whenever having, when task is completed on new virtual machine starting or virtual machine, task dispatcher is that the virtual machine distributes task;
S6, task is completed:
Whole virtual machines complete whole tasks and return to task result.
2. the dispatching method in a kind of cloud computing parallel work-flow according to claim 1, which is characterized in that optimal in step 2 Scheme determine specific method be;
When the virtual machine of starting constantly takes task execution from task pool, all virtual machines are approximate to end task simultaneously, so always Task completes the task finish time for being equal to every virtual machine at the moment, is also equal to the average start-up time and task of every virtual machine The sum of the average time of execution, formula are as follows:
Tall=Tavg+ τ/m=am+b+ τ/m
When m takesWhen, total time obtains minimum value;
Wherein, TavgFor the average used time for starting every virtual machine, TallThe general assignment deadline.
3. the dispatching method in a kind of cloud computing parallel work-flow according to claim 1, which is characterized in that task in step S5 The specific method that scheduler distributes task is that task dispatcher takes out 1 part of task from task pool, and duty mapping is virtual to this On machine, virtual machine starts execution task, when on virtual machine task be finished after check task pool, if task pool be not it is empty, Continue to assign tasks to virtual machine, if task pool is sky, which is reverted into original state.
CN201810997296.7A 2018-08-29 2018-08-29 Scheduling method in cloud computing parallel operation Active CN109684070B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810997296.7A CN109684070B (en) 2018-08-29 2018-08-29 Scheduling method in cloud computing parallel operation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810997296.7A CN109684070B (en) 2018-08-29 2018-08-29 Scheduling method in cloud computing parallel operation

Publications (2)

Publication Number Publication Date
CN109684070A true CN109684070A (en) 2019-04-26
CN109684070B CN109684070B (en) 2022-12-13

Family

ID=66184433

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810997296.7A Active CN109684070B (en) 2018-08-29 2018-08-29 Scheduling method in cloud computing parallel operation

Country Status (1)

Country Link
CN (1) CN109684070B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012104898A1 (en) * 2011-01-31 2012-08-09 トヨタ自動車株式会社 Safety control device and safety control method
JP2013152513A (en) * 2012-01-24 2013-08-08 Nippon Telegr & Teleph Corp <Ntt> Task management system, task management server, task management method and task management program
CN104536806A (en) * 2014-12-26 2015-04-22 东南大学 Workflow application flexible resource supplying method in cloud environment
CN105159752A (en) * 2015-09-22 2015-12-16 中国人民解放军国防科学技术大学 Real-time task and resource scheduling method with function of machine startup time perception in virtualized clouds
CN105912406A (en) * 2016-05-05 2016-08-31 中国人民解放军国防科学技术大学 Low-energy independent task scheduling and resource configuration method
KR101707601B1 (en) * 2015-12-31 2017-02-16 숭실대학교산학협력단 Virtual machine monitor and schduling method of virtual machine monitor
CN107357641A (en) * 2017-06-21 2017-11-17 西安电子科技大学 Method for scheduling task in a kind of cloud computing

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012104898A1 (en) * 2011-01-31 2012-08-09 トヨタ自動車株式会社 Safety control device and safety control method
JP2013152513A (en) * 2012-01-24 2013-08-08 Nippon Telegr & Teleph Corp <Ntt> Task management system, task management server, task management method and task management program
CN104536806A (en) * 2014-12-26 2015-04-22 东南大学 Workflow application flexible resource supplying method in cloud environment
CN105159752A (en) * 2015-09-22 2015-12-16 中国人民解放军国防科学技术大学 Real-time task and resource scheduling method with function of machine startup time perception in virtualized clouds
KR101707601B1 (en) * 2015-12-31 2017-02-16 숭실대학교산학협력단 Virtual machine monitor and schduling method of virtual machine monitor
CN105912406A (en) * 2016-05-05 2016-08-31 中国人民解放军国防科学技术大学 Low-energy independent task scheduling and resource configuration method
CN107357641A (en) * 2017-06-21 2017-11-17 西安电子科技大学 Method for scheduling task in a kind of cloud computing

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
YICHUAN WANG, SANTOSH CHANDRASEKHAR,MUKESH SINGHAL,JIANFENG MA: "A limited-trust capacity model for mitigating threats of internal malicious services in cloud computing", 《SPRINGERLINK》 *
赖俊凡等: "考虑处理机时间窗口的可分任务调度优化模型", 《西安交通大学学报》 *
陈黄科等: "云计算中资源延迟感知的实时任务调度方法", 《计算机研究与发展》 *

Also Published As

Publication number Publication date
CN109684070B (en) 2022-12-13

Similar Documents

Publication Publication Date Title
Singh et al. Task scheduling in cloud computing
Mattess et al. Scaling mapreduce applications across hybrid clouds to meet soft deadlines
CN102681889B (en) Scheduling method of cloud computing open platform
Mittal et al. An optimized task scheduling algorithm in cloud computing
CN106502792A (en) A kind of multi-tenant priority scheduling of resource method towards dissimilar load
CN104657220A (en) Model and method for scheduling for mixed cloud based on deadline and cost constraints
Liu et al. Resource preprocessing and optimal task scheduling in cloud computing environments
CN106406987A (en) Task execution method and apparatus in cluster
Domanal et al. Load balancing in cloud environment using a novel hybrid scheduling algorithm
CN106919449A (en) The dispatch control method and electronic equipment of a kind of calculating task
CN107122233A (en) A kind of adaptive real-time scheduling methods of many VCPU towards TSN business
Gawali et al. Standard deviation based modified cuckoo optimization algorithm for task scheduling to efficient resource allocation in cloud computing
Zhou et al. Concurrent workflow budget-and deadline-constrained scheduling in heterogeneous distributed environments
CN104112049A (en) P2P (peer-to-peer) architecture based cross-data-center MapReduce task scheduling system and P2P architecture based cross-data-center MapReduce task scheduling method
CN104035819A (en) Scientific workflow scheduling method and device
Dubey et al. QoS driven task scheduling in cloud computing
Natarajan Parallel queue scheduling in dynamic cloud environment using backfilling algorithm
Koneru et al. Resource allocation method using scheduling methods for parallel data processing in cloud
CN103678000A (en) Computational grid balance task scheduling method based on reliability and cooperative game
CN109684070A (en) A kind of dispatching method in cloud computing parallel work-flow
CN110262896A (en) A kind of data processing accelerated method towards Spark system
Kaur et al. Challenges to task and workflow scheduling in cloud environment
Salmani et al. A fuzzy-based multi-criteria scheduler for uniform multiprocessor real-time systems
Jiang et al. An optimized resource scheduling strategy for Hadoop speculative execution based on non-cooperative game schemes
Nzanywayingoma et al. Task scheduling and virtual resource optimising in Hadoop YARN-based cloud computing environment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant