CN115145734B - Data processing system for distributing GPU - Google Patents

Data processing system for distributing GPU Download PDF

Info

Publication number
CN115145734B
CN115145734B CN202211063127.9A CN202211063127A CN115145734B CN 115145734 B CN115145734 B CN 115145734B CN 202211063127 A CN202211063127 A CN 202211063127A CN 115145734 B CN115145734 B CN 115145734B
Authority
CN
China
Prior art keywords
gpu
space
value
target
micro
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202211063127.9A
Other languages
Chinese (zh)
Other versions
CN115145734A (en
Inventor
赵洲洋
于伟
靳雯
石江枫
王全修
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Rizhao Ruian Information Technology Co ltd
Beijing Rich Information Technology Co ltd
Original Assignee
Rizhao Ruian Information Technology Co ltd
Beijing Rich Information Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Rizhao Ruian Information Technology Co ltd, Beijing Rich Information Technology Co ltd filed Critical Rizhao Ruian Information Technology Co ltd
Priority to CN202211063127.9A priority Critical patent/CN115145734B/en
Publication of CN115145734A publication Critical patent/CN115145734A/en
Application granted granted Critical
Publication of CN115145734B publication Critical patent/CN115145734B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • G06F9/5072Grid computing

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention provides a data processing system for distributing a GPU (graphics processing Unit), which is characterized in that a micro-service information set is obtained, wherein the micro-service information set comprises the minimum value of the GPU storage space required to be occupied by a micro-service and the minimum value of the CPU storage space required to be occupied by the micro-service, a target GPU list is obtained, the residual storage space of the target GPU is larger than the minimum value of the GPU storage space required to be occupied by the micro-service, the residual storage space of the CPU of a processor where the target GPU is located is larger than the minimum value of the CPU storage space required to be occupied by the micro-service, a scheduling priority index list of the target GPU is obtained, the GPU corresponding to the maximum value in the scheduling priority index list is selected as the GPU to be operated with the micro-service, and in the process of obtaining the scheduling priority index list of the target GPU, therefore, the operation speed of the micro-service is ensured, the operation efficiency of the micro-service is improved, and time resources are saved.

Description

Data processing system for distributing GPU
Technical Field
The invention relates to the technical field of GPU scheduling, in particular to a data processing system for distributing a GPU.
Background
In the prior art, when a multi-core Central Processing Unit (CPU) and a Graphics Processing Unit (GPU) cooperatively process a large amount of micro services, in a conventional graphics Processing task scheduling method, a processor generally schedules a GPU to run a micro service according to conditions that a CPU storage space occupied by the micro service and a GPU storage space occupied by the micro service are important, so that only a storage condition of the GPU and a storage condition of the CPU are considered, but a current load of the GPU is not considered, when a remaining storage space of the GPU is large, but the load of the GPU is high, an operation speed of the micro service is not guaranteed, and an operation efficiency of the GPU is reduced.
Disclosure of Invention
Aiming at the technical problems, the technical scheme adopted by the invention is as follows:
a data processing system that allocates GPUs, comprising: a database, a processor, and a memory storing a computer program, wherein the database comprises: initial GPU list G = { G 1 ,……,G i ,……,G n },G i =(G i 0 ,X i ,C i ,L i ),G i 0 Is ID, X of ith initial GPU i Is G i 0 Corresponding size of the first space, C i Is G i 0 Corresponding size of the second space, L i Is G i 0 And the corresponding utilization rate, i is 1 to n, n is the number of the GPUs, the first space is the storage space of the GPU, the second space is the storage space of the CPU, and when a computer program is executed by a processor, the following steps are realized:
s100, acquiring a target micro-service information set F 0 =(X 0 min ,C 0 min ) Wherein F is 0 Is a target microservice ID, X 0 min Is F 0 Corresponding minimum value of first space required to be occupied, C 0 min Is F 0 The corresponding minimum value of the second space required to be occupied;
s200, based on X i And X 0 min Obtaining an intermediate GPU list G ʹ = { G ʹ = G ʹ 1 ,……,Gʹ e ,……,Gʹ h },Gʹ e =(Gʹ e 0 ,Xʹ e ,Cʹ e ,Lʹ e ),Gʹ e 0 Is the ID of the e-th intermediate GPU, X ʹ e Is G ʹ e 0 Corresponding to the size of the first space, C ʹ e Is G ʹ e 0 Corresponding to the size of the second space, L ʹ e Is G ʹ e 0 The value of e is 1 to h, h is the number of intermediate GPUs which satisfy X ʹ e >X 0 min An initial GPU of the condition;
s300, C ʹ -based e And C 0 min Acquiring a target GPU list A = { A = { (A) } 1 ,……,A r ,……,A s },A r =(A r 0 ,X r min ,C r min ,L r ),A r 0 Is ID, X of the r-th target GPU r Is A r 0 Corresponding size of the first space, C r Is A r 0 Corresponding size of the second space, L r Is A r 0 The corresponding utilization value, r is 1 s, s is the number of target GPUs, and the target GPUs satisfy C r >C 0 min A conditional intermediate GPU;
s400, according to the A and the micro service information set, obtaining a GUP scheduling priority list R = (R) 1 ,……,R r ,……R s ),R r Scheduling priority, R, for the R-th GPU r The following conditions are met:
Figure 132514DEST_PATH_IMAGE001
wherein E is 0 r Is A r X ʹ r Is the r critical first space value, C ʹ r Is the r critical second spatial value, X 0 max Is F 0 Corresponding maximum value of first space required to be occupied, C 0 max Is F 0 The maximum value of the corresponding second space required to be occupied, w1 is a first weight, w2 is a second weight, and w3 is a third weight;
s500, traversing R, and obtaining the maximum value R from R max Corresponding target GPU as run F 0 The key GPU of (2).
The invention has at least the following beneficial effects: the method comprises the steps of obtaining a target GPU list by obtaining a micro-service information set, wherein the micro-service information set comprises the minimum value of GPU storage space required to be occupied by micro-service and the minimum value of CPU storage space required to be occupied by the micro-service, obtaining a scheduling priority index list of the target GPU, selecting the GPU corresponding to the maximum value in the scheduling priority index list as the GPU to be operated with the micro-service, and in the process of obtaining the scheduling priority index list of the target GPU, not only considering the residual conditions of display memory and internal memory, but also considering the average utilization value of the GPU, so that the operation speed of the micro-service is guaranteed, the operation efficiency of the micro-service is improved, and time resources are saved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a schematic flowchart illustrating a computer program executed by a data processing system for allocating GPUs according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be obtained by a person skilled in the art without inventive step based on the embodiments of the present invention, are within the scope of protection of the present invention.
It should be noted that the terms "first," "second," and the like in the description and claims of the present invention and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the invention described herein are capable of operation in sequences other than those illustrated or described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or server that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
The embodiment of the invention provides a data processing system for distributing GPU, which comprises: a database, a processor, and a memory storing a computer program, wherein the database comprises: initial GPU list G = { G 1 ,……,G i ,……,G n },G i =(G i 0 ,X i ,C i ,L i ),G i 0 Is ID, X of ith initial GPU i Is G i 0 Corresponding size of the first space, C i Is G i 0 Corresponding size of the second space, L i Is G i 0 And correspondingly, the value of i is 1 to n, n is the number of the GPUs, the first space is the storage space of the GPU, and the second space is the storage space of the CPU.
Specifically, the ID of the GPU is a unique identifier of the GPU, and the identifier of the GPU may be an index number of the GPU or a user-defined serial number.
Further, the processor includes a GPU shared device plugin for pair G i 0 Query to obtain X i
Further, the GPU sharing device plug-in queries G through nvml library i 0 The remaining storage space, as known by those skilled in the art, any method for querying the remaining storage space of the GPU through the nvml library falls within the scope of the present invention, and is not described herein again.
When executed by a processor, the computer program performs the following steps, as shown in fig. 1:
s100, acquiring a target micro-service information set F 0 =(X 0 min ,C 0 min ) Wherein F is 0 Is a target microservice ID, X 0 min Is F 0 Corresponding minimum value of first space required to be occupied, C 0 min Is F 0 A corresponding minimum value of the second space required to be occupied.
Specifically, the target microservice ID is a unique identity of the target microservice, where the target microservice is a to-be-processed computing task that needs to be run using a GPU.
Further, said X 0 min Corresponding units and said C 0 min The corresponding units are the same.
Further, said X 0 min The corresponding unit is GB (Gigabyte).
Further, before S100, the following steps are also included:
s110, obtaining G i ID list P = { P } of corresponding target CPU 1 ,……,P i ,……,P n },P i Is G i The ID of the corresponding target CPU.
Specifically, the G group i ID of corresponding target CPU is G i The ID of the CPU corresponding to the processor.
Further, the ID of the target CPU is a unique identifier of the CPU, and the CPU identifier may be a factory-owned identifier of the CPU or a user-defined serial number.
S130, according to P, obtaining C i
Specifically, the processor further comprises a CPU sharing device plug-in, which is a view CPU utilization performance tool, for example, ps and top; the CPU shared device plug-in is used for inquiring P i The remaining memory space, those skilled in the art will appreciate that any method for querying the remaining memory space of the CPU falls within the scope of the present invention.
Preferably, P is queried using ps i On one hand, the occupation condition of the total storage space of the CPU can be inquired about by the residual storage space, and on the other hand, the occupation condition of the storage space of any micro-service in the CPU can be inquired about.
S200, based on X i And X 0 min Obtaining an intermediate GPU list G ʹ = { G ʹ = 1 ,……,Gʹ e ,……,Gʹ h },Gʹ e =(Gʹ e 0 ,Xʹ e ,Cʹ e ,Lʹ e ),Gʹ e 0 Is the ID of the e-th intermediate GPU, X ʹ e Is G ʹ e 0 Corresponding to the size of the first space, C ʹ e Is G ʹ e 0 Corresponding to the size of the second space, L ʹ e Is G ʹ e 0 The value of e is 1 to h, h is the number of intermediate GPUs which meet the requirement of X ʹ e >X 0 min Initial GPU of the condition.
S300, C ʹ -based e min And C 0 min Acquiring a target GPU list A = { A = { (A) } 1 ,……,A r ,……,A s },A r =(A r 0 ,X r ,C r ,L r ),A r 0 Is ID, X of the r-th target GPU r Is A r 0 Corresponding size of the first space, C r Is A r 0 Corresponding size of the second space, L r Is A r 0 The value of r is 1 s, s is the number of target GPUs, and the target GPUs satisfy C r >C 0 min Conditional intermediate GPU.
S400, according to the A and the micro service information set, obtaining a GUP scheduling priority list R = (R) 1 ,……,R r ,……R s ),R r For scheduling priority, R, corresponding to the R-th GPU r The following conditions are met:
Figure 610769DEST_PATH_IMAGE002
wherein E is 0 r Is A r X ʹ r Is the r critical first space value, C ʹ r Is the r critical second spatial value, X 0 max Is F 0 Corresponding maximum value of first space required to be occupied, C 0 max Is F 0 Corresponding second space required to be occupiedW1 is the first weight, w2 is the second weight, and w3 is the third weight.
Specifically, w1+ w2+ w3=1.
Further, w1 is greater than w2 is greater than w3, and when the first space size and the second space size of the GPU both meet the micro-service set operation condition, the current residual value of the GPU determines the operation efficiency of the GPU for operating the target micro-service, so that w1 is greater than w2 and greater than w3, the weight of the utilization value is set to be the maximum, the target micro-service can select a more appropriate GPU to operate, and the work efficiency of the GPU is improved.
Further, the step S400 further includes the following steps:
s401, obtaining A r Corresponding target first spatial value list X 0 =(X 0 1 ,……,X 0 r ,……,X 0 s ),X 0 r The following conditions are met:
Figure 533595DEST_PATH_IMAGE003
wherein, X 0 rg Is A r At a preset time period T 0 The size of the first space of the middle-g time node, the value of g is 1 to z, and z is a preset time period T 0 The number of medium time nodes.
Specifically, the A is r The corresponding target video memory is A r At a preset time T 0 Corresponding average remaining memory space.
Further, T 0 =48h。
Further, z satisfies the following condition:
z=T 0 /t;
wherein t is a first preset query time threshold.
Specifically, the first preset query time threshold refers to a preset time T 0 Once per query A r The value of t can be set by those skilled in the art according to the actual situation, and is not described herein again.
S402, acquiring F according to the micro-service information set 0 Corresponding maximum value X of first space required to be occupied 0 max
Specifically, the X 0 max Can be understood as being in operation F 0 In the process of (1), F 0 As one skilled in the art knows, any method for obtaining the maximum value of the GPU storage space required by the microservice falls within the protection scope of the present invention, and is not described herein again.
S403, mixing X 0 And X 0 max And comparing to obtain a key first space value list X '= (X' 1 ,……,X' r ,……,X' s ),X' r Is the r-th key first space value, X' r =min(X 0 r ,X 0 max );
S404, obtaining A r Corresponding target second spatial value list C 0 =(C 0 1 ,……,C 0 r ,……,C 0 s ),C 0 r The following conditions are met:
Figure 388287DEST_PATH_IMAGE004
wherein, C rx Is A r At a preset time period T 0 The size of the second space of the xth time node, the value range of x is 1 to q, and q is a preset time period T 0 The number of medium time nodes.
Specifically, q satisfies the following condition:
q=T 0 /t';
wherein t' is a second preset query time threshold.
Specifically, the second preset query time threshold refers to a preset time T 0 Once per query A r Corresponding to the interval time of the second space, a person skilled in the art may set the value of t' according to the actual situation, which is not described herein again.
Preferably, t = t', the second preset query time is set to be the same as the first preset query time, so that the query frequency of the GPU and the CPU can be the same, thereby avoiding errors caused by different query frequencies and improving the accuracy.
S405, according to the micro service information set, obtaining F 0 Corresponding maximum value C of the second space required to be occupied 0 max
In particular, said C 0 max Can be understood as being in operation F 0 In the process of (1), F 0 The upper limit value of the CPU storage space required to be occupied, and those skilled in the art know that any method for obtaining the maximum value of the CPU storage space required to be occupied by the microservice falls within the protection scope of the present invention, and is not described herein again.
S406, adding C 0 And C max j And comparing to obtain a key second space value list C '= (C' 1 ,……,C' r ,……,C' s ),C' r Is the r critical second space value, C' r =min(C 0 ,C max j );
S407, obtaining GUP scheduling priority index L r
In the above, in S401 to S407, the GUP scheduling priority index is obtained by obtaining the target video memory list and the target memory list corresponding to the GPU, the maximum value of the GPU storage space required to be occupied by the micro service, the maximum value of the CPU storage space required to be occupied by the micro service, and the average GPU utilization rate, and according to the above data, calculating.
S500, traversing R, and obtaining the maximum value R from L max Corresponding target GPU as run F 0 The key GPU of (1).
The invention provides a data processing system for distributing a GPU (graphics processing Unit), which is characterized in that a micro-service information set is obtained, wherein the micro-service information set comprises the minimum value of the GPU storage space required to be occupied by a micro-service and the minimum value of the CPU storage space required to be occupied by the micro-service, a target GPU list is obtained, the residual storage space of the target GPU is larger than the minimum value of the CPU storage space required to be occupied by the micro-service, the residual storage space of the CPU of a processor where the target GPU is located is larger than the minimum value of the CPU storage space required to be occupied by the micro-service, a scheduling priority index list of the target GPU is obtained, the GPU corresponding to the maximum value in the scheduling priority index list is selected as the GPU to be operated with the micro-service, and in the process of obtaining the scheduling priority index list of the target GPU, the residual conditions of a display memory and a memory are considered, more average utilization values of the GPU are considered, so that the operation speed of the micro-service is ensured, the operation efficiency of the micro-service is improved, and time resources are saved.
The present specification provides method steps as described in the examples or flowcharts, but may include more or fewer steps based on routine or non-inventive labor. The order of steps recited in the embodiments is merely one manner of performing the steps in a multitude of orders and does not represent the only order of execution. In actual system or server product execution, sequential execution or parallel execution (e.g., parallel processor or multithreaded processing environments) may occur according to the embodiments or methods shown in the figures.
The embodiments in the present specification are described in a progressive manner, and the same and similar parts among the embodiments are referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, for the apparatus and computer device embodiments, since they are substantially similar to the method embodiments, the description is relatively simple, and reference may be made to some descriptions of the method embodiments for relevant points.
Although some specific embodiments of the present invention have been described in detail by way of example, it should be understood by those skilled in the art that the above examples are for illustration only and are not intended to limit the scope of the invention. It will also be appreciated by those skilled in the art that various modifications may be made to the embodiments without departing from the scope and spirit of the invention. The scope of the invention is defined by the appended claims.

Claims (8)

1. Data processing system for distributing GPUCharacterized in that said system comprises: a database, a processor, and a memory storing a computer program, wherein the database comprises: initial GPU list G = { G 1 ,……,G i ,……,G n },G i =(G i 0 ,X i ,C i ,L i ),G i 0 Is ID, X of ith initial GPU i Is G i 0 Corresponding size of the first space, C i Is G i 0 Corresponding size of the second space, L i Is G i 0 And correspondingly, the utilization rate, i is 1 to n, n is the number of GPUs, the first space is the storage space of the GPU, the second space is the storage space of the CPU, and when the computer program is executed by a processor, the following steps are realized:
s100, acquiring a target micro-service information set F 0 =(X 0 min ,C 0 min ) Wherein F is 0 Is a target microservice ID, X 0 min Is F 0 Corresponding minimum value of first space required to be occupied, C 0 min Is F 0 The corresponding minimum value of the second space required to be occupied;
s200, based on X i And X 0 min Obtaining an intermediate GPU list G ʹ = { G ʹ = G ʹ 1 ,……,Gʹ e ,……,Gʹ h },Gʹ e =(Gʹ e 0 ,Xʹ e ,Cʹ e ,Lʹ e ),Gʹ e 0 Is the ID of the e-th intermediate GPU, X ʹ e Is G ʹ e 0 Corresponding to the size of the first space, C ʹ e Is G ʹ e 0 Corresponding to the size of the second space, L ʹ e Is G ʹ e 0 The value of e is 1 to h, h is the number of intermediate GPUs which meet the requirement of X ʹ e >X 0 min An initial GPU of the condition;
s300, C ʹ -based e And C 0 min Obtaining a target GPU list a = { a = 1 ,……,A r ,……,A s },A r =(A r 0 ,X r ,C r ,L r ),A r 0 Is ID, X, of the r-th target GPU r Is A r 0 Corresponding size of the first space, C r Is A r 0 Corresponding size of the second space, L r Is A r 0 The value of r is 1 to s, s is the number of target GPUs, and the target GPUs satisfy C r >C 0 min A conditional intermediate GPU;
s400, according to the A and the micro service information set, obtaining a GUP scheduling priority list R = (R) 1 ,……,R r ,……R s ),R r For scheduling priority, R, corresponding to the R-th GPU r The following conditions are met:
Figure DEST_PATH_IMAGE002
wherein E is 0 r Is A r Average utilization value of X ʹ r Is the r critical first space value, C ʹ r Is the r critical second spatial value, X 0 max Is F 0 Corresponding maximum value of first space required to be occupied, C 0 max Is F 0 The maximum value of the corresponding second space required to be occupied, w1 is a first weight, w2 is a second weight, and w3 is a third weight;
s500, traversing R, and obtaining a maximum value R from R max Corresponding target GPU as run F 0 The key GPU of (1).
2. The system according to claim 1, further comprising the following steps in S400:
s401, obtaining A r Corresponding target first spatial value list X 0 =(X 0 1 ,……,X 0 r ,……,X 0 s ),X 0 r The following conditions are met:
Figure DEST_PATH_IMAGE004
wherein, X rg Is A r At a preset time period T 0 The size of the first space of the middle-g time node, the value of g is 1 to z, and z is a preset time period T 0 The number of medium time nodes;
s402, obtaining F according to the micro service information set 0 Corresponding maximum value X of the first space required to be occupied 0 max
S403, mixing X 0 And X 0 max By contrast, a key first spatial value list X ʹ = (X ʹ) is obtained 1 ,……,Xʹ r ,……,Xʹ s ),Xʹ r For the r-th critical first space value, X ʹ r =min(X 0 ,X 0 max );
S404, obtaining A r Corresponding target second spatial value list C 0 =(C 0 1 ,……,C 0 r ,……,C 0 s ),C 0 r The following conditions are met:
Figure DEST_PATH_IMAGE006
wherein, C rx Is A r At a preset time period T 0 The size of the second space of the xth time node, the value range of x is 1 to q, and q is a preset time period T 0 The number of medium time nodes;
s405, according to the micro service information set, obtaining F 0 Corresponding maximum value C of second space required to be occupied 0 max
S406, adding C 0 And C 0 max By contrast, a key second spatial value list C ʹ = (C ʹ) is obtained 1 ,……,Cʹ r ,……,Cʹ s ),Cʹ r Is the r critical second space value, C ʹ r =min(C 0 ,C 0 max );
S407, obtaining a GUP scheduling priority index L r
3. The system of claim 1, wherein w1+ w2+ w3=1.
4. The system of claim 3, wherein w1 > w2 > w3.
5. The system of claim 2, wherein z satisfies the following condition:
z=T 0 /t;
wherein t is a first preset query time threshold.
6. The system of claim 2, wherein q satisfies the following condition:
q=T 0 /tʹ;
wherein t' is a second preset query time threshold.
7. The system of claim 6, wherein t = t ʹ.
8. The system of claim 1, wherein in S400, E 0 r The following conditions are met:
Figure DEST_PATH_IMAGE008
wherein E is rg Is A r At a preset time period T 0 The utilization value of the g-th time node.
CN202211063127.9A 2022-08-31 2022-08-31 Data processing system for distributing GPU Active CN115145734B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211063127.9A CN115145734B (en) 2022-08-31 2022-08-31 Data processing system for distributing GPU

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211063127.9A CN115145734B (en) 2022-08-31 2022-08-31 Data processing system for distributing GPU

Publications (2)

Publication Number Publication Date
CN115145734A CN115145734A (en) 2022-10-04
CN115145734B true CN115145734B (en) 2022-11-25

Family

ID=83415577

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211063127.9A Active CN115145734B (en) 2022-08-31 2022-08-31 Data processing system for distributing GPU

Country Status (1)

Country Link
CN (1) CN115145734B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110806930A (en) * 2019-10-29 2020-02-18 浙江大华技术股份有限公司 Micro-service scheduling method, device, equipment and storage device
CN111324471A (en) * 2020-01-22 2020-06-23 远景智能国际私人投资有限公司 Service adjusting method, device, equipment and storage medium
CN111367642A (en) * 2020-03-09 2020-07-03 中国铁塔股份有限公司 Task scheduling execution method and device
US11196641B1 (en) * 2021-01-08 2021-12-07 Optumi, Inc. Using cost and performance constraint information to identify networked equipment on which to place a computer software application

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10223109B2 (en) * 2016-12-22 2019-03-05 Juniper Networks, Inc. Automatic scaling of microservices applications

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110806930A (en) * 2019-10-29 2020-02-18 浙江大华技术股份有限公司 Micro-service scheduling method, device, equipment and storage device
CN111324471A (en) * 2020-01-22 2020-06-23 远景智能国际私人投资有限公司 Service adjusting method, device, equipment and storage medium
CN111367642A (en) * 2020-03-09 2020-07-03 中国铁塔股份有限公司 Task scheduling execution method and device
US11196641B1 (en) * 2021-01-08 2021-12-07 Optumi, Inc. Using cost and performance constraint information to identify networked equipment on which to place a computer software application

Also Published As

Publication number Publication date
CN115145734A (en) 2022-10-04

Similar Documents

Publication Publication Date Title
CN111046045B (en) Method, device, equipment and storage medium for processing data inclination
US20050177833A1 (en) Method and apparatus for reassigning objects to processing units
EP1253516A2 (en) Apparatus and method for scheduling processes on a fair share basis
EP3451727A1 (en) Access scheduling method and device for terminal, and computer storage medium
US7725900B2 (en) Method of assigning objects to processing units
JP6129290B1 (en) Method and system for recommending application parameter settings and system specification settings in distributed computing
US8549058B2 (en) Multi-dimensional transform for distributed memory network
CN112506650A (en) Resource allocation method, system, computer device and storage medium
US7664858B2 (en) Method for balancing load between processors in a multi-processor environment
CN114036031B (en) Scheduling system and method for resource service application in enterprise digital middleboxes
US20130166751A1 (en) Distributed resource management systems and methods for resource management thereof
CN115145734B (en) Data processing system for distributing GPU
CN114327884A (en) Automatic capacity expansion and reduction method, device, equipment and readable storage medium
US7647592B2 (en) Methods and systems for assigning objects to processing units
JP5577745B2 (en) Cluster system, process allocation method, and program
CN114338694B (en) One-stop cloud data center server scheduling method and system
US20170346889A1 (en) Co-locating application instances
Borst et al. Task allocation in a multi-server system
JP2004302525A (en) Device and method for carrying out performance balance evaluation and sizing of component of computer system
CN115629853A (en) Task scheduling method and device
CN102656564B (en) Reduce the expense in application process
CN115658292A (en) Resource scheduling method, device, computer equipment and storage medium
JP4594877B2 (en) Computer resource allocation management method and computer resource allocation management apparatus
CN110046040B (en) Distributed task processing method and system and storage medium
CN113608847A (en) Task processing method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant