CN106874113A - A kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+ - Google Patents

A kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+ Download PDF

Info

Publication number
CN106874113A
CN106874113A CN201710037124.0A CN201710037124A CN106874113A CN 106874113 A CN106874113 A CN 106874113A CN 201710037124 A CN201710037124 A CN 201710037124A CN 106874113 A CN106874113 A CN 106874113A
Authority
CN
China
Prior art keywords
gpu
cut
parallel
kernel
cpu
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710037124.0A
Other languages
Chinese (zh)
Inventor
陆娟娟
温柏坚
王毅
陆进军
郭文鑫
闪鑫
彭龙
查国强
王彬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nari Technology Co Ltd
NARI Nanjing Control System Co Ltd
Electric Power Dispatch Control Center of Guangdong Power Grid Co Ltd
Original Assignee
Nari Technology Co Ltd
NARI Nanjing Control System Co Ltd
Electric Power Dispatch Control Center of Guangdong Power Grid Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nari Technology Co Ltd, NARI Nanjing Control System Co Ltd, Electric Power Dispatch Control Center of Guangdong Power Grid Co Ltd filed Critical Nari Technology Co Ltd
Priority to CN201710037124.0A priority Critical patent/CN106874113A/en
Publication of CN106874113A publication Critical patent/CN106874113A/en
Priority to PCT/CN2017/093226 priority patent/WO2018133348A1/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/16Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
    • G06F15/163Interprocessor communication
    • G06F15/167Interprocessor communication using a common memory, e.g. mailbox
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/06Electricity, gas or water supply
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/50Indexing scheme relating to G06F9/50
    • G06F2209/5018Thread allocation

Abstract

The invention discloses a kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+,For the demand that bulk power grid static security analysis in practical engineering application are quickly scanned,On CUDA unified calculation framework platforms,According to system GPU configuring conditions and calculating demand,Its respective thread number is distributed using OpenMp multithreadings,Each thread is uniquely corresponding with single GPU,Developed based on CPU and GPU hybrid programmings,Build many GPU Heterogeneous Computings pattern coordinateds of CPU+ and complete forecast failure parallel computation,On the basis of single forecast failure Load flow calculation,Realize that multiple cut-offs trend iterative process high level of synchronization parallel,Static security analysis forecast failure is greatly improved by Element-Level fine grained parallel and scans parallel processing capability,For the scanning of interconnected power grid integrated dispatching system safety on line analysis and early warning provides strong technical support.

Description

A kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+
Technical field
It is quiet the present invention relates to Automation of Electric Systems dispatching technique field, more particularly to a kind of many GPU heterogeneous schemas of CPU+ State safety analysis computational methods.
Background technology
With the fast development of electric power system dispatching integrated technique, power network calculation scale expands increasingly.In the whole network state Estimate to carry out static security analysis calculating on the basis of section, calculating speed will be difficult to meet demand.On the one hand, it is necessary to be cut-off The forecast accident number of analysis significantly increases, even if being screened by direct current method, it is necessary to cut-off the forecast failure number of scanning in detail Mesh also will be considerable.On the other hand, each forecast accident is carried out cut-offfing Tidal Simulation calculating, and the increase of calculate node number is caused Single Load flow calculation speed is slow.Additionally, the expansion of power network scale causes to need the number of devices of out-of-limit monitoring and stable cross section number big Amplitude increases, and amount of calculation persistently increases.Therefore, modern power systems are developed rapidly, should to the safety analysis of scheduling static system Acid test is proposed with performance issue computationally.Conventional serial algorithm is combined using direct current method screening and is based on sparse square The means such as the local factors decomposition of battle array carry out serial engineering treatment, filter noncritical failure, effectively improve single and cut-off analysis Speed, but still engineer applied demand cannot be met.
Solution procedure finds, mutually decoupled between static security analysis forecast failure it is separate, can be analyzed to it is multiple solely Vertical full Load flow calculation, innate advantage is processed with parallelization, and this characteristic is gradually paid close attention to by researcher, and begins attempt to Using parallel algorithm seeking breakthrough.
In recent years, computer realm is tall and handsome proposes CUDA (Compute Unified Device up to company Architecture) framework, be GPU (Graphic Processing Unit, graphic process unit) development provide it is good and Row Computational frame, the research field can suitable for power system with parallel characteristics.Meanwhile, in CPU fields, in shared The OpenMP multithreadings of parallel system are deposited, it is automatic by loop parallelization by compiler, multiprocessor system can be effectively improved System application program capacity.
To sum up, for the quick scanning application scenarios of static security analysis, how effectively to use GPU parallel and the parallel skills of CPU Art, meets engineering site to dispatch automated system static security analysis computation scale and the higher and higher need of computational efficiency Ask, be the research direction probed into worth thinking deeply about.
The content of the invention
In view of the drawbacks described above of prior art, the technical problems to be solved by the invention are to provide a kind of many GPU of CPU+ Heterogeneous schemas static security analysis computational methods, for the need that bulk power grid static security analysis in practical engineering application are quickly scanned Ask, on CUDA unified calculation framework platforms, according to system GPU configuring conditions and calculating demand, using OpenMp multithreading skills Art distributes its respective thread number, and each thread is uniquely corresponding with single GPU, is developed based on CPU and GPU hybrid programmings, builds CPU+ Many GPU Heterogeneous Computings pattern coordinateds complete forecast failure parallel computation, real on the basis of single forecast failure Load flow calculation Existing multiple cut-offs trend iterative process high level of synchronization parallel, greatly improves static security analysis by Element-Level fine grained parallel pre- Fault scanning parallel processing capability is thought, for the scanning of interconnected power grid integrated dispatching system safety on line analysis and early warning has been provided The technical support of power.
To achieve the above object, the invention provides a kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+, Characterized in that, on CUDA unified calculation platforms, it is comprehensive using the OpenMP multithreadings based on shared drive parallel system Close and consider that GPU configuring conditions and forecast failure calculate demand, determine CPU line number of passes mesh, build many GPU heterogeneous schemas of CPU+, often Parallel scan task is completed according to the mono- GPU patterns coordinateds of CPU+ inside block GPU card, in single forecast failure Load flow calculation base On plinth, realize that multiple cut-offs trend iterative process high level of synchronization parallel, static peace is greatly improved by Element-Level fine grained parallel Complete analysis forecast failure scans parallel processing capability.
The many GPU heterogeneous schemas static security analysis computational methods of above-mentioned a kind of CPU+, it is characterised in that including following step Suddenly:
1) the real-time section of state estimation is obtained, ground state Load flow calculation is carried out based on Newton-Raphson method, there is provided one can For the shared data section of multiplexing;
2) according to user's request, topologically sweeping is carried out to whole network equipment, forecast failure collection is formed, for each forecast failure Depth topology search is carried out, branch road-nodal information is actually cut-off in formation;
3) CUDA frameworks are initialized, according to cut-off parallel computation demand carries out packing treatment by data, and is distributed in GPU more Deposit space;
4) according to GPU configuring conditions and actual amount of calculation, rationalization distribution is carried out to cut-offfing computing resource, assessment needs are opened GPU numbers, generate corresponding Thread Count, at utmost to play GPU computation capabilities using OpenMP technologies;
5) under the pool of CPU, Load flow calculation iterative process, including the solution of admittance battle array amendment Jacobian matrix, amendment side Journey is solved, quantity of state updates, Branch Power Flow is calculated and the out-of-limit check of equipment, all transfers to GPU to complete parallel, all to cut-off failure Iteration highly-parallel, the parallel characteristics according to each step calculating task construct corresponding kernel function, in fine grained parallel aspect Elements in parallel calculating task in upper completion iterative process;
6) judge whether all scanning is completed forecast failure, if it is not, then being transferred to step 4), cut-off failure weight to remaining Resource allocation newly is carried out, if be fully completed, step 7 is transferred to;
7) result displaying, according to scanning result, the out-of-limit heavily loaded information of equipment or section caused to failure on-off is opened up Show, and result statistics is scanned according to the scheduling static system practical demand of safety analysis module.
The many GPU heterogeneous schemas static security analysis computational methods of above-mentioned a kind of CPU+, it is characterised in that the step 4) Specifically include:
(1) according to break calculation demand and GPU configuring conditions, computing resource is cut-off in optimization distribution, assesses monolithic GPU singles Number is cut-off in maximum scan:
Wherein, S_max be the total memory headroom of GPU card, S memory sizes for needed for single Load flow calculation, M_max is monolithic The maximum that GPU card can be calculated cut-offs number;
According to formula (1), the GPU card number started needed for calculating is:
Wherein, α is expressed as rounding, the GPU card number that n starts for needed for epicycle, and N is the GPU card sum of system configuration, M_ What cal was calculated for needed for epicycle static security analysis cut-offs sum, and M_maxN once calculates connecting-disconnecting function for many GPU;
According to formula (2), according to Principle of Average Allocation, every piece of actual calculating of card cut-offs number:
As M_maxN≤M_cal, it is meant that GPU all participates in calculating and cannot once complete all break calculations, then will Residue cut-offs M_cal ' carries out new round distribution according to formula (2-3) again, repeatedly calculates in batches, and computing formula is:
M_cal '=M_cal-M_maxN (4)
(2) in the ground state sectional model data basis of shared multiplexing, based on OpenMP multithreadings according to actual fortune Row GPU card number n distributes corresponding CPU line number of passes, and each CPU line journey is uniquely corresponding with single GPU, builds many GPU of CPU+ different Structure pattern, concurrent operation is carried out in single GPU card according to the mono- GPU heterogeneous schemas of CPU+.
The many GPU heterogeneous schemas static security analysis computational methods of above-mentioned a kind of CPU+, it is characterised in that the step 5) Comprise the following steps:
(1) on the basis of ground state admittance battle array Y0, branch road-nodal information is cut-off according to N-1, defines kernel function fun_ Kernel_1, failure is cut-off while carrying out node admittance battle array local correction work, and change node injection to distributing to the card Information, is cut-off by the change of admittance battle array and nodal information come analog machine;
(2) formed according to (1) and cut-off admittance battle array Ym, Jacobian matrix elements in parallel calculating task is completed, for ease of calling CUDA kernel functions, each element uses and formula is calculated as below in Load flow calculation Jacobi matrix in block form:
Wherein, Gii,Bii,Gij,BijRespectively admittance battle array non-zero entry, θijIt is i, j node phase angle differences, Pi,QiFor i-node is noted Enter power, Vi,VjRespectively i, j node voltage amplitude;
Matrix in block form each element is quantity of state node voltage phase angle, and the four fundamental rules of admittance array element element and trigonometric function are transported substantially Calculate and solution procedure is independent of each other, with obvious parallel characteristics;
4 kernel function correspondences being defined respectively and calculating each element value in matrix in block form, in single deck tape-recorder, m is cut-off refined gram of failure Than H-matrix amount of calculation:
H=m × hnozero (9)
Wherein, hnozeroIt is H gusts of non-zero entry number, m is to cut-off number;
In CUDA frameworks, GPU has multiple SM stream handles, it is possible to provide multiple thread dispatchings.Do not have between h element Any coupled relation, the Element-Level fine grained parallel characteristic with height defines kernel function fun_kernel_2_H by formula (5) Perfect vector multiplies and vectorial add operation, and CPU calls fun_kernel_2_H functions, and GPU is incoming according to fun_kernel_2_H Function parameter enables h GPU multithreading, while perfect (5) parallel work-flow;
(3) admittance battle array Ym and start node injection rate are cut-off according to what (1) was formed, according to formula (10)
Carry out node injecting power residual computations:
Wherein, Pis,QisRespectively active, the idle injection rate of node i, as can be seen from the above equation, each node injection rate is residual Difference is quantity of state node voltage phase angle, and the four fundamental rules basic operation of admittance array element element and trigonometric function is residual with other node powers Difference solution procedure is mutually independent of, and defines fun_kernel_3 kernel functions, enables multiple threads in SM stream handles, completes m The parallel computation task of all calculate node unbalanced power amounts of failure is cut-off, detailed parallel computation process is similar to step (2);
(4) check whether power residual error meets convergence basis for estimation, if it is satisfied, then (8) are jumped to, if be unsatisfactory for, Then continue to be transferred to (5) and be iterated;
First, fun_kernel_4 kernel functions are defined, GPU is restrained to respectively cut-offfing node power residual error by formula (11) Check, if certain cut-offs meets the condition of convergence, record this and cut-off convergence;
||ΔPt,ΔQt| | < ε t≤T (11)
Wherein, ε is power convergence criterion, Δ Pt,ΔQtThe respectively power deviation of iteration t times, T is maximum iteration;
Again, inconsistent due to respectively cut-offfing failure convergence iterative steps, setting k >=80% cut-offs failure and has restrained, then tie Shu Chaoliu iterative process, otherwise all failures of cut-offfing proceed trend iteration, until meeting k >=80% condition of convergence or reaching To maximum iteration, step (8) is jumped to;
(5) Jacobian matrix formed to (2) carries out LU decomposition, and the node power residual error calculated with reference to (3) enters line Property solving equations, define fun_kernel_5 kernel functions, task-level parallelism solution is carried out to system of linear equations;
(6) Solution for System of Linear Equations as obtained by (5) is updated to initial state vector by formula (12)
In formula,Respectively before i-th flow equation state vector kth of system of linear equations time iteration, after iteration Value,It is kth time correction value, n is system node number, and r is PV node number;
From above formula, each knot vector is only related to the node increment before and after updating in iterative process, is independent of In other any node calculated values, with parallel natural attribute, can realize that node voltage is more newly appointed by GPU multi-threaded parallels Business;
Formula (12) is defined as the kernel function fun_kernel_6 of plus computing, is appointed parallel for M forecast failure Business, needs the individual thread synchronizations of M × (2 (n-1)-r) to perform fun_kernel_6 and completes a quantity of state renewal calculating altogether;
(7) (2) are jumped to and continues iteration;
(8) Fault load flow result of calculation is cut-off according to currently convergent, carries out Branch Power Flow calculating, due to Branch Power Flow only It is related to branch parameters and both sides node voltage value, mutually it is independent of, fun_kernel_7 kernel functions are defined, complete Branch Power Flow Parallel computation.
(9) according to (8) Branch Power Flow result of calculation, heavily loaded out-of-limit verification is carried out to each branch road or stable cross section, and preserve It is currently all to cut-off the heavily loaded out-of-limit result that failure is caused, fun_kernel_8 kernel functions are defined, complete branch road branch road or steady Determine the section out-of-limit verification parallel computation of heavy duty.
The beneficial effects of the invention are as follows:
CPU+ many GPU heterogeneous schemas static security analysis methods of the present invention based on OpenMP technologies, unify to count in CUDA Calculate on framework platform, according to system GPU configuring conditions and demand can be calculated, corresponding line is distributed using OpenMp multithreadings Number of passes, builds many GPU Heterogeneous Computings pattern coordinateds of CPU+ and completes forecast failure parallel computation, in single forecast failure trend On the basis of calculating, realize that multiple cut-offs trend iterative process high level of synchronization parallel, effectively lifting static security analysis forecast failure Scanning parallel processing capability, it is adaptable to which the integrated large scale system of scheduling quickly scans application to static security analysis forecast failure Scene, has very important application value to lifting bulk power grid static security analysis scanning computational efficiency.
The technique effect of design of the invention, concrete structure and generation is described further below with reference to accompanying drawing, with It is fully understood from the purpose of the present invention, feature and effect.
Brief description of the drawings
The many GPU heterogeneous schemas calculation flow charts of Fig. 1 CPU+;
The many GPU heterogeneous schemas structural representations of Fig. 2 CPU+;
Fig. 3 Jacobi matrix in block form H element parallel computation schematic diagrames.
Specific embodiment
It is an object of the invention to propose a kind of fast parallel method suitable for static security analysis.For Practical Project The demand that bulk power grid static security analysis are quickly scanned in, on CUDA unified calculation framework platforms, matches somebody with somebody according to system GPU Situation and calculating demand are put, its respective thread number is distributed using OpenMp multithreadings, each thread is uniquely right with single GPU Should, developed based on CPU and GPU hybrid programmings, build many GPU Heterogeneous Computings pattern coordinateds completion forecast failures of CPU+ parallel Calculate, on the basis of single forecast failure Load flow calculation, realize that multiple cut-offs trend iterative process high level of synchronization parallel, by unit Plain level fine grained parallel greatly improves static security analysis forecast failure scanning parallel processing capability, is interconnected power grid integration The scanning of scheduling system safety on line analysis and early warning provides strong technical support.
As shown in figure 1, a kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+, it is characterised in that in CUDA On unified calculation platform, using the OpenMP multithreadings based on shared drive parallel system, GPU configuring conditions are considered And forecast failure calculates demand, CPU line number of passes mesh is determined, many GPU heterogeneous schemas of CPU+ are built, according to CPU inside every piece of GPU card + mono- GPU patterns coordinated completes parallel scan task, on the basis of single forecast failure Load flow calculation, realizes that multiple is cut-off Trend iterative process high level of synchronization is parallel, and greatly improving static security analysis forecast failure by Element-Level fine grained parallel scans Parallel processing capability.
The many GPU heterogeneous schemas static security analysis computational methods of above-mentioned a kind of CPU+, it is characterised in that including following step Suddenly:
1) the real-time section of state estimation is obtained, ground state Load flow calculation is carried out based on Newton-Raphson method, there is provided one can For the shared data section of multiplexing;
2) according to user's request, topologically sweeping is carried out to whole network equipment, forecast failure collection is formed, for each forecast failure Depth topology search is carried out, branch road-nodal information is actually cut-off in formation;
3) CUDA frameworks are initialized, according to cut-off parallel computation demand carries out packing treatment by data, and is distributed in GPU more Deposit space;
4) according to GPU configuring conditions and actual amount of calculation, rationalization distribution is carried out to cut-offfing computing resource, assessment needs are opened GPU numbers, corresponding Thread Count is generated using OpenMP technologies, at utmost to play GPU computation capabilities, OpenMP technologies are a set of guiding process of compilation schemes for shared drive parallel system multithread programming;
5) under the pool of CPU, Load flow calculation iterative process, including the solution of admittance battle array amendment Jacobian matrix, amendment side Journey is solved, quantity of state updates, Branch Power Flow is calculated and the out-of-limit check of equipment, all transfers to GPU to complete parallel, all to cut-off failure Iteration highly-parallel, the parallel characteristics according to each step calculating task construct corresponding kernel function, in fine grained parallel aspect Elements in parallel calculating task in upper completion iterative process;
6) judge whether all scanning is completed forecast failure, if it is not, then being transferred to step 4), cut-off failure weight to remaining Resource allocation newly is carried out, if be fully completed, step 7 is transferred to;
7) result displaying, according to scanning result, the out-of-limit heavily loaded information of equipment or section caused to failure on-off is opened up Show, and result statistics is scanned according to the scheduling static system practical demand of safety analysis module.
In the present embodiment, the step 4) specifically include:
(1) according to break calculation demand and GPU configuring conditions, computing resource is cut-off in optimization distribution, assesses monolithic GPU singles Number is cut-off in maximum scan:
Wherein, S_max be the total memory headroom of GPU card, S memory sizes for needed for single Load flow calculation, M_max is monolithic The maximum that GPU card can be calculated cut-offs number;
According to formula (1), the GPU card number started needed for calculating is:
Wherein, α is expressed as rounding, the GPU card number that n starts for needed for epicycle, and N is the GPU card sum of system configuration, M_ What cal was calculated for needed for epicycle static security analysis cut-offs sum, and M_maxN once calculates connecting-disconnecting function for many GPU;
According to formula (2), according to Principle of Average Allocation, every piece of actual calculating of card cut-offs number:
As M_maxN≤M_cal, it is meant that GPU all participates in calculating and cannot once complete all break calculations, then will Residue cut-offs M_cal ' carries out new round distribution according to formula (2-3) again, repeatedly calculates in batches, and computing formula is:
M_cal '=M_cal-M_maxN (4)
(2) in the ground state sectional model data basis of shared multiplexing, based on OpenMP multithreadings according to actual fortune Row GPU card number n distributes corresponding CPU line number of passes, and each CPU line journey is uniquely corresponding with single GPU, builds many GPU of CPU+ different Structure pattern, as shown in Fig. 2 carrying out concurrent operation according to the mono- GPU heterogeneous schemas of CPU+ in single GPU card.
In the present embodiment, the step 5) comprise the following steps:
(1) on the basis of ground state admittance battle array Y0, branch road-nodal information is cut-off according to N-1, defines kernel function fun_ Kernel_1, failure is cut-off while carrying out node admittance battle array local correction work, and change node injection to distributing to the card Information, is cut-off by the change of admittance battle array and nodal information come analog machine;
(2) formed according to (1) and cut-off admittance battle array Ym, Jacobian matrix elements in parallel calculating task is completed, for ease of calling CUDA kernel functions, each element uses and formula is calculated as below in Load flow calculation Jacobi matrix in block form:
Wherein, Gii,Bii,Gij,BijRespectively admittance battle array non-zero entry, θijIt is i, j node phase angle differences, Pi,QiFor i-node is noted Enter power, Vi,VjRespectively i, j node voltage amplitude;
Matrix in block form each element is quantity of state node voltage phase angle, and the four fundamental rules of admittance array element element and trigonometric function are transported substantially Calculate and solution procedure is independent of each other, with obvious parallel characteristics;
4 kernel function correspondences being defined respectively and calculating each element value in matrix in block form, in single deck tape-recorder, m is cut-off refined gram of failure Than H-matrix amount of calculation:
H=m × hnozero (9)
Wherein, hnozeroIt is H gusts of non-zero entry number, m is to cut-off number;
In CUDA frameworks, GPU has multiple SM stream handles, it is possible to provide multiple thread dispatchings.Do not have between h element Any coupled relation, the Element-Level fine grained parallel characteristic with height defines kernel function fun_kernel_2_H by formula (5) Perfect vector multiplies and vectorial add operation, and H array elements element multithreads computing is as shown in figure 3, CPU calls fun_kernel_2_H Function, GPU enables h GPU multithreading according to the incoming function parameters of fun_kernel_2_H, and each thread corresponds in H gusts One non-zero entry, concurrently performs at the t0 moment according to formula (5), common to complete the work of Jacobian matrix Element generation;
(3) admittance battle array Ym and start node injection rate are cut-off according to what (1) was formed, node injection work(is carried out according to formula (10) Rate residual computations:
Wherein, Pis,QisRespectively active, the idle injection rate of node i, as can be seen from the above equation, each node injection rate is residual Difference is quantity of state node voltage phase angle, and the four fundamental rules basic operation of admittance array element element and trigonometric function is residual with other node powers Difference solution procedure is mutually independent of, and defines fun_kernel_3 kernel functions, enables multiple threads in SM stream handles, completes m The parallel computation task of all calculate node unbalanced power amounts of failure is cut-off, detailed parallel computation process is similar to step (2);
(4) check whether power residual error meets convergence basis for estimation, if it is satisfied, then (8) are jumped to, if be unsatisfactory for, Then continue to be transferred to (5) and be iterated;
First, fun_kernel_4 kernel functions are defined, GPU is restrained to respectively cut-offfing node power residual error by formula (11) Check, if certain cut-offs meets the condition of convergence, record this and cut-off convergence;
||ΔPt,ΔQt| | < ε t≤T (11)
Wherein, ε is power convergence criterion, Δ Pt,ΔQtThe respectively power deviation of iteration t times, T is maximum iteration;
Again, inconsistent due to respectively cut-offfing failure convergence iterative steps, setting k >=80% cut-offs failure and has restrained, then tie Shu Chaoliu iterative process, otherwise all failures of cut-offfing proceed trend iteration, until meeting k >=80% condition of convergence or reaching To maximum iteration, step (8) is jumped to;
(5) Jacobian matrix formed to (2) carries out LU decomposition, and the node power residual error calculated with reference to (3) enters line Property solving equations, define fun_kernel_5 kernel functions, task-level parallelism solution is carried out to system of linear equations;
(6) Solution for System of Linear Equations as obtained by (5) is updated to initial state vector by formula (12)
In formula,Respectively before i-th flow equation state vector kth of system of linear equations time iteration, after iteration Value,It is kth time correction value, n is system node number, and r is PV node number;
From above formula, each knot vector is only related to the node increment before and after updating in iterative process, is independent of In other any node calculated values, with parallel natural attribute, can realize that node voltage is more newly appointed by GPU multi-threaded parallels Business;
Formula (12) is defined as the kernel function fun_kernel_6 of plus computing, is appointed parallel for M forecast failure Business, needs the individual thread synchronizations of M × (2 (n-1)-r) to perform fun_kernel_6 and completes a quantity of state renewal calculating altogether;
(7) (2) are jumped to and continues iteration;
(8) Fault load flow result of calculation is cut-off according to currently convergent, carries out Branch Power Flow calculating, due to Branch Power Flow only It is related to branch parameters and both sides node voltage value, mutually it is independent of, fun_kernel_7 kernel functions are defined, complete Branch Power Flow Parallel computation.
(9) according to (8) Branch Power Flow result of calculation, heavily loaded out-of-limit verification is carried out to each branch road or stable cross section, and preserve It is currently all to cut-off the heavily loaded out-of-limit result that failure is caused, fun_kernel_8 kernel functions are defined, complete branch road branch road or steady Determine the section out-of-limit verification parallel computation of heavy duty.
Preferred embodiment of the invention described in detail above.It should be appreciated that one of ordinary skill in the art without Need creative work just can make many modifications and variations with design of the invention.Therefore, all technologies in the art Personnel are available by logical analysis, reasoning, or a limited experiment on the basis of existing technology under this invention's idea Technical scheme, all should be in the protection domain being defined in the patent claims.

Claims (6)

1. many GPU heterogeneous schemas static security analysis computational methods of a kind of CPU+, it is characterised in that unifiedly calculate platform in CUDA On, using the OpenMP multithreadings based on shared drive parallel system, consider GPU configuring conditions and forecast failure meter Calculation demand, determines CPU line number of passes mesh, builds many GPU heterogeneous schemas of CPU+, is assisted according to the mono- GPU patterns of CPU+ inside every piece of GPU card Parallel scan task is completed with coordinating, on the basis of single forecast failure Load flow calculation, realizes that multiple cut-offs trend iterative process High level of synchronization is parallel, and greatly improving static security analysis forecast failure by Element-Level fine grained parallel scans parallel processing energy Power.
2. a kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+ as claimed in claim 1, it is characterised in that Comprise the following steps:
1) the real-time section of state estimation is obtained, ground state Load flow calculation is carried out based on Newton-Raphson method, there is provided one is available for again With shared data section;
2) according to user's request, topologically sweeping is carried out to whole network equipment, forms forecast failure collection, carried out for each forecast failure Branch road-nodal information is actually cut-off in depth topology search, formation;
3) CUDA frameworks are initialized, according to parallel computation demands of cut-offfing data packing treatment is carried out into more, and it is empty to distribute GPU internal memories Between;
4) according to GPU configuring conditions and actual amount of calculation, rationalization distribution is carried out to cut-offfing computing resource, assessment needs what is enabled GPU numbers, generate corresponding Thread Count, at utmost to play GPU computation capabilities using OpenMP technologies;
5) under the pool of CPU, Load flow calculation iterative process, including admittance battle array amendment Jacobian matrix is solved, update equation is asked Solution, quantity of state are updated, Branch Power Flow is calculated and the out-of-limit check of equipment, all transfer to GPU to complete parallel, all to cut-off failure iteration Highly-parallel, the parallel characteristics according to each step calculating task construct corresponding kernel function, complete in fine grained parallel aspect The elements in parallel calculating task into iterative process;
6) judge whether all scanning is completed forecast failure, if it is not, then being transferred to step 4), remaining failure of cut-offfing is entered again Row resource allocation, if be fully completed, is transferred to step 7;
7) result displaying, according to scanning result, the out-of-limit heavily loaded information of equipment or section caused to failure on-off is shown, and Result statistics is scanned according to the scheduling static system practical demand of safety analysis module.
3. a kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+ as claimed in claim 1, it is characterised in that The step 4) specifically include:
(1) according to break calculation demand and GPU configuring conditions, computing resource is cut-off in optimization distribution, and assessment monolithic GPU singles are maximum Number is cut-off in scanning:
M _ m a x = S _ m a x S - - - ( 1 )
Wherein, S_max is the total memory headroom of GPU card, and S memory sizes for needed for single Load flow calculation, M_max is monolithic GPU card The maximum that can be calculated cut-offs number;
According to formula (1), the GPU card number started needed for calculating is:
n = α ( M _ c a l M _ max ) + 1 M _ max · N > M _ c a l N M _ max · N ≤ M _ c a l - - - ( 2 )
Wherein, α is expressed as rounding, the GPU card number that n starts for needed for epicycle, and N is the GPU card sum of system configuration, and M_cal is What is calculated needed for epicycle static security analysis cut-offs sum, and M_maxN once calculates connecting-disconnecting function for many GPU;
According to formula (2), according to Principle of Average Allocation, every piece of actual calculating of card cut-offs number:
m = M _ c a l n M _ max · N > M _ c a l M _ max M _ max · N ≤ M _ c a l - - - ( 3 )
As M_maxN≤M_cal, it is meant that GPU all participates in calculating and cannot once complete all break calculations, then by residue Cut-off M_cal ' carries out new round distribution according to formula (2-3) again, repeatedly calculates in batches, and computing formula is:
M_cal '=M_cal-M_maxN (4)
(2) in the ground state sectional model data basis of shared multiplexing, based on OpenMP multithreadings according to actual motion GPU Card number n distributes corresponding CPU line number of passes, and each CPU line journey is uniquely corresponding with single GPU, builds many GPU isomeries moulds of CPU+ Formula, concurrent operation is carried out in single GPU card according to the mono- GPU heterogeneous schemas of CPU+.
4. a kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+ as claimed in claim 1, it is characterised in that The step 5) comprise the following steps:
(1) on the basis of ground state admittance battle array Y0, branch road-nodal information is cut-off according to N-1, defines kernel function fun_kernel_ 1, failure is cut-off while carrying out node admittance battle array local correction work, and change node injection information to distributing to the card, lead to Cross admittance gust and the change of nodal information carrys out analog machine and cut-offs;
(2) formed according to (1) and cut-off admittance battle array Ym, Jacobian matrix elements in parallel calculating task is completed, for ease of calling CUDA Kernel function, each element uses and formula is calculated as below in Load flow calculation Jacobi matrix in block form:
H i i = V i 2 B i i + Q i H i j = - V i V j ( G i j sinθ i j - B i j cosθ i j ) , i ≠ j - - - ( 5 )
N i i = - V i 2 G i i - P i N i j = - V i V j ( G i j cosθ i j + B i j sinθ i j ) , i ≠ j - - - ( 6 )
J i i = V i 2 G i i - P i J i j = V i V j ( G i j cosθ i j + B i j sinθ i j ) , i ≠ j - - - ( 7 )
L i i = V i 2 B i i - Q i L i j = - V i V j ( G i j sinθ i j - B i j cosθ i j ) , i ≠ j - - - ( 8 )
Wherein, Gii,Bii,Gij,BijRespectively admittance battle array non-zero entry, θijIt is i, j node phase angle differences, Pi,QiFor i-node injects work( Rate, Vi,VjRespectively i, j node voltage amplitude;
Matrix in block form each element is quantity of state node voltage phase angle, the four fundamental rules basic operation of admittance array element element and trigonometric function and Solution procedure is independent of each other, with obvious parallel characteristics;
4 kernel function correspondences being defined respectively and calculating each element value in matrix in block form, in single deck tape-recorder, m is cut-off failure Jacobi H squares Battle array amount of calculation:
H=m × hnozero (9)
Wherein, hnozeroIt is H gusts of non-zero entry number, m is to cut-off number;
In CUDA frameworks, GPU has multiple SM stream handles, it is possible to provide multiple thread dispatchings.It is not any between h element Coupled relation, the Element-Level fine grained parallel characteristic with height defines kernel function fun_kernel_2_H and completes by formula (5) Formula vector multiplies and vectorial add operation, and CPU calls fun_kernel_2_H functions, and GPU is according to the incoming functions of fun_kernel_2_H Parameter enables h GPU multithreading, while perfect (5) parallel work-flow;
(3) admittance battle array Ym and start node injection rate are cut-off according to what (1) was formed, carries out node injecting power according to formula (10) residual Difference is calculated:
ΔP i = P i s - V i Σ j ∈ i V j ( G i j cosθ i j + B i j sinθ i j ) ΔQ i = Q i s - V i Σ j ∈ i V j ( G i j sinθ i j - B i j cosθ i j ) - - - ( 10 )
Wherein, Pis,QisRespectively active, the idle injection rate of node i, as can be seen from the above equation, each node injection rate residual error is equal It is quantity of state node voltage phase angle, the four fundamental rules basic operation of admittance array element element and trigonometric function is asked with other node power residual errors Solution preocess is mutually independent of, and defines fun_kernel_3 kernel functions, enables multiple threads in SM stream handles, completes m and cut-offs The parallel computation task of all calculate node unbalanced power amounts of failure, detailed parallel computation process is similar to step (2);
(4) check whether power residual error meets convergence basis for estimation, if it is satisfied, then (8) are jumped to, if be unsatisfactory for, after It is continuous to be transferred to (5) and be iterated;
(5) Jacobian matrix formed to (2) carries out LU decomposition, and the node power residual error calculated with reference to (3) carries out linear side Journey group is solved, and defines fun_kernel_5 kernel functions, and task-level parallelism solution is carried out to system of linear equations;
(6) the kernel function fun_kernel_6 of plus computing is defined as, for M forecast failure parallel task, M is needed altogether The individual thread synchronizations of × (2 (n-1)-r) perform fun_kernel_6 and complete a quantity of state renewal calculating;
(7) (2) are jumped to and continues iteration;
(8) according to it is current it is convergent cut-offs Fault load flow result of calculation, carry out Branch Power Flow calculating, due to Branch Power Flow only and branch Road parameter and both sides node voltage value are related, are mutually independent of, and define fun_kernel_7 kernel functions, complete Branch Power Flow parallel Calculate.
(9) according to (8) Branch Power Flow result of calculation, heavily loaded out-of-limit verification is carried out to each branch road or stable cross section, and preserve current It is all to cut-off the heavily loaded out-of-limit result that failure is caused, fun_kernel_8 kernel functions are defined, complete branch road branch road or stabilization is disconnected The face out-of-limit verification parallel computation of heavy duty.
5. a kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+ as claimed in claim 4, it is characterised in that The step 5) (4th) step be specially:
First, fun_kernel_4 kernel functions are defined, GPU carries out convergence inspection to respectively cut-offfing node power residual error by formula (11) Look into, if certain cut-offs meets the condition of convergence, record this and cut-off convergence;
||ΔPt,ΔQt| | < ε t≤T (11)
Wherein, ε is power convergence criterion, Δ Pt,ΔQtThe respectively power deviation of iteration t times, T is maximum iteration;
Again, inconsistent due to respectively cut-offfing failure convergence iterative steps, setting k >=80% cut-offs failure and has restrained, then terminate tide Stream iterative process, otherwise all failures of cut-offfing proceed trend iteration, until meeting k >=80% condition of convergence or reaching most Big iterations, jumps to step (8).
6. a kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+ as claimed in claim 4, it is characterised in that The step 5) (6th) step be defined as the kernel function fun_kernel_6 of plus computing and be:
x i k + 1 = x i k + Δx i k , ( i = 1 , 2 , ... , 2 ( n - 1 ) - r ) - - - ( 12 )
In formula,Respectively before i-th flow equation state vector kth of system of linear equations time iteration, after iteration it is worth,It is kth time correction value, n is system node number, and r is PV node number.
CN201710037124.0A 2017-01-19 2017-01-19 A kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+ Pending CN106874113A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201710037124.0A CN106874113A (en) 2017-01-19 2017-01-19 A kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+
PCT/CN2017/093226 WO2018133348A1 (en) 2017-01-19 2017-07-17 Static security analysis computation method, apparatus, and computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710037124.0A CN106874113A (en) 2017-01-19 2017-01-19 A kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+

Publications (1)

Publication Number Publication Date
CN106874113A true CN106874113A (en) 2017-06-20

Family

ID=59158723

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710037124.0A Pending CN106874113A (en) 2017-01-19 2017-01-19 A kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+

Country Status (2)

Country Link
CN (1) CN106874113A (en)
WO (1) WO2018133348A1 (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018133348A1 (en) * 2017-01-19 2018-07-26 国电南瑞科技股份有限公司 Static security analysis computation method, apparatus, and computer storage medium
CN108804386A (en) * 2018-07-09 2018-11-13 东北电力大学 A kind of parallelization computational methods of power system load nargin
CN108879691A (en) * 2018-06-21 2018-11-23 清华大学 A kind of method and device that extensive continuous tide calculates
CN108897616A (en) * 2018-06-04 2018-11-27 四川大学 Non-down sampling contourlet transform optimization method based on concurrent operation
CN109038543A (en) * 2018-06-27 2018-12-18 国网辽宁省电力有限公司 A kind of state estimation calculation method based on CPU+GPU mixing isomery
CN109167357A (en) * 2018-10-18 2019-01-08 国网山东省电力公司泰安供电公司 A method of the optimization power grid static security analysis time
CN109167354A (en) * 2018-10-08 2019-01-08 国网天津市电力公司电力科学研究院 A kind of power grid forecast failure parallel parsing calculation method based on exchange files
CN109388496A (en) * 2018-11-01 2019-02-26 北京视甄智能科技有限公司 A kind of image concurrent processing method, apparatus and system based on more GPU cards
CN109857564A (en) * 2019-03-05 2019-06-07 上海交通大学 The GPU of method for managing resource and its application based on fine-grained GPU
CN109871559A (en) * 2017-12-04 2019-06-11 广东电网有限责任公司电力调度控制中心 A kind of analysis method of power system information fault scanning
CN110175775A (en) * 2019-05-24 2019-08-27 浙江大学 Extensive Abnormal Load Flow of Power Systems calculation method based on graphics processor and central processing unit co-architecture
CN110543711A (en) * 2019-08-26 2019-12-06 中国原子能科学研究院 parallel implementation and optimization method for numerical reactor thermal hydraulic sub-channel simulation
CN110718919A (en) * 2019-09-25 2020-01-21 北京交通大学 GPU acceleration-based large power grid static safety analysis fault screening method
CN111355231A (en) * 2018-12-24 2020-06-30 中国电力科学研究院有限公司 Power distribution network topology identification method and system
CN111478333A (en) * 2020-04-14 2020-07-31 广东电网有限责任公司广州供电局 Parallel static security analysis method for improving power distribution network recovery after disaster
CN111930471A (en) * 2020-08-14 2020-11-13 中国科学院上海高等研究院 GPU-based parallel simulation evaluation selection method
CN112083956A (en) * 2020-09-15 2020-12-15 哈尔滨工业大学 Heterogeneous platform-oriented automatic management system for complex pointer data structure
CN113886057A (en) * 2020-07-01 2022-01-04 西南科技大学 Dynamic resource scheduling method based on parsing technology and data flow information on heterogeneous many-core
CN117555695A (en) * 2024-01-10 2024-02-13 深圳本贸科技股份有限公司 Optimization method and system for realizing heterogeneous computation based on parallel computation
CN117687779B (en) * 2023-11-30 2024-04-26 山东诚泉信息科技有限责任公司 Complex electric wave propagation prediction rapid calculation method based on heterogeneous multi-core calculation platform

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109711554B (en) * 2018-09-07 2021-06-04 天翼电子商务有限公司 Application elasticity management device based on infrastructure big data
CN109634830B (en) * 2018-12-19 2022-06-07 哈尔滨工业大学 CUDA program integration performance prediction method based on multi-feature coupling

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103677761A (en) * 2013-12-11 2014-03-26 中国资源卫星应用中心 Quick satellite remote sensing data processing system
CN103885040A (en) * 2014-03-29 2014-06-25 江西理工大学 Method for generating echoes of circular synthetic aperture radar based on CPU-GPU heterogeneous computing
CN105576648A (en) * 2015-11-23 2016-05-11 中国电力科学研究院 Static security analysis double-layer parallel method based on GPU-CUP heterogeneous computing platform
CN106157176A (en) * 2016-07-26 2016-11-23 东南大学 The LU decomposition method of the direction of energy Jacobian matrix that a kind of GPU accelerates

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106874113A (en) * 2017-01-19 2017-06-20 国电南瑞科技股份有限公司 A kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103677761A (en) * 2013-12-11 2014-03-26 中国资源卫星应用中心 Quick satellite remote sensing data processing system
CN103885040A (en) * 2014-03-29 2014-06-25 江西理工大学 Method for generating echoes of circular synthetic aperture radar based on CPU-GPU heterogeneous computing
CN105576648A (en) * 2015-11-23 2016-05-11 中国电力科学研究院 Static security analysis double-layer parallel method based on GPU-CUP heterogeneous computing platform
CN106157176A (en) * 2016-07-26 2016-11-23 东南大学 The LU decomposition method of the direction of energy Jacobian matrix that a kind of GPU accelerates

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
陆娟娟 等: "基于GPU并行的静态安全分析设计与应用", 《计算机科学与探索》 *

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018133348A1 (en) * 2017-01-19 2018-07-26 国电南瑞科技股份有限公司 Static security analysis computation method, apparatus, and computer storage medium
CN109871559A (en) * 2017-12-04 2019-06-11 广东电网有限责任公司电力调度控制中心 A kind of analysis method of power system information fault scanning
CN109871559B (en) * 2017-12-04 2023-05-02 广东电网有限责任公司电力调度控制中心 Analysis method for information fault scanning of power system
CN108897616A (en) * 2018-06-04 2018-11-27 四川大学 Non-down sampling contourlet transform optimization method based on concurrent operation
CN108897616B (en) * 2018-06-04 2021-08-24 四川大学 Non-downsampling contourlet transform optimization method based on parallel operation
CN108879691B (en) * 2018-06-21 2020-09-04 清华大学 Large-scale continuous power flow calculation method and device
CN108879691A (en) * 2018-06-21 2018-11-23 清华大学 A kind of method and device that extensive continuous tide calculates
CN109038543A (en) * 2018-06-27 2018-12-18 国网辽宁省电力有限公司 A kind of state estimation calculation method based on CPU+GPU mixing isomery
CN109038543B (en) * 2018-06-27 2021-10-15 国网辽宁省电力有限公司 State estimation calculation method based on CPU + GPU mixed isomerism
CN108804386A (en) * 2018-07-09 2018-11-13 东北电力大学 A kind of parallelization computational methods of power system load nargin
CN108804386B (en) * 2018-07-09 2022-03-29 东北电力大学 Parallelization calculation method for load margin of power system
CN109167354A (en) * 2018-10-08 2019-01-08 国网天津市电力公司电力科学研究院 A kind of power grid forecast failure parallel parsing calculation method based on exchange files
CN109167357A (en) * 2018-10-18 2019-01-08 国网山东省电力公司泰安供电公司 A method of the optimization power grid static security analysis time
CN109388496A (en) * 2018-11-01 2019-02-26 北京视甄智能科技有限公司 A kind of image concurrent processing method, apparatus and system based on more GPU cards
CN111355231A (en) * 2018-12-24 2020-06-30 中国电力科学研究院有限公司 Power distribution network topology identification method and system
CN109857564A (en) * 2019-03-05 2019-06-07 上海交通大学 The GPU of method for managing resource and its application based on fine-grained GPU
CN110175775A (en) * 2019-05-24 2019-08-27 浙江大学 Extensive Abnormal Load Flow of Power Systems calculation method based on graphics processor and central processing unit co-architecture
CN110543711B (en) * 2019-08-26 2021-07-20 中国原子能科学研究院 Parallel implementation and optimization method for numerical reactor thermal hydraulic sub-channel simulation
CN110543711A (en) * 2019-08-26 2019-12-06 中国原子能科学研究院 parallel implementation and optimization method for numerical reactor thermal hydraulic sub-channel simulation
CN110718919B (en) * 2019-09-25 2021-06-01 北京交通大学 GPU acceleration-based large power grid static safety analysis fault screening method
CN110718919A (en) * 2019-09-25 2020-01-21 北京交通大学 GPU acceleration-based large power grid static safety analysis fault screening method
CN111478333A (en) * 2020-04-14 2020-07-31 广东电网有限责任公司广州供电局 Parallel static security analysis method for improving power distribution network recovery after disaster
CN111478333B (en) * 2020-04-14 2021-11-30 广东电网有限责任公司广州供电局 Parallel static security analysis method for improving power distribution network recovery after disaster
CN113886057A (en) * 2020-07-01 2022-01-04 西南科技大学 Dynamic resource scheduling method based on parsing technology and data flow information on heterogeneous many-core
CN111930471A (en) * 2020-08-14 2020-11-13 中国科学院上海高等研究院 GPU-based parallel simulation evaluation selection method
CN111930471B (en) * 2020-08-14 2023-05-26 中国科学院上海高等研究院 Parallel simulation evaluation selection method based on GPU
CN112083956A (en) * 2020-09-15 2020-12-15 哈尔滨工业大学 Heterogeneous platform-oriented automatic management system for complex pointer data structure
CN117687779B (en) * 2023-11-30 2024-04-26 山东诚泉信息科技有限责任公司 Complex electric wave propagation prediction rapid calculation method based on heterogeneous multi-core calculation platform
CN117555695A (en) * 2024-01-10 2024-02-13 深圳本贸科技股份有限公司 Optimization method and system for realizing heterogeneous computation based on parallel computation

Also Published As

Publication number Publication date
WO2018133348A1 (en) 2018-07-26

Similar Documents

Publication Publication Date Title
CN106874113A (en) A kind of many GPU heterogeneous schemas static security analysis computational methods of CPU+
CN105576648B (en) Static security analysis double-layer parallel method based on GPU-CPU heterogeneous computing platform
Wang et al. Distributed optimization approaches for emerging power systems operation: A review
CN101976835B (en) Parallel computation method for Newton power flow of large-scale electric power system
Jiang et al. A reduced-space interior point method for transient stability constrained optimal power flow
CN100466412C (en) United accident-prevention rehearse method for provincial and city power system
CN106157176B (en) A kind of LU decomposition method for the direction of energy Jacobian matrix that GPU accelerates
CN105391057B (en) A kind of GPU threaded design methods that direction of energy Jacobi battle array calculates
CN106911140A (en) A kind of energy storage planing method
CN106972504A (en) Interval idle work optimization method based on genetic algorithm
CN107123994A (en) The solution method of linearization of interval idle work optimization model
CN106532710A (en) Microgrid power flow optimization method considering voltage stability constraint
CN106026107B (en) A kind of QR decomposition method for the direction of energy Jacobian matrix that GPU accelerates
CN106786977A (en) A kind of charging dispatching method of electric automobile charging station
CN106505575A (en) A kind of Line Flow economic load dispatching method based on Granule Computing
CN106356840A (en) Method and system for estimating states of regional electric power systems on basis of synchronous phasor measurement
CN104967121B (en) A kind of tidal current computing method of large-scale electrical power system node
CN115714382A (en) Active power distribution network real-time scheduling method and device based on security reinforcement learning
CN112134309A (en) Novel partitioning method suitable for distributed voltage control of power distribution network
CN106096183B (en) A kind of multiple parallel method based on the method for characteristic curves
CN114140022A (en) Multi-virtual power plant distributed dynamic economic dispatching method and system
CN104979840B (en) A kind of active distribution network three-phase idle work optimization method
CN106058994A (en) Electric automobile orderly charging control method for reducing network loss of power distribution network
CN111509784B (en) Uncertainty-considered virtual power plant robust output feasible region identification method and device
CN112948123A (en) Spark-based grid hydrological model distributed computing method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20170620

RJ01 Rejection of invention patent application after publication