CN111062467A - Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler - Google Patents

Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler Download PDF

Info

Publication number
CN111062467A
CN111062467A CN201911312625.0A CN201911312625A CN111062467A CN 111062467 A CN111062467 A CN 111062467A CN 201911312625 A CN201911312625 A CN 201911312625A CN 111062467 A CN111062467 A CN 111062467A
Authority
CN
China
Prior art keywords
node
neural network
graph
nodes
queue
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911312625.0A
Other languages
Chinese (zh)
Other versions
CN111062467B (en
Inventor
黄明飞
王海涛
吕春莹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Open Intelligent Machine Shanghai Co ltd
Original Assignee
Open Intelligent Machine Shanghai Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Open Intelligent Machine Shanghai Co ltd filed Critical Open Intelligent Machine Shanghai Co ltd
Priority to CN201911312625.0A priority Critical patent/CN111062467B/en
Publication of CN111062467A publication Critical patent/CN111062467A/en
Application granted granted Critical
Publication of CN111062467B publication Critical patent/CN111062467B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/063Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention discloses an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler, which comprises the steps of establishing a node list; numbering all nodes of the neural network computation graph in advance, and recording the input dependency relationship and the output dependency relationship of each node; acquiring any node without input dependency relationship from the neural network computational graph; deleting the nodes without the input dependency relationship and all the connecting edges taking the nodes without the input dependency relationship as starting points, and storing the serial numbers of the deleted nodes into a node list; repeating the step S2 and the step S3 until the neural network calculation graph is empty, and storing the number of the deleted node into a node list; and automatically segmenting according to the node attribute colors in the node list so as to process and obtain each neural network computation subgraph. Has the advantages that: the method is suitable for any neural network computation graph with a complex structure, and solves the problems that the neural network computation graph is not good in segmenting subgraphs and different computation unit processing subgraphs are not good in partitioning.

Description

Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler
Technical Field
The invention relates to the technical field of AI (Artificial Intelligence), in particular to an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler.
Background
With the continuous development of the AI technology, the neural network algorithm based on deep learning has become the mainstream mode of AI research at present. In consideration of the problems of cost, power consumption, privacy and the like, more and more application scenes migrate the calculation of the AI algorithm from the cloud to the mobile embedded terminal device.
At present, an embedded device chip often has a plurality of computing units coexisting, and includes, in addition to a general CPU (central processing Unit), an AI accelerator such as a GPU (graphic processing Unit), an NPU (network Processor Unit), an FPGA (Field Programmable gate array), and the like, which is specially used for accelerating an AI algorithm. The CPU universality and the programmability are good, but the performance and the energy efficiency ratio are poor, and the performance and the energy efficiency ratio of the AI accelerator are high, but the programmability is poor, for example, some AI accelerators cannot be programmed, some AI accelerators can be programmed, but the programming is complex, the development period is long, and some AI accelerators only support shaping calculation, so that the network precision loss is large.
By combining the characteristics of a CPU and an AI accelerator, the neural network is subjected to sub-graph segmentation, compute intensive operators on the AI accelerator (such as Convolution operators, full-connected operators, etc.), and compute some operators with complex computation logic or operators which are not suitable for shaping computation (such as operators for pre-and post-processing of detection network, etc.) on the CPU, so that the method is a solution for the AI heterogeneous computation at present.
In view of this, how to perform automatic sub-graph segmentation on a neural network computation graph is a problem to be solved at present to implement heterogeneous computation. The existing scheme is only suitable for simple operator chain-connected neural network computation graphs, and more neural network computation graphs contain more complex operator connection dependency relations, for example, a concat operator needs a plurality of inputs, for example, residual modules in a residual network all relate to non-chain neural network computation graphs. How to find a computing method capable of performing subgraph segmentation on any neural network computing graph is a great challenge to be solved by an AI heterogeneous compiler.
The problem to be solved at present is to find an algorithm for automatically segmenting subgraphs, and segment operator nodes deployed to different computing units in a neural network computational graph to different subgraphs.
Disclosure of Invention
Aiming at the problems in the prior art, an automatic neural network subgraph segmentation method applied to an AI heterogeneous compiler is provided.
The specific technical scheme is as follows:
an automatic neural network subgraph segmentation method applied to an AI heterogeneous compiler is used for segmenting operator nodes deployed to different computing units in a neural network computing graph to different neural network computing subgraphs, and comprises the following steps:
step S1, a node list is created;
step S2, numbering all nodes of the neural network computation graph in advance, and recording the input dependency relationship and the output dependency relationship of each node;
step S3, acquiring any node without input dependency relationship from the neural network computational graph;
step S4, deleting the nodes without dependency relationship and all the connecting edges taking the nodes without dependency relationship as the starting points, and storing the serial numbers of the deleted nodes into the node list;
step S5, repeating the step S2 and the step S3 until the neural network computation graph is empty, and storing the deleted node numbers into the node list;
and step S6, acquiring the node list, automatically segmenting according to the node attribute colors in the node list, and acquiring the input dependency relationship and the output dependency relationship related to the nodes to process to obtain each neural network computation subgraph.
Preferably, before the step S6, the method further includes: judging whether the neural network computation graph is empty or not;
if not, go to step S3;
if yes, the process goes to step S6.
Preferably, the step S6 includes the following steps:
step S60, creating a subgraph queue warehouse to store the subgraph node queue;
step S61, obtaining the current first node from the node list;
step S62, judging whether the color of the last sub-graph node queue in the sub-graph queue warehouse is the same as the color of the current first node in the node list;
if yes, go to step S63;
if not, go to step S64;
step S63, adding the current first node in the node list to the subgraph queue warehouse;
step S64, creating a new sub-graph node queue in the sub-graph queue warehouse, and using the current first node in the node list as the first node of the new sub-graph node queue;
step S65, repeating the steps S61-S62 until the node list is empty;
step S66, obtaining each sub-graph node queue from the sub-graph queue warehouse, and connecting the nodes of each sub-graph node queue according to the input dependency and the output dependency of each node to obtain each neural network computation sub-graph.
Preferably, before the step S61, the method further includes:
judging whether the node list is empty or not;
if not, go to step S61;
if yes, the process goes to step S66.
Preferably, in step S64, the color of the new sub-graph node queue is the same as the color of the current first node in the node list.
Preferably, in the method for automatically segmenting the neural network subgraph, the neural network computation graph is divided into a plurality of neural network computation subgraphs, and the neural network computation subgraphs are deployed on different computation units for computation.
Preferably, the different computing units comprise a central processing unit, a graphics processor and a digital signal processor.
The invention provides an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler, which has the advantages that:
(1) the nodes of the neural network computational graph are automatically arranged according to the input dependency relationship and the output dependency relationship of the nodes of the neural network computational graph, so that each neural network computational graph is obtained, the problem that the subgraph is not well segmented due to a complex dependency relationship structure in the neural network computational graph is solved, the running sequence of the nodes is ensured, and the neural network computational graph is suitable for any complex-structure neural network computational graph;
(2) the neural network computation graph is divided into a plurality of neural network computation subgraphs and is deployed on different computation units for computation, so that the problem of subgraph division of the neural network computation graph of different computation units by the AI heterogeneous compiler is solved, the method is suitable for the conditions of two computation units and the conditions of various computation units, the universality is good, the optimal computing power of the whole hardware chip is fully exerted, and the high-efficiency heterogeneous computation of the AI algorithm in actual hardware is completed.
Drawings
Embodiments of the present invention will now be described more fully hereinafter with reference to the accompanying drawings. The drawings are, however, to be regarded as illustrative and explanatory only and are not restrictive of the scope of the invention.
FIG. 1 is a diagram of a neural network computational graph in the prior art;
FIG. 2 is a flowchart illustrating steps of an automatic neural network subgraph segmentation method applied to an AI heterogeneous compiler according to an embodiment of the present invention;
FIG. 3 is a flowchart illustrating a first step of an embodiment of a method for automatically segmenting a neural network subgraph applied to an AI heterogeneous compiler according to the present invention;
FIG. 4 is a flowchart illustrating steps of a second embodiment of a method for automatically segmenting a neural network subgraph applied to an AI heterogeneous compiler according to the present invention;
fig. 5 is a segmentation graph of each neural network computation sub-graph of the neural network computation graph of the second embodiment of the automatic segmentation neural network sub-graph method applied to the AI heterogeneous compiler according to the embodiment of the present invention;
fig. 6 is a flowchart illustrating steps of a third embodiment of the method for automatically segmenting a neural network subgraph applied to an AI heterogeneous compiler according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It should be noted that the embodiments and features of the embodiments may be combined with each other without conflict.
The invention is further described with reference to the following drawings and specific examples, which are not intended to be limiting.
In the prior art, the neural network computational graph can be regarded as a directed acyclic graph, as shown in fig. 1, each node is an operator, and each arrow represents a relationship between operators. Assuming that the neural network computation graph is executed by using a central processing unit or a graphic processor, in fig. 1, the nodes 1, 2, 3, 5, and 6 may respectively use the central processing unit to perform computation processing, and the node 4 may use the graphic processor to perform computation processing.
The above is only suitable for simple operator chain-connected neural network computation graphs, while more and more neural network computation graphs contain more complex operator connection dependency relationships, for example, concat operators need multiple inputs, for example, residual modules in residual network, all relate to non-chain neural network computation graphs. How to find a computing method capable of performing subgraph segmentation on any neural network computing graph is a great challenge to be solved by an AI heterogeneous compiler.
Aiming at the defects in the prior art, the problem to be solved at present is to find an algorithm for automatically segmenting subgraphs and segment operator nodes of different colors in a neural network calculation graph onto different subgraphs.
Therefore, the invention provides an automatic neural network subgraph segmentation method applied to an AI heterogeneous compiler, which is used for deploying a neural network computation graph to operator nodes of different computation units to segment the operator nodes to different neural network computation subgraphs, wherein the method comprises the following steps:
step S1, a node list is created;
step S2, numbering all nodes of the neural network computation graph in advance, and recording the input dependency relationship and the output dependency relationship of each node;
step S3, acquiring any node without input dependency relationship from the neural network computational graph;
step S4, deleting the nodes without the input dependency relationship and all the connecting edges taking the nodes without the input dependency relationship as the starting points, and storing the serial numbers of the deleted nodes into a node list;
step S5, repeating the step S2 and the step S3 until the neural network calculation graph is empty, and storing the number of the deleted node into a node list;
and step S6, acquiring a node list, automatically segmenting according to the node attribute color (the type of the computing unit to which the node operator belongs) in the node list, and acquiring the input dependency relationship and the output dependency relationship related to the node to process to obtain each neural network computing subgraph.
According to the technical scheme of the automatic neural network subgraph segmentation method applied to the AI heterogeneous compiler, the automatic neural network subgraph segmentation method is suitable for any neural network computational graph with a complex structure, and the nodes of the neural network computational graph are automatically arranged according to the input dependency relationship and the output dependency relationship of the nodes of the neural network computational graph, so that each neural network computational subgraph is obtained, the problem that the subgraph is not well segmented due to the complex dependency relationship structure in the neural network computational graph is solved, and the running dependency sequence of the nodes is also ensured.
Further, the neural network computation graph is divided into a plurality of neural network computation subgraphs, and the computation is carried out by deploying the subgraphs on different computation units, wherein the different computation units comprise a central processing unit, a graphic processing unit and a digital signal processing unit, so that the problem of subgraph division of the neural network computation graph of different computation units by the AI heterogeneous compiler is solved, the AI heterogeneous compiler is not only suitable for the conditions of two computation units, but also suitable for the conditions of various computation units, and has good universality, thereby fully exerting the optimal computing power of the whole hardware chip and completing the high-efficiency heterogeneous computation of the AI algorithm in actual hardware.
In the above technical solution, as a preferred embodiment, as shown in fig. 3, before step S6, the method further includes: judging whether the neural network calculation graph is empty or not;
if not, go to step S3;
if yes, the process goes to step S6.
In this embodiment, the purpose of determining whether the neural network computation graph is empty is to thoroughly acquire nodes without dependency relationship input in the neural network computation graph without omission, store the numbers of the deleted nodes in the node list, automatically segment the nodes according to the colors in the node list, and acquire input dependency relationship and output dependency relationship related to the nodes to process the input dependency relationship and the output dependency relationship to obtain each neural network computation subgraph.
Furthermore, the method for automatically segmenting the neural network subgraph is suitable for any neural network computational graph with a complex structure, solves the problem that the subgraph is not well segmented due to a complex dependency structure in the neural network computational graph, and also ensures the operation sequence of the nodes.
In the above technical solution, as shown in fig. 4, the step S6 preferably includes the following steps:
step S60, creating a subgraph queue warehouse to store the subgraph node queue;
step S61, obtaining the current first node from the node list;
step S62, judging whether the color of the last sub-graph node queue in the sub-graph queue warehouse is the same as the color of the current first node in the node list;
if yes, go to step S63;
if not, go to step S64;
step S63, adding the current first node in the node list to the subgraph queue warehouse;
step S64, creating a new sub-graph node queue in the sub-graph queue warehouse, and using the current first node in the node list as the first node of the new sub-graph node queue, wherein the color of the new sub-graph node queue is the same as the color of the current first node in the node list;
step S65, repeating the steps S61-S62 until the node list is empty;
and step S66, obtaining each sub-graph node queue from the sub-graph queue warehouse, and connecting the nodes of each sub-graph node queue according to the input dependency and the output dependency of each node to obtain each neural network computation sub-graph.
In this embodiment, taking the neural network computational graph of fig. 1 as an example, after steps S1-S5, the obtained node list is shown as table one, the node 1, the node 2, the node 3, the node 5, and the node 6 are in the same color, and the node 4 is in another color, and after step S6, the colors of the node 1, the node 2, the node 3, the node 4, the node 5, and the node 6 in the node list are segmented, so that a subgraph a, a subgraph B, and a subgraph C can be obtained, as shown in fig. 5.
Watch 1
1 2 3 4 5 6
Further, the nodes of the neural network computation graph are automatically arranged according to the input dependency relationship and the output dependency relationship of the nodes of the neural network computation graph, and the node list is automatically divided into each neural network computation subgraph according to the color attributes of the nodes.
Furthermore, the method for automatically segmenting the neural network subgraph is suitable for any neural network computational graph with a complex structure, solves the problem that the subgraph is not well segmented due to a complex dependency structure in the neural network computational graph, and also ensures the operation sequence of the nodes.
In the above technical solution, as a preferred embodiment, as shown in fig. 6, before step S61, the method further includes:
judging whether the node list is empty or not;
if not, go to step S61;
if yes, the process goes to step S66.
In this embodiment, the purpose of determining whether the node list is empty is to thoroughly determine that the colors of the nodes in the node list are sequentially matched with the colors in the sub-graph node queues without omission, then obtain each sub-graph node queue from the sub-graph queue warehouse, and connect the nodes of each sub-graph node queue according to the input dependency and the output dependency of each node to obtain each neural network computation sub-graph.
Furthermore, the method for automatically segmenting the neural network subgraph is suitable for any neural network computational graph with a complex structure, solves the problem that the subgraph is not well segmented due to a complex dependency structure in the neural network computational graph, and also ensures the operation sequence of the nodes.
While the invention has been described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the invention.

Claims (7)

1. An automatic neural network subgraph segmentation method applied to an AI heterogeneous compiler is used for segmenting operator nodes deployed to different computing units in a neural network computing graph to different neural network computing subgraphs, and is characterized by comprising the following steps:
step S1, a node list is created;
step S2, numbering all nodes of the neural network computation graph in advance, and recording the input dependency relationship and the output dependency relationship of each node;
step S3, acquiring any node without input dependency relationship from the neural network computational graph;
step S4, deleting the nodes without dependency relationship and all the connecting edges taking the nodes without dependency relationship as the starting points, and storing the serial numbers of the deleted nodes into the node list;
step S5, repeating the step S2 and the step S3 until the neural network computation graph is empty, and storing the deleted node numbers into the node list;
and step S6, acquiring the node list, automatically segmenting according to the node attribute colors in the node list, and acquiring the input dependency relationship and the output dependency relationship related to the nodes to process to obtain each neural network computation subgraph.
2. The method for automatically segmenting neural network subgraph according to claim 1, further comprising, before the step S6: judging whether the neural network computation graph is empty or not;
if not, go to step S3;
if yes, the process goes to step S6.
3. The method for automatically segmenting neural network subgraph according to claim 1, wherein the step S6 comprises the following steps:
step S60, creating a subgraph queue warehouse to store the subgraph node queue;
step S61, obtaining the current first node from the node list;
step S62, judging whether the color of the last sub-graph node queue in the sub-graph queue warehouse is the same as the color of the current first node in the node list;
if yes, go to step S63;
if not, go to step S64;
step S63, adding the current first node in the node list to the subgraph queue warehouse;
step S64, creating a new sub-graph node queue in the sub-graph queue warehouse, and using the current first node in the node list as the first node of the new sub-graph node queue;
step S65, repeating the steps S61-S62 until the node list is empty;
step S66, obtaining each sub-graph node queue from the sub-graph queue warehouse, and connecting the nodes of each sub-graph node queue according to the input dependency and the output dependency of each node to obtain each neural network computation sub-graph.
4. The method for automatically segmenting a neural network subgraph according to claim 3, further comprising, before said step S61:
judging whether the node list is empty or not;
if not, go to step S61;
if yes, the process goes to step S66.
5. The method according to claim 3, wherein in step S64, the color of the new sub-graph node queue is the same as the color of the first node in the node list.
6. The method according to claim 1, wherein the neural network computational graph is divided into a plurality of neural network computational subgraphs and deployed for computation on different computational units.
7. The method according to claim 6, wherein the different computing units comprise a central processing unit, a graphics processor and a digital signal processor.
CN201911312625.0A 2019-12-18 2019-12-18 Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler Active CN111062467B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911312625.0A CN111062467B (en) 2019-12-18 2019-12-18 Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911312625.0A CN111062467B (en) 2019-12-18 2019-12-18 Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler

Publications (2)

Publication Number Publication Date
CN111062467A true CN111062467A (en) 2020-04-24
CN111062467B CN111062467B (en) 2023-05-12

Family

ID=70301038

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911312625.0A Active CN111062467B (en) 2019-12-18 2019-12-18 Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler

Country Status (1)

Country Link
CN (1) CN111062467B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111738434A (en) * 2020-06-03 2020-10-02 中国科学院计算技术研究所 Method for executing deep neural network on heterogeneous processing unit
CN112070213A (en) * 2020-08-28 2020-12-11 Oppo广东移动通信有限公司 Neural network model optimization method, device, equipment and storage medium
CN112711422A (en) * 2020-12-31 2021-04-27 北京清微智能科技有限公司 Optimization method and system for neural network compiling
WO2022022670A1 (en) * 2020-07-31 2022-02-03 北京灵汐科技有限公司 Neural network computation graph processing method and apparatus, and processing device
WO2023122854A1 (en) * 2021-12-27 2023-07-06 华为技术有限公司 Data processing method and apparatus
CN117576125A (en) * 2024-01-16 2024-02-20 芯瞳半导体技术(山东)有限公司 Neural network calculation graph segmentation method, device, equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018206261A (en) * 2017-06-08 2018-12-27 日本電信電話株式会社 Word division estimation model learning device, word division device, method and program
CN109902819A (en) * 2019-02-12 2019-06-18 Oppo广东移动通信有限公司 Neural computing method, apparatus, mobile terminal and storage medium
WO2019177824A1 (en) * 2018-03-14 2019-09-19 Microsoft Technology Licensing, Llc Hardware accelerated neural network subgraphs

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018206261A (en) * 2017-06-08 2018-12-27 日本電信電話株式会社 Word division estimation model learning device, word division device, method and program
WO2019177824A1 (en) * 2018-03-14 2019-09-19 Microsoft Technology Licensing, Llc Hardware accelerated neural network subgraphs
CN109902819A (en) * 2019-02-12 2019-06-18 Oppo广东移动通信有限公司 Neural computing method, apparatus, mobile terminal and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
方良达;余永权;: "属性依赖理论及其在神经网络中的应用" *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111738434A (en) * 2020-06-03 2020-10-02 中国科学院计算技术研究所 Method for executing deep neural network on heterogeneous processing unit
CN111738434B (en) * 2020-06-03 2023-04-07 中国科学院计算技术研究所 Method for executing deep neural network on heterogeneous processing unit
WO2022022670A1 (en) * 2020-07-31 2022-02-03 北京灵汐科技有限公司 Neural network computation graph processing method and apparatus, and processing device
CN112070213A (en) * 2020-08-28 2020-12-11 Oppo广东移动通信有限公司 Neural network model optimization method, device, equipment and storage medium
CN112711422A (en) * 2020-12-31 2021-04-27 北京清微智能科技有限公司 Optimization method and system for neural network compiling
CN112711422B (en) * 2020-12-31 2024-01-19 北京清微智能科技有限公司 Neural network compiling optimization method and system
WO2023122854A1 (en) * 2021-12-27 2023-07-06 华为技术有限公司 Data processing method and apparatus
CN117576125A (en) * 2024-01-16 2024-02-20 芯瞳半导体技术(山东)有限公司 Neural network calculation graph segmentation method, device, equipment and storage medium
CN117576125B (en) * 2024-01-16 2024-04-16 芯瞳半导体技术(山东)有限公司 Neural network calculation graph segmentation method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN111062467B (en) 2023-05-12

Similar Documents

Publication Publication Date Title
CN111062467A (en) Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler
US11151442B2 (en) Convolutional neural network processing method and device
US9665660B2 (en) Logical data flow mapping rules for (sub) graph isomorphism in a cluster computing environment
US11615090B2 (en) Method, apparatus and system for determining accessible paths between two nodes
Ediger et al. Tracking structure of streaming social networks
WO2017116924A1 (en) Neural network training performance optimization framework
EP4115342A1 (en) Parallel computing scheme generation for neural networks
CN110399387A (en) Method and device based on table incidence relation dynamic generation query SQL
CN111159577B (en) Community dividing method and device, storage medium and electronic device
Červený et al. Faster FPT algorithm for 5-path vertex cover
CN113886092A (en) Computation graph execution method and device and related equipment
CN114841309A (en) Data processing method and device and electronic equipment
CN103747033B (en) A kind of method of community discovery
CN109412149B (en) Power grid subgraph construction method based on regional division, topology analysis method and device
Anari et al. Sampling arborescences in parallel
CN114239816B (en) Reconfigurable hardware acceleration architecture of convolutional neural network-graph convolutional neural network
CN109240813A (en) Task schedule and task immigration method in a kind of mobile cloud computing
CN116933841A (en) Operator fusion method and device, electronic equipment and computer readable medium
CN110851178B (en) Inter-process program static analysis method based on distributed graph reachable computation
CN103268614A (en) Generation method for prospect spectrogram for multi-prospect co-segmentation
US10360312B2 (en) System and method for identifying clones
CN113051080A (en) Computation graph execution method and device and heterogeneous platform
CN113485848A (en) Deep neural network deployment method and device, computer equipment and storage medium
CN107564289B (en) Road network preprocessing method for merging traffic nodes
CN106991195B (en) Distributed subgraph enumeration method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant