CN111062467B - Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler - Google Patents
Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler Download PDFInfo
- Publication number
- CN111062467B CN111062467B CN201911312625.0A CN201911312625A CN111062467B CN 111062467 B CN111062467 B CN 111062467B CN 201911312625 A CN201911312625 A CN 201911312625A CN 111062467 B CN111062467 B CN 111062467B
- Authority
- CN
- China
- Prior art keywords
- node
- neural network
- graph
- sub
- queue
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- General Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Neurology (AREA)
- Stored Programmes (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses an automatic neural network subgraph segmentation method applied to an AI heterogeneous compiler, which comprises the steps of creating a node list; numbering all nodes of the neural network computational graph in advance, and recording the input dependency relationship and the output dependency relationship of each node; acquiring any node without any input dependency relationship from a neural network computational graph; deleting the input nodes without the dependency and all connecting edges taking the input nodes without the dependency as starting points, and storing the numbers of the deleted nodes into a node list; repeating the step S2 and the step S3 until the neural network calculation graph is empty, and storing the number of the deleted node into a node list; and automatically segmenting according to the node attribute colors in the node list to obtain each neural network calculation subgraph through processing. The beneficial effects are that: the method is suitable for the neural network computational graph with any complex structure, and solves the problems that the neural network computational graph is not good for splitting sub-graphs and different computational units are not good for processing sub-graphs.
Description
Technical Field
The invention relates to the technical field of AI (Artificial Intelligence ), in particular to an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler.
Background
With the continuous development of AI technology, a neural network algorithm based on deep learning has become a mainstream mode of AI research at present. In consideration of cost, power consumption, privacy and the like, more and more application scenes migrate the calculation of the AI algorithm from the cloud to the mobile embedded terminal equipment.
The current embedded device chip is often a plurality of computing units simultaneously, and besides general-purpose CPUs (Central Processing Unit/processors, central processing units), there are special GPUs (Graphics Processing Unit, graphics processors) for accelerating AI algorithms, NPUs (network process units, network processors), FPGAs (Field-Programmable GateArray, field programmable gate arrays) and other AI accelerators. The CPU has good universality and programmability, but the performance and energy efficiency ratio are poor, the AI accelerator has high performance and energy efficiency ratio, but the programmability is poor, some AI accelerators are not programmable, some AI accelerators are programmable, the programming complex development period is long, and some AI accelerators only support shaping calculation, so that the network precision loss is large.
By combining the characteristics of a CPU and an AI accelerator, by carrying out sub-graph segmentation on a neural network, calculating a computationally intensive operator on the AI accelerator (such as a Convolution operator Convolvulation Op, a full-connected operator Fully-connected Op and the like), and calculating some operators with complex calculation logic or operators unsuitable for shaping calculation on the CPU (such as operators for pre-processing and post-processing of a detection network and the like), the method is a solution of the current AI heterogeneous calculation.
In view of this, how to automatically segment the neural network computation graph to realize heterogeneous computation is a problem to be solved at present. The existing schemes are only suitable for simple operator chained neural network computational graphs, and more neural network computational graphs contain more complex operator connection dependency relationships, such as a concat operator needs a plurality of inputs, such as a residual module in a residual network, and all involve non-chained neural network computational graphs. How to find a calculation method capable of carrying out sub-graph segmentation on any neural network calculation graph is a great challenge to be solved by an AI heterogeneous compiler.
The problem to be solved at present is to find out an algorithm for automatically dividing sub-graphs, and divide operator nodes deployed to different computing units in a neural network computing graph into different sub-graphs.
Disclosure of Invention
Aiming at the problems in the prior art, an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler is provided.
The specific technical scheme is as follows:
an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler is used for segmenting operator nodes deployed to different computing units in a neural network computation graph into different neural network computation subgraphs, and comprises the following steps:
step S1, creating a node list;
s2, numbering all nodes of the neural network computational graph in advance, and recording the input dependency relationship and the output dependency relationship of each node;
s3, acquiring any node without any input dependency relationship from the neural network calculation graph;
s4, deleting the node without the input dependency and all connecting edges taking the node without the input dependency as a starting point, and storing the number of the deleted node into the node list;
step S5, repeating the step S2 and the step S3 until the neural network calculation graph is empty, and simultaneously storing the deleted number of the node into the node list;
and S6, acquiring the node list, automatically segmenting according to the attribute colors of the nodes in the node list, and acquiring the input dependency relationship and the output dependency relationship related to the nodes to process and obtain each neural network calculation subgraph.
Preferably, before the step S6, the method further includes: judging whether the neural network calculation graph is empty or not;
if not, turning to the step S3;
if yes, turning to the step S6.
Preferably, the step S6 includes the steps of:
step S60, creating a sub-graph queue warehouse to store sub-graph node queues;
step S61, acquiring a current first node from the node list;
step S62, judging whether the color of the last sub-graph node queue in the sub-graph queue warehouse is the same as the color of the current first node in the node list;
if yes, go to step S63;
if not, turning to step S64;
step S63, adding the current first node in the node list to the subgraph queue warehouse;
step S64, a new sub-graph node queue is newly built in the sub-graph queue warehouse, and a current first node in the node list is used as a first node of the new sub-graph node queue;
step S65, repeating the steps S61-S62 until the node list is empty;
and step S66, acquiring each sub-graph node queue from the sub-graph queue warehouse, and connecting the nodes of each sub-graph node queue according to the input dependency relationship and the output dependency relationship of each node to obtain each neural network calculation sub-graph.
Preferably, before the step S61, the method further includes:
judging whether the node list is empty or not;
if not, turning to the step S61;
if yes, go to step S66.
Preferably, in the step S64, the color of the new child node queue is the same as the color of the current first node in the node list.
Preferably, in the method for automatically splitting the neural network subgraph, the neural network computation graph is split into a plurality of the neural network computation subgraphs, and is deployed on different computation units for computation.
Preferably, the different computing units include a central processor, a graphics processor, and a digital signal processor.
The invention provides an automatic neural network subgraph segmentation method applied to an AI heterogeneous compiler, which has the beneficial effects that:
(1) According to the input dependency and the output dependency of the nodes of the neural network calculation graph, the nodes of the neural network calculation graph are automatically arranged, so that each neural network calculation subgraph is obtained, the problem that the subgraph is not well segmented due to a complex dependency structure in the neural network calculation graph is solved, the running sequence dependency of the nodes is ensured, and the method is applicable to the neural network calculation graph with any complex structure;
(2) The neural network computing graph is divided into a plurality of neural network computing subgraphs and is deployed on different computing units for computing, so that the problem of subgraphs of the neural network computing graph of the different computing units is solved, the method is applicable to not only the situations of two computing units but also the situations of a plurality of computing units, and the method is good in universality, so that the optimal computing force of the whole hardware chip is fully exerted, and the efficient heterogeneous computing of an AI algorithm in actual hardware is completed.
Drawings
Embodiments of the present invention will now be described more fully with reference to the accompanying drawings. The drawings, however, are for illustration and description only and are not intended as a definition of the limits of the invention.
FIG. 1 is a schematic diagram of a neural network computational graph in the prior art;
FIG. 2 is a flowchart illustrating steps of a method for automatically slicing a neural network subgraph applied to an AI heterogeneous compiler according to an embodiment of the present invention;
FIG. 3 is a flowchart illustrating steps of an embodiment one of a method for automatically slicing a neural network subgraph applied to an AI heterogeneous compiler according to an embodiment of the present invention;
FIG. 4 is a flowchart illustrating a second exemplary embodiment of a method for automatically slicing a neural network subgraph applied to an AI heterogeneous compiler;
FIG. 5 is a graph of a division of each neural network computation subgraph of a neural network computation graph of an embodiment II of an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler according to an embodiment of the present invention;
fig. 6 is a flowchart illustrating a third exemplary embodiment of a method for automatically slicing a neural network subgraph applied to an AI-heterogeneous compiler.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
It should be noted that, without conflict, the embodiments of the present invention and features of the embodiments may be combined with each other.
The invention is further described below with reference to the drawings and specific examples, which are not intended to be limiting.
In the prior art, the neural network computational graph may be considered a directed acyclic graph, as shown in FIG. 1, with each node being an operator and each arrow representing a relationship between operators. Assuming that the neural network computation graph is executed by two devices, i.e., a central processor or a graphics processor, in fig. 1, node 2, node 3, node 5, and node 6 may respectively use the central processor to compute the processing, and node 4 may use the graphics processor to compute the processing.
While the above described neural network computational graph that is suitable for simple operator chaining only involves more complex operator join dependencies for more and more neural network computational graphs, such as the concat operator requires multiple inputs, such as the residual module in the residual network, both involve non-chained neural network computational graphs. How to find a calculation method capable of carrying out sub-graph segmentation on any neural network calculation graph is a great challenge to be solved by an AI heterogeneous compiler.
Aiming at the defects in the prior art, the problem to be solved at present is to find an algorithm for automatically dividing the subgraph, and divide the operator nodes with different colors in the neural network calculation graph into different subgraphs.
Therefore, the invention provides an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler, which is used for deploying a neural network computational graph to operator nodes of different computational units to segment the operator nodes to different neural network computational subgraphs, and comprises the following steps:
step S1, creating a node list;
s2, numbering all nodes of the neural network computational graph in advance, and recording the input dependency relationship and the output dependency relationship of each node;
s3, acquiring any node without any input dependency relationship from the neural network calculation graph;
s4, deleting the input nodes without the dependency relationship and all the connecting edges taking the input nodes without the dependency relationship as starting points, and storing the numbers of the deleted nodes into a node list;
step S5, repeating the step S2 and the step S3 until the neural network calculation graph is empty, and simultaneously storing the number of the deleted node into a node list;
and S6, acquiring a node list, automatically segmenting according to the attribute color (the type of a computing unit to which a node operator belongs) of the node in the node list, and acquiring and processing the input dependency relationship and the output dependency relationship related to the node to obtain each neural network computing subgraph.
Through the technical scheme of the automatic segmentation neural network subgraph method applied to the AI heterogeneous compiler, the automatic segmentation neural network subgraph method is suitable for any neural network computational graph with a complex structure, and according to the input dependency and the output dependency of the nodes of the neural network computational graph, the nodes of the neural network computational graph are automatically arranged, so that each neural network computational subgraph is obtained, the problem that the subgraph is not segmented well due to the complex dependency structure in the neural network computational graph is solved, and the running sequence dependency of the nodes is ensured.
Further, the neural network calculation graph is divided into a plurality of neural network calculation subgraphs and is deployed on different calculation units for calculation, wherein the different calculation units comprise a central processing unit, a graphic processor and a digital signal processor, the problem that the AI heterogeneous compiler divides the subgraphs of the neural network calculation graph of the different calculation units is solved, the method is not only applicable to the situation of two calculation units, but also applicable to the situation of a plurality of calculation units, the universality is good, the optimal calculation force of the whole hardware chip is fully exerted, and the efficient heterogeneous calculation of the AI algorithm in actual hardware is completed.
In the above technical solution, as a preferred embodiment, as shown in fig. 3, before step S6, the method further includes: judging whether the neural network calculation graph is empty or not;
if not, turning to step S3;
if yes, go to step S6.
In this embodiment, the purpose of judging whether the neural network computation graph is empty is to thoroughly acquire the nodes with no input dependency relationship in the neural network computation graph, and store the numbers of the deleted nodes in the node list, automatically segment the nodes according to the colors in the node list, and acquire the input dependency relationship and the output dependency relationship related to the nodes to process and obtain each neural network computation subgraph.
Furthermore, the method for automatically splitting the neural network subgraph is suitable for any neural network computational graph with a complex structure, solves the problem that subgraphs are not split well in the neural network computational graph due to the complex dependency structure, and ensures the running sequence dependency sequence of nodes.
In the above technical solution, as a preferred embodiment, as shown in fig. 4, step S6 includes the following steps:
step S60, creating a sub-graph queue warehouse to store sub-graph node queues;
step S61, acquiring a current first node from a node list;
step S62, judging whether the color of the last sub-graph node queue in the sub-graph queue warehouse is the same as the color of the current first node in the node list;
if yes, go to step S63;
if not, turning to step S64;
step S63, adding the current first node in the node list into a subgraph queue warehouse;
step S64, a new sub-graph node queue is newly built in the sub-graph queue warehouse, and a current first node in the node list is used as a first node of the new sub-graph node queue, wherein the color of the new sub-graph node queue is the same as that of the current first node in the node list;
step S65, repeating the steps S61-S62 until the node list is empty;
and step S66, each sub-graph node queue is obtained from the sub-graph queue warehouse, and the nodes of each sub-graph node queue are connected according to the input dependency relationship and the output dependency relationship of each node so as to obtain each neural network calculation sub-graph.
In this embodiment, taking the neural network calculation diagram of fig. 1 as an example, through steps S1 to S5, the obtained node list is shown in table one, the node 1, the node 2, the node 3, the node 5, and the node 6 are in the same color, the node 4 is in another color, after step S6, the colors of the node 1, the node 2, the node 3, the node 4, the node 5, and the node 6 in the node list are segmented, so as to obtain a sub-graph a, a sub-graph B, and a sub-graph C, as shown in fig. 5.
List one
1 | 2 | 3 | 4 | 5 | 6 |
Further, according to the input dependency relationship and the output dependency relationship of the nodes of the neural network computation graph, the nodes of the neural network computation graph are automatically arranged, and according to the color attribute of the nodes, the node list is automatically divided into each neural network computation subgraph.
Furthermore, the method for automatically splitting the neural network subgraph is suitable for any neural network computational graph with a complex structure, solves the problem that subgraphs are not split well in the neural network computational graph due to the complex dependency structure, and ensures the running sequence dependency sequence of nodes.
In the above technical solution, as a preferred embodiment, as shown in fig. 6, before step S61, the method further includes:
judging whether the node list is empty or not;
if not, turning to step S61;
if yes, go to step S66.
In this embodiment, the purpose of judging whether the node list is empty is to be able to thoroughly judge that the colors of the nodes in the node list and the colors in the sub-graph node queues are sequentially matched and not missed, then obtain each sub-graph node queue from the sub-graph queue warehouse, and connect the nodes of each sub-graph node queue according to the input dependency relationship and the output dependency relationship of each node, so as to obtain each neural network calculation sub-graph.
Furthermore, the method for automatically splitting the neural network subgraph is suitable for any neural network computational graph with a complex structure, solves the problem that subgraphs are not split well in the neural network computational graph due to the complex dependency structure, and ensures the running sequence dependency sequence of nodes.
The foregoing description is only illustrative of the preferred embodiments of the present invention and is not to be construed as limiting the scope of the invention, and it will be appreciated by those skilled in the art that equivalent substitutions and obvious variations may be made using the description and illustrations of the present invention, and are intended to be included within the scope of the present invention.
Claims (7)
1. An automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler is used for segmenting operator nodes deployed to different computing units in a neural network computation graph into different neural network computation subgraphs, and is characterized by comprising the following steps:
step S1, creating a node list;
s2, numbering all nodes of the neural network computational graph in advance, and recording the input dependency relationship and the output dependency relationship of each node;
s3, acquiring any node without any input dependency relationship from the neural network calculation graph;
s4, deleting the node without the input dependency and all connecting edges taking the node without the input dependency as a starting point, and storing the number of the deleted node into the node list;
step S5, repeating the step S2 and the step S3 until the neural network calculation graph is empty, and simultaneously storing the deleted number of the node into the node list;
and S6, acquiring the node list, automatically segmenting according to the attribute colors of the nodes in the node list, and acquiring the input dependency relationship and the output dependency relationship related to the nodes to process and obtain each neural network calculation subgraph.
2. The automatic slicing neural network subgraph method of claim 1 further comprising, prior to step S6: judging whether the neural network calculation graph is empty or not;
if not, turning to the step S3;
if yes, turning to the step S6.
3. The automatic slicing neural network subgraph method of claim 1 wherein step S6 includes the steps of:
step S60, creating a sub-graph queue warehouse to store sub-graph node queues;
step S61, acquiring a current first node from the node list;
step S62, judging whether the color of the last sub-graph node queue in the sub-graph queue warehouse is the same as the color of the current first node in the node list;
if yes, go to step S63;
if not, turning to step S64;
step S63, adding the current first node in the node list to the subgraph queue warehouse;
step S64, a new sub-graph node queue is newly built in the sub-graph queue warehouse, and a current first node in the node list is used as a first node of the new sub-graph node queue;
step S65, repeating the steps S61-S62 until the node list is empty;
and step S66, acquiring each sub-graph node queue from the sub-graph queue warehouse, and connecting the nodes of each sub-graph node queue according to the input dependency relationship and the output dependency relationship of each node to obtain each neural network calculation sub-graph.
4. The automatic slicing neural network subgraph method of claim 3 further comprising, prior to step S61:
judging whether the node list is empty or not;
if not, turning to the step S61;
if yes, go to step S66.
5. The automatic slicing neural network subgraph method of claim 3 wherein in step S64 the color of the new subgraph node queue is the same as the color of the current first node in the node list.
6. The automatic slicing neural network subgraph method of claim 1 wherein in the automatic slicing neural network subgraph method, the neural network computation graph is divided into a plurality of the neural network computation subgraphs and deployed for computation on different computation units.
7. The automatic slicing neural network subgraph method of claim 6 wherein different ones of the computing units include a central processor, a graphics processor and a digital signal processor.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911312625.0A CN111062467B (en) | 2019-12-18 | 2019-12-18 | Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911312625.0A CN111062467B (en) | 2019-12-18 | 2019-12-18 | Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111062467A CN111062467A (en) | 2020-04-24 |
CN111062467B true CN111062467B (en) | 2023-05-12 |
Family
ID=70301038
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911312625.0A Active CN111062467B (en) | 2019-12-18 | 2019-12-18 | Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111062467B (en) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111738434B (en) * | 2020-06-03 | 2023-04-07 | 中国科学院计算技术研究所 | Method for executing deep neural network on heterogeneous processing unit |
CN111860820A (en) * | 2020-07-31 | 2020-10-30 | 北京灵汐科技有限公司 | Neural network operator dividing method and device and dividing equipment |
CN112070213A (en) * | 2020-08-28 | 2020-12-11 | Oppo广东移动通信有限公司 | Neural network model optimization method, device, equipment and storage medium |
CN112598121A (en) * | 2020-12-21 | 2021-04-02 | 北京时代民芯科技有限公司 | Efficient operator optimization method for deep learning compiler |
CN112711422B (en) * | 2020-12-31 | 2024-01-19 | 北京清微智能科技有限公司 | Neural network compiling optimization method and system |
WO2023122854A1 (en) * | 2021-12-27 | 2023-07-06 | 华为技术有限公司 | Data processing method and apparatus |
WO2024178522A1 (en) * | 2023-02-27 | 2024-09-06 | 华为技术有限公司 | Computational graph splitting method and related device |
CN117576125B (en) * | 2024-01-16 | 2024-04-16 | 芯瞳半导体技术(山东)有限公司 | Neural network calculation graph segmentation method, device, equipment and storage medium |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018206261A (en) * | 2017-06-08 | 2018-12-27 | 日本電信電話株式会社 | Word division estimation model learning device, word division device, method and program |
CN109902819A (en) * | 2019-02-12 | 2019-06-18 | Oppo广东移动通信有限公司 | Neural computing method, apparatus, mobile terminal and storage medium |
WO2019177824A1 (en) * | 2018-03-14 | 2019-09-19 | Microsoft Technology Licensing, Llc | Hardware accelerated neural network subgraphs |
-
2019
- 2019-12-18 CN CN201911312625.0A patent/CN111062467B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018206261A (en) * | 2017-06-08 | 2018-12-27 | 日本電信電話株式会社 | Word division estimation model learning device, word division device, method and program |
WO2019177824A1 (en) * | 2018-03-14 | 2019-09-19 | Microsoft Technology Licensing, Llc | Hardware accelerated neural network subgraphs |
CN109902819A (en) * | 2019-02-12 | 2019-06-18 | Oppo广东移动通信有限公司 | Neural computing method, apparatus, mobile terminal and storage medium |
Non-Patent Citations (1)
Title |
---|
方良达 ; 余永权 ; .属性依赖理论及其在神经网络中的应用.计算机应用.2010,(04),全文. * |
Also Published As
Publication number | Publication date |
---|---|
CN111062467A (en) | 2020-04-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111062467B (en) | Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler | |
WO2017116924A1 (en) | Neural network training performance optimization framework | |
WO2021190761A1 (en) | Parallel computing scheme generation for neural networks | |
US20200184366A1 (en) | Scheduling task graph operations | |
US11782724B2 (en) | Parallel decision system and method for distributed data processing | |
WO2019085709A1 (en) | Pooling method and system applied to convolutional neural network | |
US20150379158A1 (en) | Systems and methods for pattern matching and relationship discovery | |
CN112100450A (en) | Graph calculation data segmentation method, terminal device and storage medium | |
CN113886092A (en) | Computation graph execution method and device and related equipment | |
Duan et al. | Computation offloading scheduling for deep neural network inference in mobile computing | |
CN114626552A (en) | Segmentation method and device of machine learning model | |
CN111159577A (en) | Community division method and device, storage medium and electronic device | |
CN114341894A (en) | Hyper-parameter recommendation method for machine learning method | |
CN114003775A (en) | Graph data processing and querying method and system | |
US20160328490A1 (en) | System and method for identifying clones | |
CN113691403B (en) | Topology node configuration method, related device and computer program product | |
CN116933841A (en) | Operator fusion method and device, electronic equipment and computer readable medium | |
CN103268614A (en) | Generation method for prospect spectrogram for multi-prospect co-segmentation | |
CN113051080A (en) | Computation graph execution method and device and heterogeneous platform | |
CN109388428B (en) | Layer traversal method, control device and data processing system | |
Yazıcıoğlu et al. | Decentralized degree regularization for multi-agent networks | |
Taoka et al. | Heuristic algorithms for the marking construction problem of Petri nets | |
Alessio | Another time-complexity analysis for maximal clique enumeration algorithm CLIQUES | |
CN115774800B (en) | NUMA architecture-based time-varying graph processing method, electronic equipment and medium | |
CN111260038B (en) | Implementation method and device of convolutional neural network, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |