CN111062467B - Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler - Google Patents

Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler Download PDF

Info

Publication number
CN111062467B
CN111062467B CN201911312625.0A CN201911312625A CN111062467B CN 111062467 B CN111062467 B CN 111062467B CN 201911312625 A CN201911312625 A CN 201911312625A CN 111062467 B CN111062467 B CN 111062467B
Authority
CN
China
Prior art keywords
node
neural network
graph
sub
queue
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911312625.0A
Other languages
Chinese (zh)
Other versions
CN111062467A (en
Inventor
黄明飞
王海涛
吕春莹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Open Intelligent Machine Shanghai Co ltd
Original Assignee
Open Intelligent Machine Shanghai Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Open Intelligent Machine Shanghai Co ltd filed Critical Open Intelligent Machine Shanghai Co ltd
Priority to CN201911312625.0A priority Critical patent/CN111062467B/en
Publication of CN111062467A publication Critical patent/CN111062467A/en
Application granted granted Critical
Publication of CN111062467B publication Critical patent/CN111062467B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/063Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Neurology (AREA)
  • Stored Programmes (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses an automatic neural network subgraph segmentation method applied to an AI heterogeneous compiler, which comprises the steps of creating a node list; numbering all nodes of the neural network computational graph in advance, and recording the input dependency relationship and the output dependency relationship of each node; acquiring any node without any input dependency relationship from a neural network computational graph; deleting the input nodes without the dependency and all connecting edges taking the input nodes without the dependency as starting points, and storing the numbers of the deleted nodes into a node list; repeating the step S2 and the step S3 until the neural network calculation graph is empty, and storing the number of the deleted node into a node list; and automatically segmenting according to the node attribute colors in the node list to obtain each neural network calculation subgraph through processing. The beneficial effects are that: the method is suitable for the neural network computational graph with any complex structure, and solves the problems that the neural network computational graph is not good for splitting sub-graphs and different computational units are not good for processing sub-graphs.

Description

Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler
Technical Field
The invention relates to the technical field of AI (Artificial Intelligence ), in particular to an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler.
Background
With the continuous development of AI technology, a neural network algorithm based on deep learning has become a mainstream mode of AI research at present. In consideration of cost, power consumption, privacy and the like, more and more application scenes migrate the calculation of the AI algorithm from the cloud to the mobile embedded terminal equipment.
The current embedded device chip is often a plurality of computing units simultaneously, and besides general-purpose CPUs (Central Processing Unit/processors, central processing units), there are special GPUs (Graphics Processing Unit, graphics processors) for accelerating AI algorithms, NPUs (network process units, network processors), FPGAs (Field-Programmable GateArray, field programmable gate arrays) and other AI accelerators. The CPU has good universality and programmability, but the performance and energy efficiency ratio are poor, the AI accelerator has high performance and energy efficiency ratio, but the programmability is poor, some AI accelerators are not programmable, some AI accelerators are programmable, the programming complex development period is long, and some AI accelerators only support shaping calculation, so that the network precision loss is large.
By combining the characteristics of a CPU and an AI accelerator, by carrying out sub-graph segmentation on a neural network, calculating a computationally intensive operator on the AI accelerator (such as a Convolution operator Convolvulation Op, a full-connected operator Fully-connected Op and the like), and calculating some operators with complex calculation logic or operators unsuitable for shaping calculation on the CPU (such as operators for pre-processing and post-processing of a detection network and the like), the method is a solution of the current AI heterogeneous calculation.
In view of this, how to automatically segment the neural network computation graph to realize heterogeneous computation is a problem to be solved at present. The existing schemes are only suitable for simple operator chained neural network computational graphs, and more neural network computational graphs contain more complex operator connection dependency relationships, such as a concat operator needs a plurality of inputs, such as a residual module in a residual network, and all involve non-chained neural network computational graphs. How to find a calculation method capable of carrying out sub-graph segmentation on any neural network calculation graph is a great challenge to be solved by an AI heterogeneous compiler.
The problem to be solved at present is to find out an algorithm for automatically dividing sub-graphs, and divide operator nodes deployed to different computing units in a neural network computing graph into different sub-graphs.
Disclosure of Invention
Aiming at the problems in the prior art, an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler is provided.
The specific technical scheme is as follows:
an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler is used for segmenting operator nodes deployed to different computing units in a neural network computation graph into different neural network computation subgraphs, and comprises the following steps:
step S1, creating a node list;
s2, numbering all nodes of the neural network computational graph in advance, and recording the input dependency relationship and the output dependency relationship of each node;
s3, acquiring any node without any input dependency relationship from the neural network calculation graph;
s4, deleting the node without the input dependency and all connecting edges taking the node without the input dependency as a starting point, and storing the number of the deleted node into the node list;
step S5, repeating the step S2 and the step S3 until the neural network calculation graph is empty, and simultaneously storing the deleted number of the node into the node list;
and S6, acquiring the node list, automatically segmenting according to the attribute colors of the nodes in the node list, and acquiring the input dependency relationship and the output dependency relationship related to the nodes to process and obtain each neural network calculation subgraph.
Preferably, before the step S6, the method further includes: judging whether the neural network calculation graph is empty or not;
if not, turning to the step S3;
if yes, turning to the step S6.
Preferably, the step S6 includes the steps of:
step S60, creating a sub-graph queue warehouse to store sub-graph node queues;
step S61, acquiring a current first node from the node list;
step S62, judging whether the color of the last sub-graph node queue in the sub-graph queue warehouse is the same as the color of the current first node in the node list;
if yes, go to step S63;
if not, turning to step S64;
step S63, adding the current first node in the node list to the subgraph queue warehouse;
step S64, a new sub-graph node queue is newly built in the sub-graph queue warehouse, and a current first node in the node list is used as a first node of the new sub-graph node queue;
step S65, repeating the steps S61-S62 until the node list is empty;
and step S66, acquiring each sub-graph node queue from the sub-graph queue warehouse, and connecting the nodes of each sub-graph node queue according to the input dependency relationship and the output dependency relationship of each node to obtain each neural network calculation sub-graph.
Preferably, before the step S61, the method further includes:
judging whether the node list is empty or not;
if not, turning to the step S61;
if yes, go to step S66.
Preferably, in the step S64, the color of the new child node queue is the same as the color of the current first node in the node list.
Preferably, in the method for automatically splitting the neural network subgraph, the neural network computation graph is split into a plurality of the neural network computation subgraphs, and is deployed on different computation units for computation.
Preferably, the different computing units include a central processor, a graphics processor, and a digital signal processor.
The invention provides an automatic neural network subgraph segmentation method applied to an AI heterogeneous compiler, which has the beneficial effects that:
(1) According to the input dependency and the output dependency of the nodes of the neural network calculation graph, the nodes of the neural network calculation graph are automatically arranged, so that each neural network calculation subgraph is obtained, the problem that the subgraph is not well segmented due to a complex dependency structure in the neural network calculation graph is solved, the running sequence dependency of the nodes is ensured, and the method is applicable to the neural network calculation graph with any complex structure;
(2) The neural network computing graph is divided into a plurality of neural network computing subgraphs and is deployed on different computing units for computing, so that the problem of subgraphs of the neural network computing graph of the different computing units is solved, the method is applicable to not only the situations of two computing units but also the situations of a plurality of computing units, and the method is good in universality, so that the optimal computing force of the whole hardware chip is fully exerted, and the efficient heterogeneous computing of an AI algorithm in actual hardware is completed.
Drawings
Embodiments of the present invention will now be described more fully with reference to the accompanying drawings. The drawings, however, are for illustration and description only and are not intended as a definition of the limits of the invention.
FIG. 1 is a schematic diagram of a neural network computational graph in the prior art;
FIG. 2 is a flowchart illustrating steps of a method for automatically slicing a neural network subgraph applied to an AI heterogeneous compiler according to an embodiment of the present invention;
FIG. 3 is a flowchart illustrating steps of an embodiment one of a method for automatically slicing a neural network subgraph applied to an AI heterogeneous compiler according to an embodiment of the present invention;
FIG. 4 is a flowchart illustrating a second exemplary embodiment of a method for automatically slicing a neural network subgraph applied to an AI heterogeneous compiler;
FIG. 5 is a graph of a division of each neural network computation subgraph of a neural network computation graph of an embodiment II of an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler according to an embodiment of the present invention;
fig. 6 is a flowchart illustrating a third exemplary embodiment of a method for automatically slicing a neural network subgraph applied to an AI-heterogeneous compiler.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
It should be noted that, without conflict, the embodiments of the present invention and features of the embodiments may be combined with each other.
The invention is further described below with reference to the drawings and specific examples, which are not intended to be limiting.
In the prior art, the neural network computational graph may be considered a directed acyclic graph, as shown in FIG. 1, with each node being an operator and each arrow representing a relationship between operators. Assuming that the neural network computation graph is executed by two devices, i.e., a central processor or a graphics processor, in fig. 1, node 2, node 3, node 5, and node 6 may respectively use the central processor to compute the processing, and node 4 may use the graphics processor to compute the processing.
While the above described neural network computational graph that is suitable for simple operator chaining only involves more complex operator join dependencies for more and more neural network computational graphs, such as the concat operator requires multiple inputs, such as the residual module in the residual network, both involve non-chained neural network computational graphs. How to find a calculation method capable of carrying out sub-graph segmentation on any neural network calculation graph is a great challenge to be solved by an AI heterogeneous compiler.
Aiming at the defects in the prior art, the problem to be solved at present is to find an algorithm for automatically dividing the subgraph, and divide the operator nodes with different colors in the neural network calculation graph into different subgraphs.
Therefore, the invention provides an automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler, which is used for deploying a neural network computational graph to operator nodes of different computational units to segment the operator nodes to different neural network computational subgraphs, and comprises the following steps:
step S1, creating a node list;
s2, numbering all nodes of the neural network computational graph in advance, and recording the input dependency relationship and the output dependency relationship of each node;
s3, acquiring any node without any input dependency relationship from the neural network calculation graph;
s4, deleting the input nodes without the dependency relationship and all the connecting edges taking the input nodes without the dependency relationship as starting points, and storing the numbers of the deleted nodes into a node list;
step S5, repeating the step S2 and the step S3 until the neural network calculation graph is empty, and simultaneously storing the number of the deleted node into a node list;
and S6, acquiring a node list, automatically segmenting according to the attribute color (the type of a computing unit to which a node operator belongs) of the node in the node list, and acquiring and processing the input dependency relationship and the output dependency relationship related to the node to obtain each neural network computing subgraph.
Through the technical scheme of the automatic segmentation neural network subgraph method applied to the AI heterogeneous compiler, the automatic segmentation neural network subgraph method is suitable for any neural network computational graph with a complex structure, and according to the input dependency and the output dependency of the nodes of the neural network computational graph, the nodes of the neural network computational graph are automatically arranged, so that each neural network computational subgraph is obtained, the problem that the subgraph is not segmented well due to the complex dependency structure in the neural network computational graph is solved, and the running sequence dependency of the nodes is ensured.
Further, the neural network calculation graph is divided into a plurality of neural network calculation subgraphs and is deployed on different calculation units for calculation, wherein the different calculation units comprise a central processing unit, a graphic processor and a digital signal processor, the problem that the AI heterogeneous compiler divides the subgraphs of the neural network calculation graph of the different calculation units is solved, the method is not only applicable to the situation of two calculation units, but also applicable to the situation of a plurality of calculation units, the universality is good, the optimal calculation force of the whole hardware chip is fully exerted, and the efficient heterogeneous calculation of the AI algorithm in actual hardware is completed.
In the above technical solution, as a preferred embodiment, as shown in fig. 3, before step S6, the method further includes: judging whether the neural network calculation graph is empty or not;
if not, turning to step S3;
if yes, go to step S6.
In this embodiment, the purpose of judging whether the neural network computation graph is empty is to thoroughly acquire the nodes with no input dependency relationship in the neural network computation graph, and store the numbers of the deleted nodes in the node list, automatically segment the nodes according to the colors in the node list, and acquire the input dependency relationship and the output dependency relationship related to the nodes to process and obtain each neural network computation subgraph.
Furthermore, the method for automatically splitting the neural network subgraph is suitable for any neural network computational graph with a complex structure, solves the problem that subgraphs are not split well in the neural network computational graph due to the complex dependency structure, and ensures the running sequence dependency sequence of nodes.
In the above technical solution, as a preferred embodiment, as shown in fig. 4, step S6 includes the following steps:
step S60, creating a sub-graph queue warehouse to store sub-graph node queues;
step S61, acquiring a current first node from a node list;
step S62, judging whether the color of the last sub-graph node queue in the sub-graph queue warehouse is the same as the color of the current first node in the node list;
if yes, go to step S63;
if not, turning to step S64;
step S63, adding the current first node in the node list into a subgraph queue warehouse;
step S64, a new sub-graph node queue is newly built in the sub-graph queue warehouse, and a current first node in the node list is used as a first node of the new sub-graph node queue, wherein the color of the new sub-graph node queue is the same as that of the current first node in the node list;
step S65, repeating the steps S61-S62 until the node list is empty;
and step S66, each sub-graph node queue is obtained from the sub-graph queue warehouse, and the nodes of each sub-graph node queue are connected according to the input dependency relationship and the output dependency relationship of each node so as to obtain each neural network calculation sub-graph.
In this embodiment, taking the neural network calculation diagram of fig. 1 as an example, through steps S1 to S5, the obtained node list is shown in table one, the node 1, the node 2, the node 3, the node 5, and the node 6 are in the same color, the node 4 is in another color, after step S6, the colors of the node 1, the node 2, the node 3, the node 4, the node 5, and the node 6 in the node list are segmented, so as to obtain a sub-graph a, a sub-graph B, and a sub-graph C, as shown in fig. 5.
List one
1 2 3 4 5 6
Further, according to the input dependency relationship and the output dependency relationship of the nodes of the neural network computation graph, the nodes of the neural network computation graph are automatically arranged, and according to the color attribute of the nodes, the node list is automatically divided into each neural network computation subgraph.
Furthermore, the method for automatically splitting the neural network subgraph is suitable for any neural network computational graph with a complex structure, solves the problem that subgraphs are not split well in the neural network computational graph due to the complex dependency structure, and ensures the running sequence dependency sequence of nodes.
In the above technical solution, as a preferred embodiment, as shown in fig. 6, before step S61, the method further includes:
judging whether the node list is empty or not;
if not, turning to step S61;
if yes, go to step S66.
In this embodiment, the purpose of judging whether the node list is empty is to be able to thoroughly judge that the colors of the nodes in the node list and the colors in the sub-graph node queues are sequentially matched and not missed, then obtain each sub-graph node queue from the sub-graph queue warehouse, and connect the nodes of each sub-graph node queue according to the input dependency relationship and the output dependency relationship of each node, so as to obtain each neural network calculation sub-graph.
Furthermore, the method for automatically splitting the neural network subgraph is suitable for any neural network computational graph with a complex structure, solves the problem that subgraphs are not split well in the neural network computational graph due to the complex dependency structure, and ensures the running sequence dependency sequence of nodes.
The foregoing description is only illustrative of the preferred embodiments of the present invention and is not to be construed as limiting the scope of the invention, and it will be appreciated by those skilled in the art that equivalent substitutions and obvious variations may be made using the description and illustrations of the present invention, and are intended to be included within the scope of the present invention.

Claims (7)

1. An automatic segmentation neural network subgraph method applied to an AI heterogeneous compiler is used for segmenting operator nodes deployed to different computing units in a neural network computation graph into different neural network computation subgraphs, and is characterized by comprising the following steps:
step S1, creating a node list;
s2, numbering all nodes of the neural network computational graph in advance, and recording the input dependency relationship and the output dependency relationship of each node;
s3, acquiring any node without any input dependency relationship from the neural network calculation graph;
s4, deleting the node without the input dependency and all connecting edges taking the node without the input dependency as a starting point, and storing the number of the deleted node into the node list;
step S5, repeating the step S2 and the step S3 until the neural network calculation graph is empty, and simultaneously storing the deleted number of the node into the node list;
and S6, acquiring the node list, automatically segmenting according to the attribute colors of the nodes in the node list, and acquiring the input dependency relationship and the output dependency relationship related to the nodes to process and obtain each neural network calculation subgraph.
2. The automatic slicing neural network subgraph method of claim 1 further comprising, prior to step S6: judging whether the neural network calculation graph is empty or not;
if not, turning to the step S3;
if yes, turning to the step S6.
3. The automatic slicing neural network subgraph method of claim 1 wherein step S6 includes the steps of:
step S60, creating a sub-graph queue warehouse to store sub-graph node queues;
step S61, acquiring a current first node from the node list;
step S62, judging whether the color of the last sub-graph node queue in the sub-graph queue warehouse is the same as the color of the current first node in the node list;
if yes, go to step S63;
if not, turning to step S64;
step S63, adding the current first node in the node list to the subgraph queue warehouse;
step S64, a new sub-graph node queue is newly built in the sub-graph queue warehouse, and a current first node in the node list is used as a first node of the new sub-graph node queue;
step S65, repeating the steps S61-S62 until the node list is empty;
and step S66, acquiring each sub-graph node queue from the sub-graph queue warehouse, and connecting the nodes of each sub-graph node queue according to the input dependency relationship and the output dependency relationship of each node to obtain each neural network calculation sub-graph.
4. The automatic slicing neural network subgraph method of claim 3 further comprising, prior to step S61:
judging whether the node list is empty or not;
if not, turning to the step S61;
if yes, go to step S66.
5. The automatic slicing neural network subgraph method of claim 3 wherein in step S64 the color of the new subgraph node queue is the same as the color of the current first node in the node list.
6. The automatic slicing neural network subgraph method of claim 1 wherein in the automatic slicing neural network subgraph method, the neural network computation graph is divided into a plurality of the neural network computation subgraphs and deployed for computation on different computation units.
7. The automatic slicing neural network subgraph method of claim 6 wherein different ones of the computing units include a central processor, a graphics processor and a digital signal processor.
CN201911312625.0A 2019-12-18 2019-12-18 Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler Active CN111062467B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911312625.0A CN111062467B (en) 2019-12-18 2019-12-18 Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911312625.0A CN111062467B (en) 2019-12-18 2019-12-18 Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler

Publications (2)

Publication Number Publication Date
CN111062467A CN111062467A (en) 2020-04-24
CN111062467B true CN111062467B (en) 2023-05-12

Family

ID=70301038

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911312625.0A Active CN111062467B (en) 2019-12-18 2019-12-18 Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler

Country Status (1)

Country Link
CN (1) CN111062467B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111738434B (en) * 2020-06-03 2023-04-07 中国科学院计算技术研究所 Method for executing deep neural network on heterogeneous processing unit
CN111860820A (en) * 2020-07-31 2020-10-30 北京灵汐科技有限公司 Neural network operator dividing method and device and dividing equipment
CN112070213A (en) * 2020-08-28 2020-12-11 Oppo广东移动通信有限公司 Neural network model optimization method, device, equipment and storage medium
CN112598121A (en) * 2020-12-21 2021-04-02 北京时代民芯科技有限公司 Efficient operator optimization method for deep learning compiler
CN112711422B (en) * 2020-12-31 2024-01-19 北京清微智能科技有限公司 Neural network compiling optimization method and system
WO2023122854A1 (en) * 2021-12-27 2023-07-06 华为技术有限公司 Data processing method and apparatus
WO2024178522A1 (en) * 2023-02-27 2024-09-06 华为技术有限公司 Computational graph splitting method and related device
CN117576125B (en) * 2024-01-16 2024-04-16 芯瞳半导体技术(山东)有限公司 Neural network calculation graph segmentation method, device, equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018206261A (en) * 2017-06-08 2018-12-27 日本電信電話株式会社 Word division estimation model learning device, word division device, method and program
CN109902819A (en) * 2019-02-12 2019-06-18 Oppo广东移动通信有限公司 Neural computing method, apparatus, mobile terminal and storage medium
WO2019177824A1 (en) * 2018-03-14 2019-09-19 Microsoft Technology Licensing, Llc Hardware accelerated neural network subgraphs

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018206261A (en) * 2017-06-08 2018-12-27 日本電信電話株式会社 Word division estimation model learning device, word division device, method and program
WO2019177824A1 (en) * 2018-03-14 2019-09-19 Microsoft Technology Licensing, Llc Hardware accelerated neural network subgraphs
CN109902819A (en) * 2019-02-12 2019-06-18 Oppo广东移动通信有限公司 Neural computing method, apparatus, mobile terminal and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
方良达 ; 余永权 ; .属性依赖理论及其在神经网络中的应用.计算机应用.2010,(04),全文. *

Also Published As

Publication number Publication date
CN111062467A (en) 2020-04-24

Similar Documents

Publication Publication Date Title
CN111062467B (en) Automatic neural network subgraph segmentation method applied to AI heterogeneous compiler
WO2017116924A1 (en) Neural network training performance optimization framework
WO2021190761A1 (en) Parallel computing scheme generation for neural networks
US20200184366A1 (en) Scheduling task graph operations
US11782724B2 (en) Parallel decision system and method for distributed data processing
WO2019085709A1 (en) Pooling method and system applied to convolutional neural network
US20150379158A1 (en) Systems and methods for pattern matching and relationship discovery
CN112100450A (en) Graph calculation data segmentation method, terminal device and storage medium
CN113886092A (en) Computation graph execution method and device and related equipment
Duan et al. Computation offloading scheduling for deep neural network inference in mobile computing
CN114626552A (en) Segmentation method and device of machine learning model
CN111159577A (en) Community division method and device, storage medium and electronic device
CN114341894A (en) Hyper-parameter recommendation method for machine learning method
CN114003775A (en) Graph data processing and querying method and system
US20160328490A1 (en) System and method for identifying clones
CN113691403B (en) Topology node configuration method, related device and computer program product
CN116933841A (en) Operator fusion method and device, electronic equipment and computer readable medium
CN103268614A (en) Generation method for prospect spectrogram for multi-prospect co-segmentation
CN113051080A (en) Computation graph execution method and device and heterogeneous platform
CN109388428B (en) Layer traversal method, control device and data processing system
Yazıcıoğlu et al. Decentralized degree regularization for multi-agent networks
Taoka et al. Heuristic algorithms for the marking construction problem of Petri nets
Alessio Another time-complexity analysis for maximal clique enumeration algorithm CLIQUES
CN115774800B (en) NUMA architecture-based time-varying graph processing method, electronic equipment and medium
CN111260038B (en) Implementation method and device of convolutional neural network, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant