WO2021259039A1 - Procédé, système et dispositif de personnalisation de modèle de réseau neuronal, et support de stockage - Google Patents

Procédé, système et dispositif de personnalisation de modèle de réseau neuronal, et support de stockage Download PDF

Info

Publication number
WO2021259039A1
WO2021259039A1 PCT/CN2021/098288 CN2021098288W WO2021259039A1 WO 2021259039 A1 WO2021259039 A1 WO 2021259039A1 CN 2021098288 W CN2021098288 W CN 2021098288W WO 2021259039 A1 WO2021259039 A1 WO 2021259039A1
Authority
WO
WIPO (PCT)
Prior art keywords
model
calculation
node
neural network
calculation graph
Prior art date
Application number
PCT/CN2021/098288
Other languages
English (en)
Chinese (zh)
Inventor
熊超
蔡权雄
牛昕宇
Original Assignee
深圳鲲云信息科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳鲲云信息科技有限公司 filed Critical 深圳鲲云信息科技有限公司
Publication of WO2021259039A1 publication Critical patent/WO2021259039A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Definitions

  • This application relates to the field of neural network technology, for example, to a method, system, device, and storage medium for customizing a neural network model.
  • the deep learning neural network is realized through the data flow calculation graph.
  • the data (tensor, etc.) flows from the input node through the collection nodes, and finally the inference result is obtained at the output node.
  • Different frameworks implement data flow calculation graphs in different ways. Some use static calculation graphs, and some are dynamically created during operation. Either way, the structure calculation diagram and weights of a neural network model will eventually be obtained.
  • Different frameworks have their own model deployment methods. For artificial intelligence chip manufacturers, how to analyze and deploy multiple deep learning frameworks more easily is a key issue in their development chip tool chain. For the conversion of different deep learning frameworks, it is usually necessary to analyze the neural network models of different frameworks, generate a specific intermediate representation (Intermediate Representation), customize the intermediate representation and deploy it to the hardware device.
  • This application provides a method, system, equipment, and storage medium for customizing a neural network model to realize the generation of a custom neural network model to be suitable for different deep learning frameworks.
  • a method for customizing a neural network model which includes:
  • Obtain a preset neural network model convert the neural network model into a static calculation graph model; construct a directed acyclic calculation graph model according to the node information of the first calculation node of the static calculation graph model; pass the preset
  • the graph analysis engine converts the directed acyclic calculation graph model into an intermediate expression calculation graph; and generates a customized target neural network model according to the intermediate expression calculation graph.
  • a customized system of neural network model is also provided, which includes:
  • the model acquisition module is set to obtain a preset neural network model; the model conversion module is set to convert the neural network model into a static calculation graph model; the model construction module is set to perform the first calculation according to the static calculation graph model
  • the node information of the nodes constructs a directed acyclic calculation graph model; the calculation graph conversion module is configured to convert the directed acyclic calculation graph model into an intermediate expression calculation graph through a preset graph analysis engine; the model generation module is set To generate a customized target neural network model according to the intermediate expression calculation graph.
  • a customized device for a neural network model includes: one or more processors; a storage device configured to store one or more programs. Execution by each processor, so that the one or more processors implement the above-mentioned customized method of the neural network model.
  • a computer-readable storage medium is also provided, on which a computer program is stored, and when the program is executed by a processor, the method for customizing the above-mentioned neural network model is realized.
  • FIG. 1 is a schematic flowchart of a method for customizing a neural network model provided in Embodiment 1 of the present application;
  • FIG. 2 is a schematic flowchart of a method for customizing a neural network model provided in Embodiment 2 of the present application;
  • FIG. 3 is a schematic flowchart of step S230 in FIG. 2 according to the second embodiment of the present application;
  • FIG. 4 is a schematic flowchart of step S260 in FIG. 2 according to the second embodiment of the present application.
  • FIG. 5 is a schematic structural diagram of a customized system for a neural network model provided in Embodiment 3 of the present application.
  • Fig. 6 is a schematic structural diagram of a neural network model customized device provided in the fourth embodiment of the present application.
  • first”, “second”, etc. may be used herein to describe various directions, actions, steps or elements, etc., but these directions, actions, steps or elements are not limited by these terms. These terms are only used to distinguish a first direction, action, step or element from another direction, action, step or element.
  • the first computing node may be referred to as the second computing node, and similarly, the second computing node may be referred to as the first computing node.
  • Both the first computing node and the second computing node are computing nodes, but they are not the same computing node.
  • the terms “first”, “second”, etc. cannot be understood as indicating or implying relative importance or implicitly indicating the number of indicated technical features. Therefore, the features defined with “first” and “second” may explicitly or implicitly include one or more of the features.
  • "a plurality of” means at least two, such as two, three, etc., unless otherwise defined.
  • Embodiment 1 of the present application provides a method for customizing a neural network model, which includes:
  • the static computational graph model is a pre-defined computational graph model that cannot be modified during the inference process.
  • the static computational graph model can be a data flow-based programming that removes the training-related data and retains only the inference-related data.
  • the symbolic mathematics system (tensorflow) static calculation graph model can also be the open source Python machine learning library (pyotrch) static calculation graph model exported using pyotrch's official deployment tool (jit).
  • S140 Convert the directed acyclic calculation graph model into an intermediate expression calculation graph through a preset graph analysis engine.
  • the node information includes operators, operator attributes, model parameters, and edge relations, etc., where the operator In order to complete a kind of operation node, such as the node that completes addition, subtraction, multiplication and division, the large operator can be a combination of multiple mathematical operations, such as convolution operation and pooling operation; the attribute of the operator means that some operator nodes need to be specified Some parameters of the convolution operation, such as the kernel size (kernel_size) and stride length (strides), etc.; model parameters refer to the trainable parameters of the neural network model during training; the edge relationship refers to the input of the operator node The output relationship, the output of the predecessor node is the input of the node, which is generally represented by a string.
  • kernel size kernel size
  • stride length stride length
  • a preset tool is used to perform node mapping, corresponding to new nodes, adding node information, and adding edge relationships, thereby constructing a directed acyclic computing graph model.
  • the directed acyclic calculation graph model is obtained, multiple analysis rules are predefined in the preset graph analysis engine according to user needs, and these analysis rules are repeatedly iterated to match, write and replace the directed acyclic calculation graph model. Thereby converted into an intermediate expression calculation graph. After obtaining the intermediate expression calculation graph, you can easily generate a customized target neural network model through the preset interface.
  • the embodiment of the application obtains a preset neural network model; converts the neural network model into a static calculation graph model; constructs a directed acyclic calculation graph model according to the node information of the first calculation node of the static calculation graph model Convert the directed acyclic calculation graph model into an intermediate expression calculation graph through the preset graph analysis engine; generate a customized target neural network model according to the intermediate expression calculation graph, which solves the problem that the analytical front-end is not conducive to expansion and maintenance
  • the problem of high cost realizes the effect of generating customized neural network models to apply to different deep learning frameworks.
  • the second embodiment of the present application provides a method for customizing a neural network model.
  • the second embodiment of the present application is described on the basis of the first embodiment of the present application.
  • the method includes:
  • S230 Construct a directed acyclic calculation graph model according to the node information of the first calculation node of the static calculation graph model.
  • S240 Visually display the directed acyclic calculation graph model through a preset interface.
  • the user can visualize the directed acyclic calculation graph model in the browser through the Application Programming Interface (API), and check the directed acyclic calculation graph through visualization Whether the node mapping of the model is correct, so as to receive the user's modification of the directed acyclic computational graph model.
  • API Application Programming Interface
  • S260 Convert the directed acyclic calculation graph model into an intermediate expression calculation graph through a preset graph analysis engine.
  • the intermediate expression calculation graph can be converted into a customized intermediate expression through a preset serialization interface, and the intermediate expression is serialized into a customized target neural network model, thereby Completed the conversion of the preset neural network model into a customized target neural network model. Users can simply complete the neural network model extension to different deep learning frameworks, and it is easy to maintain.
  • step S230 in the embodiment of the present application includes:
  • the process of constructing the directed acyclic calculation graph model first obtain the node information of the first calculation node of the static calculation graph model, and use the python dictionary class to temporarily save the node information.
  • the node information is in addition to operators, Operator attributes, model parameters and edge relations, as well as the number of calculation nodes and the order of calculation nodes.
  • the number of calculation nodes includes the number of operators and the number of constant nodes.
  • the order of calculation nodes is each first calculation node in the dictionary record of python. The successor node.
  • the second calculation node after obtaining the number of computing nodes, initialize the same number of second computing nodes as the number of computing nodes, that is, add the same number of second computing nodes as the number of computing nodes, and then correspond the node information one-to-one according to the topological ordering
  • Add to the second calculation node add the operator name, operator attributes, model parameters, input and output and other node information for the second calculation node, thereby performing a one-to-one mapping of the original static calculation graph model to construct An undirected computational graph model with the same graph structure as the original static computational graph model.
  • S234 Add an edge relationship to the second computing node of the undirected computing graph model according to the sequence of computing nodes to construct a directed acyclic computing graph model.
  • the second computing node of the undirected computing graph model can add edge relationships one by one for each edge based on this information. For example, according to the order of computing nodes Obtain the successor nodes of node a of the first computing node including node b and node c, then correspondingly add two edges ab and ac to node a of the second computing node, and add all the edge relationships to construct a directed acyclic Computational graph model.
  • step S260 in the embodiment of the present application includes:
  • parsing rules defined in a preset graph parsing engine, where the parsing rules include matching rules, rewriting rules, and assignment rules.
  • the parsing rules can be pre-defined and stored in the graph parsing engine according to customized requirements.
  • the rules include matching rules, rewriting rules and assignment rules.
  • S262 Match the third computing node of the directed acyclic computing graph model according to the matching rule.
  • the third calculation node of the directed acyclic calculation graph model is a calculation node obtained by adding an edge relationship to the second calculation node of the undirected calculation graph model.
  • S264 Assign a value to the fourth computing node according to the assignment rule and the third computing node that is successfully matched, and delete the third computing node that is successfully matched to obtain an intermediate expression calculation graph.
  • parsing rules there are multiple parsing rules.
  • One parsing rule can be iterated first, and the third computing node of the directed acyclic computing graph model is matched according to the matching rule in the parsing rule, and the special attributes of the third computing node are matched. , Or match the topological connection relationship of the subgraph.
  • the matching is successful, the graphical structure of the fourth computing node is determined according to the rewriting rule in the analysis rule, and then the node connection relationship between the fourth computing node and the fourth computing node is added. Finally, the fourth computing node is assigned according to the assignment rule in the parsing rule and the successfully matched third computing node.
  • the fourth computing node performs the assignment, and the third computing node that successfully matches is deleted. Continue to iterate the next parsing rule until the iteration of all parsing rules is completed, thereby generating an intermediate expression calculation graph according to the obtained fourth calculation node.
  • the embodiment of the application obtains the node information of the first computing node of the static computing graph model.
  • the node information includes the number of computing nodes and the sequence of computing nodes; initializes the same number of second computing nodes as the number of computing nodes; and corresponds the node information according to the topological ordering
  • Add to the second calculation node to construct an undirected calculation graph model add edge relations to the second calculation node of the undirected calculation graph model according to the order of the calculation nodes to construct a directed acyclic calculation graph model, and obtain the definition in the preset
  • the analysis rules of the graph analysis engine the analysis rules include matching rules, rewriting rules and assignment rules; according to the matching rules, the third computing node of the directed acyclic computing graph model is matched; when the matching is successful, the fourth is added according to the rewriting rules
  • the node connection relationship between the computing node and the fourth computing node; the fourth computing node is assigned according to the assignment rules and the successfully matched third computing node, and the successfully matched third
  • the third embodiment of the present application provides a neural network model customization system 100.
  • the neural network model customization system 100 provided in the third embodiment of the present application can execute the system 100 provided by any embodiment of the present application.
  • the customized method of the neural network model has the corresponding functional modules and effects of the execution method.
  • the neural network model customization system 100 includes a model acquisition module 200, a model conversion module 300, a model construction module 400, a calculation graph conversion module 500, and a model generation module 600.
  • the model obtaining module 200 is set to obtain a preset neural network model; the model conversion module 300 is set to convert the neural network model into a static calculation graph model; the model building module 400 is set to obtain node information of the first calculation node of the static calculation graph model A directed acyclic calculation graph model is constructed; the calculation graph conversion module 500 is set to convert the directed acyclic calculation graph model into an intermediate expression calculation graph through a preset graph analysis engine; the model generation module 600 is set to calculate the graph based on the intermediate expression Generate customized target neural network model.
  • the model building module 400 is configured to obtain node information, where the node information includes the number of computing nodes and the sequence of computing nodes; an undirected computing graph model is constructed according to the number of computing nodes; an undirected computing graph model is constructed according to the sequence of computing nodes The second computing node of add edge relationship to construct a directed acyclic computing graph model.
  • the model building module 400 is configured to construct an undirected computing graph model according to the number of computing nodes in the following manner: initialize the same number of second computing nodes as the number of computing nodes; sort the operator names and operators included in the node information according to the topological order The attributes, model parameters, inputs, and outputs are correspondingly added to the same number of second calculation nodes as the number of calculation nodes to construct an undirected calculation graph model.
  • the neural network model customization system 100 further includes a model display module 700, which is configured to visually display the directed acyclic calculation graph model through a preset interface; and receive the user's modification of the directed acyclic calculation graph model.
  • a model display module 700 which is configured to visually display the directed acyclic calculation graph model through a preset interface; and receive the user's modification of the directed acyclic calculation graph model.
  • the calculation graph conversion module 500 is configured to obtain the analysis rules defined in the preset graph analysis engine; replace the third calculation node of the directed acyclic calculation graph model according to the analysis rules to obtain the intermediate expression calculation graph, wherein, the third calculation node of the directed acyclic calculation graph model is a calculation node obtained by adding the edge relationship to the second calculation node of the undirected calculation graph model.
  • the parsing rules include matching rules, rewriting rules, and assignment rules.
  • the calculation graph conversion module 500 is configured to replace the third calculation node of the directed acyclic calculation graph model according to the parsing rules in the following manner to obtain the intermediate expression calculation graph: according to the matching rules Match the third computing node of the directed acyclic computing graph model; when the third computing node is successfully matched, the node connection relationship between the fourth computing node and the fourth computing node is added according to the rewriting rule; according to the assignment rule and the matching success
  • the third computing node assigns a value to the fourth computing node, and deletes the third computing node that is successfully matched to obtain an intermediate expression calculation graph.
  • the model generation module 600 is configured to convert the intermediate expression calculation graph into a customized intermediate expression through a preset serialization interface; serialize the intermediate expression into a customized target neural network model.
  • FIG. 6 is a schematic structural diagram of a neural network model customized device provided in the fourth embodiment of the present application.
  • FIG. 6 shows a block diagram of an exemplary computer device 12 suitable for implementing the embodiments of the present application.
  • the computer device 12 shown in FIG. 6 is only an example, and should not bring any limitation to the function and scope of use of the embodiments of the present application.
  • the computer device 12 is represented in the form of a general-purpose computing device.
  • the components of the computer device 12 may include, but are not limited to: one or more processors or processing units 16, a system memory 28, and a bus 18 connecting different system components (including the system memory 28 and the processing unit 16).
  • the bus 18 represents one or more of several types of bus structures, including a memory bus or a memory controller, a peripheral bus, a graphics acceleration port, a processor, or a local bus using any bus structure among multiple bus structures.
  • these architectures include but are not limited to Industry Standard Architecture (ISA) bus, Micro Channel Architecture (MAC) bus, enhanced ISA bus, Video Electronics Standards Association (Video Electronics Standards) Association, VESA) local bus and Peripheral Component Interconnect (PCI) bus.
  • the computer device 12 includes a variety of computer system readable media. These media can be any available media that can be accessed by the computer device 12, including volatile and nonvolatile media, removable and non-removable media.
  • the system memory 28 may include a computer system readable medium in the form of a volatile memory, such as a random access memory (RAM) 30 and/or a cache memory 32.
  • the computer device 12 may include other removable/non-removable, volatile/nonvolatile computer system storage media.
  • the storage system 34 may be configured to read and write a non-removable, non-volatile magnetic medium (not shown in FIG. 6, usually referred to as a "hard drive").
  • a disk drive configured to read and write to a removable non-volatile disk (such as a "floppy disk"), and a removable non-volatile optical disk (such as a compact disc read-only memory (Compact Disc)) can be provided.
  • a removable non-volatile disk such as a "floppy disk”
  • a removable non-volatile optical disk such as a compact disc read-only memory (Compact Disc)
  • the memory 28 may include at least one program product having a set of (for example, at least one) program modules, and these program modules are configured to perform the functions of the embodiments of the present application.
  • a program/utility tool 40 having a set of (at least one) program module 42 may be stored in, for example, the memory 28.
  • Such program module 42 includes, but is not limited to, an operating system, one or more application programs, and other programs Modules and program data, each of these examples or a combination may include the realization of a network environment.
  • the program module 42 usually executes the functions and/or methods in the embodiments described in this application.
  • the computer device 12 can also communicate with one or more external devices 14 (such as keyboards, pointing devices, displays 24, etc.), and can also communicate with one or more devices that enable users to interact with the computer device 12, and/or communicate with Any device (such as a network card, modem, etc.) that enables the computer device 12 to communicate with one or more other computing devices. This communication can be performed through an input/output (Input/Output, I/O) interface 22.
  • the computer device 12 may also communicate with one or more networks (for example, a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet) through the network adapter 20. As shown in the figure, the network adapter 20 communicates with other modules of the computer device 12 through the bus 18.
  • LAN local area network
  • WAN wide area network
  • public network such as the Internet
  • the processing unit 16 executes a variety of functional applications and data processing by running programs stored in the system memory 28, for example, to implement the methods provided in the embodiments of the present application:
  • Obtain a preset neural network model convert the neural network model into a static calculation graph model; construct a directed acyclic calculation graph model according to the node information of the first calculation node of the static calculation graph model; pass the preset
  • the graph analysis engine converts the directed acyclic calculation graph model into an intermediate expression calculation graph; and generates a customized target neural network model according to the intermediate expression calculation graph.
  • the fifth embodiment of the present application also provides a computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, the method as provided in all the application embodiments of the present application is implemented:
  • Obtain a preset neural network model convert the neural network model into a static calculation graph model; construct a directed acyclic calculation graph model according to the node information of the first calculation node of the static calculation graph model; pass the preset
  • the graph analysis engine converts the directed acyclic calculation graph model into an intermediate expression calculation graph; and generates a customized target neural network model according to the intermediate expression calculation graph.
  • the computer storage medium of the embodiment of the present application may adopt any combination of one or more computer-readable media.
  • the computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium.
  • the computer-readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or a combination of any of the above.
  • Examples of computer-readable storage media include: electrical connections with one or more wires, portable computer disks, hard disks, RAM, ROM, Erasable Programmable Read-Only Memory, EPROM or flash memory), optical fiber, CD-ROM, optical storage device, magnetic storage device, or any suitable combination of the above.
  • the computer-readable storage medium can be any tangible medium that contains or stores a program, and the program can be used by or in combination with an instruction execution system, apparatus, or device.
  • the computer-readable signal medium may include a data signal propagated in baseband or as a part of a carrier wave, and computer-readable program code is carried therein. This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • the computer-readable signal medium may also be any computer-readable medium other than the computer-readable storage medium.
  • the computer-readable medium may send, propagate or transmit the program for use by or in combination with the instruction execution system, apparatus, or device .
  • the program code contained on the computer-readable medium can be transmitted by any suitable medium, including but not limited to wireless, wire, optical cable, radio frequency (RF), etc., or any suitable combination of the foregoing.
  • suitable medium including but not limited to wireless, wire, optical cable, radio frequency (RF), etc., or any suitable combination of the foregoing.
  • the computer program code used to perform the operations of this application can be written in one or more programming languages or a combination thereof.
  • the programming languages include object-oriented programming languages—such as Java, Smalltalk, C++, and also conventional Procedural programming language-such as "C" language or similar programming language.
  • the program code can be executed entirely on the user's computer, partly on the user's computer, executed as an independent software package, partly on the user's computer and partly executed on a remote computer, or entirely executed on the remote computer or server.
  • the remote computer may be connected to the user's computer through any kind of network including LAN or WAN, or may be connected to an external computer (for example, using an Internet service provider to connect through the Internet).

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

Procédé, système et dispositif de personnalisation de modèle de réseau neuronal, et support de stockage. Le procédé de personnalisation de modèle de réseau neuronal consiste à : acquérir un modèle de réseau neuronal prédéfini (S110) ; convertir le modèle de réseau neuronal en un modèle de graphique de calcul statique (S120) ; construire un modèle de graphique de calcul cyclique dirigé selon des informations de nœud d'un premier nœud de calcul du modèle de graphique de calcul statique (S130) ; convertir le modèle de graphique de calcul cyclique dirigé en un graphique de calcul de représentation intermédiaire au moyen d'un moteur d'analyse de graphique prédéfini (S140) ; et générer un modèle de réseau neuronal cible personnalisé en fonction du graphique de calcul de représentation intermédiaire (S150).
PCT/CN2021/098288 2020-06-22 2021-06-04 Procédé, système et dispositif de personnalisation de modèle de réseau neuronal, et support de stockage WO2021259039A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010575490.3A CN111753983B (zh) 2020-06-22 2020-06-22 神经网络模型的定制化方法、系统、设备和存储介质
CN202010575490.3 2020-06-22

Publications (1)

Publication Number Publication Date
WO2021259039A1 true WO2021259039A1 (fr) 2021-12-30

Family

ID=72676376

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/098288 WO2021259039A1 (fr) 2020-06-22 2021-06-04 Procédé, système et dispositif de personnalisation de modèle de réseau neuronal, et support de stockage

Country Status (2)

Country Link
CN (1) CN111753983B (fr)
WO (1) WO2021259039A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023246177A1 (fr) * 2022-06-20 2023-12-28 美的集团(上海)有限公司 Procédé de traitement d'image et dispositif électronique et support d'informations

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111753983B (zh) * 2020-06-22 2024-05-28 深圳鲲云信息科技有限公司 神经网络模型的定制化方法、系统、设备和存储介质
CN112508163B (zh) * 2020-11-23 2021-12-07 北京百度网讯科技有限公司 神经网络模型中子图的展示方法、装置和存储介质
CN114936631B (zh) * 2021-04-26 2023-06-09 华为技术有限公司 一种模型处理方法及装置
CN113885845B (zh) * 2021-09-30 2024-01-12 苏州浪潮智能科技有限公司 深度学习编译器的计算图的生成方法、系统、设备及介质
CN114021707B (zh) * 2022-01-06 2022-04-05 中兴通讯股份有限公司 模型获取方法、系统、电子设备和存储介质
CN114911630B (zh) * 2022-07-14 2022-11-04 小米汽车科技有限公司 数据处理方法、装置、车辆、存储介质及芯片
CN117669683A (zh) * 2023-12-21 2024-03-08 摩尔线程智能科技(北京)有限责任公司 计算图处理方法、装置和存储介质

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106650922A (zh) * 2016-09-29 2017-05-10 清华大学 硬件神经网络转换方法、计算装置、编译方法和神经网络软硬件协作系统
WO2017152990A1 (fr) * 2016-03-11 2017-09-14 Telecom Italia S.P.A. Réseaux neuronaux convolutifs, en particulier pour analyse d'images
CN109284815A (zh) * 2018-11-30 2019-01-29 上海寒武纪信息科技有限公司 神经网络模型算法编译方法、装置及相关产品
CN109543825A (zh) * 2018-11-30 2019-03-29 上海寒武纪信息科技有限公司 神经网络模型算法编译方法、装置及相关产品
CN111753983A (zh) * 2020-06-22 2020-10-09 深圳鲲云信息科技有限公司 神经网络模型的定制化方法、系统、设备和存储介质

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017083399A2 (fr) * 2015-11-09 2017-05-18 Google Inc. Formation de réseaux neuronaux représentés sous forme de graphes de calcul
CN110766147B (zh) * 2018-07-25 2022-10-11 赛灵思公司 神经网络编译器架构及编译方法
CN109740751B (zh) * 2018-12-24 2020-04-14 中科寒武纪科技股份有限公司 神经网络模型的架构融合方法及相关装置

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017152990A1 (fr) * 2016-03-11 2017-09-14 Telecom Italia S.P.A. Réseaux neuronaux convolutifs, en particulier pour analyse d'images
CN106650922A (zh) * 2016-09-29 2017-05-10 清华大学 硬件神经网络转换方法、计算装置、编译方法和神经网络软硬件协作系统
CN109284815A (zh) * 2018-11-30 2019-01-29 上海寒武纪信息科技有限公司 神经网络模型算法编译方法、装置及相关产品
CN109543825A (zh) * 2018-11-30 2019-03-29 上海寒武纪信息科技有限公司 神经网络模型算法编译方法、装置及相关产品
CN111753983A (zh) * 2020-06-22 2020-10-09 深圳鲲云信息科技有限公司 神经网络模型的定制化方法、系统、设备和存储介质

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023246177A1 (fr) * 2022-06-20 2023-12-28 美的集团(上海)有限公司 Procédé de traitement d'image et dispositif électronique et support d'informations

Also Published As

Publication number Publication date
CN111753983A (zh) 2020-10-09
CN111753983B (zh) 2024-05-28

Similar Documents

Publication Publication Date Title
WO2021259039A1 (fr) Procédé, système et dispositif de personnalisation de modèle de réseau neuronal, et support de stockage
WO2021190597A1 (fr) Procédé de traitement pour un modèle de réseau neuronal, et dispositif associé
JP2022003576A (ja) 制御パルス生成方法、装置、システム、電子デバイス、記憶媒体及びプログラム
JPWO2017159614A1 (ja) 学習サービス提供装置
WO2017159638A1 (fr) Dispositif de génération de données de communication de capacité
KR20150143473A (ko) 재계산 사용자 인터페이스 내의 신호 캡처 컨트롤
US8091070B2 (en) Deep embedding of program languages
US11861469B2 (en) Code generation for Auto-AI
WO2017159620A1 (fr) Dispositif de maintien d'extensibilité
CN112734040A (zh) 一种嵌入式人工智能计算框架及应用方法
US11144291B1 (en) Loop-oriented neural network compilation
CN116450486A (zh) 多元异构计算系统内节点的建模方法、装置、设备及介质
CN111222636A (zh) 深度学习模型的转化方法、装置、服务器及存储介质
JP2020008836A (ja) 語彙テーブルの選択方法、装置およびコンピュータ読み取り可能な記憶媒体
CN113742231A (zh) 基于Protobuf协议动态解析的造数方法及装置
van der Vlag et al. RateML: A code generation tool for brain network models
CN117573507A (zh) 测试用例生成方法、装置、电子设备及存储介质
WO2024036662A1 (fr) Procédé et appareil d'exploration parallèle de règles de graphes sur la base d'un échantillonnage de données
CN116739154A (zh) 一种故障预测方法及其相关设备
CN115795025A (zh) 一种摘要生成方法及其相关设备
CN114707070A (zh) 一种用户行为预测方法及其相关设备
WO2021077282A1 (fr) Procédé et appareil de conversion de modèle de réseau neuronal, serveur et support d'informations
US11782706B1 (en) Reconfigurable neural network processing based on subgraph recognition
Gutosk Qualitative Analysis of Deep Learning Frameworks
WO2023103690A1 (fr) Procédé et appareil de génération d'essai unitaire, et dispositif associé

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21828120

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21828120

Country of ref document: EP

Kind code of ref document: A1