CN111178517B - Model deployment method, system, chip, electronic equipment and medium - Google Patents

Model deployment method, system, chip, electronic equipment and medium Download PDF

Info

Publication number
CN111178517B
CN111178517B CN202010064768.0A CN202010064768A CN111178517B CN 111178517 B CN111178517 B CN 111178517B CN 202010064768 A CN202010064768 A CN 202010064768A CN 111178517 B CN111178517 B CN 111178517B
Authority
CN
China
Prior art keywords
model
deployment
data
training
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010064768.0A
Other languages
Chinese (zh)
Other versions
CN111178517A (en
Inventor
黄文豪
黄杰
杨忠程
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Yitu Technology Co ltd
Original Assignee
Shanghai Yitu Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Yitu Technology Co ltd filed Critical Shanghai Yitu Technology Co ltd
Priority to CN202010064768.0A priority Critical patent/CN111178517B/en
Publication of CN111178517A publication Critical patent/CN111178517A/en
Application granted granted Critical
Publication of CN111178517B publication Critical patent/CN111178517B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons

Abstract

The model deployment method, system, chip, electronic equipment and medium comprises the following steps: obtaining a training result of the model configuration to be converted; splitting the training result; searching the split result in the database, and converting the code; and reconnecting the converted model codes to obtain a deployment model. According to the application, after the input and output forms of the designated task type and the model are obtained, training data and verification data are input, model parameters can be automatically trained and adjusted during training, super parameters and structures of the model are adjusted through an optimization algorithm, and finally, the optimal model is iterated and selected.

Description

Model deployment method, system, chip, electronic equipment and medium
Technical Field
The application relates to a computer data processing system and method, in particular to a model deployment method, a system, a chip, electronic equipment and a medium.
Background
The essence of the existing Artificial Intelligence (AI) research and development system is to provide a database of software, which requires a certain programming background for users in use; if a good calculation model is to be obtained, the user needs to adjust and select parameters, and therefore, relevant deep learning optimization knowledge is needed. Meanwhile, the existing Artificial Intelligence (AI) development system generally only can support a single user and a single task, but can only manually allocate various resources under the condition of multiple users, and the utilization rate of the computing resources is poor.
In an Artificial Intelligence (AI) development system, an existing training tool generally only supports training through a preset model after data is input, and the training tool often has poor effect when facing a new task. If the model needs to be optimized, knowledge about the optimization of the model is also needed, and the optimization is realized by manual design and programming, so that the user range is narrow. The existing other automatic model optimization algorithms generally automatically design a network model through methods such as RNN, and the method has the disadvantages of slower process and poor parallelism and needs a large amount of data; the method is not applicable to the situation that the data volume is medium and the computing resources are small.
In addition, the application system is deployed and run on the cloud platform to provide services externally. At present, in a model deployment link, aiming at a deployment environment and a target format, a conversion code is manually written for conversion; and whether the verification fails or not can not be verified, and the test script needs to be manually written for testing, so that the test is very tedious and takes a lot of time.
Disclosure of Invention
Aiming at the defects of the artificial intelligence research and development system, the application aims to provide a model deployment method, a system, electronic equipment and a medium suitable for a model optimization deployment system so as to solve the existing problems.
In order to achieve the above object, the present application provides a model deployment method of a model optimization deployment system, comprising:
s1, obtaining a training result of a network model to be converted;
s2, splitting a training result;
s3, searching the split result in a database, and converting the split result into codes;
s4, reconnecting the converted model codes to obtain a deployment model.
In some embodiments, the method further includes S5, extracting feature layers from the test data for the network model to be converted and the deployment model in the hardware environment to be deployed, and correspondingly calculating given vector distances between every two pairs in sequence; if the difference is smaller than the preset threshold value, the results are considered to be aligned, and the deployable model is output as a conversion result.
In some embodiments, in S2, the splitting is performed according to a network structure of training results.
In some embodiments, the network structure is a computational layer or sub-structure.
In some embodiments, in S2, if a preset structure in the database is identified and found during splitting, the preset structure is saved; and when the method is deployed, selecting a conversion method according to the stored preset structure.
In some embodiments, in S3, the lookup is performed in a look-up table to obtain the code.
In some embodiments, in S3, the database is a database of a hardware environment-computing framework.
The application also provides a model deployment system, comprising:
the data preprocessing module receives the reflow data and then outputs a preprocessed data set;
the model training module is used for executing a model training method on the data set to obtain an optimal configuration model;
the model deployment module is used for adapting and converting the optimal configuration model according to the hardware environment to be deployed, and finally deploying the optimal configuration model in a cross-platform mode.
In some embodiments, the system further comprises a data labeling module for labeling the auxiliary data.
In some embodiments, the system further comprises a storage module for performing unified storage of the whole system data.
In addition, the application also provides a model training method based on the model deployment system, which comprises the following steps:
training a group of models on a preset basic neural network, and selecting a model T0 with the best performance on a verification set;
step two, optimizing on T0 and then obtaining a plurality of alternative experimental configurations, training the plurality of alternative experimental configurations and obtaining a model T1' with optimal performance;
retraining the T1', and ensuring that the average value of the retraining performance results is larger than the performance of T0; and obtaining the optimal model configuration.
In some embodiments, in the second step, the optimization method for T0 is an adjustment parameter.
In some embodiments, the adjustment parameters include one or a combination of the following: model width, learning rate, optimization strategy, whether data enhancement is used, parameters of data enhancement, and network unit modules.
In some embodiments, in the retraining of T1' in step three S3, if the performance result of the retraining is greater than the performance result of T0, then T0 is replaced as a new alternative experimental configuration, i.e., iterative optimization.
In some embodiments, the iterative optimization process loops repeatedly until performance is optimal.
In some embodiments, the number of iterative optimizations is two.
The application also provides a chip comprising a processor for calling and running a computer program from a memory, so that a device on which the chip is installed performs the model deployment method according to any one of the above.
The application also provides an electronic device comprising a processor and a memory for storing executable instructions of the processor, the processor executing any of the model deployment methods described above when running.
The application also provides a computer readable medium having stored thereon computer program instructions which, when processed and executed, implement any of the above described model deployment methods.
Compared with the prior art, after the input and output forms of the designated task type and the model are obtained, training data and verification data are input, the model parameters can be automatically trained and adjusted during training, simultaneously, the super parameters and the structure of the model are adjusted through an optimal adjustment algorithm, and finally, the optimal model is iterated and selected.
The application provides a system for optimizing and deploying models for personnel without programming background and deep learning algorithm background under the condition of data, which can automatically preprocess data, automatically train and optimize an AI model, evaluate and verify the AI model and automatically deploy the AI model, and store a resource module through a computing resource containerization module, so as to support multi-user and multi-task model research and development.
The application is based on a method for optimizing the deployment system by the model, which can automatically convert the trained model into caffe or TF pb or other common model formats, automatically verify whether the model can be deployed on the corresponding hardware system which is expected to be deployed, and automatically perform some optimizations.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings used in the embodiments or the description of the prior art will be briefly described below, it being obvious that the drawings in the following description are only some embodiments of the present application, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a schematic flow chart of the present application;
FIG. 2 is a schematic diagram of a framework design of a model-optimized deployment system.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present application more clear, the technical solutions of the embodiments of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application; it will be apparent that the described embodiments are some, but not all, embodiments of the application. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to be within the scope of the application.
Referring to fig. 2, the embodiment is based on a model deployment system, and the architecture design of the system is shown in fig. 2, and the system mainly comprises a training flow part and a resource scheduling part. The training flow part comprises a data labeling module for assisting in data labeling, a data preprocessing module for preprocessing, cleaning and segmenting the data after the data labeling, and a model training module for automatically training and parameter tuning the processed data.
And the scheduling part is used for automatically adapting and converting the obtained model according to the hardware environment which is expected to be deployed through the model deployment module after repeated data iteration and model optimization. And finally, the whole training process is efficiently and simply carried out.
Wherein: the data preprocessing module receives the reflow data and outputs a preprocessed data set;
the model training module is used for automatically training and optimizing parameters of the processed data set to obtain an optimal configuration model;
the model deployment module is used for automatically adapting and converting the optimal configuration model according to the environment of the hardware to be deployed and finally deploying the optimal configuration model on the hardware.
The system also comprises a data labeling module for labeling the auxiliary data. In order to support multi-user multitasking, the unified scheduling of the computing resources is preferably performed through the containerized sharing module, and the unified storage of the data is preferably performed by the storage module, so that the computing resources are efficiently utilized.
The model optimization deployment system realizes automatic optimization of algorithm configuration, and can select a better technical mode according to a planned layout environment; meanwhile, unusable structures can be avoided according to the target scene and are abreast as a blacklist.
Based on the system, in the model training module, the model training is as follows:
training a group of models on a preset basic neural network, and selecting a model T0 with the best performance on a verification set;
step two, optimizing on T0 and then obtaining a plurality of alternative experimental configurations, training the plurality of alternative experimental configurations and obtaining a model T1' with optimal performance;
preferably, the optimization method is to sequentially try to adjust various preset parameters. And the adjustment parameters include one or a combination of the following: model width, learning rate, optimization strategy, whether data enhancement is used, parameters of data enhancement, network unit module selection.
And step three, retraining the T1' to ensure that the average value of the performance results is larger than that of the T0, and the optimal model configuration is the final optimal result. Especially in the retraining of T1', if the performance result is larger than T0, substituting T0 as a new experimental configuration, and continuing the iterative optimization of the process; and stopping iterative optimization until the performance of the model configuration cannot continue to be stably improved.
Preferably, the number of iterative optimizations is generally two.
In the optimization, according to the preset hardware environment to be deployed, the optimization direction of the attempt is limited, for example, the operation speed is limited to be not lower than certain lower limits, and the network module is selected by using the support of a specific hardware platform or the optimization is better.
Referring to fig. 1, after performing iterative optimization on a network model for a plurality of times to obtain a final optimized result, a model deployment method is performed, including:
s1, obtaining a training result of the network model to be converted, namely stopping iterative optimization in the step three to obtain a result.
S2, splitting a training result; splitting is typically done at the compute layer. If the preset structure in the existing library is identified and found during splitting, the structure is saved.
S3, searching the split result in a database, and converting the split result into codes; the database is a database of hardware environment-computing frames, including the types of servers that can be deployed, and the types of computing frames. Wherein the server includes, but is not limited to, a Graphics Processor (GPU), a Central Processing Unit (CPU), an ARM processor, an AI chip, a cell phone, etc.; the computing framework includes, but is not limited to, TF, caffe, torch, and the like. As many types as possible should be included to meet different requirements.
Wherein Caffe is a deep learning framework with expressive, speed and thought modularizing functions; multiple types of deep learning architectures are supported, image classification and image segmentation oriented, as well as CNN, RCNN, LSTM and fully connected neural network designs.
TF, tensorFlow, is a symbolic mathematical system based on data stream programming, and is widely applied to the programming implementation of various machine learning algorithms; the TF framework has a multi-level structure, can be deployed on various servers, PC terminals and webpages, and supports GPU and TPU high-performance numerical computation.
Torch is a scientific computational framework supported by a large number of machine learning algorithms, and is characterized by being particularly flexible and employing the programming language Lua.
And then searching (looking up a table) the split calculation layer or the sub-structure in a database to convert according to a preset code. The substructures in the database select the optimal conversion method according to the format of the conversion plan or the platform which is expected to be deployed.
S4, reconnecting the converted model codes to obtain a deployment model.
Because of the large differences between the experimental environment and the actual deployment environment, it may be a different hardware platform, a different computing framework. The last ring in the overall model training process therefore requires that the trained network and its corresponding weights be converted to the model format of the target environment for final deployment. This step traditionally requires targeted programming of the transcoding for network architecture and source/target model formats. This environment should be automated as part of the end-to-end AI system.
And finally, respectively calculating an original model in an experimental environment and a deployment model in an environment to be deployed by using preset test data (pictures, videos and the like), respectively extracting part of characteristic layers, and correspondingly calculating given vector distances between every two pairs in sequence. If the difference is smaller than a certain acceptable threshold, the results are considered to be aligned, the conversion is successful, and the model is output as the final conversion result.
The method can automatically convert the trained model into caffe or TF pb or other common model formats, automatically verify whether the model can be deployed on a corresponding hardware platform which is expected to be deployed, and automatically perform some optimization.
Referring to fig. 2, the present application further provides a model optimized deployment system, comprising: and the data preprocessing module receives the reflow data and then outputs a preprocessed data set.
The model training module is used for executing a training method on the data set to obtain an optimal configuration model; specifically, a group of models are trained on a preset basic neural network, and a model T0 with the best performance on a verification set is selected; then optimizing on T0 and then obtaining a plurality of alternative experimental configurations, training the plurality of alternative experimental configurations and obtaining a model T1' with optimal performance, and adjusting parameters of an optimization method of T0, wherein the parameters comprise one or a combination of the following steps: model width, learning rate, optimizing strategy, whether to use data enhancement, parameters of data enhancement and a network unit module;
finally retraining the T1', and ensuring that the average value of the retraining performance results is larger than the performance of T0; and obtaining the optimal model configuration. In retraining T1', if the retrained performance result is greater than the performance result of T0, then T0 is replaced as a new alternative experimental configuration, i.e., iterative optimization. And repeatedly and circularly carrying out until the performance is optimal. And typically twice.
The model deployment module is used for adapting and converting the optimal configuration model according to the environment of the hardware to be deployed and finally deploying the optimal configuration model on the hardware. Specifically, a training result of a network model to be converted is obtained; splitting the training result according to the network structures such as the network structure of the training result; then searching the split result in a database of a hardware environment-computing framework through a table lookup, and converting the code; and finally, reconnecting the converted model codes to obtain a deployment model. Extracting feature layers from a network model to be converted and a deployment model in a hardware environment to be deployed according to test data, and correspondingly calculating given vector distances between every two pairs in sequence; if the difference is smaller than the preset threshold, the results are considered to be aligned, and the deployable model is output as a conversion result.
In addition, the application also provides electronic equipment, which comprises: at least one processor; a memory coupled to the at least one processor, the memory storing executable instructions that when executed by the at least one processor cause the method of the present application described above to be implemented.
For example, the memory may include random access memory, flash memory, read-only memory, programmable read-only memory, non-volatile memory, registers, or the like. The processor may be a central processing unit (Central Processing Unit, CPU) or the like. Or an image processor (Graphic Processing Unit, GPU) memory may store executable instructions. The processor may execute executable instructions stored in the memory to implement the various processes described herein.
It will be appreciated that the memory in this embodiment may be either volatile memory or nonvolatile memory, or may include both volatile and nonvolatile memory. The nonvolatile memory may be a ROM (Read-only memory), a PROM (programmable Read-only memory), an EPROM (erasablprom, erasable programmable Read-only memory), an EEPROM (electrically erasable EPROM), or a flash memory. The volatile memory may be a RAM (random access memory) which serves as an external cache. By way of example, and not limitation, many forms of RAM are available, such as SRAM (static RAM), DRAM (dynamic RAM), SDRAM (synchronous DRAM), ddr SDRAM (DoubleDataRate SDRAM, double data rate synchronous DRAM), ESDRAM (Enhanced SDRAM), SLDRAM (synclinkdram), and DRRAM (directrambus RAM). The memory described herein is intended to comprise, without being limited to, these and any other suitable types of memory.
In some embodiments, the memory stores the following elements, an upgrade package, an executable unit, or a data structure, or a subset thereof, or an extended set thereof: an operating system and application programs.
The operating system includes various system programs, such as a framework layer, a core library layer, a driving layer, and the like, and is used for realizing various basic services and processing hardware-based tasks. And the application programs comprise various application programs and are used for realizing various application services. The program for implementing the method of the embodiment of the application can be contained in an application program.
In the embodiment of the present application, the processor is configured to execute the above method steps by calling a program or an instruction stored in the memory, specifically, a program or an instruction stored in an application program.
The embodiment of the application also provides a chip for executing the method. Specifically, the chip includes: and a processor for calling and running the computer program from the memory, so that the device on which the chip is mounted is used for executing the above method.
The present application also provides a computer readable storage medium having a computer program stored thereon, which when executed by a processor implements the steps of the above-described method of the present application.
For example, machine-readable storage media may include, but are not limited to, various known and unknown types of non-volatile memory.
Embodiments of the present application also provide a computer program product comprising computer program instructions for causing a computer to perform the above method.
Those of skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, or combinations of software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the solution. Those skilled in the art may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
In embodiments of the present application, the disclosed systems, electronic devices, and methods may be implemented in other ways. For example, the division of the units is only one logic function division, and other division manners are also possible in actual implementation. For example, multiple units or components may be combined or may be integrated into another system. In addition, the coupling between the individual units may be direct coupling or indirect coupling. In addition, each functional unit in the embodiments of the present application may be integrated in one processing unit, or may exist alone physically, or the like.
It should be understood that, in various embodiments of the present application, the size of the sequence number of each process does not mean that the execution sequence of each process should be determined by its functions and internal logic, and should not constitute any limitation on the implementation process of the embodiments of the present application.
The functions, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored on a machine-readable storage medium. Accordingly, aspects of the present application may be embodied in a software product, which may be stored on a machine-readable storage medium, which may include instructions for causing an electronic device to perform all or part of the processes of the aspects described in embodiments of the present application. The storage medium may include a ROM, a RAM, a removable disk, a hard disk, a magnetic disk, or an optical disk, etc. various media in which program codes can be stored.
The above is merely an embodiment of the present application, and the scope of the present application is not limited thereto. Those skilled in the art can make changes or substitutions within the technical scope of the present disclosure, and such changes or substitutions should be included in the scope of the present disclosure.

Claims (8)

1. A method of model deployment, comprising:
s1, obtaining a training result of a network model to be converted;
s2, splitting the training result according to a network structure of the training result, wherein the network structure is a calculation layer;
s3, searching the split result in a hardware environment-computing framework database in a table look-up mode, and converting the code;
s4, reconnecting the converted model codes to obtain a deployment model;
s5, extracting feature layers from the network model to be converted and the deployment model in the hardware environment to be deployed according to the test data, and correspondingly calculating given vector distances between every two pairs in sequence; if the given vector distance is smaller than the preset threshold value, the results are considered to be aligned, and the deployable model is output as a conversion result.
2. The model deployment method of claim 1, wherein: s2, if a preset structure in the database is identified and found during splitting, the preset structure is stored; and when the method is deployed, selecting a conversion method according to the stored preset structure.
3. A model deployment system, comprising:
the data preprocessing module receives the reflow data and then outputs a preprocessed data set;
the model training module is used for executing training optimization on the data set to obtain an optimal configuration model;
the model deployment module executes the model deployment method of claim 1 or 2 according to the environment of the hardware to be deployed, adapts and converts the optimal configuration model, and finally deploys the optimal configuration model on the hardware.
4. A model deployment system as in claim 3, wherein: the system also comprises a data labeling module for labeling the auxiliary data.
5. The model deployment system of claim 4, wherein: the system also comprises a storage module for executing unified storage of the whole system data.
6. A chip comprising a processor for calling and running a computer program from a memory, so that a device on which the chip is mounted performs the model deployment method of claim 1 or 2.
7. An electronic device, characterized in that: a memory comprising a processor, and memory for storing executable instructions for the processor, the processor when run performing the model deployment method of claim 1 or 2.
8. A computer-readable medium, characterized by: on which computer program instructions are stored which, when processed and executed, implement the model deployment method of claim 1 or 2.
CN202010064768.0A 2020-01-20 2020-01-20 Model deployment method, system, chip, electronic equipment and medium Active CN111178517B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010064768.0A CN111178517B (en) 2020-01-20 2020-01-20 Model deployment method, system, chip, electronic equipment and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010064768.0A CN111178517B (en) 2020-01-20 2020-01-20 Model deployment method, system, chip, electronic equipment and medium

Publications (2)

Publication Number Publication Date
CN111178517A CN111178517A (en) 2020-05-19
CN111178517B true CN111178517B (en) 2023-12-05

Family

ID=70658036

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010064768.0A Active CN111178517B (en) 2020-01-20 2020-01-20 Model deployment method, system, chip, electronic equipment and medium

Country Status (1)

Country Link
CN (1) CN111178517B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112015519A (en) * 2020-08-28 2020-12-01 江苏银承网络科技股份有限公司 Model online deployment method and device
CN112015470B (en) * 2020-09-09 2022-02-01 平安科技(深圳)有限公司 Model deployment method, device, equipment and storage medium
CN112463301B (en) * 2020-11-30 2022-02-11 常州微亿智造科技有限公司 Container-based model training test tuning and deployment method and device
CN113313239A (en) * 2021-06-25 2021-08-27 展讯通信(上海)有限公司 Artificial intelligence model design optimization method and device
CN114168232A (en) * 2021-12-20 2022-03-11 南京星云数字技术有限公司 Algorithm model result conversion configuration method, device, equipment and medium
CN115187821B (en) * 2022-07-05 2024-03-22 阿波罗智能技术(北京)有限公司 Method, related device and program product for verifying correctness of model before and after conversion
CN116542344A (en) * 2023-07-05 2023-08-04 浙江大华技术股份有限公司 Model automatic deployment method, platform and system

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1601510A (en) * 2003-03-06 2005-03-30 微软公司 Architecture for distributed computing system and automated design, deployment, and management of distributed applications
CN104468716A (en) * 2014-11-04 2015-03-25 中科同德(厦门)物联网科技有限公司 Method and system for continuously collecting indoor and outdoor environment information of residential area
CN206517618U (en) * 2017-02-28 2017-09-22 通号(郑州)轨道交通科技有限公司 A kind of track group network system of two kinds of wireless signals of compatible WLAN and LTE M
CN108197633A (en) * 2017-11-24 2018-06-22 百年金海科技有限公司 Deep learning image classification based on TensorFlow is with applying dispositions method
CN108764487A (en) * 2018-05-29 2018-11-06 北京百度网讯科技有限公司 For generating the method and apparatus of model, the method and apparatus of information for identification
CN108898222A (en) * 2018-06-26 2018-11-27 郑州云海信息技术有限公司 A kind of method and apparatus automatically adjusting network model hyper parameter
CN109409528A (en) * 2018-09-10 2019-03-01 平安科技(深圳)有限公司 Model generating method, device, computer equipment and storage medium
CN109558674A (en) * 2018-11-28 2019-04-02 北京超萌国际文化有限公司 Method for Sales Forecast and its model training method, device
CN109558952A (en) * 2018-11-27 2019-04-02 北京旷视科技有限公司 Data processing method, system, equipment and storage medium
CN109685160A (en) * 2019-01-18 2019-04-26 创新奇智(合肥)科技有限公司 A kind of on-time model trained and dispositions method and system automatically
CN109740560A (en) * 2019-01-11 2019-05-10 济南浪潮高新科技投资发展有限公司 Human cellular protein automatic identifying method and system based on convolutional neural networks
CN110033091A (en) * 2018-12-13 2019-07-19 阿里巴巴集团控股有限公司 A kind of method and apparatus predicted based on model
CN110073301A (en) * 2017-08-02 2019-07-30 强力物联网投资组合2016有限公司 The detection method and system under data collection environment in industrial Internet of Things with large data sets
CN110378463A (en) * 2019-07-15 2019-10-25 北京智能工场科技有限公司 A kind of artificial intelligence model standardized training platform and automated system
CN110418354A (en) * 2019-08-06 2019-11-05 北京邮电大学 It is a kind of that propagation model wireless network planning method is exempted from based on machine learning
CN110443321A (en) * 2019-08-14 2019-11-12 北京百度网讯科技有限公司 Model structure method of adjustment and device
CN110543950A (en) * 2019-09-27 2019-12-06 宁波和利时智能科技有限公司 Industrial big data modeling platform
CN110659740A (en) * 2018-06-28 2020-01-07 国际商业机器公司 Ordering and updating machine learning models based on data input at edge nodes

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8244696B2 (en) * 2008-04-21 2012-08-14 Oracle International Corporation Methods and systems for supporting multiple deployment models of enterprise business applications
US10482389B2 (en) * 2014-12-04 2019-11-19 Sap Se Parallel development and deployment for machine learning models

Patent Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1601510A (en) * 2003-03-06 2005-03-30 微软公司 Architecture for distributed computing system and automated design, deployment, and management of distributed applications
CN104468716A (en) * 2014-11-04 2015-03-25 中科同德(厦门)物联网科技有限公司 Method and system for continuously collecting indoor and outdoor environment information of residential area
CN206517618U (en) * 2017-02-28 2017-09-22 通号(郑州)轨道交通科技有限公司 A kind of track group network system of two kinds of wireless signals of compatible WLAN and LTE M
CN110073301A (en) * 2017-08-02 2019-07-30 强力物联网投资组合2016有限公司 The detection method and system under data collection environment in industrial Internet of Things with large data sets
CN108197633A (en) * 2017-11-24 2018-06-22 百年金海科技有限公司 Deep learning image classification based on TensorFlow is with applying dispositions method
CN108764487A (en) * 2018-05-29 2018-11-06 北京百度网讯科技有限公司 For generating the method and apparatus of model, the method and apparatus of information for identification
CN108898222A (en) * 2018-06-26 2018-11-27 郑州云海信息技术有限公司 A kind of method and apparatus automatically adjusting network model hyper parameter
CN110659740A (en) * 2018-06-28 2020-01-07 国际商业机器公司 Ordering and updating machine learning models based on data input at edge nodes
CN109409528A (en) * 2018-09-10 2019-03-01 平安科技(深圳)有限公司 Model generating method, device, computer equipment and storage medium
CN109558952A (en) * 2018-11-27 2019-04-02 北京旷视科技有限公司 Data processing method, system, equipment and storage medium
CN109558674A (en) * 2018-11-28 2019-04-02 北京超萌国际文化有限公司 Method for Sales Forecast and its model training method, device
CN110033091A (en) * 2018-12-13 2019-07-19 阿里巴巴集团控股有限公司 A kind of method and apparatus predicted based on model
CN109740560A (en) * 2019-01-11 2019-05-10 济南浪潮高新科技投资发展有限公司 Human cellular protein automatic identifying method and system based on convolutional neural networks
CN109685160A (en) * 2019-01-18 2019-04-26 创新奇智(合肥)科技有限公司 A kind of on-time model trained and dispositions method and system automatically
CN110378463A (en) * 2019-07-15 2019-10-25 北京智能工场科技有限公司 A kind of artificial intelligence model standardized training platform and automated system
CN110418354A (en) * 2019-08-06 2019-11-05 北京邮电大学 It is a kind of that propagation model wireless network planning method is exempted from based on machine learning
CN110443321A (en) * 2019-08-14 2019-11-12 北京百度网讯科技有限公司 Model structure method of adjustment and device
CN110543950A (en) * 2019-09-27 2019-12-06 宁波和利时智能科技有限公司 Industrial big data modeling platform

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Andres Milioto等.Bonnet: An Open-Source Training and Deployment Framework for Semantic Segmentation in Robotics using CNNs.《2019 International Conference on Robotics and Automation (ICRA)》.2019,7094-7100. *
郭佳豪等.基于Web的目标检测神经网络协同计算方法的研究与实现.《2019中国信息通信大会论文集(CICC 2019)》.2019,372-378. *

Also Published As

Publication number Publication date
CN111178517A (en) 2020-05-19

Similar Documents

Publication Publication Date Title
CN111178517B (en) Model deployment method, system, chip, electronic equipment and medium
CN111191789B (en) Model optimization deployment system, chip, electronic equipment and medium
US8225074B2 (en) Methods and systems for managing computations on a hybrid computing platform including a parallel accelerator
US20200050443A1 (en) Optimization and update system for deep learning models
CN107958285A (en) The mapping method and device of the neutral net of embedded system
CN113128478B (en) Model training method, pedestrian analysis method, device, equipment and storage medium
US20220188645A1 (en) Using generative adversarial networks to construct realistic counterfactual explanations for machine learning models
CN113610232A (en) Network model quantization method and device, computer equipment and storage medium
CN114116236B (en) Construction method and system of heterogeneous computing system
CN109684088B (en) Remote sensing big data rapid processing task scheduling method based on cloud platform resource constraint
CN112099848B (en) Service processing method, device and equipment
US20210295158A1 (en) End-to-end optimization
KR20200052417A (en) Apparatus and method for selecting inference module of target device
CN111831355A (en) Weight precision configuration method, device, equipment and storage medium
CN111831359A (en) Weight precision configuration method, device, equipment and storage medium
CN114429208A (en) Model compression method, device, equipment and medium based on residual structure pruning
CN110889316B (en) Target object identification method and device and storage medium
Tiwari et al. NCS based ultra low power optimized machine learning techniques for image classification
US20210383258A1 (en) Systems and Methods for Dynamically Configuring Multidimensional Data Processing Pipelines in Software-controlled Hardware-Software Embedded Systems
US20220067453A1 (en) Adaptive and hierarchical convolutional neural networks using partial reconfiguration on fpga
CN114092313A (en) Model reasoning acceleration method and system based on GPU (graphics processing Unit) equipment
Tong et al. Study on mindspore deep learning framework
Chichin et al. Capability to embed deep neural networks: Study on cpu processor in avionics context
CN113297855A (en) Embedded remote sensing image text generation method for satellite in-orbit application
CN112633516A (en) Performance prediction and machine learning compilation optimization method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant