CN111582475A - Data processing method and device based on automatic lightweight neural network - Google Patents

Data processing method and device based on automatic lightweight neural network Download PDF

Info

Publication number
CN111582475A
CN111582475A CN202010351374.3A CN202010351374A CN111582475A CN 111582475 A CN111582475 A CN 111582475A CN 202010351374 A CN202010351374 A CN 202010351374A CN 111582475 A CN111582475 A CN 111582475A
Authority
CN
China
Prior art keywords
data
neural network
block
lightweight neural
mth
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010351374.3A
Other languages
Chinese (zh)
Inventor
付琨
孙显
刁文辉
李霁豪
张义
陈凯强
吴红莉
王铁平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Aerospace Information Research Institute of CAS
Original Assignee
Aerospace Information Research Institute of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Aerospace Information Research Institute of CAS filed Critical Aerospace Information Research Institute of CAS
Priority to CN202010351374.3A priority Critical patent/CN111582475A/en
Publication of CN111582475A publication Critical patent/CN111582475A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a data processing method and a device based on an automatic lightweight neural network, comprising the following steps: acquiring data to be processed; inputting the data to be processed into a pre-established automatic lightweight neural network, and acquiring a processing result of the data to be processed output by the pre-established automatic lightweight neural network; the automatic lightweight neural network in the technical scheme provided by the invention has abundant topological connection forms of the neural network, so that the searched network is more diversified, and various convolutions are matched for use, so that the network is lighter, the receptive field of the network is more flexible, and the identification, detection and segmentation of objects or scenes with larger scale change in the data processing process are facilitated.

Description

Data processing method and device based on automatic lightweight neural network
Technical Field
The invention relates to the field of computer vision, in particular to a data processing method and device based on an automatic lightweight neural network.
Background
The construction of large deep neural networks usually requires strong expert knowledge and generally consumes a great deal of time and effort of researchers, especially in the field of remote sensing image processing. Meanwhile, these artificially designed networks are usually high in computational complexity and large in memory overhead, and also pose a great challenge to deployment on edge computing devices. Neural Architecture Search (NAS) is an automatic design method specially for a deep neural network, the method can effectively reduce the degree of artificial participation, and the deep neural network is automatically built under the machine view angle.
The current NAS research mainly uses Reinforcement Learning (RL) as an optimization algorithm to select the quality of a neural network structure in a search space (search space). NAS based on reinforcement learning generally uses a Long Short Term Memory (LSTM) as a Controller (Controller), the Controller combines each structure in a search space, and feeds back the performance of the combined model as a Reward (Reward) to the Controller. Since the process is discontinuous, the process is generally optimized by Policy Gradient (Policy Gradient), and after the PG algorithm converges, the final neural network model structure can be derived by the controller, and then the network is finely tuned.
However, the network topology connection form derived by the RL-based NAS is relatively single, and the model has high computational complexity and is not suitable for large-size and multi-scale remote sensing images. Meanwhile, the RL optimization cost is also very large, generally, this method needs to perform parallel operation on several hundred computer Graphics Processing Units (GPUs) for about four weeks, and the search process is also unstable.
Disclosure of Invention
In view of the shortcomings of the prior art, the invention aims to provide a processing method suitable for identifying, detecting and segmenting an object or a scene with large scale change.
The purpose of the invention is realized by adopting the following technical scheme:
in a method of data processing based on an automated lightweight neural network, the improvement comprising:
acquiring data to be processed;
and inputting the data to be processed into a pre-established automatic lightweight neural network, and acquiring a processing result of the data to be processed output by the pre-established automatic lightweight neural network.
Preferably, the data to be processed is remote sensing data.
Preferably, the processing result includes: data recognition results, data detection results and data segmentation results.
Preferably, the process of establishing the pre-established automated lightweight neural network includes:
step 1, acquiring training set data and verification set data corresponding to data to be processed;
step 2, setting the number M of blocks of the pre-established automatic lightweight neural network, and numbering the blocks, wherein the 1 st Block is a stem layer;
step 3, setting the number of neurons in each of the blocks from the 2 nd to the Mth as N, and numbering the neurons in each of the blocks from the 2 nd to the Mth from 1 to N;
step 4, selecting a random graph which is depended by each Block from the 2 nd Block to the Mth Block;
step 5, generating a topological relation among the neurons of each Block in the 2 nd to Mth blocks based on the random graphs depended on by the 2 nd to Mth blocks, wherein the neurons of each Block in the 2 nd to Mth blocks meet the condition that the neurons with smaller sequence numbers point to the neurons with larger sequence numbers;
step 6, setting a search space of the neurons of each Block from the 2 nd Block to the Mth Block;
step 7, stacking the 1 st to Mth blocks from small to large according to the numbers, inserting all the blocks from the 2 nd to the Mth to the 1 st after each Block into a pooling layer, introducing a head network corresponding to the processing method after the Mth Block, and finally forming an automatic lightweight neural network initial model;
step 8, the training data is sent into the automatic lightweight neural network initial model in a minipatch form, and learning is carried out in a supervision mode;
and 9, setting the total verification times, verifying the performance index of the trained automatic lightweight neural network initial model by using the verification data after each T times of training, and selecting the automatic lightweight neural network initial model with the maximum performance index as the pre-established automatic lightweight neural network.
Further, the step 4 comprises:
and each of the 2 nd to Mth blocks randomly selects a random graph from a BA random graph, an ER random graph and a WS random graph as a random graph depended by the Block, wherein the probability of selecting the BA random graph, the ER random graph and the WS random graph is the same.
Further, the step 6 comprises:
the neurons in each of the 2 nd to mth blocks select mapping functions belonging to themselves at {3 × 3DWConv,5 × 5DWConv,7 × 7DWConv,9 × 9DWConv }, wherein the probabilities of sampling 3 × 3DWConv,5 × 5DWConv,7 × 7DWConv, and 9 × 9DWConv are all the same, DWConv being a depth separable convolution.
Further, when the processing method is a data identification method, the performance index of the initial model of the automatic lightweight neural network is an OA performance index, when the processing method is a data detection method, the performance index of the initial model of the automatic lightweight neural network is an mAP performance index, and when the processing method is a data segmentation method, the performance index of the initial model of the automatic lightweight neural network is an mF1 performance index.
Based on the same inventive concept, the invention also provides a data processing device based on the automatic lightweight neural network, and the improvement is that the device comprises:
the acquisition module is used for acquiring data to be processed;
and the processing module is used for inputting the data to be processed into a pre-established automatic lightweight neural network and acquiring a processing result of the data to be processed output by the pre-established automatic lightweight neural network.
Compared with the closest prior art, the invention has the following beneficial effects:
according to the technical scheme provided by the invention, firstly, data to be processed is obtained, then the data to be processed is input into a pre-established automatic lightweight neural network, and a processing result of the data to be processed output by the pre-established automatic lightweight neural network is obtained, so that the method and the device can be suitable for identifying, detecting and segmenting objects or scenes with large scale change;
furthermore, the automatic lightweight neural network searches the topological relation of the blocks based on the random graph method, is beneficial to improving the diversity of the automatically searched model, enriches the connection form inside the blocks, can adjust the number of nodes in the blocks, has certain controllability when the model is deployed on edge computing equipment, samples the mapping function of each neuron in a search space formed by deep separable convolution, can make the model lighter, enables the perception field of the neural network to be more flexible due to the matching use of convolution kernels with various sizes, is beneficial to identifying objects or scenes with large scale changes in remote sensing images, and can deepen the depth of the model and improve the final performance of the model by stacking the blocks with topology and mapping.
Drawings
FIG. 1 is a flow chart of a data processing method based on an automated lightweight neural network according to the present invention;
FIG. 2 is a schematic flow chart of an automated lightweight neural network design in an embodiment of the present invention;
FIG. 3 is a schematic diagram of a topological connection relationship of a BA random graph model in the embodiment of the present invention;
FIG. 4 is a schematic diagram of topological connection relationships of an ER random graph model in an embodiment of the present invention;
FIG. 5 is a schematic diagram illustrating a topological connection relationship of a WS random graph model in an embodiment of the present invention;
FIG. 6 is a diagram illustrating the operation of the depth separable convolution according to an embodiment of the present invention;
FIG. 7 is a schematic structural diagram of a data processing apparatus based on an automated lightweight neural network according to the present invention;
Detailed Description
The following describes embodiments of the present invention in further detail with reference to the accompanying drawings.
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
In order to solve the problems that the existing NAS topology connection form based on RL is single, the generated model is high in complexity and is not suitable for remote sensing images, the invention provides a data processing method based on an automatic lightweight neural network, as shown in FIG. 1, the method comprises the following steps:
101, acquiring data to be processed;
102, inputting the data to be processed to a pre-established automatic lightweight neural network, and acquiring a processing result of the data to be processed output by the pre-established automatic lightweight neural network.
Wherein, the data to be processed is remote sensing data, and the processing result comprises: data recognition results, data detection results and data segmentation results.
Further, the process of establishing the pre-established automated lightweight neural network includes:
step 1, training set data and verification set data corresponding to data to be processed are obtained, in the process, the training set and the verification set can be cut, and memory or video memory overflow is prevented;
step 2, setting the number M of blocks of the pre-established automatic lightweight neural network, and numbering the blocks, wherein the 1 st Block is a Stem Layer (Stem Layer) which is used for changing the size of an input image, changing the width and the height of the input image into the former half, simultaneously increasing the number of channels of input data, and changing the channels from 3 channels into C channels;
step 3, setting the number of neurons in each of the blocks from the 2 nd to the Mth as N, and numbering the neurons in each of the blocks from the 2 nd to the Mth from 1 to N;
wherein, the setting of M and N is to satisfy the constraint of edge computing device hardware, such as CPU or GPU performance, memory or storage space size, etc.;
step 4, selecting a random graph which is depended by each Block from the 2 nd Block to the Mth Block;
step 5, generating a topological relation among the neurons of each Block in the 2 nd to Mth blocks based on the random graphs depended on by the 2 nd to Mth blocks, wherein the neurons of each Block in the 2 nd to Mth blocks meet the condition that the neurons with smaller sequence numbers point to the neurons with larger sequence numbers;
step 6, setting a search space of the neurons of each Block from the 2 nd Block to the Mth Block;
step 7, stacking the 1 st to Mth blocks from small to large according to numbers, inserting a pooling layer after the 2 nd to Mth-1 th blocks, introducing a head network corresponding to a processing method after the Mth Block, and finally forming an automatic lightweight neural network initial model, taking data detection processing as an example, introducing a detection head network as shown in FIG. 2;
step 8, the training data is sent into the automatic lightweight neural network initial model in a minipatch form, and learning is carried out in a supervision mode;
furthermore, in the learning process, training data sequentially pass through each Block and Pobing Layer, and the detected target coordinates and the confidence score thereof are output by the detection head network. The cross-over ratio (IoU) of the output detection box and the real label (ground route) is then calculated, and the weight in the network is finely adjusted by using a Stochastic Gradient Descent (SGD) algorithm with a strategy of warmup learning rate (learning rate) under the constraint of the local loss function.
And 9, setting the total verification times, verifying the performance index of the trained automatic lightweight neural network initial model by using the verification data after each T times of training, and selecting the automatic lightweight neural network initial model with the maximum performance index as the pre-established automatic lightweight neural network.
In a preferred embodiment of the present invention, the step 4 includes:
each of the 2 nd to mth blocks randomly selects a random graph from a BA random graph, an ER random graph and a WS random graph as a random graph depended on by the Block, wherein the probabilities of the selection of the BA random graph, the ER random graph and the WS random graph are the same, and schematic diagrams of topological connection relationships of the BA random graph, the ER random graph and the WS random graph are respectively shown in fig. 3, 4 and 5.
Further, the step 6 comprises:
the neurons in each of the 2 nd to mth blocks select mapping functions belonging to the neurons at {3 × 3DWConv,5 × 5DWConv,7 × 7DWConv,9 × 9DWConv }, wherein the probabilities of sampling 3 × 3DWConv,5 × 5DWConv,7 × 7DWConv and 9 × 9DWConv are all the same, DWConv is depth separable convolution, and a schematic diagram of the calculation process of 3 × 3DWConv convolution is shown in fig. 6.
When the processing method is a data identification method, the performance index of the initial model of the automatic lightweight neural network is an OA performance index, when the processing method is a data detection method, the performance index of the initial model of the automatic lightweight neural network is an mAP performance index, and when the processing method is a data segmentation method, the performance index of the initial model of the automatic lightweight neural network is an mF1 performance index.
In the network generation process, the topological connection relation between each neuron of each Block in the deep neural network is constructed on the basis of a classical random graph (random graph) model, the setting of artificial rules is reduced, and the connection form of the model is enriched. This example employs three random graph models: BA random graph, ER random graph, WS random graph. The number of nodes in the random graph is related to the complexity of the finally generated model, and the model is more complex when the number of nodes is larger. The number should be set in conjunction with the actual hardware conditions of the edge computing device, such as CPU or GPU performance, memory or storage size, etc. The classical random Graph model is an undirected Graph structure, while neural networks are generally viewed as a Directed Acyclic Graph (DAG) structure. Here, in the embodiment of the present application, each node in Block is numbered, and it is specified that the connection direction can only be pointed to a node with a larger sequence number by a node with a smaller sequence number, so that confusion caused by forward propagation and backward propagation due to a searched network forming a loop is effectively avoided.
Meanwhile, the embodiment provided by the invention performs operation sampling (operation sampling) on the constructed network containing the topological connection relation, and perfects Block. After the topological connection relationship is determined, each neuron of the neural network also needs to map (mapping) data flowing in the network. The mapping function of each neuron is sampled according to some probability distribution in a search space (search space) that includes all the required mappings. In consideration of a lightweight structure and a large-size multi-scale remote sensing image, the present embodiment takes depth separable convolution (DWConv) of different sizes as a search space. The deep separable convolution is a special form of grouped convolution, the number of channels of a convolution kernel is equal to that of a feature map, and the convolution kernel of each channel only performs convolution operation with the feature map of the corresponding channel and does not perform any operation with the feature maps of other channels. DWConv not only has good expression capacity for data characteristics, but also can effectively reduce Parameters (Parameters) and floating point operations (FLOPs) of the constructed neural network, and the convolution in the form is used as a search space, so that the finally derived network model is lighter, and the deployment and application on edge computing equipment are very facilitated. Meanwhile, DWConv with various sizes can be used together in a matched mode, and the receptive field (receptive field) of the neural network can be adjusted, so that the finally generated model is more robust to recognition of objects or scenes with large scale changes in the remote sensing image.
And finally, stacking the obtained network blocks with the topological connection relation and the mapping function, removing the first Block and the last Block, inserting a Pooling Layer (Pooling Layer) into the rest blocks, and changing the size of the characteristic diagram extracted by each Block. And finally, aiming at different tasks, introducing different head networks at the top end of the network to generate a final model structure, and then finely adjusting the generated network through iterative training on training data.
Based on the same inventive concept, the present invention further provides a data processing apparatus based on an automated lightweight neural network, as shown in fig. 7, the apparatus includes:
the acquisition module is used for acquiring data to be processed;
and the processing module is used for inputting the data to be processed into a pre-established automatic lightweight neural network and acquiring a processing result of the data to be processed output by the pre-established automatic lightweight neural network.
Preferably, the data to be processed is remote sensing data.
Preferably, the processing result includes: data recognition results, data detection results and data segmentation results.
Preferably, the process of establishing the pre-established automated lightweight neural network includes:
step 1, acquiring training set data and verification set data corresponding to data to be processed;
step 2, setting the number M of blocks of the pre-established automatic lightweight neural network, and numbering the blocks, wherein the 1 st Block is a stem layer;
step 3, setting the number of neurons in each of the blocks from the 2 nd to the Mth as N, and numbering the neurons in each of the blocks from the 2 nd to the Mth from 1 to N;
step 4, selecting a random graph which is depended by each Block from the 2 nd Block to the Mth Block;
step 5, generating a topological relation among the neurons of each Block in the 2 nd to Mth blocks based on the random graphs depended on by the 2 nd to Mth blocks, wherein the neurons of each Block in the 2 nd to Mth blocks meet the condition that the neurons with smaller sequence numbers point to the neurons with larger sequence numbers;
step 6, setting a search space of the neurons of each Block from the 2 nd Block to the Mth Block;
step 7, stacking the 1 st to Mth blocks from small to large according to the numbers, inserting all the blocks from the 2 nd to the Mth to the 1 st after each Block into a pooling layer, introducing a head network corresponding to the processing method after the Mth Block, and finally forming an automatic lightweight neural network initial model;
step 8, the training data is sent into the automatic lightweight neural network initial model in a minipatch form, and learning is carried out in a supervision mode;
and 9, setting the total verification times, verifying the performance index of the trained automatic lightweight neural network initial model by using the verification data after each T times of training, and selecting the automatic lightweight neural network initial model with the maximum performance index as the pre-established automatic lightweight neural network.
Further, the step 4 comprises:
and each of the 2 nd to Mth blocks randomly selects a random graph from a BA random graph, an ER random graph and a WS random graph as a random graph depended by the Block, wherein the probability of selecting the BA random graph, the ER random graph and the WS random graph is the same.
Further, the step 6 comprises:
the neurons in each of the 2 nd to mth blocks select mapping functions belonging to themselves at {3 × 3DWConv,5 × 5DWConv,7 × 7DWConv,9 × 9DWConv }, wherein the probabilities of sampling 3 × 3DWConv,5 × 5DWConv,7 × 7DWConv, and 9 × 9DWConv are all the same, DWConv being a depth separable convolution.
Further, when the processing method is a data identification method, the performance index of the initial model of the automatic lightweight neural network is an OA performance index, when the processing method is a data detection method, the performance index of the initial model of the automatic lightweight neural network is an mAP performance index, and when the processing method is a data segmentation method, the performance index of the initial model of the automatic lightweight neural network is an mF1 performance index.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solutions of the present invention and not for limiting the same, and although the present invention is described in detail with reference to the above embodiments, those of ordinary skill in the art should understand that: modifications and equivalents may be made to the embodiments of the invention without departing from the spirit and scope of the invention, which is to be covered by the claims.

Claims (8)

1. A data processing method based on an automatic lightweight neural network is characterized by comprising the following steps:
acquiring data to be processed;
and inputting the data to be processed into a pre-established automatic lightweight neural network, and acquiring a processing result of the data to be processed output by the pre-established automatic lightweight neural network.
2. The method of claim 1, wherein the data to be processed is remotely sensed data.
3. The method of claim 1, wherein the processing results comprise: data recognition results, data detection results and data segmentation results.
4. The method of claim 1, wherein the pre-established automated lightweight neural network establishment procedure comprises:
step 1, acquiring training set data and verification set data corresponding to data to be processed;
step 2, setting the number M of blocks of the pre-established automatic lightweight neural network, and numbering the blocks, wherein the 1 st Block is a stem layer;
step 3, setting the number of neurons in each of the blocks from the 2 nd to the Mth as N, and numbering the neurons in each of the blocks from the 2 nd to the Mth from 1 to N;
step 4, selecting a random graph which is depended by each Block from the 2 nd Block to the Mth Block;
step 5, generating a topological relation among the neurons of each Block in the 2 nd to Mth blocks based on the random graphs depended on by the 2 nd to Mth blocks, wherein the neurons of each Block in the 2 nd to Mth blocks meet the condition that the neurons with smaller sequence numbers point to the neurons with larger sequence numbers;
step 6, setting a search space of the neurons of each Block from the 2 nd Block to the Mth Block;
step 7, stacking the 1 st to Mth blocks from small to large according to the numbers, inserting all the blocks from the 2 nd to the Mth to the 1 st after each Block into a pooling layer, introducing a head network corresponding to the processing method after the Mth Block, and finally forming an automatic lightweight neural network initial model;
step 8, the training data is sent into the automatic lightweight neural network initial model in a minipatch form, and learning is carried out in a supervision mode;
and 9, setting the total verification times, verifying the performance index of the trained automatic lightweight neural network initial model by using the verification data after each T times of training, and selecting the automatic lightweight neural network initial model with the maximum performance index as the pre-established automatic lightweight neural network.
5. The method of claim 4, wherein step 4 comprises:
and each of the 2 nd to Mth blocks randomly selects a random graph from a BA random graph, an ER random graph and a WS random graph as a random graph depended by the Block, wherein the probability of selecting the BA random graph, the ER random graph and the WS random graph is the same.
6. The method of claim 4, wherein the step 6 comprises:
the neurons in each of the 2 nd to mth blocks select mapping functions belonging to themselves at {3 × 3DWConv,5 × 5DWConv,7 × 7DWConv,9 × 9DWConv }, wherein the probabilities of sampling 3 × 3DWConv,5 × 5DWConv,7 × 7DWConv, and 9 × 9DWConv are all the same, DWConv being a depth separable convolution.
7. The method of claim 4, wherein the performance index of the initial model of the automated lightweight neural network is an OA performance index when the processing method is a data recognition method, an mAP performance index when the processing method is a data detection method, and an mF1 performance index when the processing method is a data segmentation method.
8. A data processing apparatus based on an automated lightweight neural network, the apparatus comprising:
the acquisition module is used for acquiring data to be processed;
and the processing module is used for inputting the data to be processed into a pre-established automatic lightweight neural network and acquiring a processing result of the data to be processed output by the pre-established automatic lightweight neural network.
CN202010351374.3A 2020-04-28 2020-04-28 Data processing method and device based on automatic lightweight neural network Pending CN111582475A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010351374.3A CN111582475A (en) 2020-04-28 2020-04-28 Data processing method and device based on automatic lightweight neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010351374.3A CN111582475A (en) 2020-04-28 2020-04-28 Data processing method and device based on automatic lightweight neural network

Publications (1)

Publication Number Publication Date
CN111582475A true CN111582475A (en) 2020-08-25

Family

ID=72126227

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010351374.3A Pending CN111582475A (en) 2020-04-28 2020-04-28 Data processing method and device based on automatic lightweight neural network

Country Status (1)

Country Link
CN (1) CN111582475A (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104820840A (en) * 2015-05-08 2015-08-05 西安电子科技大学 Nearest neighborhood hyper-spectral image classification method based on dictionary and band restructuring
CN105809185A (en) * 2015-12-31 2016-07-27 天津大学 High-spectrum image nonlinear demixing method based on neural network and differential search
CN106991397A (en) * 2017-03-31 2017-07-28 中国科学院电子学研究所 View-based access control model conspicuousness constrains the remote sensing images detection method of depth confidence network
CN109299142A (en) * 2018-11-14 2019-02-01 中山大学 A kind of convolutional neural networks search structure method and system based on evolution algorithm
CN109948029A (en) * 2019-01-25 2019-06-28 南京邮电大学 Based on the adaptive depth hashing image searching method of neural network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104820840A (en) * 2015-05-08 2015-08-05 西安电子科技大学 Nearest neighborhood hyper-spectral image classification method based on dictionary and band restructuring
CN105809185A (en) * 2015-12-31 2016-07-27 天津大学 High-spectrum image nonlinear demixing method based on neural network and differential search
CN106991397A (en) * 2017-03-31 2017-07-28 中国科学院电子学研究所 View-based access control model conspicuousness constrains the remote sensing images detection method of depth confidence network
CN109299142A (en) * 2018-11-14 2019-02-01 中山大学 A kind of convolutional neural networks search structure method and system based on evolution algorithm
CN109948029A (en) * 2019-01-25 2019-06-28 南京邮电大学 Based on the adaptive depth hashing image searching method of neural network

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
CHENXI LIU等: "《Auto-DeepLab: Hierarchical Neural Architecture Search for Semantic Image Segmentation》", 《2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR)》 *
SAINING XIE等: "《Exploring Randomly Wired Neural Networks for Image Recognition》", 《2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV)》 *
SIGAI: "《NAS(神经结构搜索)综述》", 《HTTPS://ZHUANLAN.ZHIHU.COM/P/60414004》 *

Similar Documents

Publication Publication Date Title
CN110298404B (en) Target tracking method based on triple twin Hash network learning
CN109919108B (en) Remote sensing image rapid target detection method based on deep hash auxiliary network
CN110633745B (en) Image classification training method and device based on artificial intelligence and storage medium
CN113110482B (en) Indoor environment robot exploration method and system based on priori information heuristic method
CN109902192B (en) Remote sensing image retrieval method, system, equipment and medium based on unsupervised depth regression
CN106897714A (en) A kind of video actions detection method based on convolutional neural networks
CN108805149A (en) A kind of winding detection method and device of visual synchronization positioning and map structuring
CN110472738A (en) A kind of unmanned boat Real Time Obstacle Avoiding algorithm based on deeply study
CN111709468B (en) Training method and device for directional artificial intelligence and storage medium
CN117292283B (en) Target identification method based on unmanned aerial vehicle
CN109948457A (en) The real time target recognitio algorithm accelerated based on convolutional neural networks and CUDA
CN114881225A (en) Power transmission and transformation inspection model network structure searching method, system and storage medium
KR20210027910A (en) Method and system for highlighting similar areas using similarity between images
CN108985385A (en) Based on the quick Weakly supervised object detection method for generating confrontation study
CN113887649A (en) Target detection method based on fusion of deep-layer features and shallow-layer features
CN106022293A (en) Pedestrian re-identification method of evolutionary algorithm based on self-adaption shared microhabitat
CN116189281B (en) End-to-end human behavior classification method and system based on space-time self-adaptive fusion
CN111582475A (en) Data processing method and device based on automatic lightweight neural network
CN115457269A (en) Semantic segmentation method based on improved DenseNAS
CN115661123A (en) Industrial product surface defect position detection method based on weak supervision target detection
CN115620238A (en) Park pedestrian attribute identification method based on multivariate information fusion
CN115909027A (en) Situation estimation method and device
Liu et al. Research on object detection algorithm for small object of pests based on YOLOv3
CN114038058A (en) Parallel human body posture detection tracking method based on posture guiding re-recognition features
CN114067155A (en) Image classification method, device, product and storage medium based on meta learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20200825

RJ01 Rejection of invention patent application after publication