CN112258431B - Image classification model based on mixed depth separable expansion convolution and classification method thereof - Google Patents

Image classification model based on mixed depth separable expansion convolution and classification method thereof Download PDF

Info

Publication number
CN112258431B
CN112258431B CN202011032957.6A CN202011032957A CN112258431B CN 112258431 B CN112258431 B CN 112258431B CN 202011032957 A CN202011032957 A CN 202011032957A CN 112258431 B CN112258431 B CN 112258431B
Authority
CN
China
Prior art keywords
layer
convolution
depth separable
image classification
mixed depth
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011032957.6A
Other languages
Chinese (zh)
Other versions
CN112258431A (en
Inventor
闫超
黄俊洁
陶陶
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Dongfang Tiancheng Intelligent Technology Co ltd
Original Assignee
Chengdu Dongfang Tiancheng Intelligent Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Dongfang Tiancheng Intelligent Technology Co ltd filed Critical Chengdu Dongfang Tiancheng Intelligent Technology Co ltd
Priority to CN202011032957.6A priority Critical patent/CN112258431B/en
Publication of CN112258431A publication Critical patent/CN112258431A/en
Application granted granted Critical
Publication of CN112258431B publication Critical patent/CN112258431B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration by the use of local operators
    • G06T5/30Erosion or dilatation, e.g. thinning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/2431Multiple classes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Abstract

The invention discloses an image classification model based on mixed depth separable expansion convolution, which comprises the following construction processes: packaging the depth separable expansion convolution layer, the characteristic connecting layer, the convolution layer, the batch standardization layer and the correction linear unit layer into a mixed depth separable expansion convolution module from front to back; packaging the convolution layer, the batch normalization layer, the correction linear unit layer, the mixed depth separable expansion convolution module, the maximum value pooling layer, the flattening layer, the random inactivation layer and the full connection layer from front to back to form a trunk network of the deep neural network; carrying out random initialization on the parameter weight of the backbone network, and presetting iteration times and momentum parameters of a batch normalization layer; and optimizing parameters of the network model by adopting a random gradient descent method, and repeating iterative calculation until the loss value is converged to obtain the optimal network model. Through the scheme, the method has the advantages of simple structure, less calculation workload, accurate classification and the like.

Description

Image classification model based on mixed depth separable expansion convolution and classification method thereof
Technical Field
The invention relates to the technical field of image processing, in particular to an image classification model based on mixed depth separable expansion convolution and a classification method thereof.
Background
The images can objectively show natural objects, are important resources for people to learn the world, and technicians can obtain beneficial information by analyzing the images and develop related algorithms. Image classification belongs to the computer vision direction, and is widely applied to the fields of medical treatment, food safety and the like.
At present, the main idea of an image classification algorithm in the prior art is to assign corresponding labels to an image set to be classified, and for a computer, an image is a pixel matrix, and effective information in the pixel matrix is extracted by using a related algorithm technology, which is different from a human image recognition mode. The traditional image classification algorithm mainly comprises three steps of manual feature extraction, feature coding and classifier classification; among the commonly used manual features are Local Binary Pattern (LBP), Scale-Invariant Feature Transform (SIFT), Histogram of Oriented Gradient (HOG), and so on. The feature coding is to remove redundant feature information and improve the robustness of the feature information. The general classifiers include a Support Vector Machine (SVM) and an Adaptive Boosting classifier (AdaBoost). Due to the fact that the generalization of the manual features is poor, the application of the traditional image classification algorithm in an actual scene is influenced, and the method has no great advantages in precision and speed. With the improvement of computer computing power, the deep learning algorithm is rapidly developed, and the image classification algorithm for deep learning also starts to rise.
Alex Krizhevsky et al in 2012 proposed an Alexnet network structure to learn the ImageNet data set, to obtain the champion of the ImageNet match in 2012, from which a deep learning algorithm applied to big data image classification has been continuously developed. The image classification algorithm for deep learning has richer semantic information by utilizing the convolution neural network to extract the convolution characteristics of the samples, can enable an algorithm model to learn more efficient expression capability, and is more accurate compared with manual characteristics. However, when the image classification algorithm of the deep learning algorithm is used for training a network model, feature down-sampling operation is added for reducing feature dimensionality, and overfitting phenomena of the network, such as pooling layers, are avoided to a certain extent, so that convolution features lose detail information. Jonathan Long et al propose a full convolution network structure, in which the resolution of the feature map is increased using the deconvolution layer, recovering some detailed information to some extent but roughly. In the same year, Kaiming He et al propose a residual deep neural network, and adopt a jump connection structure to fuse deep network layer features and shallow network layer features, thereby improving the utilization rate of detail information in a feature map.
In the image classification algorithm for deep learning in the prior art, a pooling layer or a convolution layer with a step length of 2 is adopted to perform down-sampling operation on a feature image, so that the number is reduced, the receptive field is enlarged, and the semantic information of the feature image is increased, but the detailed information of the feature image is lost.
In summary, the above methods can only repair the detail information of the feature map to a certain extent, but the parameters of the network model are increased, so that the network model becomes complicated and difficult to optimize, and therefore the problem of loss of the detail information of the feature map still affects the classification detection accuracy.
In addition, the chinese invention patent with patent application number "201910818758.9" and name "industrial product defect image classification method based on lightweight deep neural network" includes: 1. preparing an industrial product image data set; 2. constructing a lightweight deep neural network; 3. inputting a defect image data set of an industrial product into a built lightweight deep neural network, extracting multi-scale features of a polarizer image through network training, and inputting the extracted features into a Softmax layer for classification to obtain a classification model; 4. inputting the test image into a classification model, inputting the probability of the image belonging to a certain class and the label corresponding to the image into an Accuracy layer, and outputting the correct classification result of the image. However, this technique has the following drawbacks:
firstly, the technology mainly fuses the features after deep separable convolution processing with the convolution kernel size of 3x3 and the features after ordinary convolution processing with the convolution kernel size of 1x1, and because the two parts have different characteristic receptive fields and different scales, the purpose of increasing the receptive field is achieved after complementation, although the receptive field size is improved, the receptive field is single, and the classification of small targets is easy to ignore;
secondly, in the technology, the maximum pooling layer is used for down-sampling the feature map, but the resolution of the feature map is not repaired, so that the detail information of the features is greatly lost, and the classification precision is reduced;
thirdly, the parallel depth separable convolution module in the technology improves the semanteme of the feature map through stacking convolution, greatly increases the parameter quantity of the network structure, is not easy to optimize, and limits the classification precision to a certain extent without combining context information for classification.
Therefore, an image classification model based on mixed depth separable expansion convolution and a classification method thereof, which have a simple structure and do not increase the calculation amount of the original model, are urgently needed to be provided, so that the loss of detail information of a feature map is reduced, the generalization and robustness of a network model are improved, and the classification accuracy of the network model is improved.
Disclosure of Invention
In view of the above problems, an object of the present invention is to provide an image classification model based on mixed depth separable dilation convolution and a classification method thereof, and the technical solution adopted by the present invention is as follows:
an image classification model based on mixed depth separable dilation convolution, wherein the image classification model is constructed by the following process:
packaging the depth separable expansion convolution layer, the characteristic connecting layer, the convolution layer, the batch standardization layer and the correction linear unit layer into a mixed depth separable expansion convolution module from front to back;
packaging the convolution layer, the batch normalization layer, the correction linear unit layer, the mixed depth separable expansion convolution module, the maximum value pooling layer, the flattening layer, the random inactivation layer and the full connection layer from front to back to form a trunk network of the deep neural network;
carrying out random initialization on the parameter weight of the backbone network, and presetting iteration times and momentum parameters of a batch normalization layer; and optimizing parameters of the network model by adopting a random gradient descent method, and repeating iterative calculation until the loss value is converged to obtain the optimal network model.
Further, the depth separable expansion convolution layers in the mixed depth separable expansion convolution module are distributed in parallel by adopting 3.
Preferably, the mixed depth separable expansion convolution modules are provided with 8 blocks according to 2, 3 and 3, and a maximum value pooling layer is arranged after any block.
Preferably, the expansion rates of the 3 depth separable expansion convolution layers from top to bottom in the mixed depth separable expansion convolution module are sequentially 1, 2 and 3, the sizes of convolution kernels are all 3 × 3, and the step lengths are all 1.
Further, convolution kernels of convolution layers in the mixed depth separable dilation convolution module are all 1 × 1 in size, and step sizes are all 1.
Further, the deactivation rate of the random deactivation layer was 0.5.
Further, the sampling kernel size of any of the maximum pooling layers is 2 and the step size is 2.
Preferably, the momentum parameter of the batch normalization layer is 0.975, and the learning rate is set to 0.1.
Preferably, the number of iterations is 30000.
An image classification method adopts an image classification model based on mixed depth separable expansion convolution for classification.
Compared with the prior art, the invention has the following beneficial effects:
(1) the invention skillfully adopts the mixed depth separable expansion convolution module which can keep the detail information of the characteristics under the condition of unchanged original network parameters and quickly increase the receptive field; in addition, the invention builds a mixed depth separable expansion convolution module, and adjusts the original convolution kernel by adjusting the expansion rate, so that the parameter quantity of the original network model is not increased, the context information in the characteristic diagram can be supplemented, the information loss caused by operations such as downsampling and the like is compensated, and the expression capability of the network model is improved.
(2) The invention provides a method for expanding the receptive field of a characteristic image by using mixed depth separable expansion convolution in the field of image classification, obtaining more context information, weakening the sensitivity of a network model to the scale change of an image target, being beneficial to distinguishing the category of the image and improving the classification precision of the network model.
(3) The invention skillfully adopts 3 parallel depth separable expansion convolution layers, when the characteristic block is input into the mixed depth separable expansion convolution module, the characteristic block is divided into three parts according to the number of channels and respectively input into the parallel depth separable expansion convolution layers, and the receptive field of the characteristic diagram is expanded by utilizing the characteristics of expansion convolution.
(4) The main part network of the invention extracts the characteristic information and sends the information to the flattening layer, the multidimensional characteristic is unidimensional, and the multi-class classification is carried out from the convolution layer to the full connection layer. The classification loss function of the invention adopts a flexible maximum loss function, and the loss value calculated by continuously optimizing the loss function during training guides the predicted value calculated by the network model to gradually approach the true value.
(5) According to the invention, the expansion rates of the 3 depth separable expansion convolution layers from top to bottom are set to be 1, 2 and 3 according to the saw-tooth structure, and then three parts of feature information are spliced into one feature block so as to enlarge the extraction range of features and mutually compensate the information, so that the network effect can be effectively avoided.
(6) According to the method, the common convolution integration characteristics are set after the depth separable expansion convolution layer, and the relation between characteristic graphs is established, so that the expression capability of the characteristic blocks can be improved, and the number of channels can be increased to increase the dimensionality of the characteristic blocks.
In conclusion, the method has the advantages of simple structure, less calculation workload, accurate classification and the like, and has high practical value and popularization value in the technical field of image processing.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the embodiments will be briefly described below, it should be understood that the following drawings only illustrate some embodiments of the present invention, and therefore should not be considered as limiting the scope of protection, and it is obvious for those skilled in the art that other related drawings can be obtained according to these drawings without inventive efforts.
FIG. 1 is a schematic diagram of the hybrid depth separable dilation-convolution module of the present invention.
Fig. 2 is a schematic structural diagram of the present invention.
FIG. 3 is a diagram illustrating the classification effect of the present invention.
Detailed Description
To further clarify the objects, technical solutions and advantages of the present application, the present invention will be further described with reference to the accompanying drawings and examples, and embodiments of the present invention include, but are not limited to, the following examples. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
Examples
As shown in fig. 1 to 2, in the present embodiment, an image classification method based on mixed depth separable expansion convolution is provided, in which a mixed depth separable expansion convolution module is added to a constructed network model structure, and the module is composed of a plurality of depth separable expansion convolution layers with different expansion rates, so as to expand a range of feature extraction, supplement context information of features, improve an expression capability of a network model, and thereby improve classification accuracy of the model.
Specifically, the image classification model of the present embodiment includes a convolution layer, a batch normalization layer, a modified linear unit layer, 2 mixed-depth separable expansion convolution modules, 1 maximum pooling layer, 3 mixed-depth separable expansion convolution modules, 1 maximum pooling layer, a flattening layer, a random deactivation layer, and a full connection layer, which are packaged from front to back; in this embodiment, the flexible maximum loss function is used as the loss function, and the predicted value calculated by the network model is guided to gradually approach the true value by continuously optimizing the difference value calculated by the loss function during training. And setting a certain iteration number, calculating a loss value between the predicted value and the true value by forward propagation, optimizing network model parameters by backward propagation, and repeating the calculation until the loss value is converged to obtain the optimal network model. Any one mixed depth separable expansion convolution module is formed by packaging 3 depth separable expansion convolution layers which are arranged in parallel and a characteristic connecting layer, a convolution layer, a batch standardization layer and a correction linear unit layer which are connected in sequence.
In this embodiment, when the feature block is input to the module, the feature block is divided into three parts according to the number of channels, and the three parts are respectively input to the 3 depth separable expansion convolution layers, so as to expand the feature extraction range and supplement the context information of the feature map, and the calculation formula of the actual convolution kernel size of the expansion convolution layer is as follows:
K=k+(k-1)(r-1)
where K represents the original convolution kernel size input by the depth separable inflated convolution layer, r represents the inflation rate, and K is the actual convolution kernel size of the theoretically calculated depth separable inflated convolution layer.
In this embodiment, because the dilation convolution is sampled at equal intervals according to dilation rate, local information on the feature map is lost, correlation between feature points is affected, a grid effect occurs, and expression capability of the feature map is reduced, the dilation rate of each depth separable dilation convolution layer needs to be set differently, that is, set to 1, 2, and 3 according to a zigzag structure, and then three parts of feature information are spliced into a feature block, so that an extraction range of the features is expanded, and information is mutually compensated, so that the network effect can be effectively avoided, and feature maps with different receptive fields are spliced to form a multi-scale feature, thereby reducing sensitivity to multi-scale changes of a target. Secondly, after the characteristic information is processed by the depth separable expansion convolution, a common convolution integrating characteristic with the convolution kernel size of 1 multiplied by 1 is used for establishing the relation between characteristic graphs, improving the expression capability of the characteristic blocks, increasing the number of channels and increasing the dimensionality of the characteristic blocks.
In this embodiment, a mixed depth separable dilated convolution module is built, which adjusts the original convolution kernel by adjusting the dilation rate, for example, the depth separable dilated convolution with the convolution kernel size of 3x3, when the dilation rate is 2, the convolution kernel becomes 5x 5; at a dilation rate of 3, the convolution kernel becomes 7 × 7. In this embodiment, a random gradient descent method is used to optimize the parameters of the network model with a Neisseliverv momentum set to 0.9. In addition, the activation function layer used in the entire network structure is a modified linear unit layer suppressing the maximum value to 6. The deactivation rate of the random deactivation layer was 0.5.
In this example, the convolution component parameters and the receptive field size are compared as shown in the following table:
TABLE 1 convolution partial parameter and receptive field size comparison
Figure GDA0003071399880000071
In order to verify that the model has good classification performance, the following tests are carried out:
experiments were performed as in this example using the public data set flower _ lights data set, with 5 types of flowers, tulips (tulips), sunflowers (sunflowers), roses (roses), dandelions (dandelion), and brookfield (daisy), respectively; its data set, although less, can also be adapted to conventional classification. In the experiment of the embodiment, the parallel convolution part in the mixed depth separable expansion convolution layer is replaced by the common depth separable convolution layer, and the influence of the two on the accuracy of the network model is compared. In the present embodiment, the accuracy pair of the data set is as follows:
TABLE 2 precision contrast on flow _ Phots dataset
Figure GDA0003071399880000072
The overall network structure is the same, the convolution kernels are the same in size and are all 3x3, and the accuracy of the network structure is improved by about 1.2% by using the mixed depth separable expansion convolution module compared with the accuracy of the network structure using the common depth separable convolution layer according to the table. The recognition result of this embodiment is shown in fig. 3.
The above-mentioned embodiments are only preferred embodiments of the present invention, and do not limit the scope of the present invention, but all the modifications made by the principles of the present invention and the non-inventive efforts based on the above-mentioned embodiments shall fall within the scope of the present invention.

Claims (9)

1. The image classification model based on the mixed depth separable dilation convolution is characterized in that the construction process of the image classification model is as follows:
packaging the depth separable expansion convolution layer, the characteristic connecting layer, the convolution layer, the batch standardization layer and the correction linear unit layer into a mixed depth separable expansion convolution module from front to back; the depth separable expansion convolution layers in the mixed depth separable expansion convolution module adopt 3 and are arranged in parallel;
packaging the convolution layer, the batch normalization layer, the correction linear unit layer, the mixed depth separable expansion convolution module, the maximum value pooling layer, the flattening layer, the random inactivation layer and the full connection layer from front to back to form a trunk network of the deep neural network;
carrying out random initialization on the parameter weight of the backbone network, and presetting iteration times and momentum parameters of a batch normalization layer; and optimizing parameters of the network model by adopting a random gradient descent method, and repeating iterative calculation until the loss value is converged to obtain the optimal network model.
2. The mixed depth separable dilated convolution based image classification model of claim 1, wherein the mixed depth separable dilated convolution module is provided with 8 and is partitioned into 2, 3 and 3 blocks, and any partition is followed by a maximum pooling layer.
3. The image classification model based on the mixed depth separable dilation convolution of claim 1, wherein dilation rates of 3 depth separable dilation convolution layers from top to bottom in the mixed depth separable dilation convolution module are 1, 2, and 3 in sequence, convolution kernel sizes are all 3 × 3, and step sizes are all 1.
4. The mixed depth separable dilated convolution based image classification model of claim 3, wherein convolution layers in the mixed depth separable dilated convolution module have convolution kernel sizes of 1x1 and step sizes of 1.
5. The mixed depth separable dilated convolution based image classification model of claim 1, wherein the inactivation rate of the random inactivation layer is 0.5.
6. The mixed depth separable dilated convolution based image classification model of claim 2, wherein the sampling kernel size of any of the maximum pooling layers is 2 and the step size is 2.
7. The mixed depth separable dilated convolution based image classification model of claim 1, wherein the momentum parameter of the batch normalization layer is 0.975 and the learning rate is set to 0.1.
8. The mixed depth separable dilated convolution based image classification model of claim 1, wherein the number of iterations is 30000.
9. An image classification method, characterized in that the image classification model based on the mixed depth separable dilation convolution of any one of claims 1 to 8 is used for classification.
CN202011032957.6A 2020-09-27 2020-09-27 Image classification model based on mixed depth separable expansion convolution and classification method thereof Active CN112258431B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011032957.6A CN112258431B (en) 2020-09-27 2020-09-27 Image classification model based on mixed depth separable expansion convolution and classification method thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011032957.6A CN112258431B (en) 2020-09-27 2020-09-27 Image classification model based on mixed depth separable expansion convolution and classification method thereof

Publications (2)

Publication Number Publication Date
CN112258431A CN112258431A (en) 2021-01-22
CN112258431B true CN112258431B (en) 2021-07-20

Family

ID=74234344

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011032957.6A Active CN112258431B (en) 2020-09-27 2020-09-27 Image classification model based on mixed depth separable expansion convolution and classification method thereof

Country Status (1)

Country Link
CN (1) CN112258431B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112507982B (en) * 2021-02-02 2021-05-07 成都东方天呈智能科技有限公司 Cross-model conversion system and method for face feature codes
CN112507996B (en) * 2021-02-05 2021-04-20 成都东方天呈智能科技有限公司 Face detection method of main sample attention mechanism
CN113111889A (en) * 2021-03-10 2021-07-13 国网浙江省电力有限公司宁波供电公司 Target detection network processing method for edge computing terminal
CN112949614B (en) * 2021-04-29 2021-09-10 成都市威虎科技有限公司 Face detection method and device for automatically allocating candidate areas and electronic equipment
CN113033504B (en) * 2021-05-19 2021-08-27 广东众聚人工智能科技有限公司 Multi-scale video anomaly detection method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109544621A (en) * 2018-11-21 2019-03-29 马浩鑫 Light field depth estimation method, system and medium based on convolutional neural networks
CN109635882A (en) * 2019-01-23 2019-04-16 福州大学 Salient object detection method based on multi-scale convolution feature extraction and fusion
CN111179283A (en) * 2019-12-30 2020-05-19 深圳市商汤科技有限公司 Image semantic segmentation method and device and storage medium
CN111462124A (en) * 2020-03-31 2020-07-28 武汉卓目科技有限公司 Remote sensing satellite cloud detection method based on Deep L abV3+
CN111523459A (en) * 2020-04-22 2020-08-11 中科三清科技有限公司 Remote sensing image bare area identification method and device, electronic equipment and storage medium

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109214406B (en) * 2018-05-16 2021-07-09 长沙理工大学 Image classification method based on D-MobileNet neural network
CN111582007A (en) * 2019-02-19 2020-08-25 富士通株式会社 Object identification method, device and network
CN110232696B (en) * 2019-06-20 2024-03-08 腾讯科技(深圳)有限公司 Image region segmentation method, model training method and device
CN110660046B (en) * 2019-08-30 2022-09-30 太原科技大学 Industrial product defect image classification method based on lightweight deep neural network
CN111104962B (en) * 2019-11-05 2023-04-18 北京航空航天大学青岛研究院 Semantic segmentation method and device for image, electronic equipment and readable storage medium
CN110929602B (en) * 2019-11-09 2023-08-22 北京格镭信息科技有限公司 Foundation cloud picture cloud identification method based on convolutional neural network
CN111401361B (en) * 2020-03-06 2022-09-30 南京理工大学 End-to-end lightweight depth license plate recognition method
CN111401292B (en) * 2020-03-25 2023-05-26 成都东方天呈智能科技有限公司 Face recognition network construction method integrating infrared image training
CN111598108A (en) * 2020-04-22 2020-08-28 南开大学 Rapid salient object detection method of multi-scale neural network based on three-dimensional attention control
CN111488884A (en) * 2020-04-28 2020-08-04 东南大学 Real-time semantic segmentation method with low calculation amount and high feature fusion
CN111652129A (en) * 2020-06-02 2020-09-11 北京联合大学 Vehicle front obstacle detection method based on semantic segmentation and multi-feature fusion
AU2020101729A4 (en) * 2020-08-08 2020-09-17 Babuprasad, Shweta MS Continuous labelling assessment of products to improve efficiency of reverse logistics by deep learning model

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109544621A (en) * 2018-11-21 2019-03-29 马浩鑫 Light field depth estimation method, system and medium based on convolutional neural networks
CN109635882A (en) * 2019-01-23 2019-04-16 福州大学 Salient object detection method based on multi-scale convolution feature extraction and fusion
CN111179283A (en) * 2019-12-30 2020-05-19 深圳市商汤科技有限公司 Image semantic segmentation method and device and storage medium
CN111462124A (en) * 2020-03-31 2020-07-28 武汉卓目科技有限公司 Remote sensing satellite cloud detection method based on Deep L abV3+
CN111523459A (en) * 2020-04-22 2020-08-11 中科三清科技有限公司 Remote sensing image bare area identification method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN112258431A (en) 2021-01-22

Similar Documents

Publication Publication Date Title
CN112258431B (en) Image classification model based on mixed depth separable expansion convolution and classification method thereof
US10713563B2 (en) Object recognition using a convolutional neural network trained by principal component analysis and repeated spectral clustering
CN110321967B (en) Image classification improvement method based on convolutional neural network
CN110909801B (en) Data classification method, system, medium and device based on convolutional neural network
CN111898621B (en) Contour shape recognition method
CN109002755B (en) Age estimation model construction method and estimation method based on face image
Vijayaraghavan et al. Handwritten Tamil recognition using a convolutional neural network
Yang et al. Down image recognition based on deep convolutional neural network
CN112163114B (en) Image retrieval method based on feature fusion
CN108985442B (en) Handwriting model training method, handwritten character recognition method, device, equipment and medium
Jasitha et al. Venation based plant leaves classification using GoogLeNet and VGG
Giraddi et al. Flower classification using deep learning models
CN114818963A (en) Small sample detection algorithm based on cross-image feature fusion
CN111428758A (en) Improved remote sensing image scene classification method based on unsupervised characterization learning
Ulaganathan et al. Isolated handwritten Tamil character recognition using convolutional neural networks
Dan et al. Pf-vit: Parallel and fast vision transformer for offline handwritten chinese character recognition
James et al. Malayalam handwritten character recognition using AlexNet based architecture
CN113033345B (en) V2V video face recognition method based on public feature subspace
Elaraby et al. A Novel Siamese Network for Few/Zero-Shot Handwritten Character Recognition Tasks.
CN108805280A (en) A kind of method and apparatus of image retrieval
CN113963272A (en) Unmanned aerial vehicle image target detection method based on improved yolov3
Zhang et al. Consecutive convolutional activations for scene character recognition
CN111027570B (en) Image multi-scale feature extraction method based on cellular neural network
Ghayoumi et al. Local sensitive hashing (LSH) and convolutional neural networks (CNNs) for object recognition
CN113486175A (en) Text classification method, text classification device, computer equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant