CN112232263B - Tomato identification method based on deep learning - Google Patents

Tomato identification method based on deep learning Download PDF

Info

Publication number
CN112232263B
CN112232263B CN202011169184.6A CN202011169184A CN112232263B CN 112232263 B CN112232263 B CN 112232263B CN 202011169184 A CN202011169184 A CN 202011169184A CN 112232263 B CN112232263 B CN 112232263B
Authority
CN
China
Prior art keywords
layers
tomato
convolution
training
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011169184.6A
Other languages
Chinese (zh)
Other versions
CN112232263A (en
Inventor
梁喜凤
顾鹏程
赵力勤
余文胜
孙立峰
徐学珍
谢文兵
王永维
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Jiliang University
Original Assignee
China Jiliang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Jiliang University filed Critical China Jiliang University
Priority to CN202011169184.6A priority Critical patent/CN112232263B/en
Publication of CN112232263A publication Critical patent/CN112232263A/en
Application granted granted Critical
Publication of CN112232263B publication Critical patent/CN112232263B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a tomato identification method based on deep learning. Firstly, collecting images of tomatoes under natural conditions, carrying out image data enhancement on the collected images, increasing data samples, and then carrying out artificial annotation of target tomatoes on all the image data, wherein the image data is divided into a training set and a verification set. All pooling layers are removed from the VGG network framework, the residual network is changed, and the convolution layer of the residual network uses hole convolution. According to the method for introducing the cavity convolution layer in the image recognition stage, the tomato recognition accuracy in the complex environment can be improved, and the overall working efficiency of the tomato picking robot can be improved.

Description

Tomato identification method based on deep learning
Technical Field
The invention relates to a tomato identification method, in particular to a tomato identification method based on deep learning.
Background
China is the country of the largest tomato production and consumption in the world, but the lack of labor for the production data activities is serious. The tomato picking robot has important significance for reducing production cost and improving picking efficiency. The accuracy of identification and positioning determines the working efficiency of the tomato picking robot. The growth forms of the tomatoes are different, the fruits are overlapped, the leaves, the branches and the illumination intensity of the tomatoes are different, the fruits are shielded, and the recognition and the positioning of the tomatoes under natural conditions are studied, so that the method has important significance for improving the efficiency of the picking robot.
The tomato identification and detection under the natural environment is to identify target tomatoes in a complex environment by utilizing a computer vision technology, and transmit the obtained position information to a mechanical arm of a tomato picking robot, so that the subsequent tomato picking work can be accurately performed.
The traditional tomato recognition methods in the greenhouse environment are all based on characteristic information of colors or shapes to extract and classify, including histograms of colors or threshold segmentation, classifier classification based on a support vector machine and the like, but the methods do not consider environmental influence factors under the natural complex condition, and are difficult to meet actual requirements.
The convolutional neural network based on the deep learning provides a new method for object recognition, the detection method based on the deep learning can be divided into two types, namely an R-CNN system algorithm based on Region Propos sal and one-stage algorithm such as Yolo and SSD, and the types and positions of different targets can be directly predicted by using only one CNN network.
The tomato can be effectively identified through the deep learning algorithms such as SSD and YOLO, and the tomato identification can be performed on complex natural conditions which cannot be acted by the traditional image algorithm after learning and training of a large number of samples.
In summary, a novel method is provided for tomato identification in natural environment through the SSD network integrated with the cavity convolution.
Disclosure of Invention
The invention aims to provide an SSD convolution neural network fused with cavity convolution, which improves the target recognition rate of a tomato picking robot under complex conditions. A tomato identification and positioning method is provided to solve the problem that a target tomato is difficult to identify in a complex environment.
The technical scheme adopted by the invention is as follows:
a tomato identification method based on deep learning comprises the following steps:
s1: collecting color images of tomatoes under outdoor natural illumination by using a color camera, and constructing a training sample;
s2: carrying out data enhancement on the color image acquired in the step S1 to form a training set;
s3: labeling the tomato position in each sample in the training set in the S2 to obtain a labeling file comprising real frame coordinate parameters;
s4: improving an SSD network architecture, respectively replacing 5 layers of the max pooling layer in the VGG16 network with hole convolution layers, setting the hole convolution layers to be 2 times of downsampling, respectively replacing FC6 layers and FC7 layers in the SSD network with hole convolution layers, and simultaneously respectively replacing conv8_2 layers and conv9_2 layers with hole convolution layers to obtain the improved SSD network architecture;
s5: training the improved SSD network architecture by using the training set, so that the tomato position in the color image can be identified;
s6: and (3) inputting the color image to be detected into the SSD network after training in the step (S5), marking the boundary box with the score larger than the threshold value, and selecting the tomato region as the box selection region.
Preferably, in the step S1, tomato images under various complex conditions need to be collected, and the training samples include different illumination, angles, changes of fruit sizes, single fruits, a plurality of fruits, and situations that the fruits are covered by tomato leaves.
Preferably, the data enhancement includes flipping, panning, cropping, color dithering, and noise enhancement.
Preferably, in S3, the tomato pictures in the training set are manually marked by using a marking file provided by a kit Labelimg.
Preferably, in S4, the downsampling of the 5 layers of the maximum pooling layer in the original VGG16 network is replaced by downsampling of 2 times by 5 layers of the hole convolution layers of which the layers are 2 layers, 3 layers, 6 layers, 9 layers and 14 layers, so as to form 4 residual blocks.
Preferably, in S4, the convolution kernel size of the two-layer hole convolution layer used for replacing FC6 and FC7 is 3×3, the padding is set to 0, the step size is set to 1, the number of intervals between the points of the convolution kernel is 1, and the final receptive field size is 7×7.
Preferably, in S4, the convolution kernel size of the hole convolution layer for replacing the conv8_2 layer is 3×3, and the hole ratio is 2; the convolution kernel size of the hole convolution layer for replacing the conv9_2 layer is 3×3, and the hole rate is 4.
Preferably, in the step S5, frame parameters are required to be configured in advance when training the improved SSD network architecture, the detection type is set to be 1, the matched threshold is set to be 0.5, and the number of training steps is set to be 20000.
According to the invention, all pooling layers are removed from a VGG network frame in an SSD network architecture, a residual network is changed, and a convolution layer of the residual network uses hole convolution. Because the size of the low-level feature map used in the SSD algorithm is larger, the size of the high-level feature map is smaller, the size of the bottom-level feature map is reduced by utilizing the downsampling operation, and the cavity convolution replaces the maximum pooling layer, so that the feeling of the feature map is enlarged, and more effective global information is obtained by the feature map elements. Therefore, the method for introducing the cavity convolution layer in the image recognition stage can improve the tomato recognition accuracy in a complex environment, and is beneficial to improving the overall working efficiency of the tomato picking robot.
Drawings
Fig. 1 is a flow chart of recognition of tomatoes based on deep learning.
Fig. 2 is a modified view of an SSD network frame.
Fig. 3 is a residual block network diagram.
FIG. 4 is a graph showing the detection results in one embodiment.
Detailed Description
The invention will be further described with reference to the accompanying drawings and specific examples.
Referring to fig. 1, a basic operation flowchart of a tomato recognition method based on deep learning according to a preferred embodiment of the present invention is shown, wherein the method specifically comprises the following steps:
step 1, image acquisition
And collecting color images of the tomatoes under outdoor natural illumination by using a color camera, and constructing a training sample. When the training samples are acquired, the image samples under different conditions should be selected as much as possible under the natural illumination condition so as to enrich the sample types. The method can select different times of the day and different illumination degrees, collect images, select different illumination, angles, changes of fruit sizes, single fruits, a plurality of fruits, and the fruits are shielded by tomato leaves to collect images as much as possible.
Step 2, image data enhancement
And (3) carrying out data enhancement on the color image acquired in the step (1) to form a training set. The data enhancement is a method capable of improving algorithm robustness under the condition of not reducing detection precision, and the collected image data of tomatoes is often too small in sample under natural illumination, and the method comprising horizontal overturning, translation, shaking and noise addition is adopted for enhancing the image data.
Step 3, sample labeling
And (3) manually labeling all the enhanced image samples in the step (2), and labeling 4-tuple parameters (Xmin, ymin, xmax, ymax) containing the region of interest (namely the tomato position), wherein the 4-tuple parameters respectively represent the upper left corner coordinates (Xmin, ymax) and the lower right corner coordinates (Xmax, ymin) of the labeling frame, so that a labeling file comprising real frame coordinate parameters is obtained. The annotation file is used for judging the matching score according to the coincidence rate of the output frame and the annotation frame when evaluating the accuracy of the model. In this embodiment, the tomato pictures in the training set are manually marked by the marking file provided by the tool box Labelimg.
Step 4, improved SSD network architecture construction
The conventional SSD object detection framework is based on a feedforward neural network, the structure of which is described in the prior art in Liu W, anguelov D, erhan D, et al SSD Single Shot MultiBox Detector [ C ]// European Conference on Computer Vision Springer International Publishing,2016.
The SSD frame uses the VGG16 of the full convolution layer as a base network to directly predict multi-destination categories and peripheral frames with feature maps, but in the present invention improvements to the SSD network architecture are needed so that it can accurately identify tomatoes in an image. In this embodiment, the basic structure of the SSD network is unchanged, and only a part of layers in the SSD network are replaced by hole convolution layers, which is specifically as follows:
in one aspect, 5 layers of the max-pooling layer in the VGG16 network are replaced with the hole convolution layers, respectively, and the hole convolution layers are set to 2 times downsampling. On the other hand, the FC6 layer and the FC7 layer in the SSD network are replaced with the hole convolution layers, respectively, while the conv8_2 layer and the conv9_2 layer are replaced with the hole convolution layers, respectively. After the replacement, an improved SSD network architecture is obtained, which can be seen in fig. 2.
The specific parameters of the cavity convolution layer can be adjusted according to the needs. In this embodiment, the number of layers of the hole convolution layer for replacing the 5-layer maximum pooling layer in the VGG16 network is 2,3,6,9, and 14 in order. In the improved SSD network frame, 5 layers of maximum pooling layers are removed, 2 times downsampling is carried out by using 2,3,6,9 and 14 layers of cavity convolution layers to replace downsampling of the 5 layers of maximum pooling layers in the original VGG16 network, a depth residual network with 4 residual blocks is formed, and the structure is that 'building block' reduces convolution parameters needing to be calculated. The residual network structure can be seen in fig. 3.
In addition, the convolution kernel size of the two-layer hole convolution layers for replacing FC6 and FC7 is 3×3, the padding is set to 0, the step size is set to 1, the number of intervals of the points of the convolution kernel is set to 1, and thus the final receptive field size is 7×7. The convolution kernel size of the hole convolution layer for replacing the conv8_2 layer is 3×3, and the hole rate is set to 2; the convolution kernel size of the hole convolution layer for replacing the conv9_2 layer is 3×3, and the hole rate is set to 4.
In the invention, the SSD network combined with the cavity convolution is adopted, so that the information loss of the pooling operation can be avoided, and under the condition of the same calculation condition, the receptive field is enlarged, and each convolution output contains a larger range of information. The feature map size calculation method obtained after cavity convolution is as follows:
wherein i is the size of the input feature map; p is the padding value size; k is the convolution kernel size; d is an expansion factor, and the hole convolution operation is completed on the elements with d-l intervals; s is the step size.
In the improved SSD network, the form of the loss function is substantially similar to a traditional SSD network. Generating a set of bounding boxes of fixed size and confidence in the object class in the boxes in the multi-layer feature map:
wherein c is the confidence of the Softmax function on each category, N is the number of matching default frames, the weight term alpha is set to 1 through cross validation, x is defined as the default frame matching value, l is defined as the predicted frame, and g is defined as the real label value; l (L) conf Defined as confidence loss, L loc Defined as a loss of position.
Wherein the position loss is a smooth L1 loss between the prediction box (L) and the real label box (g), as shown in formula (2):
pos is defined as a default box in this formula; the box is defined as the central coordinate of the prediction frame and the width and height of the prediction frame; x is x ij Defined as the matching value of the i default box to the j real label box of class k.
In this embodiment, the SSD network extracts 4 or 6 default frames according to different aspect ratios for each unit of the conv4_3, fc7, conv8_2, conv9_2, conv10_2 and conv11_2 feature layers, and finally obtains 8732 default frames.
Step 5, model training
After the improved SSD network architecture is built, training the improved SSD network architecture by utilizing the training set and the labeling file obtained in the steps, so that the tomato position in the input color image can be identified. When training the network, frame parameters are required to be configured in advance, and in the embodiment, only tomatoes need to be detected, so that the detection type can be set to be 1, the matched threshold value is 0.5, and the training step number is 20000.
In the training stage, a formula (1) is used for carrying out regression analysis of coordinate offset for a predicted frame, each default frame in the testing stage is matched with a labeling frame in an overlapping rate, and the matching scores are ranked from high to low.
Because the size of the low-level feature map used in the SSD algorithm is larger, the size of the high-level feature map is smaller, the size of the bottom-level feature map is reduced by utilizing the downsampling operation, the cavity convolution replaces the maximum pooling layer, the receptive field of the feature map is enlarged, and more effective global information is obtained by the feature map elements.
Step 6, tomato identification
And inputting the color image to be detected into the trained SSD network, marking the boundary box with the score being more than 50% of the threshold value, and selecting the tomato region as the box selection region. In this embodiment, the test set is input into the network model, the detection result is obviously better than that of the traditional SSD network model, and fig. 4 shows a tomato bounding box detected in one of the pictures, so that it can be seen that the position of the tomato is accurately identified in general.
Therefore, the convolution layer is used for downsampling instead of the maximum pooling layer, more characteristic information can be reserved, and the characteristic multiplexing and fusion of the front layer and the rear layer are realized by introducing a residual structure into the main network, so that the rapidness and the accuracy of extracting the characteristics of the tomato fruits are simultaneously met.
The above embodiment is only a preferred embodiment of the present invention, but it is not intended to limit the present invention. Various changes and modifications may be made by one of ordinary skill in the pertinent art without departing from the spirit and scope of the present invention. Therefore, all the technical schemes obtained by adopting the equivalent substitution or equivalent transformation are within the protection scope of the invention.

Claims (6)

1. The tomato identification method based on deep learning is characterized by comprising the following steps:
s1: collecting color images of tomatoes under outdoor natural illumination by using a color camera, and constructing a training sample;
s2: carrying out data enhancement on the color image acquired in the step S1 to form a training set;
s3: labeling the tomato position in each sample in the training set in the S2 to obtain a labeling file comprising real frame coordinate parameters;
s4: improving an SSD network architecture, respectively replacing 5 layers of the max pooling layer in the VGG16 network with hole convolution layers, setting the hole convolution layers to be 2 times of downsampling, respectively replacing FC6 layers and FC7 layers in the SSD network with hole convolution layers, and simultaneously respectively replacing conv8_2 layers and conv9_2 layers with hole convolution layers to obtain the improved SSD network architecture;
s5: training the improved SSD network architecture by using the training set, so that the tomato position in the color image can be identified;
s6: inputting the color image to be detected into the SSD network trained in the step S5, marking a boundary box with a score larger than a threshold value, wherein a box selection area is a detected tomato area;
in the step S4, the downsampling of 5 layers of maximum pooling layers in the original VGG16 network is replaced by downsampling of 2 times of 5 layers of cavity convolution layers with the layers of 2 layers, 3 layers, 6 layers, 9 layers and 14 layers respectively to form 4 residual blocks;
in the step S4, the convolution kernel size of the cavity convolution layer for replacing the conv8_2 layer is 3×3, and the cavity rate is 2; the convolution kernel size of the hole convolution layer for replacing the conv9_2 layer is 3×3, and the hole rate is 4.
2. A deep learning-based tomato recognition method as claimed in claim 1, wherein in S1, a plurality of complex tomato images are acquired, and the training samples include the changes of different illumination, angles and fruit sizes, single fruit, a plurality of fruits, and the situation that the fruits are covered by tomato leaves.
3. A deep learning based tomato recognition method as claimed in claim 1, wherein the data enhancement includes flipping, panning, clipping, color dithering and noise enhancement.
4. The method for recognizing tomatoes based on deep learning as claimed in claim 1, wherein in S3, the tomato pictures in the training set are manually marked by using a marking file provided by a kit Labelimg.
5. A tomato recognition method based on deep learning as claimed in claim 1, wherein in S4, the convolution kernel size of the two-layer hole convolution layer for replacing FC6 and FC7 is 3 x 3, the padding is set to 0, the step size is set to 1, the number of intervals of the points of the convolution kernel is 1, and the final receptive field size is 7 x 7.
6. The tomato recognition method based on deep learning as claimed in claim 1, wherein in the step S5, frame parameters are required to be configured in advance when training the improved SSD network architecture, the detection type is set to be 1, the matched threshold is 0.5, and the training step number is 20000.
CN202011169184.6A 2020-10-28 2020-10-28 Tomato identification method based on deep learning Active CN112232263B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011169184.6A CN112232263B (en) 2020-10-28 2020-10-28 Tomato identification method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011169184.6A CN112232263B (en) 2020-10-28 2020-10-28 Tomato identification method based on deep learning

Publications (2)

Publication Number Publication Date
CN112232263A CN112232263A (en) 2021-01-15
CN112232263B true CN112232263B (en) 2024-03-19

Family

ID=74109139

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011169184.6A Active CN112232263B (en) 2020-10-28 2020-10-28 Tomato identification method based on deep learning

Country Status (1)

Country Link
CN (1) CN112232263B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112836623B (en) * 2021-01-29 2024-04-16 北京农业智能装备技术研究中心 Auxiliary method and device for agricultural decision of facility tomatoes
CN113076819A (en) * 2021-03-17 2021-07-06 山东师范大学 Fruit identification method and device under homochromatic background and fruit picking robot
CN113326808A (en) * 2021-06-26 2021-08-31 西北农林科技大学 Night tomato identification system and method based on improved yolo
CN115424247B (en) * 2022-06-24 2023-04-25 中国农业科学院农业信息研究所 Greenhouse tomato identification and detection method adopting CBAM and octave convolution to improve YOLOV5
CN116071423A (en) * 2023-02-09 2023-05-05 哈尔滨市科佳通用机电股份有限公司 Method, system and medium for positioning railway wagon brake adjuster component

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109684967A (en) * 2018-12-17 2019-04-26 东北农业大学 A kind of soybean plant strain stem pod recognition methods based on SSD convolutional network
CN109784190A (en) * 2018-12-19 2019-05-21 华东理工大学 A kind of automatic Pilot scene common-denominator target Detection and Extraction method based on deep learning
CN110991511A (en) * 2019-11-26 2020-04-10 中原工学院 Sunflower crop seed sorting method based on deep convolutional neural network
CN111079822A (en) * 2019-12-12 2020-04-28 哈尔滨市科佳通用机电股份有限公司 Method for identifying dislocation fault image of middle rubber and upper and lower plates of axle box rubber pad
CN111626993A (en) * 2020-05-07 2020-09-04 武汉科技大学 Image automatic detection counting method and system based on embedded FEFnet network
CN111652869A (en) * 2020-06-02 2020-09-11 中冶赛迪重庆信息技术有限公司 Slab void identification method, system, medium and terminal based on deep learning

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109684967A (en) * 2018-12-17 2019-04-26 东北农业大学 A kind of soybean plant strain stem pod recognition methods based on SSD convolutional network
CN109784190A (en) * 2018-12-19 2019-05-21 华东理工大学 A kind of automatic Pilot scene common-denominator target Detection and Extraction method based on deep learning
CN110991511A (en) * 2019-11-26 2020-04-10 中原工学院 Sunflower crop seed sorting method based on deep convolutional neural network
CN111079822A (en) * 2019-12-12 2020-04-28 哈尔滨市科佳通用机电股份有限公司 Method for identifying dislocation fault image of middle rubber and upper and lower plates of axle box rubber pad
CN111626993A (en) * 2020-05-07 2020-09-04 武汉科技大学 Image automatic detection counting method and system based on embedded FEFnet network
CN111652869A (en) * 2020-06-02 2020-09-11 中冶赛迪重庆信息技术有限公司 Slab void identification method, system, medium and terminal based on deep learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
MULTI-SCALE CONTEXT AGGREGATION BY DILATED CONVOLUTIONS;Fisher Yu 等;arXiv;全文 *
改进SSD 卷积神经网络在钢板边裂纹检测中的应用;刘继丹 等;冶金自动化;第44卷(第4期);第43-47,80页 *

Also Published As

Publication number Publication date
CN112232263A (en) 2021-01-15

Similar Documents

Publication Publication Date Title
CN112232263B (en) Tomato identification method based on deep learning
Jia et al. Detection and segmentation of overlapped fruits based on optimized mask R-CNN application in apple harvesting robot
CN111126325B (en) Intelligent personnel security identification statistical method based on video
CN109800628A (en) A kind of network structure and detection method for reinforcing SSD Small object pedestrian detection performance
CN111275082A (en) Indoor object target detection method based on improved end-to-end neural network
CN110298291A (en) Ox face and ox face critical point detection method based on Mask-RCNN
CN110853015A (en) Aluminum profile defect detection method based on improved Faster-RCNN
CN111340141A (en) Crop seedling and weed detection method and system based on deep learning
CN111862119A (en) Semantic information extraction method based on Mask-RCNN
CN110927171A (en) Bearing roller chamfer surface defect detection method based on machine vision
CN111563452A (en) Multi-human body posture detection and state discrimination method based on example segmentation
CN112464766A (en) Farmland automatic identification method and system
CN113191334B (en) Plant canopy dense leaf counting method based on improved CenterNet
CN114140665A (en) Dense small target detection method based on improved YOLOv5
CN111178177A (en) Cucumber disease identification method based on convolutional neural network
CN110929746A (en) Electronic file title positioning, extracting and classifying method based on deep neural network
CN115035082B (en) Method for detecting defects of transparent parts of aircraft based on YOLOv4 improved algorithm
CN116052222A (en) Cattle face recognition method for naturally collecting cattle face image
CN112465820A (en) Semantic segmentation based rice disease detection method integrating global context information
CN113610035A (en) Rice tillering stage weed segmentation and identification method based on improved coding and decoding network
CN112396655A (en) Point cloud data-based ship target 6D pose estimation method
CN115205626A (en) Data enhancement method applied to field of coating defect detection
CN109615610B (en) Medical band-aid flaw detection method based on YOLO v2-tiny
CN116524344A (en) Tomato string picking point detection method based on RGB-D information fusion
CN112686872A (en) Wood counting method based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant