CN112036437A - Rice seedling detection model based on improved YOLOV3 network and method thereof - Google Patents

Rice seedling detection model based on improved YOLOV3 network and method thereof Download PDF

Info

Publication number
CN112036437A
CN112036437A CN202010736191.3A CN202010736191A CN112036437A CN 112036437 A CN112036437 A CN 112036437A CN 202010736191 A CN202010736191 A CN 202010736191A CN 112036437 A CN112036437 A CN 112036437A
Authority
CN
China
Prior art keywords
scale
seedling
feature
fusion
feature extraction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010736191.3A
Other languages
Chinese (zh)
Other versions
CN112036437B (en
Inventor
王姗姗
余山山
张文毅
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Research Institute for Agricultural Mechanization Ministry of Agriculture
Original Assignee
Nanjing Research Institute for Agricultural Mechanization Ministry of Agriculture
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Research Institute for Agricultural Mechanization Ministry of Agriculture filed Critical Nanjing Research Institute for Agricultural Mechanization Ministry of Agriculture
Priority to CN202010736191.3A priority Critical patent/CN112036437B/en
Publication of CN112036437A publication Critical patent/CN112036437A/en
Application granted granted Critical
Publication of CN112036437B publication Critical patent/CN112036437B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a rice seedling detection model based on an improved YOLOV3 network and a method thereof, which solve the problems of poor adaptability and robustness of the traditional image processing method to the rice seedling detection and poor image processing algorithm effect; the technical scheme is characterized by comprising a feature extraction module for performing multi-scale feature extraction on an input rice seedling image and a multi-scale prediction module for predicting the position of the rice seedling according to a feature map of the multi-scale feature extraction; the invention discloses a rice seedling detection model based on an improved YOLOV3 network and a method thereof, and solves the problems of poor robustness and accuracy of rice seedling detection in complex paddy field environments, such as high and low weed density, different illumination conditions and seedling loss.

Description

Rice seedling detection model based on improved YOLOV3 network and method thereof
Technical Field
The invention relates to the technical field of artificial intelligence, in particular to a rice seedling detection model based on an improved YOLOV3 network and a method thereof.
Background
Rice is one of three major food crops in the world, the cultivation area and the total yield of the rice are second to wheat, and the rice is staple food for more than half of the population in the world. The weeding mode that paddy field adopted at present mainly is to spray chemical agent and artifical weeding, and chemical agent's use causes easily that the pesticide residue volume exceeds standard, environmental pollution and ecological chain's destruction scheduling problem, and artifical weeding is consuming time hard, inefficiency and with high costs. Therefore, the demand for mechanized weeding of rice is more and more urgent. The mechanized weeding technology for the rice has great significance for reducing the use of chemical agents, reducing the pollution of ecological environment and improving the grain safety.
The rice seedling detection is one of key links of rice mechanized weeding, and has important guiding function for accurate weeding and fertilization in accurate agriculture. The position information of the rice seedlings is obtained through the detection of the rice seedlings, and the inter-row weeds before the row closing of the rice can be removed by guiding the action of the weeding mechanism.
Crop and weed detection is mainly classified into 2 methods, the first method is based on image processing, the crop is segmented from the background, and then the crop row center line detection is carried out. The second category is crop and weed detection based on deep learning methods. Because the paddy field environment is very complex, the detection influence factors of the rice seedlings are more, and the traditional image processing method has poor adaptability and robustness to the detection of the rice seedlings. Especially when the paddy field image has insignificant variation in R, G and B pixel intensities, the conventional image processing algorithm is not effective when there are many weeds or similar to the size of crops.
Disclosure of Invention
The invention aims to provide a rice seedling detection model based on an improved YOLOV3 network and a method thereof, which solve the problem of poor robustness and accuracy of rice seedling detection under various conditions in a complex paddy field environment.
The technical purpose of the invention is realized by the following technical scheme:
a rice seedling detection model based on an improved YOLOV3 network comprises a feature extraction module for performing multi-scale feature extraction on an input rice seedling image to obtain a multi-scale seedling feature map, and a multi-scale prediction module for predicting the position of a rice seedling according to the multi-scale seedling feature map;
the multi-scale prediction module comprises a multi-scale fusion characteristic construction module for performing fusion construction processing on the multi-scale seedling characteristic diagram to obtain a multi-scale fusion characteristic diagram, and a multi-scale seedling position prediction module for predicting the seedling position according to the corresponding fusion characteristic in the fusion characteristic diagram.
Preferably, the feature extraction module comprises a 52-scale feature extraction submodule, a 26-scale feature extraction submodule and a 13-scale feature extraction submodule which are used for sequentially carrying out magnification sampling extraction processing on the input rice seedling image so as to respectively obtain a 52-scale seedling feature map, a 26-scale seedling feature map and a 13-scale seedling feature map.
Preferably, the multi-scale fusion feature construction module comprises a 52-scale fusion feature construction submodule for performing fusion processing on seedling feature maps with different scales, a 26-scale fusion feature construction submodule and a 13-scale fusion feature construction submodule.
Preferably, the 52-scale feature extraction sub-module performs feature extraction on the input seedling feature map of the 52-scale feature extraction sub-module to obtain a 52-scale seedling feature map containing 52-scale seedling features; the 26-scale feature extraction sub-module performs feature extraction on the input seedling feature map of the 26-scale feature extraction sub-module to obtain a 26-scale seedling feature map containing 26-scale seedling features; and the 13-scale feature extraction sub-module performs feature extraction on the input seedling feature map of the 13-scale feature extraction sub-module to obtain a 13-scale seedling feature map containing 13-scale seedling features.
A detection method of a rice seedling detection model based on an improved YOLOV3 network comprises the following steps:
acquiring an input rice seedling image;
respectively carrying out down-sampling with different multiplying powers on the rice seedling images, and obtaining a 52-scale seedling characteristic diagram, a 26-scale seedling characteristic diagram and a 13-scale seedling characteristic diagram through a characteristic extraction module;
sequentially performing feature fusion on the obtained 13-scale seedling feature map, 26-scale seedling feature map and 52-scale seedling feature map through a multi-scale fusion feature construction module to obtain a 13-scale seedling fusion feature map, a 26-scale seedling fusion feature map and a 52-scale seedling fusion feature map;
and predicting the obtained 13-scale seedling fusion characteristic graph, 26-scale seedling fusion characteristic graph and 52-scale seedling fusion characteristic graph in sequence through a multi-scale seedling position prediction module to obtain a 13-scale seedling prediction result, a 26-scale seedling prediction result and a 52-scale seedling prediction result.
Preferably, the specific steps of the feature extraction module for feature extraction are as follows:
the input convolution layer of the 52-scale feature extraction submodule is obtained by passing the input rice seedling graph through 10 convolution layers and 3 residual error layers, and 52-scale seedling features are obtained by extraction of the 52-scale feature extraction submodule;
convolving the output layer of the 52-scale feature extraction submodule by a convolution kernel of 3 × 3/2 × 512 to obtain an input convolution layer of the 26-scale feature extraction submodule, and extracting by the 26-scale feature extraction submodule to obtain 26-scale seedling features;
and (3) convolving the output layer of the 26-scale feature extraction submodule by a convolution kernel of 3 × 3/2 × 512 to obtain an input convolution layer of the 13-scale feature extraction submodule, and extracting by the 13-scale feature extraction submodule to obtain 13-scale seedling features.
Preferably, the multi-scale fusion feature construction module performs feature fusion construction by the following specific steps:
carrying out down-sampling on the output layer of the 26-scale feature extraction sub-module to obtain a 13 x 512 feature map, and carrying out tensor splicing on the feature map and the output layer of the 13-scale feature extraction sub-module to obtain a 13-scale seedling fusion feature map;
carrying out down-sampling on an output layer of the 52-scale feature extraction sub-module to obtain 26 × 256 feature maps, carrying out up-sampling on an output layer of the 13-scale feature extraction sub-module to obtain 26 × 512 feature maps, and carrying out tensor splicing on the 26 × 256 feature maps obtained by down-sampling, the 26 × 512 feature maps obtained by up-sampling and the output layer of the 26-scale feature extraction sub-module to obtain 26-scale seedling fusion feature maps;
and carrying out down-sampling on 104-scale output seedling feature maps obtained by carrying out feature extraction on the input rice seedling pictures to obtain 52 x 128 feature maps, carrying out up-sampling on output layers of the 26-scale feature extraction sub-modules to obtain 52 x 512 feature maps, and carrying out tensor splicing on the 52 x 128 feature maps obtained by down-sampling, the 52 x 512 feature maps obtained by up-sampling and the output layers of the 52-scale feature extraction sub-modules to obtain 52-scale seedling fusion feature maps.
Preferably, the specific prediction steps of the multi-scale seedling position prediction model are as follows:
the 13-scale seedling fusion features constructed by the 13-scale fusion feature construction submodule are subjected to convolution of a set of convolution kernels, 13 × 512 convolution kernels and 1 × 18 convolution kernel to obtain a 13-scale prediction result 13 × 18;
the 26-scale seedling fusion features constructed by the 26-scale fusion feature construction submodule are subjected to convolution of a set of convolution kernels, 13 × 512 convolution kernels and 1 × 18 convolution kernel to obtain a 26-scale prediction result 26 × 18;
the 52-scale seedling fusion features constructed by the 52-scale fusion feature construction submodule are subjected to convolution of a set of convolution kernels, 13 × 256 convolution kernel and 1 × 18 convolution kernel to obtain a 52-scale prediction result 52 × 18.
In conclusion, the invention has the following beneficial effects:
the rice seedling detection model based on the improved Yolov3 network can automatically extract rice seedling characteristics, the model adopts a 52-scale characteristic extraction submodule to replace an original residual error unit to widen the network on a 52-scale characteristic graph close to the rice seedling size in an original characteristic extraction network, sampling characteristics with different scales are added when fusion characteristics of all scales are constructed, the combination of position information of a low layer and semantic information of a high layer is added to the fused characteristics, the obtained rice seedling fusion characteristic levels of all scales are richer, and the improved rice seedling detection model has better accuracy and robustness on high and low weed density distribution in a complex paddy field environment, different illumination conditions and rice seedling detection under the condition of rice seedling loss.
Drawings
FIG. 1 is a block diagram schematically illustrating the structure of a detection model;
FIG. 2 is a network architecture diagram of a detection model;
FIG. 3 is a schematic flow chart of the detection method of the present invention;
FIG. 4 is a test p-r plot based on the modified Yolov3 and Yolov3 rice seedling detection models.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings.
According to one or more embodiments, a rice seedling detection model based on an improved YOLOV3 network is disclosed, as shown in fig. 1, which includes a feature extraction module for performing multi-scale rice seedling feature extraction on an input rice seedling image to obtain a multi-scale seedling feature map, and a multi-scale prediction module for predicting the position of a rice seedling according to the multi-scale seedling feature map extracted by the feature extraction module.
As shown in fig. 1, the feature extraction module includes a 52-scale feature extraction submodule, a 26-scale feature extraction submodule, and a 13-scale feature extraction submodule, which sequentially perform magnification sampling extraction processing on an input rice seedling image. For clarity, the rice seedling image input by way of example is an image of 416 × 416, and when sampling extraction is performed, 8-time, 16-time and 32-time down-sampling are performed respectively.
The multi-scale prediction module comprises a multi-scale fusion characteristic construction module for performing fusion construction processing on seedling characteristic graphs of different scales extracted by the characteristic extraction module to obtain a multi-scale fusion characteristic graph, and a multi-scale seedling position prediction module for predicting the seedling position according to the multi-scale fusion characteristic in the fusion characteristic graph. The multi-scale fusion feature construction module specifically comprises a 52-scale fusion feature construction submodule, a 26-scale fusion feature construction submodule and a 13-scale fusion feature construction submodule, and the multi-scale seedling position prediction module specifically comprises a 52-scale position prediction submodule, a 26-scale position prediction submodule and a 13-scale position prediction submodule.
The input seedling characteristic diagram extracted by the 52-scale characteristic extraction submodule specifically comprises: and (3) passing the input rice seedling image through 10 convolution layers and 3 residual layers to obtain an input convolution layer of the 52-scale feature extraction submodule as an input rice seedling feature map of the 52-scale feature extraction submodule. And the 52-scale feature extraction sub-module performs feature extraction on the input seedling feature map to obtain a 52-scale seedling feature map containing 52-scale seedling features.
The input seedling characteristic diagram extracted by the 26-scale characteristic extraction submodule specifically comprises: and (3) convolving the output layer of the 52-scale feature extraction submodule, namely the output 52-scale seedling feature map, by the convolution kernel of 3 × 3/2 × 512 to obtain the input convolution layer of the 26-scale feature extraction submodule, so as to obtain the input seedling feature map of the 26-scale feature extraction submodule. And the 26-scale feature extraction sub-module performs feature extraction on the input seedling feature map to obtain a 26-scale seedling feature map containing 26-scale seedling features.
The input seedling characteristic diagram extracted by the 13-scale characteristic extraction submodule specifically comprises: and (3) convolving the output layer of the 26-scale feature extraction submodule, namely the output 26-scale seedling feature map, by using a convolution kernel of 3 × 3/2 × 512 to obtain an input convolution layer of the 13-scale feature extraction submodule, so as to obtain the input seedling feature map of the 13-scale feature extraction submodule. And extracting a 13-scale seedling feature map containing 13-scale seedling features by using a 13-scale feature extraction submodule. And the feature extraction modules of the three scales sequentially extract.
The feature map constructed by the fusion of the 13-scale fusion feature construction submodule comprises a 13-scale seedling feature map obtained by extraction of an output layer of the 13-scale feature extraction submodule, a 13 x 512 feature map obtained by down-sampling of a 26-scale seedling feature map obtained by extraction of an output layer of the 26-scale feature extraction submodule, and the 13-scale seedling feature map and the 13 x 512 feature map obtained by tensor splicing are subjected to tensor splicing to obtain the 13-scale seedling fusion feature map.
Similarly, the seedling feature map constructed by the 26-scale fusion feature construction submodule in a fusion mode comprises an output layer of the 26-scale feature extraction submodule, the output layer of the 52-scale feature extraction submodule, namely the 52-scale seedling feature map obtained by extraction, is subjected to down-sampling to obtain 26 x 256 feature maps, the output layer of the 13-scale feature extraction submodule is subjected to up-sampling to obtain 26 x 512 feature maps, and the 26 x 256 feature maps obtained by down-sampling, the 26 x 512 feature maps obtained by up-sampling and the output layer of the 26-scale feature extraction submodule are spliced to obtain the 26-scale seedling fusion feature map.
Similarly, the seedling feature map constructed by the fusion of the 52-scale fusion feature construction submodule comprises an output layer of the 52-scale feature extraction submodule, a 104-scale output seedling feature map obtained by performing feature extraction on the input rice seedling picture is subjected to down sampling to obtain a 52 x 128 feature map, an output layer of the 26-scale feature extraction submodule is subjected to up sampling to obtain a 52 x 512 feature map, and the 52 x 128 feature map obtained by down sampling, a 52 x 512 feature map obtained by up sampling and an output layer of the 52-scale feature extraction submodule are spliced to obtain a 52-scale seedling fusion feature map.
The 13-scale location prediction sub-module of the multi-scale seedling location prediction module obtains 13-scale prediction results 13 x 18 by convolving the 13-scale seedling fusion features with 13 x 512 convolution kernel and 1 x 18 convolution kernel through a set of convolution kernels.
The 26-scale location prediction sub-module obtains 26 x 18 prediction results at 26 scales by convolving the 26-scale seedling fusion features with 13 x 512 convolution kernel and 1 x 18 convolution kernel.
The 52-scale location prediction sub-module yields 52-scale prediction results 52 x 18 by convolving the 52-scale seedling fusion features with 13 x 256 convolution kernel and 1 x 18 convolution kernel.
The original Yolov3 network model adopts a mode of sampling on a small-scale feature map and fusing the small-scale feature map and the scale feature map when each scale feature is fused, the invention adopts a mode of sampling on the small-scale feature map, sampling on a large-scale feature map and fusing the scale feature map based on the improved Yolov3 model, the fused features increase the combination of the position information of a lower layer and the semantic information of a higher layer, and the obtained rice seedling fusion feature levels of all scales are richer.
According to one or more embodiments, a method for detecting a rice seedling detection model based on an improved YOLOV3 network is disclosed, as shown in fig. 2 and 3, and comprises the following steps:
acquiring an input rice seedling image;
respectively carrying out down-sampling with different multiplying powers on the rice seedling images, and obtaining a 52-scale seedling characteristic diagram, a 26-scale seedling characteristic diagram and a 13-scale seedling characteristic diagram through a characteristic extraction module;
sequentially performing feature fusion on the obtained 13-scale seedling feature map, 26-scale seedling feature map and 52-scale seedling feature map through a multi-scale fusion feature construction module to obtain a 13-scale seedling fusion feature map, a 26-scale seedling fusion feature map and a 52-scale seedling fusion feature map;
and predicting the obtained 13-scale seedling fusion characteristic graph, 26-scale seedling fusion characteristic graph and 52-scale seedling fusion characteristic graph in sequence through a multi-scale seedling position prediction module to obtain a 13-scale seedling prediction result, a 26-scale seedling prediction result and a 52-scale seedling prediction result.
Specifically, the specific steps of the feature extraction module for feature extraction are as follows:
the input convolution layer of the 52-scale feature extraction submodule is obtained by passing the input rice seedling graph through 10 convolution layers and 3 residual error layers, and 52-scale seedling features are obtained by extraction of the 52-scale feature extraction submodule;
convolving the output layer of the 52-scale feature extraction submodule by a convolution kernel of 3 × 3/2 × 512 to obtain an input convolution layer of the 26-scale feature extraction submodule, and extracting by the 26-scale feature extraction submodule to obtain 26-scale seedling features;
and (3) convolving the output layer of the 26-scale feature extraction submodule by a convolution kernel of 3 × 3/2 × 512 to obtain an input convolution layer of the 13-scale feature extraction submodule, and extracting by the 13-scale feature extraction submodule to obtain 13-scale seedling features.
Specifically, the multi-scale fusion feature construction module performs feature fusion construction specifically including the following steps:
carrying out down-sampling on the output layer of the 26-scale feature extraction sub-module to obtain a 13 x 512 feature map, and carrying out tensor splicing on the feature map and the output layer of the 13-scale feature extraction sub-module to obtain a 13-scale seedling fusion feature map;
carrying out down-sampling on an output layer of the 52-scale feature extraction sub-module to obtain 26 × 256 feature maps, carrying out up-sampling on an output layer of the 13-scale feature extraction sub-module to obtain 26 × 512 feature maps, and carrying out tensor splicing on the 26 × 256 feature maps obtained by down-sampling, the 26 × 512 feature maps obtained by up-sampling and the output layer of the 26-scale feature extraction sub-module to obtain 26-scale seedling fusion feature maps;
and carrying out down-sampling on 104-scale output seedling feature maps obtained by carrying out feature extraction on the input rice seedling pictures to obtain 52 x 128 feature maps, carrying out up-sampling on output layers of the 26-scale feature extraction sub-modules to obtain 52 x 512 feature maps, and carrying out tensor splicing on the 52 x 128 feature maps obtained by down-sampling, the 52 x 512 feature maps obtained by up-sampling and the output layers of the 52-scale feature extraction sub-modules to obtain 52-scale seedling fusion feature maps.
The 52-scale feature extraction sub-module is an inclusion module, and a three-way convolution broadening network is adopted: the first convolution reduces the number of channels of the input seedling feature map by adopting 1 × 64 convolution kernels, and then convolution extraction features are carried out by adopting 3 × 128 and 3 × 64 convolution kernels; in the second path, 1 × 64 convolution kernels are adopted to reduce the number of channels of the input seedling characteristic graph, and then 3 × 128 convolution kernels are adopted to perform convolution to extract characteristics; the third path reduces the number of paths for inputting the seedling characteristic diagram by 1 x 64 convolution. And carrying out tensor splicing on the 52-scale seedling feature map obtained after the three-way convolution processing by the inclusion module and the 52 x 128 feature map and the 52 x 512 feature map obtained by up/down sampling to complete 52-scale feature fusion. The Inception modules are preferably set to be 4, the original residual error unit is replaced to widen the network, more features can be extracted, and parameters and training time can be effectively reduced.
Further, the specific prediction steps of the multi-scale seedling position prediction model are as follows:
performing convolution on the 13-scale seedling fusion features of the 13-scale seedling fusion feature map through a group of convolution kernels, 1 convolution kernel of 3 × 512 and 1 convolution kernel of 1 × 18 to obtain a prediction result of 13 × 18 of the 13 scales;
performing convolution on the 26-scale seedling fusion features of the 26-scale seedling fusion feature map through a group of convolution kernels, 1 convolution kernel of 3 × 512 and 1 convolution kernel of 1 × 18 to obtain a 26-scale prediction result of 26 × 18;
and (3) performing convolution on the 52-scale seedling fusion features of the 52-scale seedling fusion feature map through a group of convolution kernels, 1 convolution kernel of 3 × 256 and 1 convolution kernel of 1 × 18 to obtain a 52-scale prediction result 52 × 18.
For clarity, taking an example, the rice seedling detection models based on the improved Yolov3 and Yolov3 were tested using a rice seedling test set, and after the test, p-r curves were plotted, as shown in fig. 4(a), (b), (a) being a p-r curve of the improved Yolov3 and (b) being a p-r curve of the Yolov3, when the confidence threshold was 0.5, the accuracies of the rice seedling detection models based on the improved Yolov3 and Yolov3 networks were 0.82 and 0.48, respectively, so that the rice seedling detection algorithm based on the improved Yolov3 was improved in accuracy by about 34%.
The present embodiment is only for explaining the present invention, and it is not limited to the present invention, and those skilled in the art can make modifications of the present embodiment without inventive contribution as needed after reading the present specification, but all of them are protected by patent law within the scope of the claims of the present invention.

Claims (8)

1. A rice seedling detection model based on an improved YOLOV3 network is characterized in that: the system comprises a feature extraction module for performing multi-scale feature extraction on an input rice seedling image to obtain a multi-scale seedling feature map, and a multi-scale prediction module for predicting the position of a rice seedling according to the multi-scale seedling feature map;
the multi-scale prediction module comprises a multi-scale fusion characteristic construction module for performing fusion construction processing on the multi-scale seedling characteristic diagram to obtain a multi-scale fusion characteristic diagram, and a multi-scale seedling position prediction module for predicting the seedling position according to the corresponding fusion characteristic in the fusion characteristic diagram.
2. The improved YOLOV3 network-based rice seedling detection model of claim 1, wherein: the characteristic extraction module comprises a 52-scale characteristic extraction submodule, a 26-scale characteristic extraction submodule and a 13-scale characteristic extraction submodule which are used for sequentially carrying out multiplying factor sampling extraction processing on the input rice seedling image so as to respectively obtain a 52-scale seedling characteristic diagram, a 26-scale seedling characteristic diagram and a 13-scale seedling characteristic diagram.
3. The improved YOLOV3 network-based rice seedling detection model of claim 2, wherein: the multi-scale fusion feature construction module comprises a 52-scale fusion feature construction submodule for performing fusion processing on seedling feature maps with different scales, a 26-scale fusion feature construction submodule and a 13-scale fusion feature construction submodule.
4. The rice seedling detection model based on the improved YOLOV3 network as claimed in claim 3, wherein: the 52-scale feature extraction sub-module performs feature extraction on the input seedling feature map of the 52-scale feature extraction sub-module to obtain a 52-scale seedling feature map containing 52-scale seedling features; the 26-scale feature extraction sub-module performs feature extraction on the input seedling feature map of the 26-scale feature extraction sub-module to obtain a 26-scale seedling feature map containing 26-scale seedling features; and the 13-scale feature extraction sub-module performs feature extraction on the input seedling feature map of the 13-scale feature extraction sub-module to obtain a 13-scale seedling feature map containing 13-scale seedling features.
5. A detection method of a rice seedling detection model based on an improved YOLOV3 network is characterized by comprising the following steps:
acquiring an input rice seedling image;
respectively carrying out down-sampling with different multiplying powers on the rice seedling images, and obtaining a 52-scale seedling characteristic diagram, a 26-scale seedling characteristic diagram and a 13-scale seedling characteristic diagram through a characteristic extraction module;
sequentially performing feature fusion on the obtained 13-scale seedling feature map, 26-scale seedling feature map and 52-scale seedling feature map through a multi-scale fusion feature construction module to obtain a 13-scale seedling fusion feature map, a 26-scale seedling fusion feature map and a 52-scale seedling fusion feature map;
and predicting the obtained 13-scale seedling fusion characteristic graph, 26-scale seedling fusion characteristic graph and 52-scale seedling fusion characteristic graph in sequence through a multi-scale seedling position prediction module to obtain a 13-scale seedling prediction result, a 26-scale seedling prediction result and a 52-scale seedling prediction result.
6. The improved YOLOV3 network-based rice seedling detection method as claimed in claim 5, wherein the feature extraction module performs the following steps:
the input convolution layer of the 52-scale feature extraction submodule is obtained by passing the input rice seedling image through 10 convolution layers and 3 residual error layers, and the 52-scale seedling features are obtained by extraction of the 52-scale feature extraction submodule;
convolving the output layer of the 52-scale feature extraction submodule by a convolution kernel of 3 × 3/2 × 512 to obtain an input convolution layer of the 26-scale feature extraction submodule, and extracting by the 26-scale feature extraction submodule to obtain 26-scale seedling features;
and (3) convolving the output layer of the 26-scale feature extraction submodule by a convolution kernel of 3 × 3/2 × 512 to obtain an input convolution layer of the 13-scale feature extraction submodule, and extracting by the 13-scale feature extraction submodule to obtain 13-scale seedling features.
7. The improved YOLOV3 network-based rice seedling detection method as claimed in claim 6, wherein the multi-scale fusion feature construction module performs the specific steps of feature fusion construction as follows:
carrying out down-sampling on the output layer of the 26-scale feature extraction sub-module to obtain a 13 x 512 feature map, and carrying out tensor splicing on the feature map and the output layer of the 13-scale feature extraction sub-module to obtain a 13-scale seedling fusion feature map;
carrying out down-sampling on an output layer of the 52-scale feature extraction sub-module to obtain 26 × 256 feature maps, carrying out up-sampling on an output layer of the 13-scale feature extraction sub-module to obtain 26 × 512 feature maps, and carrying out tensor splicing on the 26 × 256 feature maps obtained by down-sampling, the 26 × 512 feature maps obtained by up-sampling and the output layer of the 26-scale feature extraction sub-module to obtain 26-scale seedling fusion feature maps;
and carrying out down-sampling on 104-scale output seedling feature maps obtained by carrying out feature extraction on the input rice seedling pictures to obtain 52 x 128 feature maps, carrying out up-sampling on output layers of the 26-scale feature extraction sub-modules to obtain 52 x 512 feature maps, and carrying out tensor splicing on the 52 x 128 feature maps obtained by down-sampling, the 52 x 512 feature maps obtained by up-sampling and the output layers of the 52-scale feature extraction sub-modules to obtain 52-scale seedling fusion feature maps.
8. The improved YOLOV3 network-based rice seedling testing method as claimed in claim 7, wherein the multi-scale seedling position prediction model comprises the following steps:
the 13-scale seedling fusion features constructed by the 13-scale fusion feature construction submodule are subjected to convolution of a set of convolution kernels, 13 × 512 convolution kernels and 1 × 18 convolution kernel to obtain a 13-scale prediction result 13 × 18;
the 26-scale seedling fusion features constructed by the 26-scale fusion feature construction submodule are subjected to convolution of a set of convolution kernels, 13 × 512 convolution kernels and 1 × 18 convolution kernel to obtain a 26-scale prediction result 26 × 18;
the 52-scale seedling fusion features constructed by the 52-scale fusion feature construction submodule are subjected to convolution of a set of convolution kernels, 13 × 256 convolution kernel and 1 × 18 convolution kernel to obtain a 52-scale prediction result 52 × 18.
CN202010736191.3A 2020-07-28 2020-07-28 Rice seedling detection model based on improved YOLOV network and method thereof Active CN112036437B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010736191.3A CN112036437B (en) 2020-07-28 2020-07-28 Rice seedling detection model based on improved YOLOV network and method thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010736191.3A CN112036437B (en) 2020-07-28 2020-07-28 Rice seedling detection model based on improved YOLOV network and method thereof

Publications (2)

Publication Number Publication Date
CN112036437A true CN112036437A (en) 2020-12-04
CN112036437B CN112036437B (en) 2024-06-07

Family

ID=73583325

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010736191.3A Active CN112036437B (en) 2020-07-28 2020-07-28 Rice seedling detection model based on improved YOLOV network and method thereof

Country Status (1)

Country Link
CN (1) CN112036437B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113435309A (en) * 2021-06-24 2021-09-24 农业农村部南京农业机械化研究所 Rice seedling row identification method based on row vector grid classification
CN114119640A (en) * 2022-01-27 2022-03-01 广东皓行科技有限公司 Model training method, image segmentation method and image segmentation system

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019144575A1 (en) * 2018-01-24 2019-08-01 中山大学 Fast pedestrian detection method and device
CN110648316A (en) * 2019-09-07 2020-01-03 创新奇智(成都)科技有限公司 Steel coil end face edge detection algorithm based on deep learning
WO2020047738A1 (en) * 2018-09-04 2020-03-12 安徽中科智能感知大数据产业技术研究院有限责任公司 Automatic pest counting method based on combination of multi-scale feature fusion network and positioning model

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019144575A1 (en) * 2018-01-24 2019-08-01 中山大学 Fast pedestrian detection method and device
WO2020047738A1 (en) * 2018-09-04 2020-03-12 安徽中科智能感知大数据产业技术研究院有限责任公司 Automatic pest counting method based on combination of multi-scale feature fusion network and positioning model
CN110648316A (en) * 2019-09-07 2020-01-03 创新奇智(成都)科技有限公司 Steel coil end face edge detection algorithm based on deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
孟本成;: "基于YOLOV3算法的行人检测方法", 电视技术, no. 09 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113435309A (en) * 2021-06-24 2021-09-24 农业农村部南京农业机械化研究所 Rice seedling row identification method based on row vector grid classification
CN113435309B (en) * 2021-06-24 2022-04-26 农业农村部南京农业机械化研究所 Rice seedling row identification method based on row vector grid classification
CN114119640A (en) * 2022-01-27 2022-03-01 广东皓行科技有限公司 Model training method, image segmentation method and image segmentation system
CN114119640B (en) * 2022-01-27 2022-04-22 广东皓行科技有限公司 Model training method, image segmentation method and image segmentation system

Also Published As

Publication number Publication date
CN112036437B (en) 2024-06-07

Similar Documents

Publication Publication Date Title
CN109800736B (en) Road extraction method based on remote sensing image and deep learning
Arya et al. Deep learning: As the new frontier in high-throughput plant phenotyping
CN111340141A (en) Crop seedling and weed detection method and system based on deep learning
Wen et al. Wheat spike detection and counting in the field based on SpikeRetinaNet
CN109344738A (en) The recognition methods of crop diseases and pest crop smothering and device
CN114841961B (en) Wheat scab detection method based on image enhancement and improved YOLOv5
CN112036437A (en) Rice seedling detection model based on improved YOLOV3 network and method thereof
Agarwal et al. Efficient NetB3 for Automated Pest Detection in Agriculture
CN115797781A (en) Crop identification method and device, computer equipment and storage medium
CN113051983B (en) Method for training field crop disease recognition model and field crop disease recognition
Ahmed et al. Recognizing a spatial extreme dependence structure: A deep learning approach
CN114299379A (en) Shadow area vegetation coverage extraction method based on high dynamic image
CN114120359A (en) Method for measuring body size of group-fed pigs based on stacked hourglass network
CN117557914A (en) Crop pest identification method based on deep learning
Wang et al. High‐throughput measurement of plant fitness traits with an object detection method using Faster R‐CNN
Menezes et al. Pseudo-label semi-supervised learning for soybean monitoring
Chadœuf et al. Modelling unbiased dispersal kernels over continuous space by accounting for spatial heterogeneity in marking and observation efforts
Wang et al. Sugarcane stem node detection and localization for cutting using deep learning
CN117392668A (en) Wheat scab state evaluation method and system and electronic equipment
Jurado-Ruiz et al. LeTra: a leaf tracking workflow based on convolutional neural networks and intersection over union
CN114549958B (en) Night and camouflage target detection method based on context information perception mechanism
CN116188855A (en) Multi-scale plant disease identification method, device, storage medium and apparatus
CN116206210A (en) NAS-Swin-based remote sensing image agricultural greenhouse extraction method
Kamei Effect of image resolution on automatic detection of whitefly (Hemiptera: Aleyrodidae) species on tomato leaflets using deep learning
CN114882362A (en) Natural gas pipeline high consequence area building change detection method based on remote sensing image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant