CN113762190B - Method and device for detecting package stacking based on neural network - Google Patents

Method and device for detecting package stacking based on neural network Download PDF

Info

Publication number
CN113762190B
CN113762190B CN202111078544.6A CN202111078544A CN113762190B CN 113762190 B CN113762190 B CN 113762190B CN 202111078544 A CN202111078544 A CN 202111078544A CN 113762190 B CN113762190 B CN 113762190B
Authority
CN
China
Prior art keywords
package
detection
stacking
neural network
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111078544.6A
Other languages
Chinese (zh)
Other versions
CN113762190A (en
Inventor
朱强
唐金亚
杜萍
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhongke Weizhi Technology Co ltd
Original Assignee
Zhongke Weizhi Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhongke Weizhi Technology Co ltd filed Critical Zhongke Weizhi Technology Co ltd
Priority to CN202111078544.6A priority Critical patent/CN113762190B/en
Publication of CN113762190A publication Critical patent/CN113762190A/en
Application granted granted Critical
Publication of CN113762190B publication Critical patent/CN113762190B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a package stacking detection method and device based on a neural network. The method comprises the following steps: step 101, training to obtain a package detection neural network model, wherein the package detection neural network is based on a YOLOV3 target detection model, a backbone network adopts a MobileNet V2-based architecture, and a detection head part uses a two-layer YOLO layer to output a detection frame; 102, acquiring belt images when conveying packages through a top scanning camera; step 103, extracting the package information in the belt image through the package detection neural network model, wherein the package information extracted by the package detection neural network model comprises the positions of packages and the relation among the packages; and 104, judging the package stacking state according to the extracted package information. The invention can effectively realize the detection of package stacking and improve the accuracy of automatic package feeding and the sorting efficiency.

Description

Method and device for detecting package stacking based on neural network
Technical Field
The invention relates to a detection method and a detection device, in particular to a package stacking detection method and a package stacking detection device based on a neural network.
Background
The full-automatic package supply system consists of a cross belt type sorting system and a package separation system. When entering the cross-belt sortation system, it must be ensured that packages are sequentially entered one by one. If there are stacked packages entering the cross-belt sortation system, package sort failures and reduced sort efficiency will result.
Currently, for package separation, the package separation is controlled by changing the belt speed, and the package with serious stacking cannot be separated. Therefore, to ensure package sorting accuracy and efficiency, how to detect stacked packages before entering the cross-belt sorting system is a critical issue to be addressed.
Disclosure of Invention
The invention aims to overcome the defects in the prior art and provides a parcel stacking detection method and device based on a neural network, which can effectively realize the detection of parcel stacking and improve the accuracy of automatic parcel supply and the sorting efficiency.
According to the technical scheme provided by the invention, the package stacking detection method based on the neural network comprises the following steps of:
step 101, training to obtain a package detection neural network model, wherein the package detection neural network is based on a YOLOV3 target detection model, a backbone network adopts a MobileNet V2-based architecture, and a detection head part uses a two-layer YOLO layer to output a detection frame;
102, acquiring belt images when conveying packages through a top scanning camera;
step 103, extracting the package information in the belt image through the package detection neural network model, wherein the package information extracted by the package detection neural network model comprises the positions of packages and the relation among the packages;
and 104, judging the package stacking state according to the extracted package information.
In step 101, when the YOLOV 3-based target detection model is adopted, training of the neural network includes a pre-training step and a stacking detection training step; wherein,
in the pre-training process, the pre-training data set is a COCO data set, and a He Kaiming initialization method is used for initializing the network weight; the network is fixedly iterated for M1 times, and the optimal weight of MAP is selected as a pre-training weight;
during stacking detection training, performing stacking detection training by using an acquired stacking detection data set, and initializing network weights by using weights obtained by pre-training; when selecting the verification set, only selecting pictures which only contain one single packet or only contain one multi-packet; classifying each verification set picture after non-maximum suppression, dividing the verification set picture into single packets or multiple packets, calculating an F1-score, recording an F1-score curve, and selecting the optimal weight of the F1-score when training is finished; the initial learning rate is 0.0005 and the learning rate decay strategy is cosine annealing.
When the package information in the belt image is extracted through the package detection neural network model, the non-maximum suppression is used for carrying out post-processing on the extracted package target, and the suppression is only carried out among the same classes.
When the non-maximum suppression is used for processing and suppressing the extracted parcel target, the method comprises the following steps:
ordering the target frames in the target frame set { B } from high to low according to the confidence;
traversing the set of target boxes { B }, for each target box B i From object frame B i+1 Traversing backward if target frame B i And target frame B j Belonging to the same class and having a cross-over ratio greater than a threshold value, and the target frame B j Marking and removing;
if the target frame B is traversed while traversing the target frame set { B }, the target frame B i If the mark is removed, skipping the subsequent operation of the round of traversal, and selecting the target frame B i+1 Starting traversing; wherein, the value range of i is [0, N-2 ]]N is the number of target frames, and the value range of j is [ i+1, N-1]。
In step 104, the package stacking state is determined according to the extracted package information, where the package stacking state includes:
when only one single packet appears, judging that the single packet is in a non-stacking condition;
when a plurality of packages appear, judging that the stacking situation exists when the head-to-tail distance between the first package and the second package positive rectangular frame is smaller than a threshold value L;
when a plurality of packages appear, judging that the package is in a non-stacking condition when the first package type is a single package and the head-tail distance between the first package type and the positive rectangular frame of the second package is larger than a threshold value L;
whether there are multiple packages or not, when the first package category is multiple packages, then a stacking situation is determined.
The threshold L is not greater than 250mm.
The utility model provides a parcel stacks detection device based on neural network which characterized in that: comprising
The image acquisition module can acquire belt images when the package is conveyed;
the stacking detection module is connected with the image acquisition module, a package detection neural network model is arranged in the stacking detection module, the package detection neural network is based on a YOLOV3 target detection model, a backbone network adopts a MobileNet V2-based architecture, and a detection head part outputs a detection frame by using two layers of YOLO layers; the stacking detection module can extract package information in the belt image according to the belt image acquired by the image acquisition module, and judges package stacking state corresponding to the belt image according to the extracted package information;
the signal sending module is connected with the stacking detection module and the package conveying control system and can send the package stacking state judged by the stacking detection module to the package conveying control system.
The image acquisition module comprises an area array color camera, and the distance between the image acquisition module and a belt for conveying the package is 1000mm to 1200mm.
Judging the state of the package stacking according to the extracted package information, wherein the state of the package stacking comprises the following steps:
when only one single packet appears, judging that the single packet is in a non-stacking condition;
when a plurality of packages appear, judging that the stacking situation exists when the head-to-tail distance between the first package and the second package positive rectangular frame is smaller than a threshold value L;
when a plurality of packages appear, judging that the package is in a non-stacking condition when the first package type is a single package and the head-tail distance between the first package type and the positive rectangular frame of the second package is larger than a threshold value L;
whether there are multiple packages or not, when the first package category is multiple packages, then a stacking situation is determined.
A virtual transmission line is arranged in the signal transmission module, and a stacking signal is transmitted to the control system only when a package touches the transmission line; and the same package is restrained from triggering and sending signals for multiple times through target tracking.
The invention has the advantages that: acquiring a belt image through the image acquisition module, wherein the belt is used for conveying packages; a package detection neural network model is arranged in the stack detection module, the package detection neural network is based on a YOLOV3 target detection model, a backbone network adopts a MobileNet V2-based architecture, and a detection head part uses a two-layer YOLO layer to output a detection frame; the stacking detection module can extract package information in the belt image according to the belt image acquired by the image acquisition module, and judge package stacking state corresponding to the belt image according to the extracted package information, namely, the stacking detection module extracts package information and judges the stacking condition at the current moment, and the signal transmission module transmits the stacking condition to the control system; the control system can change the running direction of the packages according to the stacking condition, so that the stacked packages are prevented from entering a subsequent sorting system, and the package feeding accuracy and the sorting efficiency are effectively improved.
Drawings
FIG. 1 is a flow chart of the detection method of the present invention.
Fig. 2 is a flow chart of the present invention for controlling package delivery.
Fig. 3 is a schematic diagram of a MobileNetV 2-based architecture according to the present invention.
Detailed Description
The invention will be further described with reference to the following specific drawings and examples.
As shown in fig. 1, in order to effectively realize package stacking detection and improve automatic package feeding accuracy and sorting efficiency, the package stacking method of the invention comprises the following steps:
step 101, training to obtain a package detection neural network model, wherein the package detection neural network is based on a YOLOV3 target detection model, a backbone network adopts a MobileNet V2-based architecture, and a detection head part uses a two-layer YOLO layer to output a detection frame;
in specific implementation, the YOLOV3 target detection model is a conventional target detection model, which is well known to those skilled in the art, and will not be described herein. In order to adapt to package detection, a backbone network adopts a MobileNet V2-based architecture, and a detection head part uses a two-layer YOLO layer output detection frame, wherein the specific situation based on the MobileNet V2 architecture is consistent with the prior art. In fig. 3, a case based on a MobileNetV2 architecture is shown, specifically, input represents the input tensor size, operator represents the tensor operation method, c represents the number of output channels, n represents the number of iterations of Operator, s represents the step size of output relative to input, conv2D is 2D convolution, bottleck is an anti-residual module, and the anti-residual module bottleck may specifically take a conventional commonly used form, i.e. the specific embodiment of the MobileNetV2 architecture is well known to those skilled in the art, and will not be repeated herein.
In addition, the detection head portion uses a two-layer YOLO layer output detection frame, which is specifically related to the YOLOV3 target detection model, and is specifically well known in the art, and is not described herein.
In the embodiment of the invention, when the parcel detection neural network model is based on a YOLOV3 target detection model, training of the neural network comprises a pre-training step and a stacking detection training step; in particular, the size of the picture resolution during pre-training and stacking detection training affects the size of the picture resolution required during forward reasoning. The picture resolution of training is obtained according to the picture resolution required by reasoning, and specifically, the picture resolution required by reasoning is between the maximum value and the minimum value of the picture resolution of training. For example, the picture resolution at the time of reasoning is 320x320, the training picture resolution is 224 at the minimum and 512 at the maximum, i.e. the selected input picture resolution is a multiple of 32 between 224 and 512.
In the pre-training process, the pre-training data set is a COCO data set, and a He Kaiming initialization method is used for initializing the network weight; the network is fixedly iterated for M1 times, and the optimal weight of MAP is selected as a pre-training weight;
in order to adapt to the detection of packages, after selecting the YOLOV 3-based target detection model, a corresponding training step is further required, where the training step specifically includes a pre-training step and a stacking detection step, and the purpose of the specific step is consistent with that of the existing one, and will not be repeated here.
In the embodiment of the invention, training is performed on a COCO data set in the pre-training process, and a He Kaiming initialization method is used for initializing the network weight; in particular, the COCO data set is in particular an MS-COCO-2017 data set, which is obtainable from an MS COCO functional network. In pre-training, the network is set to detect class 80. In specific implementation, setting the network fixed iteration M1 as 800020 times, and selecting the weight with the optimal MAP value as the pre-training weight; the initial learning rate is 0.001, the learning rate attenuation strategy is a sectional attenuation, and the attenuation is 0.9 times when the 400000 th iteration and the 650000 th iteration are carried out; in the first 4000 iterations, the learning rate rises from 0 to 0.001 at a constant rate.
After the conditions such as the data set and the network weight initialization method in the pre-training process are selected, the pre-training of the Y0LOV3 target detection model can be realized, and the specific training process is well known to those skilled in the art and is not repeated here.
During stacking detection training, performing stacking detection training by using an acquired stacking detection data set, and initializing network weights by using weights obtained by pre-training; when selecting the verification set, only selecting pictures which only contain one single packet or only contain one multi-packet; classifying each verification set picture after non-maximum suppression, dividing the verification set picture into single packets or multiple packets, calculating an F1-score, recording an F1-score curve, and selecting the optimal weight of the F1-score when training is finished; the initial learning rate is 0.0005 and the learning rate decay strategy is cosine annealing.
In the embodiment of the invention, the stacking detection training data set can be obtained in a snap shot mode in an actual production line. Specifically, the camera is fixed on the belt to collect pictures. Under the condition that no sample exists, obtaining pictures in a mode of equiinterval snapshot, and manually picking out an effective sample; under the condition that a small-scale sample is obtained, preliminary stacking detection training is carried out, the camera captures pictures in real time, and the pictures containing packages are automatically stored by using a trained model to serve as effective samples. When stacking detection training is performed, the network is set to be in detection class 2, namely, the network is divided into a single-packet sample and a multi-packet sample, and the specific process of acquiring the stacking detection training data set can be selected and determined according to actual needs, which is well known to those skilled in the art, and is not repeated here.
In specific implementation, training is carried out on the collected stacked detection data set, and the weight obtained by pre-training is used for initializing the network weight; when selecting the verification set, only selecting pictures which only contain one single packet or only contain one multi-packet; classifying each verification set picture after non-maximum suppression, dividing the verification set picture into single packets or multiple packets, calculating an F1-score, recording an F1-score curve, and selecting the optimal weight of the F1-score when training is finished; the initial learning rate is 0.0005, and the learning rate attenuation strategy is cosine annealing; the total number of training rounds is determined by the data set size and should not be less than 2000 iteration lots. In the specific training process, when more picture samples are collected, 150 training wheels can be fixed.
In the training process, the specific processes of calculating the F1-score, recording the F1-score curve, and selecting the optimal weight of the F1-score at the end of training are consistent with the existing processes, and are well known to those skilled in the art, and will not be repeated here. After the pre-training and stacking detection training, the required parcel detection neural network model based on the YOLOV3 target detection model can be obtained.
102, acquiring belt images when conveying packages through a top scanning camera;
specifically, the manner and process of obtaining the belt image by the top scanning camera are consistent with the prior art, and are well known to those skilled in the art, and are not described herein.
Step 103, extracting the package information in the belt image through the package detection neural network model, wherein the package information extracted by the package detection neural network model comprises the positions of packages and the relation among the packages;
in the embodiment of the invention, after the package detection neural network model is obtained, for an arbitrarily acquired belt image, the package information in the leather strip image can be extracted by using the package detection neural network model, and the process of obtaining the package information by using the package detection neural network model is consistent with the prior art, and is not repeated here. In specific implementation, the extracted package information includes the locations of packages and the relationships between packages.
When the method is implemented, the non-maximum suppression is used for carrying out post-treatment on the extracted parcel target and only suppressing among the same class when parcel information in the belt image is extracted through the parcel detection neural network model.
In the embodiment of the invention, when the non-maximum suppression is used for processing and suppressing the extracted parcel target, the method comprises the following steps:
ordering the target frames in the target frame set { B } from high to low according to the confidence;
traversing the set of object boxes { B }, for each objectFrame B i From object frame B i+1 Traversing backward if target frame B i And target frame B j Belonging to the same class and having a cross-over ratio greater than a threshold value, and the target frame B j Marking and removing;
if the target frame B is traversed while traversing the target frame set { B }, the target frame B i If the mark is removed, skipping the subsequent operation of the round of traversal, and selecting the target frame B i+1 Starting traversing; wherein, the value range of i is [0, N-2 ]]N is the number of target frames, and the value range of j is [ i+1, N-1]。
In the embodiment of the present invention, the target frame set { B } includes a plurality of target frames, and the number of target frames is related to the acquired belt image, which is well known in the art and will not be described herein. After the package information is extracted by using the package detection neural network model, the confidence coefficient of each target frame is also given at the same time, and is in accordance with the existing, and details are not repeated here. After the non-maximum value suppression processing, the interference between the single-packet information and the multi-packet information can be reduced.
And 104, judging the package stacking state according to the extracted package information.
In specific implementation, the package stacking state is judged according to the extracted package information, and the package stacking state comprises the following steps:
when only one single packet appears, judging that the single packet is in a non-stacking condition;
when a plurality of packages appear, judging that the stacking situation exists when the head-to-tail distance between the first package and the second package positive rectangular frame is smaller than a threshold value L;
when a plurality of packages appear, judging that the package is in a non-stacking condition when the first package type is a single package and the head-tail distance between the first package type and the positive rectangular frame of the second package is larger than a threshold value L;
whether there are multiple packages or not, when the first package category is multiple packages, then a stacking situation is determined.
As can be seen from the above description, after the belt image is extracted to obtain the package information, the location of the package and the relationship between the packages can be obtained, so that the package stacking state can be determined according to the package information. In specific implementation, the threshold L is not greater than 250mm, and the specific situation of the threshold L may be selected according to situations such as actual package delivery, which are well known in the art, and will not be described herein.
In summary, a neural network-based package stack detection device is available, specifically comprising
The image acquisition module can acquire belt images when the package is conveyed;
the stacking detection module is connected with the image acquisition module, a package detection neural network model is arranged in the stacking detection module, the package detection neural network is based on a YOLOV3 target detection model, a backbone network adopts a MobileNet V2-based architecture, and a detection head part outputs a detection frame by using two layers of YOLO layers; the stacking detection module can extract package information in the belt image according to the belt image acquired by the image acquisition module, and judges package stacking state corresponding to the belt image according to the extracted package information;
the signal sending module is connected with the stacking detection module and the package conveying control system and can send the package stacking state judged by the stacking detection module to the package conveying control system.
In the embodiment of the present invention, for the image acquisition module, an area-array camera, a color or gray sensor may be used, and specifically may be selected according to actual needs, which will not be described herein. When an area-array color camera is used to acquire the target image, the suspension height depends on the actual field of view, and specifically, the suspension height of the area-array color camera is preferably between 1000mm and 1200mm.
In a real scene, the package may move at a high speed, and the exposure time should be reduced to avoid motion blur of the photographed image. Therefore, a light supplementing module should also be included. The light supplementing module can adopt a mainstream LED light source and has the characteristics of small heating value, high brightness, long service life and the like. The device adopts a white light LED lamp bead with a fixed emission angle.
For the stacking detection module, a package detection neural network is arranged in the stacking detection module, the specific conditions of the package detection neural network, detection of the belt image and the like are consistent with the above description, and the above digital-analog can be referred to for details, which are not repeated here.
A virtual transmission line is arranged in the signal transmission module, and a stacking signal is transmitted to the control system only when a package touches the transmission line; and the same package is restrained from triggering and sending signals for multiple times through target tracking.
In the embodiment of the invention, each package is tracked, and when the package enters the visual field, a tracking list is added; when the virtual transmission line is set, the stacking condition is sent to the package conveying control system when the package touches the virtual transmission line, the package conveying control system can control the conveying process of the package, the package conveying control system can adopt the conventional common mode, the conveying control process of the package is consistent with the conventional mode, and the like, and the package conveying control system is well known to the person skilled in the art, and is not used anytime.
In particular, the virtual transmission line should be set with the image acquisition frame rate and the parcel motion rate taken into consideration. It should be ensured that the package does not move directly outside the field of view of the image from a position where the transmission line is not touched, within the interval of two frame image acquisitions. When the moving speed of the package is 1800mm/s, the photographing interval is 100ms, and the virtual sending line is arranged at a position 250mm away from the edge of the visual field outlet. In particular, the virtual transmission line is set 250mm from the view outlet edge, which refers to the side of the package that is touched when leaving in the image. For the specific position of the virtual transmission line, the specific position needs to be adjusted according to the movement speed of the belt and the image acquisition interval, and the distance of the belt movement in one acquisition interval needs to be slightly larger than that of the belt movement in one acquisition interval. When the system works, each frame of image judges the position relation between the front-end package and the sending line, and when the front-end package touches the sending line, the current judging result is sent to the package conveying control system.
The target tracking can be implemented by matching the current frame package detection result with the previous frame result in a track-by-detection manner well known to those skilled in the art, so that the tracking is implemented, and the specific target tracking process is consistent with the existing one, and is well known to those skilled in the art and will not be repeated.
As shown in fig. 2, a workflow diagram of the package stack inspection device of the present invention is shown, specifically: acquiring a belt image through the image acquisition module, wherein the belt is used for conveying packages; extracting package information and judging the stacking condition at the current moment through a stacking detection module, wherein the package information comprises package positions, package intervals and package types; and sending the stacking condition to a control system through a signal sending module. The problem of wrong division caused by package stacking is solved, and the automatic package feeding efficiency and accuracy are effectively improved.
The technical features of the above-described embodiments may be arbitrarily combined, and all possible combinations of the technical features in the above-described embodiments are not described for brevity of description, however, as long as there is no contradiction between the combinations of the technical features, they should be considered as the scope of the description.
The above examples merely represent a few embodiments of the present application, which are described in more detail and are not to be construed as limiting the scope of the invention. It should be noted that it would be apparent to those skilled in the art that various modifications and improvements could be made without departing from the spirit of the present application, which would be within the scope of the present application. Accordingly, the scope of protection of the present application is to be determined by the claims appended hereto.

Claims (9)

1. The package stacking detection method based on the neural network is characterized by comprising the following steps of:
step 101, training to obtain a package detection neural network model, wherein the package detection neural network is based on a YOLOV3 target detection model, a backbone network adopts a MobileNet V2-based architecture, and a detection head part uses a two-layer YOLO layer to output a detection frame;
102, acquiring belt images when conveying packages through a top scanning camera;
step 103, extracting the package information in the belt image through the package detection neural network model, wherein the package information extracted by the package detection neural network model comprises the positions of packages and the relation among the packages;
104, judging the package stacking state according to the extracted package information;
in step 101, when the YOLOV 3-based target detection model is adopted, training of the neural network includes a pre-training step and a stacking detection training step; wherein,
in the pre-training process, the pre-training data set is a COCO data set, and a He Kaiming initialization method is used for initializing the network weight; the network is fixedly iterated for M1 times, and the optimal weight of MAP is selected as a pre-training weight;
during stacking detection training, performing stacking detection training by using an acquired stacking detection data set, and initializing network weights by using weights obtained by pre-training; when selecting the verification set, only selecting pictures which only contain one single packet or only contain one multi-packet; classifying each verification set picture after non-maximum suppression, dividing the verification set picture into single packets or multiple packets, calculating an F1-score, recording an F1-score curve, and selecting the optimal weight of the F1-score when training is finished; the initial learning rate is 0.0005 and the learning rate decay strategy is cosine annealing.
2. The neural network-based parcel stacking detection method of claim 1, wherein when parcel information in a belt image is extracted by a parcel detection neural network model, non-maximum suppression is used to post-process the extracted parcel target, and suppression is performed only between the same classes.
3. The neural network-based parcel stacking detection method of claim 2, wherein when using non-maximum suppression to process and suppress extracted parcel targets, comprising the following steps:
ordering the target frames in the target frame set { B } from high to low according to the confidence;
traversing the set of target boxes { B }, for each target box B i From object frame B i+1 Traversing backward if target frame B i And target frame B j Belonging to the same class and having a cross-over ratio greater than a threshold value, and the target frame B j Marking and removing;
if the target frame B is traversed while traversing the target frame set { B }, the target frame B i If the mark is removed, skipping the subsequent operation of the round of traversal, and selecting the target frame B i+1 Starting traversing; wherein, the value range of i is [0, N-2 ]]N is the number of target frames, and the value range of j is [ i+1, N-1]。
4. A method of detecting a package stack based on a neural network according to any one of claims 1 to 3, wherein in step 104, the status of the package stack is determined according to the extracted package information, and the status of the package stack includes:
when only one single packet appears, judging that the single packet is in a non-stacking condition;
when a plurality of packages appear, judging that the stacking situation exists when the head-to-tail distance between the first package and the second package positive rectangular frame is smaller than a threshold value L;
when a plurality of packages appear, judging that the package is in a non-stacking condition when the first package type is a single package and the head-tail distance between the first package type and the positive rectangular frame of the second package is larger than a threshold value L;
whether there are multiple packages or not, when the first package category is multiple packages, then a stacking situation is determined.
5. The neural network-based parcel stacking detection method of claim 4, wherein said threshold L is no greater than 250mm.
6. The utility model provides a parcel stacks detection device based on neural network which characterized in that: comprising
The image acquisition module can acquire belt images when the package is conveyed;
the stacking detection module is connected with the image acquisition module, a package detection neural network model is arranged in the stacking detection module, the package detection neural network is based on a YOLOV3 target detection model, a backbone network adopts a MobileNet V2-based architecture, and a detection head part outputs a detection frame by using two layers of YOLO layers; the stacking detection module can extract package information in the belt image according to the belt image acquired by the image acquisition module, and judges package stacking state corresponding to the belt image according to the extracted package information;
the signal sending module is connected with the stacking detection module and the package conveying control system and can send the package stacking state judged by the stacking detection module to the package conveying control system;
when the detection neural network model is arranged in the stacking detection module, the method comprises the step of training to obtain the package detection neural network model, wherein,
when the YOLOV 3-based target detection model is adopted, training of the neural network comprises a pre-training step and a stacking detection training step; wherein,
in the pre-training process, the pre-training data set is a COCO data set, and a He Kaiming initialization method is used for initializing the network weight; the network is fixedly iterated for M1 times, and the optimal weight of MAP is selected as a pre-training weight;
during stacking detection training, performing stacking detection training by using an acquired stacking detection data set, and initializing network weights by using weights obtained by pre-training; when selecting the verification set, only selecting pictures which only contain one single packet or only contain one multi-packet; classifying each verification set picture after non-maximum suppression, dividing the verification set picture into single packets or multiple packets, calculating an F1-score, recording an F1-score curve, and selecting the optimal weight of the F1-score when training is finished; the initial learning rate is 0.0005 and the learning rate decay strategy is cosine annealing.
7. The neural network-based parcel stack detection apparatus of claim 6, wherein: the image acquisition module comprises an area array color camera, and the distance between the image acquisition module and a belt for conveying the package is 1000mm to 1200mm.
8. The neural network-based parcel stack detection apparatus of claim 6 or 7, wherein: judging the state of the package stacking according to the extracted package information, wherein the state of the package stacking comprises the following steps:
when only one single packet appears, judging that the single packet is in a non-stacking condition;
when a plurality of packages appear, judging that the stacking situation exists when the head-to-tail distance between the first package and the second package positive rectangular frame is smaller than a threshold value L;
when a plurality of packages appear, judging that the package is in a non-stacking condition when the first package type is a single package and the head-tail distance between the first package type and the positive rectangular frame of the second package is larger than a threshold value L;
whether there are multiple packages or not, when the first package category is multiple packages, then a stacking situation is determined.
9. The neural network-based parcel stack detection apparatus of claim 6 or 7, wherein: a virtual transmission line is arranged in the signal transmission module, and a stacking signal is transmitted to the control system only when a package touches the transmission line; and the same package is restrained from triggering and sending signals for multiple times through target tracking.
CN202111078544.6A 2021-09-15 2021-09-15 Method and device for detecting package stacking based on neural network Active CN113762190B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111078544.6A CN113762190B (en) 2021-09-15 2021-09-15 Method and device for detecting package stacking based on neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111078544.6A CN113762190B (en) 2021-09-15 2021-09-15 Method and device for detecting package stacking based on neural network

Publications (2)

Publication Number Publication Date
CN113762190A CN113762190A (en) 2021-12-07
CN113762190B true CN113762190B (en) 2024-03-29

Family

ID=78795635

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111078544.6A Active CN113762190B (en) 2021-09-15 2021-09-15 Method and device for detecting package stacking based on neural network

Country Status (1)

Country Link
CN (1) CN113762190B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116188767B (en) * 2023-01-13 2023-09-08 湖北普罗格科技股份有限公司 Neural network-based stacked wood board counting method and system

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108875595A (en) * 2018-05-29 2018-11-23 重庆大学 A kind of Driving Scene object detection method merged based on deep learning and multilayer feature
CN109784204A (en) * 2018-12-25 2019-05-21 江苏大学 A kind of main carpopodium identification of stacking string class fruit for parallel robot and extracting method
CN111310622A (en) * 2020-02-05 2020-06-19 西北工业大学 Fish swarm target identification method for intelligent operation of underwater robot
CN111369540A (en) * 2020-03-06 2020-07-03 西安电子科技大学 Plant leaf disease identification method based on mask convolutional neural network
CN111428558A (en) * 2020-02-18 2020-07-17 东华大学 Vehicle detection method based on improved YO L Ov3 method
CN111461224A (en) * 2020-04-01 2020-07-28 西安交通大学 Phase data unwrapping method based on residual self-coding neural network
CN111553200A (en) * 2020-04-07 2020-08-18 北京农业信息技术研究中心 Image detection and identification method and device
CN112163545A (en) * 2020-10-12 2021-01-01 北京易华录信息技术股份有限公司 Head feature extraction method and device, electronic equipment and storage medium
CN112668663A (en) * 2021-01-05 2021-04-16 南京航空航天大学 Aerial photography car detection method based on YOLOv4

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102009011300B4 (en) * 2009-03-02 2022-08-11 Kuka Roboter Gmbh Loading of loading equipment with packages using a manipulator
US11544900B2 (en) * 2019-07-25 2023-01-03 General Electric Company Primitive-based 3D building modeling, sensor simulation, and estimation
CN111461127B (en) * 2020-03-30 2023-06-06 华南理工大学 Instance segmentation method based on one-stage target detection framework
CN112257799A (en) * 2020-10-30 2021-01-22 电子科技大学中山学院 Method, system and device for detecting household garbage target
CN112487915B (en) * 2020-11-25 2024-04-23 江苏科技大学 Pedestrian detection method based on Embedded YOLO algorithm

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108875595A (en) * 2018-05-29 2018-11-23 重庆大学 A kind of Driving Scene object detection method merged based on deep learning and multilayer feature
CN109784204A (en) * 2018-12-25 2019-05-21 江苏大学 A kind of main carpopodium identification of stacking string class fruit for parallel robot and extracting method
CN111310622A (en) * 2020-02-05 2020-06-19 西北工业大学 Fish swarm target identification method for intelligent operation of underwater robot
CN111428558A (en) * 2020-02-18 2020-07-17 东华大学 Vehicle detection method based on improved YO L Ov3 method
CN111369540A (en) * 2020-03-06 2020-07-03 西安电子科技大学 Plant leaf disease identification method based on mask convolutional neural network
CN111461224A (en) * 2020-04-01 2020-07-28 西安交通大学 Phase data unwrapping method based on residual self-coding neural network
CN111553200A (en) * 2020-04-07 2020-08-18 北京农业信息技术研究中心 Image detection and identification method and device
CN112163545A (en) * 2020-10-12 2021-01-01 北京易华录信息技术股份有限公司 Head feature extraction method and device, electronic equipment and storage medium
CN112668663A (en) * 2021-01-05 2021-04-16 南京航空航天大学 Aerial photography car detection method based on YOLOv4

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"基于YOLO的深度学习目标检测算法研究";熊咏平;《中国优秀硕士学位论文全文数据库 (基础科学辑)》;I138-984 *

Also Published As

Publication number Publication date
CN113762190A (en) 2021-12-07

Similar Documents

Publication Publication Date Title
CN111145545B (en) Road traffic behavior unmanned aerial vehicle monitoring system and method based on deep learning
CN109460709B (en) RTG visual barrier detection method based on RGB and D information fusion
CN103824066B (en) A kind of licence plate recognition method based on video flowing
CN111709935B (en) Real-time coal gangue positioning and identifying method for ground moving belt
CN106707296A (en) Dual-aperture photoelectric imaging system-based unmanned aerial vehicle detection and recognition method
CN111191611B (en) Traffic sign label identification method based on deep learning
CN113762190B (en) Method and device for detecting package stacking based on neural network
CN111723854B (en) Expressway traffic jam detection method, equipment and readable storage medium
US9216869B2 (en) Device and method for controlling the tracking of a value document stack
CN110570454A (en) Method and device for detecting foreign matter invasion
CN205788226U (en) Container number based on image procossing and truck car number identification system
CN104361323B (en) Recognition method of container number in channel and system
CN113807466B (en) Logistics package autonomous detection method based on deep learning
CN113128507B (en) License plate recognition method and device, electronic equipment and storage medium
CN109934108A (en) The vehicle detection and range-measurement system and implementation method of a kind of multiple target multiple types
CN107506753B (en) Multi-vehicle tracking method for dynamic video monitoring
CN111127520B (en) Vehicle tracking method and system based on video analysis
CN116434088A (en) Lane line detection and lane auxiliary keeping method based on unmanned aerial vehicle aerial image
CN115841633A (en) Power tower and power line associated correction power tower and power line detection method
CN117274967A (en) Multi-mode fusion license plate recognition algorithm based on convolutional neural network
CN113591643A (en) Underground vehicle station entering and exiting detection system and method based on computer vision
CN117215327A (en) Unmanned aerial vehicle-based highway inspection detection and intelligent flight control method
CN115082834B (en) Engineering vehicle black smoke emission monitoring method and system based on deep learning
CN114511832B (en) Lane line analysis method and device, electronic device and storage medium
CN116342632A (en) Depth information-based matting method and matting network training method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information

Country or region after: China

Address after: No. 979, Antai Third Road, Xishan District, Wuxi City, Jiangsu Province, 214000

Applicant after: Zhongke Weizhi Technology Co.,Ltd.

Address before: 299 Dacheng Road, Xishan District, Wuxi City, Jiangsu Province

Applicant before: Zhongke Weizhi intelligent manufacturing technology Jiangsu Co.,Ltd.

Country or region before: China

CB02 Change of applicant information
GR01 Patent grant
GR01 Patent grant