CN111914815A - Machine vision intelligent recognition system and method for garbage target - Google Patents

Machine vision intelligent recognition system and method for garbage target Download PDF

Info

Publication number
CN111914815A
CN111914815A CN202010924563.5A CN202010924563A CN111914815A CN 111914815 A CN111914815 A CN 111914815A CN 202010924563 A CN202010924563 A CN 202010924563A CN 111914815 A CN111914815 A CN 111914815A
Authority
CN
China
Prior art keywords
target
data
garbage
layer
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010924563.5A
Other languages
Chinese (zh)
Inventor
黄纯根
冼海仪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Kunpeng Intelligent Machine Equipment Co ltd
Original Assignee
Guangdong Kunpeng Intelligent Machine Equipment Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Kunpeng Intelligent Machine Equipment Co ltd filed Critical Guangdong Kunpeng Intelligent Machine Equipment Co ltd
Priority to CN202010924563.5A priority Critical patent/CN111914815A/en
Publication of CN111914815A publication Critical patent/CN111914815A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C22/00Measuring distance traversed on the ground by vehicles, persons, animals or other moving solid bodies, e.g. using odometers, using pedometers
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/86Combinations of lidar systems with systems other than lidar, radar or sonar, e.g. with direction finders
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/26Government or public services

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Data Mining & Analysis (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Tourism & Hospitality (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Development Economics (AREA)
  • Educational Administration (AREA)
  • Biophysics (AREA)
  • Economics (AREA)
  • Human Resources & Organizations (AREA)
  • Marketing (AREA)
  • Primary Health Care (AREA)
  • Strategic Management (AREA)
  • General Business, Economics & Management (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Electromagnetism (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a machine vision intelligent recognition system of a garbage target, which comprises a milemeter module, a target recognition module and a mapping module, wherein the system of the invention establishes a dynamic map of garbage target distribution, reduces the monitoring pressure of managers and improves the working efficiency of cleaning personnel, and also discloses a machine vision intelligent recognition method of the garbage target, which comprises the following steps: the method comprises the steps of calculating pose data, identifying garbage targets, establishing a point cloud map, establishing a training set and a testing set, establishing an improved yolov3 model, training a network and testing the network, wherein the method uses a deep separable convolution to replace a part of traditional convolution in darknet-53, the calculated amount of the model is greatly reduced, the garbage targets can be identified more accurately based on an improved yolov3 target identification algorithm, the parameters of the model are fewer, the time for training and optimizing is reduced, the identification speed of the model is improved, and the requirement on hardware is reduced.

Description

Machine vision intelligent recognition system and method for garbage target
[ technical field ]
The invention relates to the field of machine vision and machine learning, in particular to a machine vision intelligent recognition system and method for a garbage target.
[ background art ]
With the increase of the consumption level of residents, the quantity of household garbage generated by people in the community is continuously increased. The waste domestic garbage not only seriously affects the environmental sanitation of communities, but also is easy to breed mosquitoes and spread diseases so as to harm the health of people. At present, domestic garbage in the community is cleaned regularly by cleaning workers, the mode has good effect on treating the domestic garbage in the downstairs garbage can of residents, but the effect is poor and manpower is wasted for the domestic garbage which is nearly randomly generated in other areas in the community. Therefore, there is a need to develop a visual system capable of monitoring the changing conditions of the domestic waste in the community.
The existing garbage identification system is mainly provided with fixed cameras, and when the garbage change condition of the whole community needs to be monitored, more cameras need to be arranged; meanwhile, the identification system cannot integrate the identification results from all paths of monitoring videos into one map, so that huge challenges are brought to monitoring management personnel in the background.
In addition, the current garbage machine vision identification method mostly adopts two-step methods (R-CNN and false-RCNN) based on a deep learning target identification algorithm, but the method has the problems of more complicated training process, more difficult optimization result, large calculated amount, longer time for processing one frame of image, higher requirement on a hardware platform and the like. .
The present invention has been made in view of the above problems.
[ summary of the invention ]
The invention aims to provide a machine vision intelligent recognition system and a machine vision intelligent recognition method for garbage targets, which aim to overcome the defects of the prior art.
In order to solve the technical problems, the technical scheme adopted by the invention is as follows:
a machine vision intelligent recognition system for garbage targets comprises
The odometer module is used for calculating pose data of the robot;
the target recognition module is used for searching the position of the junk target in the image by using a target recognition algorithm and calculating the position data and the depth data of the junk target;
and the mapping module is used for receiving and processing the pose data of the odometer module, the garbage target position data and the depth data of the target identification module to obtain the coordinates of the garbage target in a world coordinate system, and splicing and constructing to obtain a point cloud map of the garbage target.
The machine vision intelligent recognition system for the garbage target comprises the wheel type odometer and the laser odometer.
The machine vision intelligent recognition system for the garbage target comprises a high-definition camera and a target recognition module.
A machine vision intelligent identification method of a rubbish target comprises the following steps:
step A, calculating pose data:
a1, wheel type odometer calibration: collecting a plurality of data segments, each data segment comprising angular velocities W of two wheelsLAnd WRCalculating the distance b between the two wheels and the radius of the two wheels by a model-based method according to the duration of the data and the matching value of the laser radar;
a2, removing the motion distortion of the laser radar data: the industrial personal computer reads laser data, the STM32 uploads odometer integral data, whether the position and attitude data in an odometer queue are synchronous with the laser radar data in time is checked, if so, the processing is not carried out, otherwise, secondary interpolation is carried out on the position and attitude data, and the laser radar data is transformed according to the position and attitude data to obtain the laser radar data with motion distortion removed;
a3, using a front-end registration algorithm on the acquired lidar data: firstly, searching a matching point in the point cloud, then carrying out pose change and error calculation on the point cloud according to a matching calculation rotation matrix R and a translation matrix T, and continuously iterating until the error is smaller than a tolerance value.
A4, optimizing the pose graph based on a graph optimization technology;
step B, identifying a garbage target:
the target recognition module finds the position of the junk target in the image by using a target recognition algorithm and then issues position data and depth data of the junk target;
step C, establishing a point cloud map:
the mapping module processes data from the odometer module and the target identification module, and performs coordinate transformation according to the following formula:
z=d
Figure BDA0002667897520000031
and transforming the coordinates of the target under the camera coordinate system by using the optimized pose to obtain the coordinates of the target under the world coordinate system, and finally splicing and constructing a point cloud map.
The machine vision intelligent identification method of the rubbish target further comprises the following steps:
step D, establishing a training set and a testing set:
d1, refining the definition of the garbage target: selecting a plurality of waste products as definitions of garbage targets;
d2, collecting photographs of the waste: acquiring a corresponding sample image by adopting a field acquisition mode and a crawling mode;
d3, preprocessing of the image: cropping the collected image data and transforming to 416 x 416 standard sample data;
d4, amplification of sample data: sample data is expanded in the following way:
d4-1. magnification of width and height of sample image 1.5 times;
d4-2. width reduction 1/3 and height reduction 1/2 of the sample image, and ensuring that the image is a multiple of 32;
d4-3, brightness enhancement of the sample image;
d4-4, the brightness of the sample image is reduced;
d4-5, turning the sample image by 90 degrees or 180 degrees;
d4-6, adding noise to the sample image;
d5, manual labeling: according to the definition in the step D1, labeling the image data in the sample set by using an image labeling tool to obtain label data (x, y, w, h, c);
d6, partitioning of sample set: dividing a sample set into a training set and a test set according to a ratio of 98: 2;
step E, constructing an improved yolov3 model:
respectively creating a convolutional layer, a depth separable convolutional layer, an upsampling layer, a residual layer, a splicing layer and a prediction layer, wherein the upsampling layer uses a bilinear interpolation algorithm to expand a feature map, the residual layer adds the output of a lower layer and the output of an upper layer, the splicing layer superposes the output of the upper layer containing rich semantics and the output of a lower layer with higher resolution, the prediction layer comprises 3 feature maps, each grid of each feature map also comprises 3 anchor boxes with different length-width ratios, and each anchor box is a1 × 20 vector and comprises the following prediction information: t is tx,ty,tw,thThe bounding box contains the confidence of the object and the probability that the object belongs to the garbage target;
step F, training the network:
initializing the weight of the model by using a random initialization strategy, then, throwing training data into the model, predicting each grid of the layer feature diagram after forward propagation to obtain a prediction result of 3 x (4+1+15), and finally, calculating the error between the prediction result and the real result according to the following cost function formula:
Figure BDA0002667897520000051
calculating an error between a predicted value and a true value through the cost function, then reversely propagating the error, updating the weight and deviation of the network through a momentum gradient descent strategy, and stopping training when the loss value of the cost function is small and stable, namely, convergence;
step G, network testing:
test set data is lost into the model, a prediction result is obtained, and a confidence threshold of a boundary box is set to be ThscoresAnd (5) rejecting the prediction result with low confidence coefficient, setting a non-maximum suppression threshold value of 0.5, executing non-maximum suppression, preventing the same target from being identified in multiple ways, and finally outputting the corrected prediction result.
Compared with the prior art, the invention has the beneficial effects that:
1. the invention establishes a dynamic map of garbage target distribution based on the pose of the odometer module, reduces the monitoring pressure of managers and improves the working efficiency of cleaning personnel.
2. According to the method, a part of traditional convolution in darknet-53 is replaced by the deep separable convolution, the calculated amount of the model is greatly reduced, the garbage target can be identified more accurately based on the improved yolov3 target identification algorithm, the parameters of the model are fewer, the time for training and optimizing is shortened, the identification speed of the model is increased, and the requirement on hardware is lowered.
[ description of the drawings ]
FIG. 1 is a schematic structural view of the present invention;
FIG. 2 is a computing flow diagram of the odometer module;
FIG. 3 is a flow chart of the present invention for performing object recognition;
FIG. 4 is a flow chart of the present invention for establishing a point cloud map;
fig. 5 is a schematic structural diagram of the modified yolov3 model of the present invention.
[ detailed description of the invention ]
Reference will now be made in detail to the present preferred embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like elements throughout.
As shown in fig. 1 to 5, a machine vision intelligent recognition system for a garbage target includes:
the odometer module 1 is used for calculating pose data of the robot;
the target recognition module 2 is used for searching the position of the junk target in the image by using a target recognition algorithm and calculating the position data and the depth data of the junk target;
and the mapping module 3 is used for receiving the pose data of the mileage meter processing module 1, the garbage target position data and the depth data of the target identification module 2, obtaining the coordinates of the garbage target in a world coordinate system, and splicing and constructing to obtain a point cloud map of the garbage target.
The odometer module 1 comprises a wheel type odometer 11 and a laser odometer 12, the laser odometer 12 comprises a 16-line laser radar, the wheel type odometer 11 comprises 2 photoelectric encoders, and the odometer module 1 estimates the pose of the robot in real time.
The target recognition module 2 comprises a high-definition camera.
The invention establishes a dynamic map of garbage target distribution based on the pose of the odometer module 1, reduces the monitoring pressure of managers and improves the working efficiency of cleaning personnel.
A machine vision intelligent identification method of a rubbish target comprises the following steps:
step A, calculating pose data:
a1, calibration of the wheeled odometer 11: collecting n data, each data including angular velocities W of two wheelsLAnd WRCalculating the distance b between the two wheels and the radius of the two wheels by a model-based method according to the duration of the data and the matching value of the laser radar;
a2, removing the motion distortion of the laser radar data: the industrial personal computer reads laser data, the STM32 uploads odometer integral data, whether the position and attitude data in an odometer queue are synchronous with the laser radar data in time is checked, if so, the processing is not carried out, otherwise, secondary interpolation is carried out on the position and attitude data, and the laser radar data is transformed according to the position and attitude data to obtain the laser radar data with motion distortion removed;
a3, using a front-end registration algorithm on the acquired lidar data: firstly, searching a matching point in the point cloud, then carrying out pose change and error calculation on the point cloud according to a matching calculation rotation matrix R and a translation matrix T, and continuously iterating until the error is smaller than a tolerance value.
A4, optimizing the pose graph based on a graph optimization technology;
step B, identifying a garbage target:
the target recognition module 2 finds the position of the junk target in the image by using a target recognition algorithm, and then issues position data and depth data of the junk target;
step C, establishing a point cloud map:
the mapping module 3 processes the data from the odometer module 1 and the target identification module 2, and performs coordinate transformation according to the following formula:
z=d
Figure BDA0002667897520000081
wherein, (x, z) is the coordinate of the garbage target in the camera coordinate system, u is the position of the target in the image coordinate system, d is the distance between the target and the camera, fxAnd cxAll the parameters are internal parameters of the camera, coordinates of the target under a camera coordinate system are transformed by using the optimized pose to obtain coordinates of the target under a world coordinate system, point cloud splicing and filtering (an outlier removal filter and a down-sampling filter) are performed, and finally a point cloud map is spliced and constructed;
step D, establishing a training set and a testing set:
d1, refining the definition of the garbage target: selecting a plurality of waste products as definitions of garbage targets, specifically selecting 14 common waste products in communities as definitions of garbage targets, wherein the waste products comprise garbage bags, express packaging bags, take-out packing boxes, cat and dog excrement, orange peel, banana peel, shaddock peel, mashed apples, cigarette boxes, thick paper boards, waste newspapers, plastic beverage bottles, beer bottles and pop cans;
d2, collecting photographs of the waste: acquiring a corresponding sample image by adopting a field acquisition mode and a crawling mode;
d3, preprocessing of the image: cropping the collected image data and transforming to 416 x 416 standard sample data;
d4, amplification of sample data: sample data is expanded in the following way:
d4-1. magnification of width and height of sample image 1.5 times;
d4-2. width reduction 1/3 and height reduction 1/2 of the sample image, and ensuring that the image is a multiple of 32;
d4-3, brightness enhancement of the sample image;
d4-4, the brightness of the sample image is reduced;
d4-5, turning the sample image by 90 degrees or 180 degrees;
d4-6, adding noise to the sample image;
d5, manual labeling: labeling the image data in the sample set by using an image labeling tool imageLabel according to the definition in the step D1 to obtain label data (x, y, w, h, c), wherein (x, y) represents the center coordinates of the garbage target in the image, w and h represent the width and height of the garbage target respectively, c represents the category of the garbage target, c is 0 to represent the background, and c is 1 to 14 to sequentially represent the garbage targets defined in the step D1;
d6, partitioning of sample set: dividing a sample set into a training set and a test set according to a ratio of 98: 2;
step E, constructing an improved yolov3 model:
respectively creating a convolutional layer, a depth separable convolutional layer, an upsampling layer, a residual layer, a splicing layer and a prediction layer, wherein the improved yolov3 model is composed of the 6 basic levels, the depth separable convolutional layer is used for replacing a3 multiplied by 3 convolutional layer in an original yolov3 residual module, the upsampling layer uses a bilinear interpolation algorithm to expand a feature map, the residual layer adds the output of a lower layer with the output of a higher layer, and the splicing layer adds the output of the higher layer containing rich semantics with the output of a lower layer with higher resolutionFor superposition, the prediction layer contains 3 feature maps, each grid of each feature map also contains 3 anchors box with different aspect ratios, each anchor box is a1 × 20 vector and contains the following prediction information: t is tx,ty,tw,thThe bounding box contains the confidence of the object and the probability that the object belongs to the garbage target;
step F, training the network:
initializing the weight of the model by using a random initialization strategy, then, throwing training data into the model, predicting each grid of the layer feature diagram after forward propagation to obtain a prediction result of 3 x (4+1+15), and finally, calculating the error between the prediction result and the real result according to the following cost function formula:
Figure BDA0002667897520000101
calculating an error between a predicted value and a true value through the cost function, then reversely propagating the error, updating the weight and deviation of the network through a momentum gradient descent strategy, and stopping training when the loss value of the cost function is small and stable, namely, convergence;
step G, network testing:
test set data is lost into the model, a prediction result is obtained, and a confidence threshold of a boundary box is set to be ThscoresAnd (5) rejecting the prediction result with low confidence coefficient, setting a non-maximum suppression threshold value of 0.5, executing non-maximum suppression, preventing the same target from being identified in multiple ways, and finally outputting the corrected prediction result.
According to the method, a part of traditional convolution in darknet-53 is replaced by the deep separable convolution, the calculated amount of the model is greatly reduced, the garbage target can be identified more accurately based on the improved yolov3 target identification algorithm, the parameters of the model are fewer, the time for training and optimizing is shortened, the identification speed of the model is increased, and the requirement on hardware is lowered.
The above examples are merely preferred embodiments of the present invention, and other embodiments of the present invention are possible, such as a reasonable combination of the technical solutions described in the examples. Those skilled in the art can make equivalent changes or substitutions without departing from the spirit of the present invention, and such equivalent changes or substitutions are included in the scope set forth in the claims of the present application.

Claims (5)

1. The utility model provides a machine vision intelligent recognition system of rubbish target which characterized in that: comprises that
The odometer module (1) is used for calculating pose data of the robot;
the target recognition module (2) searches the position of the junk target in the image by using a target recognition algorithm and calculates the position data and the depth data of the junk target;
and the mapping module (3) is used for receiving and processing the pose data of the odometer module (1), the garbage target position data and the depth data of the target identification module (2) to obtain the coordinates of the garbage target under a world coordinate system, and splicing and constructing to obtain a point cloud map of the garbage target.
2. The machine vision intelligent recognition system of a spam target according to claim 1, wherein: the odometer module (1) comprises a wheeled odometer (11) and a laser odometer (12).
3. The machine vision intelligent recognition system of a spam target according to claim 1, wherein: the target recognition module (2) comprises a high-definition camera.
4. A machine vision intelligent identification method of a rubbish target is characterized by comprising the following steps:
step A, calculating pose data:
a1, calibration of wheeled odometer (11): collecting a plurality of data segments, each data segment comprising angular velocities W of two wheelsLAnd WRCalculating the distance b between the two wheels and the radius of the two wheels by a model-based method according to the duration of the data and the matching value of the laser radar;
a2, removing the motion distortion of the laser radar data: the industrial personal computer reads laser data, the STM32 uploads odometer integral data, whether the position and attitude data in an odometer queue are synchronous with the laser radar data in time is checked, if so, the processing is not carried out, otherwise, secondary interpolation is carried out on the position and attitude data, and the laser radar data is transformed according to the position and attitude data to obtain the laser radar data with motion distortion removed;
a3, using a front-end registration algorithm on the acquired lidar data: firstly, searching a matching point in the point cloud, then carrying out pose change and error calculation on the point cloud according to a matching calculation rotation matrix R and a translation matrix T, and continuously iterating until the error is smaller than a tolerance value.
A4, optimizing the pose graph based on a graph optimization technology;
step B, identifying a garbage target:
the target recognition module (2) finds the position of the junk target in the image by using a target recognition algorithm, and then releases the position data and the depth data of the junk target;
step C, establishing a point cloud map:
the mapping module (3) processes data from the odometer module (1) and the target identification module (2), and performs coordinate transformation according to the following formula:
z=d
Figure FDA0002667897510000021
and transforming the coordinates of the target under the camera coordinate system by using the optimized pose to obtain the coordinates of the target under the world coordinate system, and finally splicing and constructing a point cloud map.
5. The machine vision intelligent recognition method of the junk target according to claim 4, further comprising the steps of:
step D, establishing a training set and a testing set:
d1, refining the definition of the garbage target: selecting a plurality of waste products as definitions of garbage targets;
d2, collecting photographs of the waste: acquiring a corresponding sample image by adopting a field acquisition mode and a crawling mode;
d3, preprocessing of the image: cropping the collected image data and transforming to 416 x 416 standard sample data;
d4, amplification of sample data: sample data is expanded in the following way:
d4-1. magnification of width and height of sample image 1.5 times;
d4-2. width reduction 1/3 and height reduction 1/2 of the sample image, and ensuring that the image is a multiple of 32;
d4-3, brightness enhancement of the sample image;
d4-4, the brightness of the sample image is reduced;
d4-5, turning the sample image by 90 degrees or 180 degrees;
d4-6, adding noise to the sample image;
d5, manual labeling: according to the definition in the step D1, labeling the image data in the sample set by using an image labeling tool to obtain label data (x, y, w, h, c);
d6, partitioning of sample set: dividing a sample set into a training set and a test set according to a ratio of 98: 2;
step E, constructing an improved yolov3 model:
respectively creating a convolutional layer, a depth separable convolutional layer, an upsampling layer, a residual layer, a splicing layer and a prediction layer, wherein the upsampling layer uses a bilinear interpolation algorithm to expand a feature map, the residual layer adds the output of a lower layer and the output of an upper layer, the splicing layer superposes the output of the upper layer containing rich semantics and the output of a lower layer with higher resolution, the prediction layer comprises 3 feature maps, each grid of each feature map also comprises 3 anchor boxes with different length-width ratios, and each anchor box is a1 × 20 vector and comprises the following prediction information: t is tx,ty,tw,thThe bounding box contains the confidence of the object and the probability that the object belongs to the garbage target;
step F, training the network:
initializing the weight of the model by using a random initialization strategy, then, throwing training data into the model, predicting each grid of the layer feature diagram after forward propagation to obtain a prediction result of 3 x (4+1+15), and finally, calculating the error between the prediction result and the real result according to the following cost function formula:
Figure FDA0002667897510000041
calculating an error between a predicted value and a true value through the cost function, then reversely propagating the error, updating the weight and deviation of the network through a momentum gradient descent strategy, and stopping training when the loss value of the cost function is small and stable, namely, convergence;
step G, network testing:
test set data is lost into the model, a prediction result is obtained, and a confidence threshold of a boundary box is set to be ThscoresAnd (5) rejecting the prediction result with low confidence coefficient, setting a non-maximum suppression threshold value of 0.5, executing non-maximum suppression, preventing the same target from being identified in multiple ways, and finally outputting the corrected prediction result.
CN202010924563.5A 2020-09-05 2020-09-05 Machine vision intelligent recognition system and method for garbage target Pending CN111914815A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010924563.5A CN111914815A (en) 2020-09-05 2020-09-05 Machine vision intelligent recognition system and method for garbage target

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010924563.5A CN111914815A (en) 2020-09-05 2020-09-05 Machine vision intelligent recognition system and method for garbage target

Publications (1)

Publication Number Publication Date
CN111914815A true CN111914815A (en) 2020-11-10

Family

ID=73267259

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010924563.5A Pending CN111914815A (en) 2020-09-05 2020-09-05 Machine vision intelligent recognition system and method for garbage target

Country Status (1)

Country Link
CN (1) CN111914815A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112560755A (en) * 2020-12-24 2021-03-26 中再云图技术有限公司 Target detection method for identifying urban exposed garbage
CN113033672A (en) * 2021-03-29 2021-06-25 西安电子科技大学 Multi-class optical image rotating target self-adaptive detection method based on feature enhancement
CN113436251A (en) * 2021-06-24 2021-09-24 东北大学 Pose estimation system and method based on improved YOLO6D algorithm
CN117421692A (en) * 2023-12-18 2024-01-19 深圳火眼智能有限公司 Garbage illegal delivery identification method, device and equipment for garbage delivery station

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110263675A (en) * 2019-06-03 2019-09-20 武汉联一合立技术有限公司 Garbage target recognition system and recognition method of community security robot
CN110796186A (en) * 2019-10-22 2020-02-14 华中科技大学无锡研究院 Dry and wet garbage identification and classification method based on improved YOLOv3 network
WO2020037960A1 (en) * 2018-08-21 2020-02-27 深圳大学 Sar target recognition method and apparatus, computer device, and storage medium
CN111360780A (en) * 2020-03-20 2020-07-03 北京工业大学 Garbage picking robot based on visual semantic SLAM

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020037960A1 (en) * 2018-08-21 2020-02-27 深圳大学 Sar target recognition method and apparatus, computer device, and storage medium
CN110263675A (en) * 2019-06-03 2019-09-20 武汉联一合立技术有限公司 Garbage target recognition system and recognition method of community security robot
CN110796186A (en) * 2019-10-22 2020-02-14 华中科技大学无锡研究院 Dry and wet garbage identification and classification method based on improved YOLOv3 network
CN111360780A (en) * 2020-03-20 2020-07-03 北京工业大学 Garbage picking robot based on visual semantic SLAM

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112560755A (en) * 2020-12-24 2021-03-26 中再云图技术有限公司 Target detection method for identifying urban exposed garbage
CN112560755B (en) * 2020-12-24 2022-08-19 中再云图技术有限公司 Target detection method for identifying urban exposed garbage
CN113033672A (en) * 2021-03-29 2021-06-25 西安电子科技大学 Multi-class optical image rotating target self-adaptive detection method based on feature enhancement
CN113033672B (en) * 2021-03-29 2023-07-28 西安电子科技大学 Multi-class optical image rotation target self-adaptive detection method based on feature enhancement
CN113436251A (en) * 2021-06-24 2021-09-24 东北大学 Pose estimation system and method based on improved YOLO6D algorithm
CN113436251B (en) * 2021-06-24 2024-01-09 东北大学 Pose estimation system and method based on improved YOLO6D algorithm
CN117421692A (en) * 2023-12-18 2024-01-19 深圳火眼智能有限公司 Garbage illegal delivery identification method, device and equipment for garbage delivery station
CN117421692B (en) * 2023-12-18 2024-04-09 深圳火眼智能有限公司 Garbage illegal delivery identification method, device and equipment for garbage delivery station

Similar Documents

Publication Publication Date Title
CN110263675B (en) Garbage target identification system and method of community security robot
CN111914815A (en) Machine vision intelligent recognition system and method for garbage target
CN111461245B (en) Wheeled robot semantic mapping method and system fusing point cloud and image
CN106845408B (en) Street garbage identification method under complex environment
CN107203781B (en) End-to-end weak supervision target detection method based on significance guidance
CN109858569A (en) Multi-tag object detecting method, system, device based on target detection network
CN102096821B (en) Number plate identification method under strong interference environment on basis of complex network theory
CN113076871B (en) Fish shoal automatic detection method based on target shielding compensation
CN113408584B (en) RGB-D multi-modal feature fusion 3D target detection method
CN112488025B (en) Double-temporal remote sensing image semantic change detection method based on multi-modal feature fusion
CN104156731A (en) License plate recognition system based on artificial neural network and method
CN110675421B (en) Depth image collaborative segmentation method based on few labeling frames
CN110032952B (en) Road boundary point detection method based on deep learning
CN112991534B (en) Indoor semantic map construction method and system based on multi-granularity object model
CN113469264A (en) Construction method of automatic garbage classification model, garbage sorting method and system
WO2022247045A1 (en) Laser radar information-based mobile robot location re-identification method
CN113065486A (en) Floater identification method, system, equipment and medium suitable for river channel
CN114022837A (en) Station left article detection method and device, electronic equipment and storage medium
CN111611956B (en) Rail detection method and system for subway visual image
CN115761674A (en) Road edge positioning detection method, equipment and medium
CN113129370B (en) Semi-supervised object pose estimation method combining generated data and label-free data
CN110853058B (en) High-resolution remote sensing image road extraction method based on visual saliency detection
CN113222025B (en) Feasible region label generation method based on laser radar
CN115359411B (en) Substation environment understanding method based on improved deep Lab V < 3+ > network
CN212624084U (en) Machine vision intelligent recognition device of rubbish target

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination