CN112784717B - Automatic pipe fitting sorting method based on deep learning - Google Patents
Automatic pipe fitting sorting method based on deep learning Download PDFInfo
- Publication number
- CN112784717B CN112784717B CN202110039092.4A CN202110039092A CN112784717B CN 112784717 B CN112784717 B CN 112784717B CN 202110039092 A CN202110039092 A CN 202110039092A CN 112784717 B CN112784717 B CN 112784717B
- Authority
- CN
- China
- Prior art keywords
- area
- pipe fitting
- network
- pipe
- camera
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 29
- 238000013135 deep learning Methods 0.000 title claims abstract description 15
- 238000002372 labelling Methods 0.000 claims description 12
- 230000009466 transformation Effects 0.000 claims description 9
- 238000001914 filtration Methods 0.000 claims description 8
- 230000007246 mechanism Effects 0.000 claims description 8
- 238000012549 training Methods 0.000 claims description 8
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 claims description 7
- 230000000694 effects Effects 0.000 claims description 7
- 238000011156 evaluation Methods 0.000 claims description 7
- 230000002146 bilateral effect Effects 0.000 claims description 4
- 239000011159 matrix material Substances 0.000 claims description 4
- 238000013461 design Methods 0.000 claims description 3
- 238000011176 pooling Methods 0.000 claims description 3
- 238000013519 translation Methods 0.000 claims description 3
- 230000009967 tasteless effect Effects 0.000 abstract description 2
- 238000001514 detection method Methods 0.000 description 9
- 238000010586 diagram Methods 0.000 description 4
- 230000011218 segmentation Effects 0.000 description 4
- 238000012360 testing method Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 210000000078 claw Anatomy 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B07—SEPARATING SOLIDS FROM SOLIDS; SORTING
- B07C—POSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
- B07C5/00—Sorting according to a characteristic or feature of the articles or material being sorted, e.g. by control effected by devices which detect or measure such characteristic or feature; Sorting by manually actuated devices, e.g. switches
- B07C5/36—Sorting apparatus characterised by the means used for distribution
- B07C5/361—Processing or control devices therefor, e.g. escort memory
- B07C5/362—Separating or distributor mechanisms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/02—Affine transformations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B07—SEPARATING SOLIDS FROM SOLIDS; SORTING
- B07C—POSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
- B07C2501/00—Sorting according to a characteristic or feature of the articles or material to be sorted
- B07C2501/0063—Using robots
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30108—Industrial image inspection
- G06T2207/30164—Workpiece; Machine component
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Biophysics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
Abstract
The invention belongs to the technical field of machine vision, and particularly relates to a method for automatically sorting pipe fittings based on deep learning. Aiming at the problems that the current manual sorting is high in cost, low in efficiency, complicated, tasteless and low in recognition degree by using a traditional method. The invention improves the MaskR-CNN algorithm, improves the recognition rate and mask judgment of the algorithm, and improves the recognition rate while ensuring the speed. And putting the pictures shot by the camera into a network to obtain a classification result and a mask. Judging the type and the size of the pipe fitting, and recording information. The pipe fitting grabbing point is positioned in a mode of Zhang Zhengyou calibration and eye-to-eye calibration outside the hand. The mechanical arm is used for grabbing and stacking the pipe fittings. The automatic sorting method for the pipe fittings has high efficiency and higher robustness for identifying and grabbing the pipe fittings in different environments; but wide application in mill's letter sorting, object classification and snatch.
Description
Technical Field
The invention belongs to the technical field of machine vision, and particularly relates to a method for automatically sorting pipe fittings based on deep learning.
Background
The visual technology of the research robot can improve the sensing ability of the robot to the external environment and greatly reduce the burden of workers, so that the workers are liberated from the complicated and tasteless working environment. At present, the labor demand of China is very high, the robot can be applied to the fields, the production cost is reduced, the working efficiency is improved, and the market competitiveness of enterprises is improved, so that industrial robots are used as core workers of the enterprises in many industries, and the production efficiency of the enterprises is improved.
In industrial robot's pipe fitting discernment field, the mode of traditional study is replaced in the accessible degree of depth study, makes the recognition effect of robot better, improves the recognition rate of part, and the mask profile through degree of depth study training is influenced less by external factors (illumination, shade), consequently can use the mask area to judge the yardstick of pipe fitting, carries out the pipe fitting more accurately and snatchs and route planning, improves and snatchs efficiency, guarantees the accuracy, stability and the practicality of operation better, has compensatied the not high defect of traditional approach recognition rate.
Conventional industrial robots generally can only recognize the same type and size of pipe fittings, and if various types and sizes of pipe fittings are to be sorted, a plurality of robots are required. The invention can use a single robot to grab the pipe fittings with multiple classes and scales.
Therefore, the automatic pipe sorting research based on deep learning has extremely important practical significance.
Disclosure of Invention
Aiming at the boring and fussy sorting of the pipe fittings, deep learning is introduced to classify the categories and the sizes of the pipe fittings. The invention provides a method for automatically sorting pipe fittings based on deep learning. The method is suitable for both static and dynamic pipe sorting (grabbing and stacking of pipes on a conveyor belt).
In order to achieve the purpose, the invention adopts the following technical scheme:
a method for automatically sorting pipes based on deep learning comprises the following steps:
step 1, labeling categories and mask outlines of collected pipe fitting images by using a labelme tool to obtain a pipe fitting image of a json type labeling result;
step 3, designing a network structure, and reading the data set in the step 2 into the network for training;
step 5, judging different sizes of the same type of pipe fittings according to the area of the mask, and calculating to obtain a grabbing point;
step 6, converting a camera and a robot coordinate system, and calculating a pipe fitting grasping point of the coordinate system at the tail end of the mechanical arm;
and 7, connecting the computer and the camera through a gige gigabit network port, setting a port number of the network port of the computer connected with the camera to be 192.169.1.X which is the same as that of the camera, connecting the computer and the robot mechanical arm through a TCP/IP protocol, compiling a RAPID language by using Robots studio software, controlling shooting and grabbing operations of the camera and the robot, and receiving feedback of the robot.
Further, in the step 2, data enhancement is performed on the pipe fitting image with the labeling result, and the specific steps are as follows: and rotating, overturning, blurring, Gaussian filtering, bilateral filtering and white noise adding are carried out on the pipe fitting image with the labeling result.
Further, the step 3 designs a network structure, specifically:
forming a backbone network by Resnet-101-FPN, using a 7 multiplied by 7 convolution kernel with the step length of 2, wherein the number of channels is 64; followed by a 3 x 3 convolution kernel with step size 2 and maximum pooling operation; the second layer convolution is composed of step length of 1, three convolution kernels of 1 × 1, 3 × 3 and 1 × 1, and channel numbers of 64, 64 and 256; the third layer of convolution comprises the steps of 1, four convolution kernels of 1 × 1, 3 × 3 and 1 × 1, and the number of channels of 128, 128 and 512; the fourth layer of convolution consists of 1 step length, twenty three convolution kernels of 1 multiplied by 1, 3 multiplied by 3 and 1 multiplied by 1, and the number of channels of 256, 256 and 1024; the last layer of convolution has step length of 1, three convolution kernels of 1 × 1, 3 × 3 and 1 × 1 and channel numbers of 512, 512 and 2048;
secondly, using an area generation network to adopt 5 scales with different area sizes, wherein each scale comprises 3 different aspect ratios, generating 15 different candidate frames by using each pixel point in a feature map in total, and generating 9 different candidate frames by considering that the sizes of the pipe fittings adopt the scales of 32 × 32, 64 × 64 and 128 × 128 and the aspect ratios of 0.8:1, 1:1 and 1:1.25 according to the generated candidate frames; scanning the image through a sliding window so as to find a regional target; the types and the pixels of the searched target areas are classified respectively, and a mixed attention mechanism is added before the pixels of the searched target areas are predicted to improve the prediction effect.
Still further, the mixed attention mechanism is a mixture of channel attention and spatial attention.
Further, judge the different sizes of the pipe fittings of the same type through the mask area in step 5, the concrete step is:
calculating the area of the marked mask outline in the step 1, sequentially placing the area into an m x n-dimensional array, and calculating area [ ] [ ]; obtaining the type of the pipe fitting through the network in the step 4, finding the row position of the area array corresponding to the pipe fitting, marking as q, calculating the mask outline obtained in the network, calculating the pixel area of the mask, comparing the pixel area with the data in the area [ q ] [ ], finding the position of the value which is most similar to the area array, and counting as p, so the area [ q ] [ p ] is the obtained position, and judging the size of the pipe fitting.
Further, the step 5 of calculating the grabbing point includes the following specific steps:
firstly, calculating the end points of all four polygons of the pipe fitting, namely xminP, xmaxP, yminP and ymaxP, and constructing four cuts of P through the four pointsDetermining two 'shell clamping' sets, if one or two lines coincide with one edge, calculating the area of a rectangle determined by the four lines, and storing the area as the current minimum value; otherwise, defining the current minimum value as infinity, rotating the line clockwise until one line is coincident with one edge of the polygon, calculating the area of a new rectangle, comparing the area with the current minimum value, updating if the area is smaller than the current minimum value, storing the rectangle information for determining the minimum value, repeating the operation until the angle of the line rotation is larger than 90 degrees, obtaining the minimum area of the external rectangle of the pipe fitting, namely obtaining the minimum external rectangle of the pipe fitting, and counting two adjacent end points of the external rectangle as (X)1,Y1)、(X2,Y2) So that the center point of the pipe isThis point is the robot's grasping point.
Further, in step 6, the coordinate system of the camera and the robot is converted, and a pipe gripping point of the coordinate system at the tail end of the robot arm is calculated, and the method specifically comprises the following steps:
the camera calibration method of the planar target of Zhangyingyou is used for obtaining the position information of the pipe in a camera coordinate system, the position corresponding relation between the robot coordinate system and the camera coordinate system is obtained through hand-eye calibration, affine transformation is used, and a matrix of the affine transformation is as follows:
X=m11×x+m12×y+m13
Y=m21×x+m22×y+m23
Z=m31×x+m32×y+m33
where, (x, y) is a coordinate point of an image pixel, (x ', y') is a coordinate converted to the end of the robot arm, m11, m12, m21, m22, m31, m32 are rotation amounts, and m13, m23, m33 are translation amounts.
And planning the path of the grasping point of the robot, and calculating by using a random path graph algorithm, wherein the robot obtains an optimal path while avoiding colliding with the barrier between the initial position and the target position.
Compared with the prior art, the invention has the following advantages: present prior art, the overwhelming majority of robot letter sorting is applicable to the pipe fitting and neatly puts, and fixed position snatchs the work. And the problems of single grabbing type, low identification precision and the like exist. The pipe fitting sorting machine can sort pipe fittings of different types and sizes through the same mechanical claw, and the pipe fittings do not need to be placed in order, so that the use cost and the time cost in the actual process are saved. In terms of precision, compared with other network structures, the network structure provided by the invention has the advantages that the recognition precision is improved, and the problems of mistaken grabbing, mistaken grabbing and the like in the grabbing process are further reduced.
Drawings
FIG. 1 is a data set annotation interface diagram;
FIG. 2 is a network layout of the present method;
FIG. 3 is a model diagram of a channel attention mechanism;
FIG. 4 is a model diagram of a spatial attention mechanism;
FIG. 5 is a pipe image artwork;
FIG. 6 is a result of network instance segmentation in accordance with the present invention;
fig. 7 is a schematic of the robot and conveyor belt.
Detailed Description
A method for automatically sorting pipes based on deep learning comprises the following steps:
step 1, shooting images of pipe fitting types such as a tee joint, an outer wire, an inner wire, an elbow and the like, and carrying out type and mask outline labeling on the collected pipe fitting images by using a labelme tool to obtain a pipe fitting image of a json type labeling result;
And 3, designing a network structure, reading the data set in the step 2 into the network for training, taking 3000 pipes as a training set and 1000 pipes as a testing set, wherein the batch size is 3, and the learning rate is 0.0001.
The invention designs a network structure as shown in figure 2, a backbone network is formed by Resnet-101-FPN, 7 multiplied by 7 convolution kernels with the step length of 2 are used, and the number of channels is 64; followed by a 3 x 3 convolution kernel with step size 2 and maximum pooling; the second layer convolution is composed of 1 step length, 1 × 1, 3 × 3 and 1 × 1 convolution kernels, 64 and 256 channels; the third layer of convolution has the step length of 1, four convolution kernels of 1 × 1, 3 × 3 and 1 × 1 and the number of channels of 128, 128 and 512; the fourth layer of convolution consists of 1 step length, twenty three convolution kernels of 1 multiplied by 1, 3 multiplied by 3 and 1 multiplied by 1, and the number of channels of 256, 256 and 1024; the last layer of convolution has step length of 1, three convolution kernels of 1 × 1, 3 × 3 and 1 × 1 and channel numbers of 512, 512 and 2048; because the low-level characteristic semantic information is less, the target position is accurate; the feature semantic information of the high layer is rich, but the target position is rough. If the information of the upper layer and the information of the lower layer are fused, the method has better effect in identification. Features in the image are fused in a top-down manner. Finally, a characteristic diagram with five scales is achieved.
Using a regional generation network to adopt 5 scales with different area sizes, wherein each scale comprises 3 different aspect ratios, generating 15 different candidate frames by using each pixel point in a feature map, and generating 9 different candidate frames by considering that the sizes of the pipe fittings adopt the scales of 32 × 32, 64 × 64 and 128 × 128 and the aspect ratios of 0.8:1, 1:1 and 1:1.25 according to the generated candidate frames; through the improvement, the training can be better converged, the occupied video memory is smaller, and the classification, regression and segmentation effects on small targets are improved. Scanning the image through a sliding window so as to find a regional target; the types and the pixels of the searched target areas are classified respectively, and a mixed attention mechanism is added before the pixels of the searched target areas are predicted to improve the prediction effect. As shown in fig. 3 and 4, the basic idea of the attention mechanism is to let the model ignore irrelevant information and focus more on the important information that we want it to focus on.
accuracy (Precision): in all the targets of model prediction, the Precision calculation formula is as follows:
recall (Recall): in all real targets, the model predicts the correct target proportion of the pipe fitting, namely Recall ratio, and the Recall calculation formula is as follows:
IoU, the ratio of the intersection and union of the "predicted bounding box" and the "true bounding box" is calculated. The calculation formula is as follows:
in the formula, the number of detection frames of TP (true Positive) IoU >0.5 is calculated only once by the same GT (ground Truth) frame.
FP (false Positive): IoU < 0.5, i.e. the number of redundant detection boxes for detecting the same GT box.
Fn (false negative): there is no detected number of GT boxes.
And acquiring and recording the category and the mask outline of the pipe fitting with the highest score. The predicted image and the result are shown in fig. 5 and 6.
Step 5, judging different sizes of the same type of pipe fittings according to the area of the mask: calculating the area of the marked mask outline in the step 1, sequentially placing the area into an m x n-dimensional array, and calculating area [ ] [ ]; obtaining the type of the pipe fitting through the network in the step 4, finding the row position of the area array corresponding to the pipe fitting, marking as q, calculating the mask outline obtained in the network, obtaining the pixel area of the mask, comparing with the data in the area [ q ] [ ], finding the position of the value which is most similar to the area array, and counting as p, so the area [ q ] [ p ] is the obtained position, and judging the size of the pipe fitting.
Calculating to obtain a grabbing point: firstly, calculating end points of all four polygons of the pipe fitting, namely xminP, xmaxP, yminP and ymaxP, constructing four tangent lines of P through the four points, determining two 'clamping shell' sets, if one or two lines coincide with one edge, calculating the area of a rectangle determined by the four lines, and storing the area as the current minimum value; otherwise, defining the current minimum value as infinity, rotating the line clockwise until one line is coincident with one edge of the polygon, calculating the area of a new rectangle, comparing the area with the current minimum value, updating if the area is smaller than the current minimum value, storing the rectangle information for determining the minimum value, repeating the operation until the angle of the line rotation is larger than 90 degrees, obtaining the minimum area of the external rectangle of the pipe fitting, namely obtaining the minimum external rectangle of the pipe fitting, and counting two adjacent end points of the external rectangle as (X)1,Y1)、(X2,Y2) So that the center point of the pipe isThis point is the robot's grasping point.
And 6, converting a camera and a robot coordinate system, and calculating a pipe gripping point of the coordinate system at the tail end of the mechanical arm: the camera calibration method of the planar target of Zhangyingyou is used for obtaining the position information of the pipe in a camera coordinate system, the position corresponding relation between the robot coordinate system and the camera coordinate system is obtained through hand-eye calibration, affine transformation is used, and a matrix of the affine transformation is as follows:
X=m11×x+m12×y+m13
Y=m21×x+m22×y+m23
Z=m31×x+m32×y+m33
where, (x, y) is a coordinate point of an image pixel, (x ', y') is a coordinate converted to the end of the arm, m11, m12, m21, m22, m31, m32 are rotation amounts, and m13, m23, m33 are translation amounts.
The method for acquiring the transformation parameters comprises the following steps: four spots on the tray were randomly found and marked with a vivid pen. The conveyer belt is started, and when the tray on the conveyer belt reaches the lower part of the camera, the camera shoots the tray. And (4) extracting the position information of the picture marked by the bright pen from the photographed picture. And manually operating the mechanical arm, respectively moving the tail end of the mechanical arm to four positions marked by using bright-colored pens, and respectively recording x and y values in a coordinate system of the mechanical arm at the moment. And obtaining a transformation parameter m through an affine transformation matrix. It is written as a function. And calling a function each time, and converting the camera coordinate system of the previously calculated pipe fitting central point into coordinates in the coordinate system of the mechanical arm.
And 7, connecting the computer and the camera through a gige gigabit network port, setting a port number of the network port of the computer connected with the camera to be 192.169.1.X which is the same as that of the camera, connecting the computer and the robot mechanical arm through a TCP/IP protocol, compiling a RAPID language by using Robots studio software, controlling shooting and grabbing operations of the camera and the robot, and receiving feedback of the robot.
The robot and conveyor belt schematic 7 shows that the conveyor belt and the mechanical arm are controlled, and the state of the conveyor belt, i.e. starting, stopping and whether the gripping point is reached, is controlled and judged through IO flow. And when the program judges that the tray on the conveyor belt reaches the shooting point, giving a computer signal to control the camera to take pictures. And (4) storing the pictures, reading the pictures into the trained network, and acquiring the position of the pipe fitting in the terminal coordinate system of the mechanical arm by using the steps 3 and 4. As TCP/IP can only transmit one character string each time, the grabbing point and the stacking point of the pipe fitting are added with "/" in the middle to form a long character string, and the long character string is transmitted to the robot by using a computer. The robot receives the character string transmitted by the computer and decodes the character string by judging '/'. And recording the grabbing points and the stacking points of the pipe fittings, and planning the path of the mechanical arm by using a random path graph algorithm. And executing the grabbing task to finish the pipe fitting grabbing task once. And starting the transmission belt to wait for the next grabbing task.
In this embodiment, the network designed by the present invention is evaluated by using the map (mean average precision) and mrecall (mean call), and the maximum number of detections per picture (maxDets) is 100. The target detection performance evaluation is shown in table 1, and the mask detection performance evaluation is shown in table 2. Experimental results show that the network model provided by the invention has better target detection and Mask segmentation effects compared with the current popular segmentation networks (Solo, Mask R-CNN and Deepmask). Compared with the basic MaskR-CNN model, the accuracy and the regression rate of target detection are respectively improved by 3.5 percent and 7.1 percent, and the accuracy and the regression rate of mask detection are improved by 2.8 percent and 2.4 percent.
TABLE 1 comparison of target detection Performance evaluation
TABLE 2 mask test Performance evaluation comparison
According to the method, different sizes of the same type of pipe fittings are distinguished by calculating the area of the mask. Selecting a three-way pipe fitting for testing, photographing the three-way pipe fitting at the same height by the same camera, calculating the area of a pixel point of the three-way pipe fitting, and comparing the areas of the three-way pipe fitting with each other as shown in table 3. In table 3, the pixel point areas calculated by the network of the present invention in the half inch, one inch and one inch half tee pipe are all closer to the manually labeled areas than the basic Mask R-CNN. Therefore, there is a great advantage in determining the tube size by using the tube mask pixels.
TABLE 3 tee pixel area comparison
Those skilled in the art will appreciate that the invention may be practiced without these specific details. Although illustrative embodiments of the present invention have been described above to facilitate the understanding of the present invention by those skilled in the art, it should be understood that the present invention is not limited to the scope of the embodiments, and various changes may be made apparent to those skilled in the art as long as they are within the spirit and scope of the present invention as defined and defined by the appended claims, and all matters of the invention which utilize the inventive concepts are protected.
Claims (7)
1. A method for automatically sorting pipe fittings based on deep learning is characterized by comprising the following steps:
step 1, labeling categories and mask outlines of collected pipe fitting images by using a labelme tool to obtain a pipe fitting image of a json type labeling result;
step 2, performing data enhancement on the pipe fitting image with the labeling result to obtain a data set;
step 3, designing a network structure, and reading the data set in the step 2 into the network for training;
step 4, reading the network weight trained in the step 3 into a network by using the accuracy, the recall rate and the intersection ratio evaluation judgment, and reading a to-be-predicted image into the network for prediction to obtain the category and the mask outline of the pipe fitting;
step 5, judging different sizes of the same type of pipe fittings according to the area of the mask, and calculating to obtain a grabbing point;
step 6, converting a camera and a robot coordinate system, and calculating a pipe fitting grasping point of the coordinate system at the tail end of the mechanical arm;
and 7, connecting the computer and the camera through a gige gigabit network port, setting a port number of the network port of the computer connected with the camera to be 192.169.1.X which is the same as that of the camera, connecting the computer and the robot mechanical arm through a TCP/IP protocol, compiling a RAPID language by using Robots studio software, controlling shooting and grabbing operations of the camera and the robot, and receiving feedback of the robot.
2. The method for automatically sorting the pipes based on the deep learning as claimed in claim 1, wherein the step 2 of performing data enhancement on the pipe image with the labeled result comprises the following specific steps: and rotating, overturning, blurring, Gaussian filtering, bilateral filtering and white noise adding are carried out on the pipe fitting image with the labeling result.
3. The method for automatically sorting pipes based on deep learning according to claim 1, wherein the step 3 is to design a network structure, specifically:
forming a backbone network by Resnet-101-FPN, using a 7 multiplied by 7 convolution kernel with the step length of 2, wherein the number of channels is 64; followed by a 3 x 3 convolution kernel with step size 2 and maximum pooling; the second layer convolution is composed of step length of 1, three convolution kernels of 1 × 1, 3 × 3 and 1 × 1, and channel numbers of 64, 64 and 256; the third layer of convolution comprises the steps of 1, four convolution kernels of 1 × 1, 3 × 3 and 1 × 1, and the number of channels of 128, 128 and 512; the fourth layer of convolution consists of 1 step length, twenty three convolution kernels of 1 multiplied by 1, 3 multiplied by 3 and 1 multiplied by 1, and the number of channels of 256, 256 and 1024; the last layer of convolution has step length of 1, three convolution kernels of 1 × 1, 3 × 3 and 1 × 1 and channel numbers of 512, 512 and 2048;
secondly, using an area generation network to adopt 5 scales with different area sizes, wherein each scale comprises 3 different aspect ratios, generating 15 different candidate frames by using each pixel point in a feature map in total, and generating 9 different candidate frames by considering that the sizes of the pipe fittings adopt the scales of 32 × 32, 64 × 64 and 128 × 128 and the aspect ratios of 0.8:1, 1:1 and 1:1.25 according to the generated candidate frames; scanning the image through a sliding window so as to find a regional target; the types and the pixels of the searched target areas are classified respectively, and a mixed attention mechanism is added before the pixels of the searched target areas are predicted to improve the prediction effect.
4. The method of claim 3, wherein the hybrid attention mechanism is a hybrid of channel attention and spatial attention.
5. The method for automatically sorting pipes based on deep learning as claimed in claim 1, wherein the step 5 is to judge different sizes of pipes of the same category by mask area, and comprises the following specific steps:
calculating the area of the marked mask outline in the step 1, sequentially placing the area into an m x n-dimensional array, and calculating area [ ] [ ]; obtaining the type of the pipe fitting through the network in the step 4, finding the row position of the area array corresponding to the pipe fitting, marking as q, calculating the mask outline obtained in the network, calculating the pixel area of the mask, comparing the pixel area with the data in the area [ q ] [ ], finding the position of the value which is most similar to the area array, and counting as p, so the area [ q ] [ p ] is the obtained position, and judging the size of the pipe fitting.
6. The method for automatically sorting pipes based on deep learning according to claim 1, wherein the grasping point is calculated in the step 5, and the method comprises the following specific steps:
firstly, calculating end points of all four polygons of the pipe fitting, namely xminP, xmaxP, yminP and ymaxP, constructing four tangent lines of P through the four points, determining two 'clamping shell' sets, if one or two lines coincide with one edge, calculating the area of a rectangle determined by the four lines, and storing the area as the current minimum value; otherwise, defining the current minimum value as infinity, rotating the line clockwise until one line is coincident with one edge of the polygon, calculating the area of a new rectangle, comparing the area with the current minimum value, updating if the area is smaller than the current minimum value, storing the rectangle information for determining the minimum value, repeating the operation until the angle of the line rotation is larger than 90 degrees, obtaining the minimum area of the external rectangle of the pipe fitting, namely obtaining the minimum external rectangle of the pipe fitting, and counting two adjacent end points of the external rectangle as (X)1,Y1)、(X2,Y2) So that the center point of the pipe isThis point is the grasping point of the robot.
7. The method for automatically sorting the pipes based on the deep learning as claimed in claim 1, wherein the step 6 is to convert a coordinate system of the camera and the robot and calculate a pipe grabbing point of the coordinate system at the tail end of the robot arm, and the method comprises the following specific steps:
the camera calibration method of the planar target of Zhangyingyou is used for obtaining the position information of the pipe in a camera coordinate system, the position corresponding relation between the robot coordinate system and the camera coordinate system is obtained through hand-eye calibration, affine transformation is used, and a matrix of the affine transformation is as follows:
X=m11×x+m12×y+m13
Y=m21×x+m22×y+m23
Z=m31×x+m32×y+m33
where, (x, y) is a coordinate point of an image pixel, (x ', y') is a coordinate converted to the end of the robot arm, m11, m12, m21, m22, m31, m32 are rotation amounts, and m13, m23, m33 are translation amounts.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110039092.4A CN112784717B (en) | 2021-01-13 | 2021-01-13 | Automatic pipe fitting sorting method based on deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110039092.4A CN112784717B (en) | 2021-01-13 | 2021-01-13 | Automatic pipe fitting sorting method based on deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112784717A CN112784717A (en) | 2021-05-11 |
CN112784717B true CN112784717B (en) | 2022-05-13 |
Family
ID=75755550
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110039092.4A Active CN112784717B (en) | 2021-01-13 | 2021-01-13 | Automatic pipe fitting sorting method based on deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112784717B (en) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113894058B (en) * | 2021-09-06 | 2024-03-01 | 东莞职业技术学院 | Quality detection and sorting method, system and storage medium based on deep learning |
CN113731857B (en) * | 2021-09-15 | 2023-04-25 | 重庆大学 | Automatic sorting system for small package medicine bags and control method thereof |
CN116188559A (en) * | 2021-11-28 | 2023-05-30 | 梅卡曼德(北京)机器人科技有限公司 | Image data processing method, device, electronic equipment and storage medium |
CN114648702B (en) * | 2022-04-02 | 2024-08-09 | 湖北工业大学 | Lobster sorting method, system, medium, equipment and terminal |
CN116071605B (en) * | 2023-03-07 | 2023-09-01 | 超音速人工智能科技股份有限公司 | Deep learning-based labeling method, device and storage medium |
CN116386028B (en) * | 2023-04-06 | 2023-10-03 | 扬州市管件厂有限公司 | Image layering identification method and device for processing tee pipe fitting |
CN117132828B (en) * | 2023-08-30 | 2024-03-19 | 常州润来科技有限公司 | Automatic classification method and system for solid waste in copper pipe machining process |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107024989A (en) * | 2017-03-24 | 2017-08-08 | 中北大学 | A kind of husky method for making picture based on Leap Motion gesture identifications |
CN108109174A (en) * | 2017-12-13 | 2018-06-01 | 上海电气集团股份有限公司 | A kind of robot monocular bootstrap technique sorted at random for part at random and system |
CN108126914A (en) * | 2017-11-24 | 2018-06-08 | 上海发那科机器人有限公司 | More object robots method for sorting at random in a kind of material frame based on deep learning |
CN108171748A (en) * | 2018-01-23 | 2018-06-15 | 哈工大机器人(合肥)国际创新研究院 | A kind of visual identity of object manipulator intelligent grabbing application and localization method |
CN108550169A (en) * | 2018-04-24 | 2018-09-18 | 中北大学 | The computational methods of the determination of pieces of chess position and its height in three dimensions |
CN110751691A (en) * | 2019-09-24 | 2020-02-04 | 同济大学 | Automatic pipe fitting grabbing method based on binocular vision |
CN111368852A (en) * | 2018-12-26 | 2020-07-03 | 沈阳新松机器人自动化股份有限公司 | Article identification and pre-sorting system and method based on deep learning and robot |
CN112170233A (en) * | 2020-09-01 | 2021-01-05 | 燕山大学 | Small part sorting method and system based on deep learning |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10788836B2 (en) * | 2016-02-29 | 2020-09-29 | AI Incorporated | Obstacle recognition method for autonomous robots |
-
2021
- 2021-01-13 CN CN202110039092.4A patent/CN112784717B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107024989A (en) * | 2017-03-24 | 2017-08-08 | 中北大学 | A kind of husky method for making picture based on Leap Motion gesture identifications |
CN108126914A (en) * | 2017-11-24 | 2018-06-08 | 上海发那科机器人有限公司 | More object robots method for sorting at random in a kind of material frame based on deep learning |
CN108109174A (en) * | 2017-12-13 | 2018-06-01 | 上海电气集团股份有限公司 | A kind of robot monocular bootstrap technique sorted at random for part at random and system |
CN108171748A (en) * | 2018-01-23 | 2018-06-15 | 哈工大机器人(合肥)国际创新研究院 | A kind of visual identity of object manipulator intelligent grabbing application and localization method |
CN108550169A (en) * | 2018-04-24 | 2018-09-18 | 中北大学 | The computational methods of the determination of pieces of chess position and its height in three dimensions |
CN111368852A (en) * | 2018-12-26 | 2020-07-03 | 沈阳新松机器人自动化股份有限公司 | Article identification and pre-sorting system and method based on deep learning and robot |
CN110751691A (en) * | 2019-09-24 | 2020-02-04 | 同济大学 | Automatic pipe fitting grabbing method based on binocular vision |
CN112170233A (en) * | 2020-09-01 | 2021-01-05 | 燕山大学 | Small part sorting method and system based on deep learning |
Non-Patent Citations (2)
Title |
---|
一种面向机器人分拣的杂乱工件视觉检测识别方法;谢先武等;《高技术通讯》;20180415(第04期);344-353 * |
基于感兴趣区域的机器人抓取系统;马世超等;《科学技术与工程》;20200418(第11期);4395-4403 * |
Also Published As
Publication number | Publication date |
---|---|
CN112784717A (en) | 2021-05-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112784717B (en) | Automatic pipe fitting sorting method based on deep learning | |
CN112170233B (en) | Small part sorting method and system based on deep learning | |
US11878433B2 (en) | Method for detecting grasping position of robot in grasping object | |
CN108010078B (en) | Object grabbing detection method based on three-level convolutional neural network | |
CN108827316B (en) | Mobile robot visual positioning method based on improved Apriltag | |
CN112949564A (en) | Pointer type instrument automatic reading method based on deep learning | |
CN105046252B (en) | A kind of RMB prefix code recognition methods | |
CN110598609A (en) | Weak supervision target detection method based on significance guidance | |
CN110991435A (en) | Express waybill key information positioning method and device based on deep learning | |
CN108126914B (en) | Deep learning-based robot sorting method for scattered multiple objects in material frame | |
CN115816460B (en) | Mechanical arm grabbing method based on deep learning target detection and image segmentation | |
CN104268602A (en) | Shielded workpiece identifying method and device based on binary system feature matching | |
CN114693661A (en) | Rapid sorting method based on deep learning | |
CN115272204A (en) | Bearing surface scratch detection method based on machine vision | |
CN116228854B (en) | Automatic parcel sorting method based on deep learning | |
CN113657423A (en) | Target detection method suitable for small-volume parts and stacked parts and application thereof | |
CN109615610B (en) | Medical band-aid flaw detection method based on YOLO v2-tiny | |
CN114663382A (en) | Surface defect detection method for electronic component based on YOLOv5 convolutional neural network | |
CN116863463A (en) | Egg assembly line rapid identification and counting method | |
CN114800533B (en) | Sorting control method and system for industrial robot | |
CN114708234B (en) | Method and device for identifying number of detonators on automatic bayonet coding all-in-one machine | |
CN110334818A (en) | A kind of method and system of pipeline automatic identification | |
CN115753765A (en) | Method for identifying position and posture of cheese | |
CN114842335A (en) | Slotting target identification method and system for construction robot | |
CN111783537A (en) | Two-stage rapid grabbing detection method based on target detection characteristics |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |