CN116128960A - Automatic workpiece grabbing method, system and device based on machine learning - Google Patents

Automatic workpiece grabbing method, system and device based on machine learning Download PDF

Info

Publication number
CN116128960A
CN116128960A CN202310253672.2A CN202310253672A CN116128960A CN 116128960 A CN116128960 A CN 116128960A CN 202310253672 A CN202310253672 A CN 202310253672A CN 116128960 A CN116128960 A CN 116128960A
Authority
CN
China
Prior art keywords
workpiece
neural network
machine learning
convolutional neural
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN202310253672.2A
Other languages
Chinese (zh)
Inventor
梁建安
王虹鑫
李凯
池家聪
张旭
宋苗苗
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanxi University
Original Assignee
Shanxi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanxi University filed Critical Shanxi University
Priority to CN202310253672.2A priority Critical patent/CN116128960A/en
Publication of CN116128960A publication Critical patent/CN116128960A/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1612Programme controls characterised by the hand, wrist, grip control
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1694Programme controls characterised by use of sensors other than normal servo-feedback from position, speed or acceleration sensors, perception control, multi-sensor controlled systems, sensor fusion
    • B25J9/1697Vision controlled systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Mechanical Engineering (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Robotics (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Orthopedic Medicine & Surgery (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to the technical field of machine learning, image recognition and binocular vision, and discloses a method, a system and a device for automatically grabbing workpieces based on machine learning, aiming at the problems that the existing automatic grabbing of workpieces is complex in system composition, inaccurate in workpiece category recognition, difficult to expand the types of grabbed targets, inefficient, inaccurate and the like in the application process. The device comprises a system and a mechanical arm control assembly, wherein the system comprises a binocular vision assembly, a data preprocessing frame, a deep convolutional neural network and an information flow; according to the method, target detection information output by the deep convolutional neural network is fused with image information of the binocular vision component through an information flow, the complexity of the system is reduced, the data set is randomly extracted by the data preprocessing frame to form a training set and a verification set, the management of the workpiece library types is completed, the recognition of multiple types of workpieces is realized by the deep neural network, the system is more efficient, easier to expand and more accurate to classify and position, and the automatic grabbing task of the workpieces is completed.

Description

Automatic workpiece grabbing method, system and device based on machine learning
The application is a divisional application of a patent application named as an automatic workpiece grabbing system and method based on machine learning, and the application date of the original application is 2021, 09, 17 and 202111095314.0.
Technical Field
The invention belongs to the technical field of machine learning, image recognition and binocular vision, and particularly relates to a workpiece automatic grabbing method, system and device based on machine learning.
Background
The task of the automatic workpiece grabbing technology is to finish the identification of the tools with the specified types, provide depth information and grabbing point information of target workpieces for a mechanical grabbing arm, effectively identify all workpiece types in a workpiece library and manage the identified types as required.
At present, the automatic workpiece grabbing technology mainly relies on means such as target feature extraction and feature matching to classify and position, and meanwhile, the measurement of depth information of a workpiece generally adopts a laser ranging or multi-graph triangulation mode, so that the following defects exist:
firstly, it is: in the partial automatic grabbing technology, an active visual scene measurement method is adopted for the spatial positioning of a workpiece, a visual component realizes the construction of a three-dimensional space by emitting laser outwards, and a point cloud image formed by a laser sensor is utilized to acquire the spatial positioning of the workpiece, so that a system is complex and the hardware requirement is high;
secondly, it is: the target detection of the workpiece adopts a characteristic matching and lightweight neural network mode, so that the effect of target classification and positioning is poor, the type expansion of the identifiable workpiece is difficult, and the multi-target detection capability is weak;
thirdly, the method comprises the following steps: the gripping of workpieces is generally performed by a geometric guiding method, so that the gripping efficiency of the workpieces is low under a dynamic condition.
Disclosure of Invention
Aiming at the problems that the existing automatic workpiece grabbing is complex in system composition, inaccurate in workpiece category identification, difficult to expand the type of the grabbed target, inefficient, inaccurate and the like in the application process, the invention provides an automatic workpiece grabbing method, an automatic workpiece grabbing system and an automatic workpiece grabbing device based on machine learning.
In order to achieve the above purpose, the present invention adopts the following technical scheme:
the invention provides a machine learning-based automatic workpiece grabbing method, which comprises the following steps of:
step 1, image acquisition and pretreatment;
collecting multiple gesture images of a workpiece in different environments; obtaining a data set and a label according to the gesture image;
step 2, training a deep convolutional neural network Net 1;
training a deep convolutional neural network by adopting the data set and the label;
obtaining parameters of the deep convolutional neural network by taking the minimum value of the loss function of the deep convolutional neural network as a target;
updating the weight of the deep convolutional neural network by adopting an Adam optimizer to obtain an updated weight;
calculating a loss variation value of the loss function according to the updated weight and the parameter;
if the loss change value is smaller than a set threshold value, taking the depth convolution neural network corresponding to the loss value as a converged depth convolution neural network;
step 3, adopting K-fold to cross and verify the weight of the deep convolutional neural network Net1 obtained by training in the step 2;
step 4, detecting a workpiece target in real time;
inputting the real-time video image of the workpiece into the depth convolution neural network Net1 verified in the step 3, and outputting target detection information;
step 5, performing space positioning on the workpiece;
calculating spatial depth information according to the target detection information;
step 6, completing a grabbing task on the corresponding workpiece;
and grabbing corresponding workpieces in real time according to the space depth information.
Optionally, the specific process of image acquisition in the step 1 is: the binocular vision assembly is provided with a left camera and a right camera, and the left camera is used for collecting images of various postures of a workpiece in different environments.
Optionally, the preprocessing includes:
marking different targets in the acquired gesture images by adopting an image segmentation method to obtain a data set and corresponding labels; the label is workpiece category information; the data set is an image acquired by the type of the workpiece library; the data set includes a training set and a validation set.
Optionally, the input layer of the deep convolutional neural network Net1 is a visible light image, the size of the image is W.H, various types of convolutional kernels are adopted to carry out convolutional operation, the characteristic image is downsampled by adjusting the convolutional step length, the excitation function is a ReLU function, and the target detection is completed under three scales through the characteristic pooling pyramid FPN network and the convolutional layer; the process may map the image to a one-dimensional vector of object detection information.
Optionally, the loss function is:
L c (o,c,O,C,l,g)=λ 1 L conf (o,c)+λ 2 L cla (O,C)+λ 3 L loc (l,g)
wherein L is conf (o, c) is confidenceLoss of degree; l (L) cla (O, C) is a classification loss; l (L) loc (l, g) is a loss of positioning; lambda (lambda) 1 A balance coefficient that is a confidence loss; lambda (lambda) 2 Balance coefficient for classification loss; lambda (lambda) 3 To locate the balance of the losses.
Optionally, updating the weight of the deep convolutional neural network by adopting an Adam optimizer to obtain an updated weight, which specifically comprises:
m t =β 1 ·m t-1 +(1-β 1 )·g(ω t )
ν t =β 2 ·ν t-1 +(1-β 2 )·g(ω t )·g(ω t )
Figure BDA0004133066560000031
Figure BDA0004133066560000032
wherein m is t To weight first order momentum, m t-1 Is the first order momentum before weight update, beta 1 (0.9)、β 2 (0.999) to control the attenuation coefficient, g (ω) t ) For time t to parameter omega t Is a loss gradient of v t Is the second order momentum, v t-1 Is the second order momentum before updating the network weight, alpha is the learning rate, epsilon (10 -7 ) To prevent the denominator from being zero decimal, ω t Is the weight parameter before updating omega t+1 Is the weight parameter after the update of the weight parameter,
Figure BDA0004133066560000033
all are calculated intermediate values. />
Optionally, calculating spatial depth information according to the target detection information specifically includes:
and processing by using a correlation matching algorithm and a triangulation principle, and calculating the spatial depth information of the workpiece.
The invention also provides a machine learning-based workpiece automatic grabbing system, which comprises a binocular vision component, a data preprocessing frame, a deep convolutional neural network and an information flow;
the binocular vision assembly is provided with a left high-definition camera and a right high-definition camera and is used for collecting images of a workpiece in various postures under different environments;
the data preprocessing framework is used for:
image data integration is carried out on multiple types of workpieces, and an image data set is established for each type of workpiece by using a manual labeling mode;
randomly extracting the data set to form a training set and a verification set, and managing the class of the workpiece library;
the deep convolutional neural network detects a workpiece target;
the information flow fuses the target detection information obtained by the depth convolution neural network with the image information of the left camera and the right camera of the binocular vision component, and performs space positioning on a workpiece to obtain space depth information; the spatial depth information is used to enable the corresponding workpiece to be grasped in real time.
The invention also provides a machine learning-based automatic workpiece grabbing device, which comprises a mechanical arm control assembly and a machine learning-based automatic workpiece grabbing system;
the mechanical arm control assembly is used for grabbing corresponding workpieces in real time according to the space depth information.
The information flow applies the target detection information obtained by the deep convolutional neural network to the spatial positioning of the binocular vision component on the workpiece, solves the problems that the workpiece type identification is inaccurate, the grabbed target type is difficult to expand, the efficiency is low, the accuracy is poor and the like, fuses the target detection information output by the deep convolutional neural network with the image information of the left camera and the right camera of the binocular vision component through the information flow, successfully reduces the complexity of the workpiece grabbing system, and enables the system to be more suitable for complex conditions.
Compared with the prior art, the invention has the following advantages:
1. the invention provides a machine learning-based workpiece automatic grabbing system, which comprises a binocular vision component, a data preprocessing frame, a deep convolutional neural network and an information flow, wherein the binocular vision component is used for preprocessing data; according to the system, the target detection information output by the deep convolutional neural network is fused with the image information of the left camera and the right camera of the binocular vision component through the information flow, so that the complexity of the workpiece grabbing system is successfully reduced, and the system can be more suitable for complex conditions.
2. The invention provides a machine learning-based automatic workpiece grabbing method, which adopts a method of combining a data preprocessing frame and a deep convolutional neural network to use: the data prediction management framework completes random extraction of the data set to form a training set and a verification set, and simultaneously completes management of the class of the workpiece library, and the deep neural network realizes identification of multiple classes of workpieces. The automatic grabbing task of the workpiece can be completed with higher efficiency, a system which is easier to expand and more accurate classification and positioning.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions of the prior art, the drawings that are needed in the embodiments will be briefly described below, it being obvious that the drawings in the following description are only some embodiments of the present invention, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a flow chart of the method of the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
In order that the above-recited objects, features and advantages of the present invention will become more readily apparent, a more particular description of the invention will be rendered by reference to the appended drawings and appended detailed description.
A machine learning-based automatic workpiece grabbing system comprises a binocular vision component, a data preprocessing frame, a deep convolutional neural network and an information flow.
The binocular vision assembly is provided with left and right high-definition cameras and is used for collecting images of a workpiece in different environments and in various postures.
The data preprocessing framework firstly integrates image data for multiple types of workpieces, establishes image data sets for various workpieces by using a manual labeling mode, and simultaneously enables the deep neural network to successfully realize accurate identification of the multiple types of workpieces, the data preprocessing framework can complete random extraction of the data sets to form training sets and verification sets, and the framework can complete management of workpiece library types.
And the deep convolution neural network performs workpiece target detection.
The information flow applies the target detection information obtained by the deep convolutional neural network to the spatial positioning of the binocular vision component on the workpiece, and solves the problems that the workpiece type identification is inaccurate, the grabbed target type is difficult to expand, the efficiency is low, the accuracy is low and the like.
The invention is described with reference to fig. 1, in which the process of spatial positioning of a workpiece by a vision component in an automatic workpiece grabbing system is resolved into the detection problem of a workpiece target in a real-time video and the acquisition problem of depth information of the workpiece by a binocular vision component, the real-time target detection is performed on the workpiece in the real-time video of a left camera by constructing a depth convolution neural network, the detection information is transmitted to the binocular vision component through an information flow, and the component combines the target detection information, performs correlation matching and triangulation on the real-time video images of the left and right cameras, and outputs the real-time spatial position of the workpiece to be detected in real time. In particular to the construction of a deep neural network, the optimization of network training and the spatial positioning of workpieces.
(1) A deep neural network for multi-type workpiece recognition is trained, the network is denoted as Net1, and the network type is a convolutional neural network.
(1-1) setting an identification network Net1: the structure of the network is a CNN network, the input layer is a visible light image, the size of the image is W.H, convolution operation is carried out by adopting various convolution kernels, the characteristic image is downsampled by adjusting the convolution step length, the excitation function is a ReLU function, the target detection is completed under three scales by the characteristic pooling pyramid FPN network and the convolution layer, and the physical meaning of the process is that the image is mapped into a target detection information one-dimensional vector.
(1-2) the binocular vision component is provided with a left camera and a right camera, the left camera is used for collecting images of various postures of a workpiece in different environments, and different targets are marked by adopting an image segmentation method. Resulting in a dataset d= { D 1 ,d 2 ,...d n Corresponding tag s= { [ S } 11 ,s 12 ...],[s 21 ,s 22 ...],...[s n1 ,s n2 ...]}。
(1-3) defining a Net1 deep convolutional neural network loss function of L for real classification and localization of workpiece targets c (o,c,O,C,l,g)=λ 1 L conf (o,c)+λ 2 L cla (O,C)+λ 3 L loc (l, g) wherein the three components are confidence, classification and location loss, lambda, respectively 1 、λ 2 、λ 3 Is the balance coefficient between the three partial losses. Training convergence to deep neural networks to L c For the purpose of minimization, parameters that optimize the convolutional neural network (minimizing the gradient of the loss function to the convolutional neural network parameters) are calculated.
(1-4) updating the weight of the convolutional neural network by adopting an Adam optimizer, wherein the Adam optimizer can adapt to the learning rate, and the parameter updating effect is further improved:
m t =β 1 ·m t-1 +(1-β 1 )·g(ω t )
ν t =β 2 ·ν t-1 +(1-β 2 )·g(ω t )·g(ω t )
Figure BDA0004133066560000071
Figure BDA0004133066560000072
wherein m is t To weight first order momentum, m t-1 Is the first order momentum before weight update, beta 1 (0.9)、β 2 (0.999) to control the attenuation coefficient, g (ω) t ) For time t to parameter omega t Is a loss gradient of v t Is the second order momentum, v t-1 Is the second order momentum before updating the network weight, alpha is the learning rate, epsilon (10 -7 ) To prevent the denominator from being zero decimal, ω t Is the weight parameter before updating omega t+1 Is the weight parameter after the update of the weight parameter,
Figure BDA0004133066560000073
all are calculated intermediate values.
(1-5) repeating steps 1-3 to 1-4, wherein the loss function variation is smaller than the threshold eta after N continuous optimizations, and the Net1 network converges.
(2) And optimizing the weight of the iterative deep neural network by adopting a K-fold cross validation strategy.
(3) Under the dynamic condition, inputting a real-time video image acquired by a left camera of the binocular vision component into a depth convolutional neural network Net1, detecting a real-time target of the environment where the workpiece is located, and outputting target detection information.
(4) The binocular vision component acquires target detection information of a workpiece to be grabbed through an information transmission flow, and image information acquired by the left camera and the right camera of the binocular vision component is processed by utilizing a correlation matching algorithm and a triangulation principle to calculate the spatial depth information of the workpiece.
(5) And in a state of real-time detection, the space depth information of the workpiece is transmitted to the mechanical arm control assembly through an information transmission flow, so that real-time grabbing of the corresponding workpiece is completed.
In the present specification, each embodiment is described in a progressive manner, and each embodiment is mainly described in a different point from other embodiments, and identical and similar parts between the embodiments are all enough to refer to each other.
The principles and embodiments of the present invention have been described herein with reference to specific examples, the description of which is intended only to assist in understanding the methods of the present invention and the core ideas thereof; also, it is within the scope of the present invention to be modified by those of ordinary skill in the art in light of the present teachings. In view of the foregoing, this description should not be construed as limiting the invention.

Claims (9)

1. The automatic workpiece grabbing method based on machine learning is characterized by comprising the following steps of:
step 1, image acquisition and pretreatment;
collecting multiple gesture images of a workpiece in different environments; obtaining a data set and a label according to the gesture image;
step 2, training a deep convolutional neural network Net 1;
training a deep convolutional neural network by adopting the data set and the label;
obtaining parameters of the deep convolutional neural network by taking the minimum value of the loss function of the deep convolutional neural network as a target;
updating the weight of the deep convolutional neural network by adopting an Adam optimizer to obtain an updated weight;
calculating a loss variation value of the loss function according to the updated weight and the parameter;
if the loss change value is smaller than a set threshold value, taking the depth convolution neural network corresponding to the loss value as a converged depth convolution neural network;
step 3, adopting K-fold to cross and verify the weight of the deep convolutional neural network Net1 obtained by training in the step 2;
step 4, detecting a workpiece target in real time;
inputting the real-time video image of the workpiece into the depth convolution neural network Net1 verified in the step 3, and outputting target detection information;
step 5, performing space positioning on the workpiece;
calculating spatial depth information according to the target detection information;
step 6, completing a grabbing task on the corresponding workpiece;
and grabbing corresponding workpieces in real time according to the space depth information.
2. The automatic workpiece grabbing method based on machine learning according to claim 1, wherein the specific process of image acquisition in step 1 is as follows: the binocular vision assembly is provided with a left camera and a right camera, and the left camera is used for collecting images of various postures of a workpiece in different environments.
3. The machine learning based automatic workpiece gripping method of claim 1, wherein the preprocessing comprises:
marking different targets in the acquired gesture images by adopting an image segmentation method to obtain a data set and corresponding labels; the label is workpiece category information; the data set is an image acquired by the type of the workpiece library; the data set includes a training set and a validation set.
4. The automatic workpiece grabbing method based on machine learning according to claim 1, wherein an input layer of the deep convolutional neural network Net1 is a visible light image, the size of the image is w×h, convolution operation is performed by adopting a plurality of types of convolution kernels, feature images are downsampled by adjusting convolution step sizes, an excitation function is a ReLU function, and target detection is completed under three scales through a feature pooling pyramid FPN network and a convolution layer.
5. The machine learning based automatic workpiece gripping method of claim 1, wherein the loss function is:
L c (o,c,O,C,l,g)=λ 1 L conf (o,c)+λ 2 L cla (O,C)+λ 3 L loc (l,g)
wherein L is conf (o, c) is a confidence loss; l (L) cla (O, C) is a classification loss; l (L) loc (l, g) is a loss of positioning; lambda (lambda) 1 A balance coefficient that is a confidence loss; lambda (lambda) 2 Balance coefficient for classification loss; lambda (lambda) 3 To locate the balance of the losses.
6. The machine learning based automatic workpiece grabbing method of claim 1, wherein the weight of the deep convolutional neural network is updated by an Adam optimizer to obtain an updated weight, and the method specifically comprises:
m t =β 1 ·m t-1 +(1-β 1 )·g(ω t )
ν t =β 2 ·ν t-1 +(1-β 2 )·g(ω t )·g(ω t )
Figure FDA0004133066550000021
Figure FDA0004133066550000022
wherein m is t To weight first order momentum, m t-1 Is the first order momentum before weight update, beta 1 (0.9)、β 2 (0.999) to control the attenuation coefficient, g (ω) t ) For time t to parameter omega t Is a loss gradient of v t Is the second order momentum, v t-1 Is the second order momentum before updating the network weight, alpha is the learning rate, epsilon (10 -7 ) To prevent the denominator from being zero decimal, ω t Is the weight parameter before updating omega t+1 Is the weight parameter after the update of the weight parameter,
Figure FDA0004133066550000023
all are calculated intermediate values.
7. The machine learning based automatic workpiece gripping method according to claim 1, wherein the calculating of the spatial depth information based on the target detection information specifically includes:
and processing by using a correlation matching algorithm and a triangulation principle, and calculating the spatial depth information of the workpiece.
8. The automatic workpiece grabbing system based on machine learning is characterized by comprising a binocular vision component, a data preprocessing frame, a deep convolutional neural network and an information flow;
the binocular vision assembly is provided with a left high-definition camera and a right high-definition camera and is used for collecting images of a workpiece in various postures under different environments;
the data preprocessing framework is used for:
image data integration is carried out on multiple types of workpieces, and an image data set is established for each type of workpiece by using a manual labeling mode;
randomly extracting the data set to form a training set and a verification set, and managing the class of the workpiece library;
the deep convolutional neural network detects a workpiece target;
the information flow fuses the target detection information output by the depth convolution neural network with the image information of the left camera and the right camera of the binocular vision component, and performs space positioning on a workpiece to obtain space depth information; the spatial depth information is used to enable the corresponding workpiece to be grasped in real time.
9. A machine learning based automatic workpiece gripping apparatus, comprising a robotic arm control assembly and the machine learning based automatic workpiece gripping system of claim 8;
the mechanical arm control assembly is used for grabbing corresponding workpieces in real time according to the space depth information.
CN202310253672.2A 2021-09-17 2021-09-17 Automatic workpiece grabbing method, system and device based on machine learning Withdrawn CN116128960A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310253672.2A CN116128960A (en) 2021-09-17 2021-09-17 Automatic workpiece grabbing method, system and device based on machine learning

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202310253672.2A CN116128960A (en) 2021-09-17 2021-09-17 Automatic workpiece grabbing method, system and device based on machine learning
CN202111095314.0A CN113808197A (en) 2021-09-17 2021-09-17 Automatic workpiece grabbing system and method based on machine learning

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
CN202111095314.0A Division CN113808197A (en) 2021-09-17 2021-09-17 Automatic workpiece grabbing system and method based on machine learning

Publications (1)

Publication Number Publication Date
CN116128960A true CN116128960A (en) 2023-05-16

Family

ID=78895844

Family Applications (2)

Application Number Title Priority Date Filing Date
CN202111095314.0A Pending CN113808197A (en) 2021-09-17 2021-09-17 Automatic workpiece grabbing system and method based on machine learning
CN202310253672.2A Withdrawn CN116128960A (en) 2021-09-17 2021-09-17 Automatic workpiece grabbing method, system and device based on machine learning

Family Applications Before (1)

Application Number Title Priority Date Filing Date
CN202111095314.0A Pending CN113808197A (en) 2021-09-17 2021-09-17 Automatic workpiece grabbing system and method based on machine learning

Country Status (1)

Country Link
CN (2) CN113808197A (en)

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108161931A (en) * 2016-12-07 2018-06-15 广州映博智能科技有限公司 The workpiece automatic identification of view-based access control model and intelligent grabbing system
CN106874914B (en) * 2017-01-12 2019-05-14 华南理工大学 A kind of industrial machinery arm visual spatial attention method based on depth convolutional neural networks
CN108171748B (en) * 2018-01-23 2021-12-07 哈工大机器人(合肥)国际创新研究院 Visual identification and positioning method for intelligent robot grabbing application
CN109344772B (en) * 2018-09-30 2021-01-26 中国人民解放军战略支援部队信息工程大学 Ultrashort wave specific signal reconnaissance method based on spectrogram and deep convolutional network
CN111368852A (en) * 2018-12-26 2020-07-03 沈阳新松机器人自动化股份有限公司 Article identification and pre-sorting system and method based on deep learning and robot
CN111768449B (en) * 2019-03-30 2024-05-14 北京伟景智能科技有限公司 Object grabbing method combining binocular vision with deep learning
CN110322515A (en) * 2019-07-02 2019-10-11 工极智能科技(苏州)有限公司 Workpiece identification and grabbing point extraction method based on binocular stereo vision
CN110363253A (en) * 2019-07-25 2019-10-22 安徽工业大学 A kind of Surfaces of Hot Rolled Strip defect classification method based on convolutional neural networks
CN110728223A (en) * 2019-10-08 2020-01-24 济南东朔微电子有限公司 Helmet wearing identification method based on deep learning
CN110969158B (en) * 2019-11-06 2023-07-25 中国科学院自动化研究所 Target detection method, system and device based on underwater operation robot vision
CN110942142B (en) * 2019-11-29 2021-09-17 广州市百果园信息技术有限公司 Neural network training and face detection method, device, equipment and storage medium
CN111340132B (en) * 2020-03-10 2024-02-02 南京工业大学 Machine olfaction mode identification method based on DA-SVM
CN111723782A (en) * 2020-07-28 2020-09-29 北京印刷学院 Deep learning-based visual robot grabbing method and system
CN113052005B (en) * 2021-02-08 2024-02-02 湖南工业大学 Garbage sorting method and garbage sorting device for household service
CN113222940B (en) * 2021-05-17 2022-07-12 哈尔滨工业大学 Method for automatically grabbing workpiece by robot based on RGB-D image and CAD model

Also Published As

Publication number Publication date
CN113808197A (en) 2021-12-17

Similar Documents

Publication Publication Date Title
CN109685141B (en) Robot article sorting visual detection method based on deep neural network
CN107833236B (en) Visual positioning system and method combining semantics under dynamic environment
CN109559320B (en) Method and system for realizing visual SLAM semantic mapping function based on hole convolution deep neural network
CN108520203B (en) Multi-target feature extraction method based on fusion of self-adaptive multi-peripheral frame and cross pooling feature
CN110287907B (en) Object detection method and device
CN114693661A (en) Rapid sorting method based on deep learning
CN110310305B (en) Target tracking method and device based on BSSD detection and Kalman filtering
CN111027555B (en) License plate recognition method and device and electronic equipment
CN110827321B (en) Multi-camera collaborative active target tracking method based on three-dimensional information
CN107025657A (en) A kind of vehicle action trail detection method based on video image
CN114677323A (en) Semantic vision SLAM positioning method based on target detection in indoor dynamic scene
CN110838145A (en) Visual positioning and mapping method for indoor dynamic scene
JP2024528419A (en) Method and apparatus for updating an object detection model
CN105976397A (en) Target tracking method based on half nonnegative optimization integration learning
CN113657423A (en) Target detection method suitable for small-volume parts and stacked parts and application thereof
CN110147724B (en) Method, apparatus, device, and medium for detecting text region in video
CN109544632B (en) Semantic SLAM object association method based on hierarchical topic model
CN114863201A (en) Training method and device of three-dimensional detection model, computer equipment and storage medium
CN113139432A (en) Industrial packaging behavior identification method based on human body skeleton and local image
CN116105721B (en) Loop optimization method, device and equipment for map construction and storage medium
CN117237411A (en) Pedestrian multi-target tracking method based on deep learning
CN116128960A (en) Automatic workpiece grabbing method, system and device based on machine learning
CN115205806A (en) Method and device for generating target detection model and automatic driving vehicle
CN115273219A (en) Yoga action evaluation method and system, storage medium and electronic equipment
CN112200850A (en) ORB extraction method based on mature characteristic points

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication

Application publication date: 20230516