CN114952809A - Workpiece recognition and pose detection method, system and grasping control method of robotic arm - Google Patents

Workpiece recognition and pose detection method, system and grasping control method of robotic arm Download PDF

Info

Publication number
CN114952809A
CN114952809A CN202210732860.9A CN202210732860A CN114952809A CN 114952809 A CN114952809 A CN 114952809A CN 202210732860 A CN202210732860 A CN 202210732860A CN 114952809 A CN114952809 A CN 114952809A
Authority
CN
China
Prior art keywords
workpiece
pose
image
scene
point cloud
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210732860.9A
Other languages
Chinese (zh)
Other versions
CN114952809B (en
Inventor
徐刚
赵有港
崔玥
周翔
许允款
曾晶
肖江剑
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ningbo Institute of Material Technology and Engineering of CAS
Original Assignee
Ningbo Institute of Material Technology and Engineering of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ningbo Institute of Material Technology and Engineering of CAS filed Critical Ningbo Institute of Material Technology and Engineering of CAS
Priority to CN202210732860.9A priority Critical patent/CN114952809B/en
Publication of CN114952809A publication Critical patent/CN114952809A/en
Application granted granted Critical
Publication of CN114952809B publication Critical patent/CN114952809B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/08Programme-controlled manipulators characterised by modular constructions
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J13/00Controls for manipulators
    • B25J13/08Controls for manipulators by means of sensing devices, e.g. viewing or touching devices
    • B25J13/087Controls for manipulators by means of sensing devices, e.g. viewing or touching devices for sensing other physical parameters, e.g. electrical or chemical properties
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J15/00Gripping heads and other end effectors
    • B25J15/08Gripping heads and other end effectors having finger members
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1602Programme controls characterised by the control system, structure, architecture
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1656Programme controls characterised by programming, planning systems for manipulators
    • B25J9/1661Programme controls characterised by programming, planning systems for manipulators characterised by task planning, object-oriented languages
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1694Programme controls characterised by use of sensors other than normal servo-feedback from position, speed or acceleration sensors, perception control, multi-sensor controlled systems, sensor fusion
    • B25J9/1697Vision controlled systems
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Robotics (AREA)
  • Mechanical Engineering (AREA)
  • Automation & Control Theory (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

本发明公开了一种工件识别和位姿检测方法、系统及机械臂的抓取控制方法。所述工件识别和位姿检测方法包括:采集待识别场景中的2D图像和3D点云图像;基于所述2D图像识别所述待识别场景中的目标工件,并基于映射关系进行实例分割,获得所述目标工件对应的点云区域;基于深度学习算法,在所述点云区域中进行位姿检测,获取所述目标工件的位姿信息。本发明所提供的工件识别和位姿检测方法在小工件散乱堆叠的抓取场景中,规避了跨模态数据特征提取及匹配难题,同时避免了过于复杂的数据处理计算,通过结合2D图像和3D点云图像,为工件堆叠识别和抓取的应用场景在有效提升识别效率和提升抓取效率这个方向上提供了优化的解决方案。

Figure 202210732860

The invention discloses a workpiece identification and pose detection method, a system and a grasping control method of a mechanical arm. The workpiece recognition and pose detection method includes: collecting 2D images and 3D point cloud images in the scene to be recognized; recognizing the target workpiece in the scene to be recognized based on the 2D image, and performing instance segmentation based on the mapping relationship to obtain The point cloud area corresponding to the target workpiece; based on a deep learning algorithm, pose detection is performed in the point cloud area to obtain the pose information of the target workpiece. The workpiece identification and pose detection method provided by the present invention avoids the problem of cross-modal data feature extraction and matching in the grasping scene where small workpieces are scattered and stacked, and at the same time avoids overly complicated data processing calculations. The 3D point cloud image provides an optimized solution for the application scenarios of workpiece stack recognition and grasping in the direction of effectively improving the recognition efficiency and improving the grasping efficiency.

Figure 202210732860

Description

工件识别和位姿检测方法、系统及机械臂的抓取控制方法Workpiece recognition and pose detection method, system and grasping control method of robotic arm

技术领域technical field

本发明涉及图像识别与机械控制技术领域,尤其涉及一种工件识别和位姿检测方法、系统及机械臂的抓取控制方法。The invention relates to the technical field of image recognition and mechanical control, in particular to a workpiece recognition and pose detection method, system and grasping control method of a mechanical arm.

背景技术Background technique

基于二维/三维视觉的机器人抓取技术,已经在物流快递、仓库搬运、拆码垛等简单场景中广泛应用,视觉引导的机器人增强了面对复杂环境的感知能力。在工业抓取场景中,二维图像能够提供致密丰富的纹理信息,经过图像处理和识别,获取被抓取工件的位置(二维坐标),但无法获取深度信息;三维图像能够提供抓取场景中的距离信息,但无法获得丰富的细节信息,导致抓取精度的降低。两类数据具有较好的互补性,融合两种模态的数据,可以实现对工件抓取场景进行更全面的感知。近年来,随着工件6D姿态估计算法的研究日趋增加,设备的计算能力日益提高,机器人抓取系统已经在工件无序散乱堆叠、无序工件装配、柔性抓取等相关领域取得创造性的突破。Robotic grasping technology based on 2D/3D vision has been widely used in simple scenarios such as logistics and express delivery, warehouse handling, and depalletizing. Vision-guided robots enhance the perception ability of complex environments. In industrial grabbing scenarios, two-dimensional images can provide dense and rich texture information, and after image processing and recognition, the position (two-dimensional coordinates) of the grabbed workpiece can be obtained, but depth information cannot be obtained; three-dimensional images can provide grabbing scenes However, rich detailed information cannot be obtained, which leads to the reduction of the grasping accuracy. The two types of data have good complementarity, and the fusion of the data of the two modalities can realize a more comprehensive perception of the workpiece grasping scene. In recent years, with the increasing research on 6D pose estimation algorithms of workpieces and the increasing computing power of equipment, robotic grasping systems have made creative breakthroughs in related fields such as disordered workpiece stacking, disordered workpiece assembly, and flexible grasping.

其中,目标物体的识别与位姿检测是机器人抓取任务的关键先决条件。自计算机视觉早期以来,目标6D位姿检测与估计是相对于给定参考系的固定坐标系,通过一个平移向量t∈R3和一个旋转矩阵R∈SO(3)来描述物体对象位姿信息,是一个长期存在的挑战和一个开放的研究领域。Among them, target object recognition and pose detection are key prerequisites for robot grasping tasks. Since the early days of computer vision, target 6D pose detection and estimation is a fixed coordinate system relative to a given reference frame, and the object pose information is described by a translation vector t ∈ R3 and a rotation matrix R ∈ SO(3), is a long-standing challenge and an open field of research.

由于现实世界中物体的多样性、潜在的物体对称性、场景中的杂波和遮挡以及变化的光照条件,目标6D位姿检测与估计的任务的核心步骤就是首先通过各种算法获取目标物体在相机坐标系下的质心位置坐标(x,y,z),然后将模型匹配到该质心位置,得到当前目标物体在相机坐标系的旋转位姿(Rx,Ry,Rz),基于手眼标定矩阵转换,获得目标物体在机械臂基坐标下的位置,最后控制机械臂运动进行抓取作业,具有一定的挑战性。从技术的角度来看,三维点云数据与二维图像属于不同模态的数据,如何巧妙的融合两者的数据,并根据抓取场景内工件的散乱堆叠情况分析出可靠的几何特征,最后识别出场景中可抓取的工件,并获取位姿信息,是国内外科技工作者着力探索一个研究方向。Due to the diversity of objects in the real world, potential object symmetry, clutter and occlusion in the scene, and changing lighting conditions, the core step of the task of target 6D pose detection and estimation is to first obtain the target object through various algorithms. The centroid position coordinates (x, y, z) in the camera coordinate system, and then match the model to the centroid position to obtain the rotation pose (R x , R y , R z ) of the current target object in the camera coordinate system, based on the hand-eye It is challenging to calibrate the matrix transformation, obtain the position of the target object in the base coordinates of the manipulator, and finally control the motion of the manipulator for grasping. From a technical point of view, 3D point cloud data and 2D image data belong to different modalities. How to skillfully integrate the two data, and analyze reliable geometric features according to the scattered stacking of workpieces in the captured scene, and finally Identifying graspable workpieces in the scene and obtaining pose information is a research direction that domestic and foreign scientific and technological workers focus on exploring.

现有的工件识别与位姿检测方法,根据不同输入的图像数据类型,可以分为以下两类:基于2D视觉数据(以RGB或者RGBD数据为输入)、基于3D视觉数据(以点云数据为输入)。单纯的基于2D数据的识别方法,由于缺乏场景深度信息,往往只能进行平面物体抓取,无法处理堆叠场景,因此基于3D视觉数据的工件识别与位姿检测方法渐渐成为主流。基于3D视觉数据的检测方法根据实现原理的差异,大致可分为以下两类:模板匹配法、深度学习法。第一类模板匹配法,通常基于PPF(Point Pair Future)算法(例如Drost B,Ulrich M,Navab N.et al.Model globally,match locally:Efficient and robust 3D objectrecognition[C].IEEE computer society conference on computer vision andpattern recognition.Piscataway:IEEE Press,2010:998-1005.),该算法是一种基于点对特征的描述方法,根据目标的3D模型数据,提取点对特征并训练目标的模型,基于PPF特征描述子在目标场景中检测3D特征点并匹配,求得姿态的一个初始估计并迭代投票,最后利用ICP算法对结果进行Refinement操作,得到一个更准确的位姿结果并输出。模板匹配法在具有最大的缺陷是会出现误匹配的现象,然而,上述方法当工件过于简单,特征不明显时,往往会得出错误的识别结果。第二类深度学习法,通过在仿真场景中制作生成仿真数据集,然后在网络中学习数据特征,最后在测试数据集中得出位姿检测的结果。例如文献(Dong Z,Liu S,Zhou T.et al.PPR-Net:point-wise pose regression network forinstance segmentation and 6d pose estimation in bin-picking scenarios[C].IEEE/RSJ International Conference on Intelligent Robots and Systems(IROS).Piscataway:IEEE Press,2019:1773-1780.)提出了一种新颖的逐点姿态回归网络PPR-Net(Point-wise Pose Regression Network)。该方法是IROS2019“bin-pick姿态估计挑战”中的优胜者,它以PointNet++为主干网络,为它所属的对象实例的点云中的每个点进行6D姿态估计,然后在空间中基于聚类法对每个识别的预测位姿进行平均,得到最终的位姿假设。但是该方法存在的不足在于:工件场景全域3D点云图像的处理效率较低,分析检测时间较长。The existing workpiece recognition and pose detection methods can be divided into the following two categories according to different input image data types: based on 2D visual data (using RGB or RGBD data as input), and based on 3D visual data (using point cloud data as the input). enter). Due to the lack of scene depth information, the simple recognition method based on 2D data can only grasp plane objects and cannot handle stacked scenes. Therefore, workpiece recognition and pose detection methods based on 3D visual data have gradually become the mainstream. Detection methods based on 3D visual data can be roughly divided into the following two categories according to the difference in implementation principle: template matching method and deep learning method. The first type of template matching method is usually based on the PPF (Point Pair Future) algorithm (such as Drost B, Ulrich M, Navab N. et al. Model globally, match locally: Efficient and robust 3D objectrecognition [C]. IEEE computer society conference on computer vision and pattern recognition. Piscataway: IEEE Press, 2010: 998-1005.), the algorithm is a description method based on point-to-point features. According to the 3D model data of the target, the point-to-point features are extracted and the model of the target is trained. Based on PPF The feature descriptor detects and matches 3D feature points in the target scene, obtains an initial estimate of the pose and iteratively votes, and finally uses the ICP algorithm to perform a Refinement operation on the result to obtain a more accurate pose result and output. The biggest defect of the template matching method is the phenomenon of mismatching. However, when the workpiece is too simple and the features are not obvious, the above methods often get wrong recognition results. The second type of deep learning method generates a simulation data set in a simulation scene, then learns the data features in the network, and finally obtains the result of pose detection in the test data set. For example, literature (Dong Z, Liu S, Zhou T. et al. PPR-Net: point-wise pose regression network for instance segmentation and 6d pose estimation in bin-picking scenarios [C]. IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). Piscataway: IEEE Press, 2019: 1773-1780.) proposes a novel point-wise pose regression network PPR-Net (Point-wise Pose Regression Network). This method, the winner in the IROS2019 "bin-pick pose estimation challenge", uses PointNet++ as the backbone network to perform 6D pose estimation for each point in the point cloud of the object instance to which it belongs, and then cluster-based in space The method averages each identified predicted pose to obtain the final pose hypothesis. However, the shortcomings of this method are: the processing efficiency of the global 3D point cloud image of the workpiece scene is low, and the analysis and detection time is long.

因此,如何提供一种适用于小工件堆叠遮挡的场景的较高效率的目标物体的识别与位姿检测方法,是一个急需解决的问题。Therefore, how to provide a high-efficiency target object recognition and pose detection method suitable for the scene where small workpieces are stacked and occluded is an urgent problem to be solved.

发明内容SUMMARY OF THE INVENTION

针对现有技术的不足,本发明的目的在于提供一种工件识别和位姿检测方法、系统及机械臂的抓取控制方法。In view of the deficiencies of the prior art, the purpose of the present invention is to provide a workpiece identification and pose detection method, a system and a grasping control method of a mechanical arm.

为实现前述发明目的,本发明采用的技术方案包括:In order to realize the foregoing invention purpose, the technical scheme adopted in the present invention includes:

第一方面,本发明提供一种工件识别和位姿检测方法,包括:In a first aspect, the present invention provides a workpiece identification and pose detection method, including:

S1,采集待识别场景中的2D图像和3D点云图像;S1, collect 2D images and 3D point cloud images in the scene to be recognized;

S2,基于所述2D图像识别所述待识别场景中的目标工件,并基于所述2D图像与3D点云图像之间的映射关系,对所述3D点云图像中的目标工件所在区域进行实例分割,获得所述目标工件对应的点云区域;S2, identify the target workpiece in the scene to be identified based on the 2D image, and based on the mapping relationship between the 2D image and the 3D point cloud image, perform an example on the area where the target workpiece is located in the 3D point cloud image Segmentation to obtain the point cloud area corresponding to the target workpiece;

S3,基于深度学习算法,在所述点云区域中进行位姿检测,获取所述目标工件的位姿信息。S3, based on a deep learning algorithm, perform pose detection in the point cloud area, and obtain pose information of the target workpiece.

第二方面,本发明还提供一种工件识别和位姿检测系统,包括:In a second aspect, the present invention also provides a workpiece recognition and pose detection system, including:

图像采集模块,用于采集待识别场景中的2D图像和3D点云图像;The image acquisition module is used to collect 2D images and 3D point cloud images in the scene to be recognized;

区域获取模块,用于基于所述2D图像识别所述待识别场景中的目标工件,并基于所述2D图像与3D点云图像之间的映射关系,对所述3D点云图像中的目标工件所在区域进行实例分割,获得所述目标工件对应的点云区域;an area acquisition module, used for identifying the target workpiece in the scene to be identified based on the 2D image, and based on the mapping relationship between the 2D image and the 3D point cloud image, for the target workpiece in the 3D point cloud image Instance segmentation is performed in the area where it is located, and the point cloud area corresponding to the target workpiece is obtained;

位姿获取模块,用于基于深度学习算法,在所述点云区域中进行位姿检测,获取所述目标工件的位姿信息。The pose acquisition module is used to perform pose detection in the point cloud area based on a deep learning algorithm, and obtain pose information of the target workpiece.

第三方面,本发明还提供一种机械臂的抓取控制方法,包括:In a third aspect, the present invention also provides a grasping control method for a robotic arm, comprising:

基于上述工件识别和位姿检测方法获取待识别场景中的目标工件及其位姿信息;Obtain the target workpiece in the scene to be recognized and its pose information based on the workpiece recognition and pose detection methods;

选定需抓取的所述目标工件,并基于所述位姿信息控制机械臂进行抓取动作。The target workpiece to be grasped is selected, and the robotic arm is controlled to perform grasping action based on the pose information.

基于上述技术方案,与现有技术相比,本发明的有益效果至少包括:Based on the above technical solutions, compared with the prior art, the beneficial effects of the present invention at least include:

本发明所提供的工件识别和位姿检测方法在小工件散乱堆叠的抓取场景中,规避了跨模态数据特征提取及匹配难题,同时避免了过于复杂的数据处理计算,通过结合2D图像和3D点云图像,为工件堆叠识别和抓取的应用场景在有效提升识别效率和提升抓取效率这个方向上提供了优化的解决方案。The workpiece identification and pose detection method provided by the present invention avoids the problem of cross-modal data feature extraction and matching in the grab scene of scattered stacking of small workpieces, and at the same time avoids overly complicated data processing and calculation. The 3D point cloud image provides an optimized solution in the direction of effectively improving the recognition efficiency and improving the grasping efficiency for the application scene of workpiece stack recognition and grasping.

上述说明仅是本发明技术方案的概述,为了能够使本领域技术人员能够更清楚地了解本申请的技术手段,并可依照说明书的内容予以实施,以下以本发明的较佳实施例并配合详细附图说明如后。The above description is only an overview of the technical solutions of the present invention. In order to enable those skilled in the art to understand the technical means of the present application more clearly, and to implement them in accordance with the contents of the description, the following preferred embodiments of the present invention are used in conjunction with the detailed descriptions. The accompanying drawings are described below.

附图说明Description of drawings

图1是本发明一典型实施方案提供的工件识别和位姿检测方法的流程示意图;1 is a schematic flowchart of a workpiece identification and pose detection method provided by an exemplary embodiment of the present invention;

图2是本发明一典型实施方案提供的工件识别和位姿检测方法的部分流程示意图;FIG. 2 is a partial schematic flowchart of a workpiece identification and pose detection method provided by an exemplary embodiment of the present invention;

图3是本发明一典型实施方案提供的工件识别和位姿检测方法的部分流程示意图;FIG. 3 is a partial flowchart of a workpiece identification and pose detection method provided by an exemplary embodiment of the present invention;

图4是本发明一典型实施方案提供的工件识别和位姿检测方法的部分流程示意图;FIG. 4 is a partial schematic flowchart of a workpiece identification and pose detection method provided by an exemplary embodiment of the present invention;

图5是本发明一典型实施方案提供的工件识别和位姿检测系统的结构示意图5 is a schematic structural diagram of a workpiece recognition and pose detection system provided by an exemplary embodiment of the present invention

图6是本发明一典型实施方案提供的仿真数据集生成系统的结构示意图6 is a schematic structural diagram of a simulation data set generation system provided by an exemplary embodiment of the present invention

图7是本发明一典型实施方案提供的2D/3D深度学习网络的结构示意图;7 is a schematic structural diagram of a 2D/3D deep learning network provided by an exemplary embodiment of the present invention;

图8是本发明一典型实施方案提供的工件识别和位姿检测方法的识别与检测效果示例图。FIG. 8 is an example diagram of the recognition and detection effects of the workpiece recognition and pose detection method provided by an exemplary embodiment of the present invention.

具体实施方式Detailed ways

鉴于现有技术中的不足,本案发明人经长期研究和大量实践,得以提出本发明的技术方案。如下将对该技术方案、其实施过程及原理等作进一步的解释说明。In view of the deficiencies in the prior art, the inventor of the present application was able to propose the technical solution of the present invention after long-term research and extensive practice. The technical solution, its implementation process and principle will be further explained as follows.

在下面的描述中阐述了很多具体细节以便于充分理解本发明,但是,本发明还可以采用其他不同于在此描述的方式来实施,因此,本发明的保护范围并不受下面公开的具体实施例的限制。Many specific details are set forth in the following description to facilitate a full understanding of the present invention. However, the present invention can also be implemented in other ways different from those described herein. Therefore, the protection scope of the present invention is not limited by the specific implementation disclosed below. example limitations.

参见图1-图4,本发明实施例提供一种工件识别和位姿检测方法,具体包括如下的步骤S1-S3:Referring to FIG. 1 to FIG. 4 , an embodiment of the present invention provides a workpiece identification and pose detection method, which specifically includes the following steps S1-S3:

S1,采集待识别场景中的2D图像和3D点云图像。S1, collect 2D images and 3D point cloud images in the scene to be recognized.

具体地,本实施例中,所采用的图像信息包括2D图像和3D图像,具体采用2D可见光相机对堆叠场景中工件进行拍照,获得2D图像,及采用3D可见光相机对堆叠场景中工件进行拍照,获得3D图像。Specifically, in this embodiment, the used image information includes a 2D image and a 3D image. Specifically, a 2D visible light camera is used to photograph the workpiece in the stacked scene to obtain a 2D image, and a 3D visible light camera is used to photograph the workpiece in the stacked scene. Get 3D images.

因此,在一些实施方案中,步骤Sl具体可以包括:利用2D相机在所述待识别场景中获取所述2D图像,利用3D相机在所述待识别场景中获取所述3D图像信息。Therefore, in some embodiments, step S1 may specifically include: using a 2D camera to acquire the 2D image in the to-be-recognized scene, and using a 3D camera to acquire the 3D image information in the to-be-recognized scene.

在一些实施方案中,所述待识别场景优选可以包括工件堆叠场景,进一步优选为工件散乱堆叠场景。上述方法尤其适应小工件散乱堆叠的应用场景,这些场景例如是,一些较小的工件无规则地散落在托盘上或容器内,例如工业生产时,常见的半成品在托盘上的无序堆叠,或一些物资转运时,例如扳手等工件在料框中散乱堆积等场景。In some embodiments, the to-be-identified scene may preferably include a workpiece stacking scene, more preferably a workpiece scattered stacking scene. The above method is especially suitable for the application scenarios where small workpieces are stacked randomly, for example, some smaller workpieces are scattered randomly on the pallet or in the container, such as the disorderly stacking of common semi-finished products on the pallet during industrial production, or When some materials are transported, for example, workpieces such as wrenches are scattered and piled up in the material frame.

S2,基于所述2D图像识别所述待识别场景中的目标工件,并基于所述2D图像与3D点云图像之间的映射关系,对所述3D点云图像中的目标工件所在区域进行实例分割,获得所述目标工件对应的点云区域。该步骤基于2D图像识别场景中的目标工件,利用2D图像与3D点云图像之间的映射关系,对识别出的目标工件区域进行实例分割。S2, identify the target workpiece in the scene to be identified based on the 2D image, and based on the mapping relationship between the 2D image and the 3D point cloud image, perform an example on the area where the target workpiece is located in the 3D point cloud image Segmentation to obtain the point cloud area corresponding to the target workpiece. This step recognizes the target workpiece in the scene based on the 2D image, and uses the mapping relationship between the 2D image and the 3D point cloud image to segment the identified target workpiece area.

具体的,如图2所示,所述的步骤S2可以包括步骤S21-S24:Specifically, as shown in FIG. 2, the step S2 may include steps S21-S24:

S21,采集若干张2D图像,标记工件轮廓封闭区域,制作成训练数据集。S21 , collecting several 2D images, marking the closed area of the contour of the workpiece, and making a training data set.

具体地,在本实施例中,可以由2D图像获取到工件的轮廓数据,将工件相同朝上面的轮廓标记为同一类,记录标注工件区域在全局图像中的位置,采集若干张图像并进行标注,制作完成训练数据集。上述标注可以由人工进行标注,也可以通过人工标注和机器对抗学习中的机器标注等方法进行标注。Specifically, in this embodiment, the contour data of the workpiece can be obtained from the 2D image, the contours of the workpiece with the same upward facing are marked as the same class, the position of the marked workpiece area in the global image is recorded, and several images are collected and marked. , to complete the training data set. The above-mentioned annotations can be manually annotated, or can be annotated by methods such as manual annotation and machine annotation in machine adversarial learning.

S22,搭建深度学习2D图像目标分割网络模型,并基于所述训练数据集,对网络模型进行训练。S22, build a deep learning 2D image target segmentation network model, and train the network model based on the training data set.

具体地,本实施例中,可以搭建基于Mask R-CNN卷积神经网络的2D图像目标分割网路模型,并基于所述制作的训练数据集,对网络模型进行训练学习。Specifically, in this embodiment, a 2D image target segmentation network model based on the Mask R-CNN convolutional neural network can be built, and the network model can be trained and learned based on the prepared training data set.

S23,将实时采集的2D图像导入训练好的所述网络模型中进行识别,获取识别工件的区域位置。S23, import the 2D image collected in real time into the trained network model for identification, and obtain the region position of the identified workpiece.

具体地,本实施例中,可以利用已经学习完成的所述2D图像目标分割网路模型,将实时采集的2D图像导入所述网络模型,利用Mask R-CNN使2D图像中识别的物体实现像素级分割,获取所述分割出的2D图像中的区域位置。Specifically, in this embodiment, the 2D image target segmentation network model that has been learned can be used, the 2D image collected in real time can be imported into the network model, and Mask R-CNN can be used to make the objects recognized in the 2D image realize pixel stage segmentation, and obtain the location of the region in the segmented 2D image.

S24,将所述识别出的2D工件区域位置映射到3D点云场景区域,并对所述3D点云场景中目标工件区域进行实例分割。S24, map the identified 2D workpiece area position to the 3D point cloud scene area, and perform instance segmentation on the target workpiece area in the 3D point cloud scene.

具体地,本实施例中,可以利用2D图像和3D点云之间的映射关系,基于所述实例分割后的目标工件区域,在3D点云数据场景中对同一位置区域进行点云实例分割,获得识别工件的点云集。Specifically, in this embodiment, the mapping relationship between the 2D image and the 3D point cloud can be used to perform point cloud instance segmentation on the same location area in the 3D point cloud data scene based on the target workpiece area after the instance segmentation, Obtain a point cloud set of identified artifacts.

因此,在一些实施方式中,步骤S2具体包括如下的步骤:Therefore, in some embodiments, step S2 specifically includes the following steps:

将所述2D图像导入目标分割模型中进行识别,获取所述目标工件的区域位置。The 2D image is imported into the target segmentation model for identification, and the region position of the target workpiece is obtained.

将所述区域位置映射到3D点云图像中的对应区域,并所述3D点云图像中的目标工件所在区域进行实例分割。The region position is mapped to the corresponding region in the 3D point cloud image, and instance segmentation is performed on the region where the target workpiece is located in the 3D point cloud image.

在一些实施方式中,所述目标分割模型的训练方法具体包括如下的步骤:In some embodiments, the training method of the target segmentation model specifically includes the following steps:

提供2D训练数据集,所述2D训练数据集包括多个用于训练的2D图像及其对应的标记信息,所述标记信息至少指示所述2D图像中的工件的轮廓封闭区域。A 2D training data set is provided, the 2D training data set includes a plurality of 2D images used for training and their corresponding marking information, the marking information indicating at least the contour enclosed area of the workpiece in the 2D image.

构建目标分割初始模型,并基于所述2D训练数据集,对所述目标分割初始模型进行训练,获得所述目标分割模型。An initial target segmentation model is constructed, and based on the 2D training data set, the initial target segmentation model is trained to obtain the target segmentation model.

S3,基于深度学习算法,在所述点云区域中进行位姿检测,获取所述目标工件的位姿信息。S3, based on a deep learning algorithm, perform pose detection in the point cloud area, and obtain pose information of the target workpiece.

具体地,本实施例中,可以在所述经过2D图像分割并映射到3D点云,获得识别工件的点云集中,基于搭建完成并且经过训练学习的PPR-Net深度学习网络,对工件位姿进行检测,从而获取工件的位姿信息。Specifically, in this embodiment, after the 2D image is segmented and mapped to a 3D point cloud, a collection of point clouds for identifying the workpiece can be obtained, and based on the PPR-Net deep learning network that has been built and trained and learned, the pose of the workpiece can be determined. Perform detection to obtain the pose information of the workpiece.

继续参见图3,步骤S3具体可以包括以下步骤S31-S34:Continuing to refer to FIG. 3, step S3 may specifically include the following steps S31-S34:

S31,基于V-REP搭建深度学习训练仿真数据集生成系统。S31, build a deep learning training simulation data set generation system based on V-REP.

具体地,本实施例中,可以基于V-REP仿真软件,搭建深度学习训练仿真数据集生成系统,其中包括搭建Kinect仿真视觉传感器、导入工件3D模型、导入料框3D模型以及编写工件掉落和图像数据采集程序等内容,搭建的仿真系统如图6所示。Specifically, in this embodiment, a deep learning training simulation data set generation system can be built based on the V-REP simulation software, which includes building a Kinect simulation vision sensor, importing a 3D model of the workpiece, importing a 3D model of the material frame, and writing the drop and drop of the workpiece. The image data acquisition program and other content, the built simulation system is shown in Figure 6.

S32,制作生成仿真3D训练数据集。S32, producing and generating a simulated 3D training data set.

S33,搭建深度学习3D位姿检测神经网络模型,并基于所述训练数据集,对网络模型进行训练。S33, build a deep learning 3D pose detection neural network model, and train the network model based on the training data set.

具体地,本实施例中,可以搭建基于PPR-Net深度学习网络的3D位姿检测神经网络模型,并基于所述制作的仿真训练数据集,对网络模型进行训练学习。Specifically, in this embodiment, a 3D pose detection neural network model based on the PPR-Net deep learning network can be built, and the network model can be trained and learned based on the produced simulation training data set.

S34,将实例分割后的3D点云图像导入所述训练好的网络模型中进行工件位姿检测,并获取工件位姿信息。S34, import the 3D point cloud image after instance segmentation into the trained network model for workpiece pose detection, and obtain workpiece pose information.

具体地,如图4所示,本实施例中,步骤S32可以包括以下步骤S321-S326:Specifically, as shown in FIG. 4, in this embodiment, step S32 may include the following steps S321-S326:

S321,设定场景中存在n个工件。S321, set that there are n workpieces in the scene.

在一优选实施例中,所述工件数量例如可以是n=27。In a preferred embodiment, the number of workpieces may be n=27, for example.

S322,基于域随机化思想在工作区域内将i个工件随机从某个位置掉落(设定初始整数i=0),并对不同的工件赋予不同的颜色信息。S322 , randomly drop i workpieces from a certain position in the work area based on the idea of domain randomization (set an initial integer i=0), and assign different color information to different workpieces.

S323,基于仿真视觉传感器采集并保存场景中的深度图像和rgb图像。S323, collect and save the depth image and the rgb image in the scene based on the simulated vision sensor.

在一优选实施例中,所述仿真视觉传感器可以为Kinect深度相机。In a preferred embodiment, the simulated vision sensor may be a Kinect depth camera.

S324,记录并保存V-REP中获取的每个工件掉落的位姿信息。S324, record and save the pose information of each workpiece dropped obtained in the V-REP.

S325,基于所述采集的rgb图像信息,对每个工件进行可视化程度分析,并记录可视化程度数据。S325, based on the collected rgb image information, perform visualization degree analysis on each workpiece, and record visualization degree data.

在一优选实施例中,所述可视化程度分析方法可以为:引入工件可见性程度v∈[0,1],该参数反映了预测对象的遮挡程度,v=0时完全不可见,v=1时完全无遮挡,以此类推。某一工件在场景中的可视化程度为:v=N/NmaxIn a preferred embodiment, the visualization degree analysis method may be: introducing the visibility degree v∈[0,1] of the workpiece, this parameter reflects the occlusion degree of the predicted object, when v=0, it is completely invisible, and v=1 completely unobstructed, and so on. The degree of visualization of a workpiece in the scene is: v=N/N max .

其中,N为某一实例工件颜色区域面积值,Nmax为全域工件中,颜色区域面积值最大的数值。Among them, N is the area value of the color area of a certain instance workpiece, and N max is the value of the largest area value of the color area in the global workpiece.

S326,若整数i<n,重复S322至S326步骤;若整数i=n,停止工件掉落,制作生成数据集标签文件。S326, if the integer i<n, repeat the steps from S322 to S326; if the integer i=n, stop the workpiece from dropping, and create a data set label file.

更为具体地,本实施例中,可以在所述经过2D图像分割并映射到3D点云,获得识别工件的点云集中,基于搭建完成并且经过训练学习的PPR-Net深度学习网络,对工件位姿进行检测,从而获取工件的位姿信息,基于Mask R-CNN卷积神经网络和PPR-Net深度学习网络的2D/3D深度学习网络结构示意图如图7所示,2D图像检测实例分割效果和3D点云位姿检测效果示意图如图8所示。More specifically, in this embodiment, after the 2D image is segmented and mapped to the 3D point cloud, the point cloud collection for identifying the workpiece can be obtained. The pose is detected to obtain the pose information of the workpiece. The schematic diagram of the 2D/3D deep learning network based on the Mask R-CNN convolutional neural network and the PPR-Net deep learning network is shown in Figure 7. The 2D image detection instance segmentation effect And the schematic diagram of the 3D point cloud pose detection effect is shown in Figure 8.

如图5所示,本发明所揭示的一种深度学习训练仿真数据集生成系统可以包括:As shown in Figure 5, a deep learning training simulation data set generation system disclosed in the present invention may include:

图像获取装置,用于获取堆叠场景中工件的图像信息。The image acquisition device is used for acquiring image information of the workpiece in the stacking scene.

其中,图像获取装置包括2D图像获取单元和3D图像获取单元,2D图像获取单元用于采用2D相机对堆叠场景中工件进行拍照,获得2D图像。3D图像获取单元用于采用3D相机对堆叠场景中工件进行拍照,获得3D图像。Wherein, the image acquisition device includes a 2D image acquisition unit and a 3D image acquisition unit, and the 2D image acquisition unit is configured to use a 2D camera to take pictures of the workpieces in the stacked scene to acquire a 2D image. The 3D image acquisition unit is used for using a 3D camera to take pictures of the workpieces in the stacked scene to obtain a 3D image.

区域限制装置,用于根据所述图像获取装置的视野大小,对工件的掉落范围进行物理限制。The area limiting device is used for physically limiting the falling range of the workpiece according to the size of the field of view of the image acquisition device.

其中,区域限制装置主要包括料框设置单元和相机设置单元,料框设置单元用于绘制和导入料框3D模型,调整合适位置,便于对工件掉落范围进行物理限制,相机设置单元用于调整相机内部参数和外部参数,保证与真实场景的相机参数一致,从而生成有效数据集。Among them, the area limiting device mainly includes a material frame setting unit and a camera setting unit. The material frame setting unit is used to draw and import the 3D model of the material frame, adjust the appropriate position, and facilitate the physical limitation of the drop range of the workpiece. The camera setting unit is used for adjustment. The camera internal parameters and external parameters are guaranteed to be consistent with the camera parameters of the real scene, thereby generating an effective data set.

工件位姿采集装置,用于在工件随机掉落动作完成后,记录当时的工件位姿信息。The workpiece pose acquisition device is used to record the workpiece pose information at that time after the random drop of the workpiece is completed.

工件可视化程度分析装置,用于在所述仿真场景中,对每个工件的可视化程度进行计算和分析。A workpiece visualization degree analysis device is used to calculate and analyze the visualization degree of each workpiece in the simulation scene.

其中,工件可视化程度分析装置主要包括颜色像素采集单元和可视化程度计算单元,颜色像素采集单元用于对场景中每个分割出的实例颜色像素面积值进行统计,可视化程度计算单元利用颜色像素采集单元提供的统计数值进行可视化程度计算,并输出可视化程度值。Among them, the workpiece visualization degree analysis device mainly includes a color pixel acquisition unit and a visualization degree calculation unit. The color pixel acquisition unit is used to count the area value of each segmented instance color pixel in the scene, and the visualization degree calculation unit uses the color pixel acquisition unit. The provided statistical value is used to calculate the visualization degree and output the visualization degree value.

数据集标签整合装置,用于对图像信息、工件位姿信息、可视化程度等信息进行整合,制作为数据集标签。The data set label integration device is used to integrate information such as image information, workpiece pose information, and visualization degree, and make it into a data set label.

因此,在一些实施方案中,步骤S3具体可以包括如下的步骤:Therefore, in some embodiments, step S3 may specifically include the following steps:

将所述点云区域导入位姿检测模型中进行所述目标工件的位姿检测,并获取所述目标工件的位姿信息。The point cloud area is imported into a pose detection model to perform pose detection of the target workpiece, and the pose information of the target workpiece is acquired.

在一些实施方案中,所述位姿检测模型的训练方法可以包括如下的步骤:In some embodiments, the training method of the pose detection model may include the following steps:

提供3D训练数据集,所述3D训练数据集至少包括3D训练图像及其对应的工件位姿标签和可视化程度标签。A 3D training data set is provided, and the 3D training data set includes at least 3D training images and their corresponding workpiece pose labels and visualization degree labels.

构建位姿检测初始模型,并基于所述3D训练数据集,对所述位姿检测初始模型进行训练,获得所述位姿检测模型。An initial pose detection model is constructed, and based on the 3D training data set, the pose detection initial model is trained to obtain the pose detection model.

在一些实施方案中,所述3D训练数据集由仿真数据集生成系统仿真生成。In some embodiments, the 3D training dataset is simulated by a simulation dataset generation system.

在一些实施方案中,所述仿真生成具体可以包括如下的步骤:In some embodiments, the simulation generation may specifically include the following steps:

构建仿真场景,并设定所述仿真场景中存在n个虚拟工件。A simulation scene is constructed, and it is assumed that there are n virtual workpieces in the simulation scene.

基于域随机化方法,在所述仿真场景的工作区域内将i个虚拟工件随机从选定位置掉落,并对不同的虚拟工件赋予不同的颜色信息,其中,i从零递增迭代。例如是从0依次+1递增。Based on the domain randomization method, i virtual workpieces are randomly dropped from a selected position in the working area of the simulation scene, and different color information is assigned to different virtual workpieces, wherein i is iterated incrementally from zero. For example, it is incremented from 0 to +1.

基于仿真视觉传感器采集并保存场景中的深度图像和rgb图像。The depth image and rgb image in the scene are collected and saved based on the simulated vision sensor.

记录并保存所述仿真场景中每个虚拟工件掉落的位姿信息,作为所述工件位姿标签。Record and save the pose information of each virtual workpiece dropped in the simulation scene as the workpiece pose label.

基于所述采集的深度图像和/或rgb图像,对每个虚拟工件进行可视化程度分析,并记录可视化程度数据作为所述可视化程度标签。Based on the acquired depth image and/or rgb image, a visualization degree analysis is performed on each virtual workpiece, and the visualization degree data is recorded as the visualization degree label.

当迭代至整数i不小于n时,停止虚拟工件掉落,基于所述工件位姿标签和可视化程度标签生成所述3D训练数据集。When iterating until the integer i is not less than n, the virtual workpiece is stopped from falling, and the 3D training data set is generated based on the workpiece pose label and the visualization degree label.

在一些实施方案中,所述仿真数据集生成系统可以包括:In some embodiments, the simulation dataset generation system may include:

图像获取装置,用于获取仿真场景中虚拟工件的3D训练图像。The image acquisition device is used for acquiring the 3D training image of the virtual workpiece in the simulation scene.

区域限制装置,用于根据所述3D训练图像获取视野大小,对虚拟工件的掉落范围进行限制。The area limiting device is used for obtaining the size of the field of view according to the 3D training image, and limiting the falling range of the virtual workpiece.

工件位姿采集装置,用于在虚拟工件随机掉落动作完成后,记录当时的所述目标工件的位姿信息。The workpiece pose acquisition device is used to record the pose information of the target workpiece at that time after the random drop action of the virtual workpiece is completed.

工件可视化程度分析装置,用于在所述仿真场景中,对每个虚拟工件的可视化程度进行计算和分析。A workpiece visualization degree analysis device is used to calculate and analyze the visualization degree of each virtual workpiece in the simulation scene.

数据集标签整合装置,用于对3D训练图像、工件位姿标签、可视化程度标签进行整合,生成所述3D训练数据集。The data set label integration device is used for integrating 3D training images, workpiece pose labels, and visualization degree labels to generate the 3D training data set.

基于上述方法,本发明的另一实施例还提供一种工件识别和位姿检测系统,其包括:Based on the above method, another embodiment of the present invention also provides a workpiece recognition and pose detection system, which includes:

图像采集模块,用于采集待识别场景中的2D图像和3D点云图像。The image acquisition module is used to collect 2D images and 3D point cloud images in the scene to be recognized.

区域获取模块,用于基于所述2D图像识别所述待识别场景中的目标工件,并基于所述2D图像与3D点云图像之间的映射关系,对所述3D点云图像中的目标工件所在区域进行实例分割,获得所述目标工件对应的点云区域。an area acquisition module, used for identifying the target workpiece in the scene to be identified based on the 2D image, and based on the mapping relationship between the 2D image and the 3D point cloud image, for the target workpiece in the 3D point cloud image Instance segmentation is performed in the region where it is located, and the point cloud region corresponding to the target workpiece is obtained.

位姿获取模块,用于基于深度学习算法,在所述点云区域中进行位姿检测,获取所述目标工件的位姿信息。The pose acquisition module is used to perform pose detection in the point cloud area based on a deep learning algorithm, and obtain pose information of the target workpiece.

同理,本发明实施例还提供一种可能应用于上述系统的电子装置,其包括处理器和存储器,所述存储器存储有计算机程序,所述计算机程序被运行时执行上述工件识别和位姿检测方法的步骤。Similarly, an embodiment of the present invention also provides an electronic device that may be applied to the above-mentioned system, which includes a processor and a memory, and the memory stores a computer program, and the computer program is executed to perform the above-mentioned workpiece recognition and pose detection. steps of the method.

同时,本发明实施例还提供一种可读存储介质,其中存储有计算机程序,该计算机程序被运行时执行上述工件识别和位姿检测方法的步骤。Meanwhile, an embodiment of the present invention also provides a readable storage medium, in which a computer program is stored, and when the computer program is run, the steps of the above method for workpiece identification and pose detection are executed.

上述实施例提供了工件识别和位姿检测方法与系统,以及其所应用到的一种仿真数据集生成系统,作为上述方法和系统的进一步应用,本发明另一实施例还提供一种机械臂的抓取控制方法,包括如下的步骤:The above embodiment provides a method and system for workpiece recognition and pose detection, and a simulation data set generation system to which the method and system are applied. As a further application of the above method and system, another embodiment of the present invention also provides a robotic arm The grab control method includes the following steps:

基于上述任一实施方式中的工件识别和位姿检测方法获取待识别场景中的目标工件及其位姿信息。The target workpiece in the scene to be recognized and its pose information are acquired based on the workpiece identification and pose detection method in any of the above embodiments.

选定需抓取的所述目标工件,并基于所述位姿信息控制机械臂进行抓取动作。The target workpiece to be grasped is selected, and the robotic arm is controlled to perform grasping action based on the pose information.

即:获得所述工件位姿信息,控制机械臂进行抓取作业。That is: to obtain the position and attitude information of the workpiece, and control the robotic arm to perform the grasping operation.

需要说明的是,本发明的主要技术构思在于如何高效且精准地获取工件的位姿信息,至于如何根据该位姿信息进行机械臂的路径和/或动作规划,并非本发明的重点,相关的技术方案已多见于多个现有技术,本领域技术人员可以无障碍地进行组合或适应性研发,可以理解的,无论基于本发明所提供的工件识别和位姿检测方法组合怎样具体的机械臂控制方法,均应属于本发明的保护范围之内。It should be noted that the main technical idea of the present invention is how to efficiently and accurately obtain the pose information of the workpiece. As for how to plan the path and/or motion of the robotic arm according to the pose information, it is not the focus of the present invention. The technical solutions have been found in many existing technologies, and those skilled in the art can carry out combination or adaptive research and development without obstacles. The control method should all fall within the protection scope of the present invention.

本发明所揭示的一种工件识别和位姿检测方法及所运用到的深度学习训练仿真数据集生成系统,在小工件散乱堆叠的抓取场景中,规避了跨模态数据特征提取及匹配难题,同时避免了过于复杂的数据处理计算。通过结合2D图像和3D点云数据,为工件堆叠识别和抓取的应用场景,在有效提升识别效率和提升抓取效率这个方向上,提供了优化的解决方案,同时,避免了常规样本手动标注,自动制作和生成训练数据集,极大提升了工作效率。The method for workpiece identification and pose detection and the applied deep learning training simulation data set generation system disclosed by the present invention avoids the problem of cross-modal data feature extraction and matching in the grabbing scene where small workpieces are scattered and stacked. , while avoiding overly complex data processing calculations. By combining 2D images and 3D point cloud data, it provides an optimized solution for the application scenario of workpiece stack recognition and grasping, in the direction of effectively improving the recognition efficiency and improving the grasping efficiency, and at the same time, it avoids the manual labeling of conventional samples. , automatically make and generate training data sets, which greatly improves work efficiency.

尽管已参考说明性实施例描述了本发明,但所属领域的技术人员将理解,在不背离本发明的精神及范围的情况下可做出各种其它改变、省略及/或添加且可用实质等效物替代所述实施例的元件。另外,可在不背离本发明的范围的情况下做出许多修改以使特定情形或材料适应本发明的教示。因此,本文并不打算将本发明限制于用于执行本发明的所揭示特定实施例,而是打算使本发明将包含归属于所附权利要求书的范围内的所有实施例。Although the present invention has been described with reference to illustrative embodiments, those skilled in the art will understand that various other changes, omissions and/or additions and the like may be made without departing from the spirit and scope of the invention Effects replace elements of the described embodiments. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the invention without departing from its scope. Therefore, it is not intended herein to limit the invention to the particular embodiments disclosed for carrying out the invention, but it is intended that this invention include all embodiments falling within the scope of the appended claims.

应当理解,上述实施例仅为说明本发明的技术构思及特点,其目的在于让熟悉此项技术的人士能够了解本发明的内容并据以实施,并不能以此限制本发明的保护范围。凡根据本发明精神实质所作的等效变化或修饰,都应涵盖在本发明的保护范围之内。It should be understood that the above-mentioned embodiments are only intended to illustrate the technical concept and characteristics of the present invention, and the purpose thereof is to enable those who are familiar with the art to understand the content of the present invention and implement it accordingly, and cannot limit the protection scope of the present invention. All equivalent changes or modifications made according to the spirit of the present invention should be included within the protection scope of the present invention.

Claims (10)

1.一种工件识别和位姿检测方法,其特征在于包括:1. a workpiece recognition and pose detection method, is characterized in that comprising: S1,采集待识别场景中的2D图像和3D点云图像;S1, collect 2D images and 3D point cloud images in the scene to be recognized; S2,基于所述2D图像识别所述待识别场景中的目标工件,并基于所述2D图像与3D点云图像之间的映射关系,对所述3D点云图像中的目标工件所在区域进行实例分割,获得所述目标工件对应的点云区域;S2, identify the target workpiece in the scene to be identified based on the 2D image, and based on the mapping relationship between the 2D image and the 3D point cloud image, perform an example on the area where the target workpiece is located in the 3D point cloud image Segmentation to obtain the point cloud area corresponding to the target workpiece; S3,基于深度学习算法,在所述点云区域中进行位姿检测,获取所述目标工件的位姿信息。S3, based on a deep learning algorithm, perform pose detection in the point cloud area, and obtain pose information of the target workpiece. 2.根据权利要求1所述的工件识别和位姿检测方法,其特征在于,步骤S1具体包括:2. workpiece identification and pose detection method according to claim 1, is characterized in that, step S1 specifically comprises: 利用2D相机在所述待识别场景中获取所述2D图像,利用3D相机在所述待识别场景中获取所述3D图像信息;Use a 2D camera to obtain the 2D image in the to-be-identified scene, and use a 3D camera to obtain the 3D image information in the to-be-identified scene; 优选的,所述待识别场景包括工件堆叠场景,进一步优选为工件无序堆叠场景。Preferably, the to-be-identified scene includes a workpiece stacking scene, more preferably a workpiece stacking scene in an orderly manner. 3.根据权利要求1所述的工件识别和位姿检测方法,其特征在于,步骤S2具体包括:3. The workpiece identification and pose detection method according to claim 1, wherein step S2 specifically comprises: 将所述2D图像导入目标分割模型中进行识别,获取所述目标工件的区域位置;The 2D image is imported into the target segmentation model for identification, and the regional position of the target workpiece is obtained; 将所述区域位置映射到3D点云图像中的对应区域,并所述3D点云图像中的目标工件所在区域进行实例分割。The region position is mapped to the corresponding region in the 3D point cloud image, and instance segmentation is performed on the region where the target workpiece is located in the 3D point cloud image. 4.根据权利要求1所述的工件识别和位姿检测方法,其特征在于,所述目标分割模型的训练方法包括:4. workpiece identification and pose detection method according to claim 1, is characterized in that, the training method of described target segmentation model comprises: 提供2D训练数据集,所述2D训练数据集包括多个用于训练的2D图像及其对应的标记信息,所述标记信息至少指示所述2D图像中的工件的轮廓封闭区域;Provide a 2D training data set, the 2D training data set includes a plurality of 2D images for training and their corresponding marking information, the marking information at least indicates the contour enclosed area of the workpiece in the 2D image; 构建目标分割初始模型,并基于所述2D训练数据集,对所述目标分割初始模型进行训练,获得所述目标分割模型。An initial target segmentation model is constructed, and based on the 2D training data set, the initial target segmentation model is trained to obtain the target segmentation model. 5.根据权利要求1所述的工件识别和位姿检测方法,其特征在于,步骤S3具体包括:5. The workpiece identification and pose detection method according to claim 1, wherein step S3 specifically comprises: 将所述点云区域导入位姿检测模型中进行所述目标工件的位姿检测,并获取所述目标工件的位姿信息。The point cloud area is imported into a pose detection model to perform pose detection of the target workpiece, and the pose information of the target workpiece is acquired. 6.根据权利要求5所述的工件识别和位姿检测方法,其特征在于,所述位姿检测模型的训练方法包括:6. The workpiece identification and pose detection method according to claim 5, wherein the training method of the pose detection model comprises: 提供3D训练数据集,所述3D训练数据集至少包括3D训练图像及其对应的工件位姿标签和可视化程度标签;Provide a 3D training data set, the 3D training data set includes at least 3D training images and their corresponding workpiece pose labels and visualization degree labels; 构建位姿检测初始模型,并基于所述3D训练数据集,对所述位姿检测初始模型进行训练,获得所述位姿检测模型。An initial pose detection model is constructed, and based on the 3D training data set, the pose detection initial model is trained to obtain the pose detection model. 7.根据权利要求6所述的工件识别和位姿检测方法,其特征在于,所述3D训练数据集由仿真数据集生成系统仿真生成;7. The workpiece identification and pose detection method according to claim 6, wherein the 3D training data set is simulated and generated by a simulation data set generation system; 优选的,所述仿真生成具体包括:Preferably, the simulation generation specifically includes: 构建仿真场景,并设定所述仿真场景中存在n个虚拟工件;constructing a simulation scene, and setting that there are n virtual artifacts in the simulation scene; 基于域随机化方法,在所述仿真场景的工作区域内将i个虚拟工件随机从选定位置掉落,并对不同的虚拟工件赋予不同的颜色信息;Based on the domain randomization method, i virtual workpieces are randomly dropped from the selected position in the working area of the simulation scene, and different color information is given to different virtual workpieces; 基于仿真视觉传感器采集并保存场景中的深度图像和rgb图像;Collect and save the depth image and rgb image in the scene based on the simulated vision sensor; 记录并保存所述仿真场景中每个虚拟工件掉落的位姿信息,作为所述工件位姿标签;Record and save the pose information of each virtual workpiece dropped in the simulation scene, as the workpiece pose label; 基于所述深度图像和/或rgb图像,对每个虚拟工件进行可视化程度分析,并记录可视化程度数据作为所述可视化程度标签;Based on the depth image and/or the rgb image, perform visualization degree analysis on each virtual workpiece, and record the visualization degree data as the visualization degree label; 当迭代至整数i不小于n时,停止虚拟工件掉落,基于所述工件位姿标签和可视化程度标签生成所述3D训练数据集。When iterating until the integer i is not less than n, the virtual workpiece is stopped from falling, and the 3D training data set is generated based on the workpiece pose label and the visualization degree label. 8.根据权利要求7所述的工件识别和位姿检测方法,其特征在于,所述仿真数据集生成系统包括:8. The workpiece identification and pose detection method according to claim 7, wherein the simulation data set generation system comprises: 图像获取装置,用于获取仿真场景中虚拟工件的3D训练图像;an image acquisition device for acquiring a 3D training image of a virtual workpiece in a simulation scene; 区域限制装置,用于根据所述3D训练图像获取视野大小,对虚拟工件的掉落范围进行限制;an area limiting device, used for obtaining the size of the field of view according to the 3D training image, and limiting the falling range of the virtual workpiece; 工件位姿采集装置,用于在虚拟工件随机掉落动作完成后,记录当时的所述目标工件的位姿信息;The workpiece pose acquisition device is used to record the pose information of the target workpiece at that time after the random drop action of the virtual workpiece is completed; 工件可视化程度分析装置,用于在所述仿真场景中,对每个虚拟工件的可视化程度进行计算和分析;A workpiece visualization degree analysis device, used for calculating and analyzing the visualization degree of each virtual workpiece in the simulation scene; 数据集标签整合装置,用于对3D训练图像、工件位姿标签、可视化程度标签进行整合,生成所述3D训练数据集。The data set label integration device is used for integrating 3D training images, workpiece pose labels, and visualization degree labels to generate the 3D training data set. 9.一种工件识别和位姿检测系统,其特征在于包括:9. A workpiece identification and pose detection system is characterized in that comprising: 图像采集模块,用于采集待识别场景中的2D图像和3D点云图像;The image acquisition module is used to collect 2D images and 3D point cloud images in the scene to be recognized; 区域获取模块,用于基于所述2D图像识别所述待识别场景中的目标工件,并基于所述2D图像与3D点云图像之间的映射关系,对所述3D点云图像中的目标工件所在区域进行实例分割,获得所述目标工件对应的点云区域;an area acquisition module, used for identifying the target workpiece in the scene to be identified based on the 2D image, and based on the mapping relationship between the 2D image and the 3D point cloud image, for the target workpiece in the 3D point cloud image Instance segmentation is performed in the area where it is located, and the point cloud area corresponding to the target workpiece is obtained; 位姿获取模块,用于基于深度学习算法,在所述点云区域中进行位姿检测,获取所述目标工件的位姿信息。The pose acquisition module is used to perform pose detection in the point cloud area based on a deep learning algorithm, and obtain pose information of the target workpiece. 10.一种机械臂的抓取控制方法,其特征在于,包括:10. A grasping control method for a robotic arm, comprising: 基于权利要求1-8中任意一项所述的工件识别和位姿检测方法获取待识别场景中的目标工件及其位姿信息;Obtain the target workpiece and its pose information in the scene to be recognized based on the workpiece identification and pose detection method according to any one of claims 1-8; 选定需抓取的所述目标工件,并基于所述位姿信息控制机械臂进行抓取动作。The target workpiece to be grasped is selected, and the robotic arm is controlled to perform grasping action based on the pose information.
CN202210732860.9A 2022-06-24 2022-06-24 Workpiece recognition and pose detection method, system, and grasping control method of a robotic arm Active CN114952809B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210732860.9A CN114952809B (en) 2022-06-24 2022-06-24 Workpiece recognition and pose detection method, system, and grasping control method of a robotic arm

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210732860.9A CN114952809B (en) 2022-06-24 2022-06-24 Workpiece recognition and pose detection method, system, and grasping control method of a robotic arm

Publications (2)

Publication Number Publication Date
CN114952809A true CN114952809A (en) 2022-08-30
CN114952809B CN114952809B (en) 2023-08-01

Family

ID=82965231

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210732860.9A Active CN114952809B (en) 2022-06-24 2022-06-24 Workpiece recognition and pose detection method, system, and grasping control method of a robotic arm

Country Status (1)

Country Link
CN (1) CN114952809B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115359112A (en) * 2022-10-24 2022-11-18 爱夫迪(沈阳)自动化科技有限公司 Stacking control method of high-level material warehouse robot
CN116038697A (en) * 2022-12-29 2023-05-02 中山大学 A method and system for automatic spraying of jeans based on manual teaching
CN116320357A (en) * 2023-05-17 2023-06-23 浙江视觉智能创新中心有限公司 3D structured light camera system, method, electronic device and readable storage medium
CN116330306A (en) * 2023-05-31 2023-06-27 之江实验室 Object grabbing method and device, storage medium and electronic equipment
CN116363085A (en) * 2023-03-21 2023-06-30 江苏共知自动化科技有限公司 Industrial part target detection method based on small sample learning and virtual synthesized data
CN117095002A (en) * 2023-10-19 2023-11-21 深圳市信润富联数字科技有限公司 Hub defect detection method and device and storage medium
CN117697768A (en) * 2024-02-05 2024-03-15 季华实验室 A target grabbing method, robot, electronic device and storage medium
CN118003340A (en) * 2024-04-08 2024-05-10 厦门熠明机器人自动化有限公司 Visual mechanical arm material grabbing control method and system based on deep learning
CN118288292A (en) * 2024-05-10 2024-07-05 中原动力智能机器人有限公司 A method, device, equipment and storage medium for randomly grasping rod-shaped workpieces

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7177459B1 (en) * 1999-04-08 2007-02-13 Fanuc Ltd Robot system having image processing function
CN108171748A (en) * 2018-01-23 2018-06-15 哈工大机器人(合肥)国际创新研究院 A kind of visual identity of object manipulator intelligent grabbing application and localization method
CN108555908A (en) * 2018-04-12 2018-09-21 同济大学 A kind of identification of stacking workpiece posture and pick-up method based on RGBD cameras
CN109102547A (en) * 2018-07-20 2018-12-28 上海节卡机器人科技有限公司 Robot based on object identification deep learning model grabs position and orientation estimation method
US20190091869A1 (en) * 2017-09-25 2019-03-28 Fanuc Corporation Robot system and workpiece picking method
CN110472534A (en) * 2019-07-31 2019-11-19 厦门理工学院 3D object detection method, device, equipment and storage medium based on RGB-D data
CN111445524A (en) * 2020-03-31 2020-07-24 清华大学 Recognition method of unsafe behavior of construction site workers based on scene understanding
CN111507334A (en) * 2019-01-30 2020-08-07 中国科学院宁波材料技术与工程研究所 Example segmentation method based on key points
CN112116551A (en) * 2019-06-20 2020-12-22 腾讯科技(深圳)有限公司 Camera shielding detection method and device, electronic equipment and storage medium
CN112347887A (en) * 2020-10-28 2021-02-09 深圳市优必选科技股份有限公司 Object detection method, object detection device and electronic equipment
CN112837371A (en) * 2021-02-26 2021-05-25 梅卡曼德(北京)机器人科技有限公司 Object grabbing method and device based on 3D matching and computing equipment
CN113538486A (en) * 2021-07-13 2021-10-22 长春工业大学 Method for improving identification and positioning accuracy of automobile sheet metal workpiece
CN114140526A (en) * 2021-11-19 2022-03-04 浙江大学 Disordered workpiece three-dimensional visual pose estimation method based on deep learning
CN114155265A (en) * 2021-12-01 2022-03-08 南京林业大学 3D LiDAR road point cloud segmentation method based on YOLACT
CN114494276A (en) * 2022-04-18 2022-05-13 成都理工大学 Two-stage multi-modal three-dimensional instance segmentation method

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7177459B1 (en) * 1999-04-08 2007-02-13 Fanuc Ltd Robot system having image processing function
US20190091869A1 (en) * 2017-09-25 2019-03-28 Fanuc Corporation Robot system and workpiece picking method
CN108171748A (en) * 2018-01-23 2018-06-15 哈工大机器人(合肥)国际创新研究院 A kind of visual identity of object manipulator intelligent grabbing application and localization method
CN108555908A (en) * 2018-04-12 2018-09-21 同济大学 A kind of identification of stacking workpiece posture and pick-up method based on RGBD cameras
CN109102547A (en) * 2018-07-20 2018-12-28 上海节卡机器人科技有限公司 Robot based on object identification deep learning model grabs position and orientation estimation method
CN111507334A (en) * 2019-01-30 2020-08-07 中国科学院宁波材料技术与工程研究所 Example segmentation method based on key points
CN112116551A (en) * 2019-06-20 2020-12-22 腾讯科技(深圳)有限公司 Camera shielding detection method and device, electronic equipment and storage medium
CN110472534A (en) * 2019-07-31 2019-11-19 厦门理工学院 3D object detection method, device, equipment and storage medium based on RGB-D data
CN111445524A (en) * 2020-03-31 2020-07-24 清华大学 Recognition method of unsafe behavior of construction site workers based on scene understanding
CN112347887A (en) * 2020-10-28 2021-02-09 深圳市优必选科技股份有限公司 Object detection method, object detection device and electronic equipment
CN112837371A (en) * 2021-02-26 2021-05-25 梅卡曼德(北京)机器人科技有限公司 Object grabbing method and device based on 3D matching and computing equipment
CN113538486A (en) * 2021-07-13 2021-10-22 长春工业大学 Method for improving identification and positioning accuracy of automobile sheet metal workpiece
CN114140526A (en) * 2021-11-19 2022-03-04 浙江大学 Disordered workpiece three-dimensional visual pose estimation method based on deep learning
CN114155265A (en) * 2021-12-01 2022-03-08 南京林业大学 3D LiDAR road point cloud segmentation method based on YOLACT
CN114494276A (en) * 2022-04-18 2022-05-13 成都理工大学 Two-stage multi-modal three-dimensional instance segmentation method

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115359112A (en) * 2022-10-24 2022-11-18 爱夫迪(沈阳)自动化科技有限公司 Stacking control method of high-level material warehouse robot
CN115359112B (en) * 2022-10-24 2023-01-03 爱夫迪(沈阳)自动化科技有限公司 Stacking control method of high-level material warehouse robot
CN116038697A (en) * 2022-12-29 2023-05-02 中山大学 A method and system for automatic spraying of jeans based on manual teaching
CN116363085B (en) * 2023-03-21 2024-01-12 江苏共知自动化科技有限公司 Industrial part target detection method based on small sample learning and virtual synthesized data
CN116363085A (en) * 2023-03-21 2023-06-30 江苏共知自动化科技有限公司 Industrial part target detection method based on small sample learning and virtual synthesized data
CN116320357A (en) * 2023-05-17 2023-06-23 浙江视觉智能创新中心有限公司 3D structured light camera system, method, electronic device and readable storage medium
CN116330306B (en) * 2023-05-31 2023-08-15 之江实验室 Object grabbing method and device, storage medium and electronic equipment
CN116330306A (en) * 2023-05-31 2023-06-27 之江实验室 Object grabbing method and device, storage medium and electronic equipment
CN117095002A (en) * 2023-10-19 2023-11-21 深圳市信润富联数字科技有限公司 Hub defect detection method and device and storage medium
CN117095002B (en) * 2023-10-19 2024-02-06 深圳市信润富联数字科技有限公司 Hub defect detection method and device and storage medium
CN117697768A (en) * 2024-02-05 2024-03-15 季华实验室 A target grabbing method, robot, electronic device and storage medium
CN117697768B (en) * 2024-02-05 2024-05-07 季华实验室 A target grabbing method, robot, electronic device and storage medium
CN118003340A (en) * 2024-04-08 2024-05-10 厦门熠明机器人自动化有限公司 Visual mechanical arm material grabbing control method and system based on deep learning
CN118288292A (en) * 2024-05-10 2024-07-05 中原动力智能机器人有限公司 A method, device, equipment and storage medium for randomly grasping rod-shaped workpieces
CN118288292B (en) * 2024-05-10 2024-12-13 中原动力智能机器人有限公司 Unordered grabbing method, unordered grabbing device, unordered grabbing equipment and unordered grabbing storage medium for rod-shaped workpieces

Also Published As

Publication number Publication date
CN114952809B (en) 2023-08-01

Similar Documents

Publication Publication Date Title
CN114952809B (en) Workpiece recognition and pose detection method, system, and grasping control method of a robotic arm
CN109255813B (en) Man-machine cooperation oriented hand-held object pose real-time detection method
CN108656107B (en) Mechanical arm grabbing system and method based on image processing
CN113112504B (en) A plant point cloud data segmentation method and system
Guo et al. Object discovery and grasp detection with a shared convolutional neural network
CN109986560B (en) Mechanical arm self-adaptive grabbing method for multiple target types
CN111046948B (en) Workpiece pose recognition and robot loading method based on point cloud simulation and deep learning
CN112906797B (en) Plane grabbing detection method based on computer vision and deep learning
CN116416444B (en) Object grabbing point estimation, model training and data generation method, device and system
CN110246127A (en) Workpiece identification and localization method and system, sorting system based on depth camera
CN112509063A (en) Mechanical arm grabbing system and method based on edge feature matching
TW201027288A (en) Method of teaching robotic system
CN115147488B (en) A workpiece pose estimation method and grasping system based on dense prediction
Huang et al. A case study of cyber-physical system design: Autonomous pick-and-place robot
CN112801988B (en) Object grabbing pose detection method based on RGBD and deep neural network
WO2023092519A1 (en) Grabbing control method and apparatus, and electronic device and storage medium
CN110969660A (en) Robot feeding system based on three-dimensional stereoscopic vision and point cloud depth learning
CN115330734A (en) Automatic robot repair welding system based on three-dimensional target detection and point cloud defect completion
Wu et al. Learning affordance space in physical world for vision-based robotic object manipulation
Peng et al. A self-supervised learning-based 6-DOF grasp planning method for manipulator
CN113192128A (en) Mechanical arm grabbing planning method and system combined with self-supervision learning
CN115578460A (en) Robot Grasping Method and System Based on Multimodal Feature Extraction and Dense Prediction
Torii et al. Model-less estimation method for robot grasping parameters using 3D shape primitive approximation
CN117325170A (en) Method for grasping hard disk rack by robotic arm guided by depth vision
CN115319739A (en) A method of grasping workpieces based on vision manipulator

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant