WO2023083273A1 - Grip point information acquisition method and apparatus, electronic device, and storage medium - Google Patents

Grip point information acquisition method and apparatus, electronic device, and storage medium Download PDF

Info

Publication number
WO2023083273A1
WO2023083273A1 PCT/CN2022/131217 CN2022131217W WO2023083273A1 WO 2023083273 A1 WO2023083273 A1 WO 2023083273A1 CN 2022131217 W CN2022131217 W CN 2022131217W WO 2023083273 A1 WO2023083273 A1 WO 2023083273A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
point
item
mask
grabbed
Prior art date
Application number
PCT/CN2022/131217
Other languages
French (fr)
Chinese (zh)
Inventor
李辉
司林林
丁有爽
邵天兰
Original Assignee
梅卡曼德(北京)机器人科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN202111329097.7A external-priority patent/CN114022342A/en
Priority claimed from CN202111329083.5A external-priority patent/CN114022341A/en
Priority claimed from CN202111338191.9A external-priority patent/CN114092428A/en
Priority claimed from CN202111329085.4A external-priority patent/CN114037595A/en
Application filed by 梅卡曼德(北京)机器人科技有限公司 filed Critical 梅卡曼德(北京)机器人科技有限公司
Publication of WO2023083273A1 publication Critical patent/WO2023083273A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods

Definitions

  • the embodiments of the present application relate to the field of automatic control and program control of a robot arm or a fixture, and in particular to an order processing method, device, equipment, system, medium and product.
  • the automatic control of the robotic arm or fixture needs to be based on the point cloud data of the object to be grasped.
  • the usual practice is to first use a 3D camera to collect point cloud data, and then determine the trajectory point or grasping point of the robot arm based on the point cloud data, and then control the robot arm to execute at an appropriate position based on the trajectory point or grasping point and other information. position to perform the grab operation with the appropriate trajectory.
  • the robot arm operates in a three-dimensional space, and its motion track is a three-dimensional track in the three directions of length, width, and height. Therefore, the track point or grab point should also have three-dimensional information of the X-axis, Y-axis, and Z-axis.
  • the point cloud data collected by the camera includes information in three dimensions.
  • it is difficult to obtain clear point cloud information of grasped objects such as in the cosmetics industry, where glass, especially black glass bottles, are used to hold liquids.
  • This kind of glass bottle has a weak light signal, and the reflective material is easily interfered by multiple reflections from surrounding objects.
  • the glass is transparent, so there will be a lot of diffuse reflection and multiple reflection data, making it difficult to collect suitable point cloud data.
  • the present application is proposed in order to overcome the above-mentioned problems or at least partially solve the above-mentioned problems. Specifically, firstly, when the point cloud of the object to be grasped cannot be obtained, the application can obtain the grasping point information of the object to be grasped with the help of other image data such as color pictures, so that the robot or the gripper can directly rely on the grasping point The grabbing of the item to be grabbed can be achieved by obtaining point information without the help of the point cloud of the item, which effectively solves the problem of item grabbing in the absence of point cloud; secondly, this application proposes to correct the mask and obtain the The method of obtaining point-related information makes it possible to correct the inaccurate mask when the accurate item mask cannot be obtained, so as to obtain an accurate mask as much as possible and further obtain accurate grab point information, effectively avoiding the The inaccurate grasping point caused by the inaccurate mask leads to the problem of inaccurate grasping or falling when grasping; thirdly, this application proposes a method for
  • the present application provides a method, device, electronic device, and storage medium for acquiring grabbing point information.
  • the grasping point information used to control the robot to grasp the item to be grasped is obtained, wherein the grasping point association information is the parameter information used to determine the grasping point information, and the grasping point information is used to represent The parameter information needed by the robot to grab the item to be grabbed.
  • an image data processing method for mask correction and capture point related information acquisition including:
  • the mask of the item to be grasped is further processed to obtain the correction mask of the object to be grasped, and the grasping point association information is obtained based on the correction mask.
  • a method for obtaining a correction mask of an item and obtaining associated information of grabbing points including:
  • the correction mask of the item to be grasped is obtained based on the inscribed circle, and the relevant information of the grasping point is obtained based on the correction mask.
  • another method for obtaining a correction mask of an item and obtaining associated information of grabbing points including:
  • the correction mask of the object to be grasped is obtained based on the processing result of the circle detection algorithm, and the relevant information of the grasping point is obtained based on the correction mask.
  • yet another method for obtaining a correction mask of an item and obtaining associated information of grabbing points including:
  • the template matching algorithm is used to process the mask of the item to be grabbed
  • the correction mask of the object to be grasped is obtained based on the processing result of the template matching algorithm, and the relevant information of the grasping point is obtained based on the correction mask.
  • a method for converting grabbing point association information into grabbing point information including:
  • the grabbing point information of the item to be grabbed is generated.
  • An image acquisition module configured to acquire non-point cloud images containing items to be captured
  • a mask generation module is used to process the non-point cloud image to obtain the mask of the object to be grabbed
  • the mask processing module is used to process the mask of the item to be grabbed, so as to obtain the associated information of the grabbing point;
  • the grasping point information generation module is used to obtain the grasping point information for controlling the robot to grasp the item to be grasped based on the grasping point association information, wherein the grasping point association information is a parameter for determining the grasping point information Information, the grasping point information is used to indicate the parameter information required by the robot to grasp the object to be grasped.
  • the electronic device includes a memory, a processor, and a computer program stored on the memory and operable on the processor.
  • the processor executes the computer program, the method for obtaining grabbing point information in any of the above embodiments is implemented.
  • the computer-readable storage medium in the embodiments of the present application stores a computer program thereon, and when the computer program is executed by a processor, the method for acquiring grabbing point information in any of the above-mentioned embodiments is implemented.
  • FIG. 1 is a schematic flow diagram of a method for acquiring information on grabbing points in a scene where the point cloud is not good in some embodiments of the present application;
  • Fig. 2 is a schematic flowchart of a method for obtaining correction mask and information related to grabbing points in some embodiments of the present application;
  • Fig. 3 is a schematic flowchart of a method for converting capture point associated information into capture point information in some embodiments of the present application
  • Fig. 4 is a schematic diagram of an item to be grabbed and an item mask acquired using a non-dedicated deep learning network in some embodiments of the present application;
  • Fig. 5 is a schematic structural diagram of a capture point information acquisition device in some embodiments of the present application.
  • Fig. 6 is a schematic structural diagram of an electronic device according to some embodiments of the present application.
  • Fig. 1 shows a schematic flow chart of a method for acquiring information on an item grabbing point according to an embodiment of the present application. As shown in Fig. 1, the method includes:
  • Step S100 acquiring a non-point cloud image containing an item to be captured
  • Step S110 processing the non-point cloud image to obtain a mask of the object to be captured
  • Step S120 process the mask of the item to be captured to obtain the information related to the captured point, where the associated information of the captured point is parameter information for determining the information of the captured point;
  • Step S130 based on the grasping point associated information, acquiring grasping point information for controlling the robot to grasp the object to be grasped.
  • step S100 in an industrial scene, it is usually necessary to use a robot to grab a large number of items, so there may be multiple items to be grabbed instead of one.
  • the items to be grabbed can be a group of black cosmetic bottles placed in a material frame, and the robot needs to grab the cosmetic bottles from the material frame and transport them to other locations.
  • the light signal of black glass itself is weak, and it is easy to be interfered by multiple reflections of surrounding objects, there are a lot of diffuse reflection and multiple reflection data, so when shooting with an industrial camera, it is likely that it is impossible to obtain a point cloud, which is difficult to collect. Control the required 3D point cloud data. Therefore, the method of the present application is particularly suitable for use in such a scenario, to identify objects to be grasped by means of image data other than point cloud data, calculate grasping point information, and control the robot to perform grasping.
  • the non-point cloud image data in this embodiment may preferably be a color picture of the item, such as a 2D color picture. Different from the point cloud, the 2D color image can clearly identify the items contained in it, even if the item is an item such as a black transparent glass bottle that cannot generate a point cloud, it can be clearly photographed and identified. It can be photographed by an industrial camera, and the group of objects to be captured is placed under the visual sensor to obtain the image data of the objects to be captured.
  • step S110 any existing method can be used to calculate the mask of the item.
  • the mask of the item can be obtained based on the deep learning network to generate the mask of the item to be grabbed.
  • Image recognition is a conventional application of a deep learning network, and a general deep learning network capable of performing image recognition already exists in the prior art.
  • an existing pre-built deep learning network for identifying items or extracting item masks can be used, and non-point cloud images can be input into the deep learning network for processing to obtain masks.
  • the network can be trained in advance.
  • Input the acquired 2D color image into the deep learning network process the color image through the deep learning network, identify the area of interest in the image to generate a mask, generate an image mask in this area, and use the image mask to cover the original image area .
  • the mask of the object to be grasped can be obtained on the basis of the color image.
  • the associated information of the grabbing point of the item may be calculated in the mask area.
  • the grasping point related information is information that lacks some information compared with the grasping point information and is difficult to achieve the purpose of robot grasping, or information that cannot be directly used by the robot but can be used to calculate the grasping point information.
  • the specific position of the grabbing point is related to the clamp used and the item to be grabbed, for example, the grabbing point may include the center point of the grabable area of the item.
  • the center point of the mouth of the cosmetic bottle can be selected as the grabbing point.
  • a non-customized deep learning network that can use and identify items in various situations.
  • identifying items in the photos and extracting item masks the generated masks usually do not match the captured items. Together, they may sometimes be slightly beyond the item's graspable area, sometimes may be slightly smaller than the item's graspable area, and the shape of the mask often does not coincide with the item's graspable area.
  • Figure 4 shows the item to be grabbed and the typical mask of the item to be grabbed generated by the deep learning network
  • Figure 4 shows the scene where the item to be grabbed is the above-mentioned cosmetic bottle, where the The shaped part is the bottleneck area of the glass bottle to be grasped, that is, the grasping area of the robotic arm.
  • the shaded part is the mask extracted by the general deep learning network. It can be seen that in this case, the mask of the item is not accurate, and the grasping point calculated from this is naturally inaccurate. The inaccurate grasping point is It may cause problems such as not being able to catch the bottle or dropping the bottle when grabbing.
  • One solution is to design a deep learning network dedicated to this scenario, and through repeated training to improve the accuracy of the processing results after the image is input into the network, so that an accurate mask can be obtained after the image is input into the deep learning network and grab points.
  • Another solution is to process the inaccurate mask and obtain grab point information from the processed mask.
  • This application preferably uses the latter solution. Specifically, this application provides a lower-cost and more general method for mask correction and grabbing point acquisition, which is also one of the key points of this application.
  • Fig. 2 shows a schematic flowchart of an image data processing method for mask correction and capture point related information acquisition according to an embodiment of the present application. As shown in Figure 2, the methods include:
  • Step S200 performing morphological processing on the mask of the object to be grabbed
  • step S210 the mask of the item to be grasped is further processed to obtain a correction mask of the object to be grasped, and information related to the grasping point is obtained based on the correction mask.
  • the item mask obtained by a general deep learning network usually cannot perfectly fit the object outline (the grasping area in Figure 4 is the area where the bottle mouth is located, it is not difficult to see that the mask There is a large gap between the film and the actual bottle mouth), showing a crooked state, and there may be many holes inside, which has little impact on general object recognition applications, but this application is applied to the industry of grabbing objects In the scene, the precision requirement is high, and such errors are intolerable. Therefore, it is necessary to process the obtained mask.
  • the first thing to do is morphological processing, which changes the graphic form of the mask area.
  • the morphological treatment may be a swelling treatment.
  • the image is expanded to fill in the defects such as lack and irregularity of the image. For example, for each pixel point on the mask, a certain number of points around the point, such as 8-25 points, can be set to have the same color as the point. This step is equivalent to filling the surroundings of each pixel, so if there is a missing part in the item mask, this operation will fill in all the missing parts. After this process, the item mask will become complete without missing, and at the same time The overall mask will also become slightly "fat” due to expansion, and proper expansion will help subsequent further image processing operations.
  • the morphological processing can also be opening operation processing or closing operation processing.
  • Both the opening operation processing and the closing operation processing are processing methods that combine expansion processing and erosion processing. The processing removes the overfilled part, so as to better improve the graphic accuracy of the mask area.
  • step S210 the present application discloses three processing methods to obtain the correction mask of the item and obtain the relevant information of the grabbing point.
  • the first treatment method includes:
  • Step S220 obtaining the circumscribed rectangle of the mask of the item to be grabbed
  • Step S221 based on the circumscribed rectangle of the mask of the item to be captured, an inscribed circle of the circumscribed rectangle is generated;
  • step S222 the correction mask of the object to be grasped is obtained based on the inscribed circle, and the relevant information of the grasping point is obtained based on the correction mask.
  • any circumscribing rectangle algorithm may be used to obtain a circumscribing rectangle for the mask.
  • the four corner points of the circumscribed rectangle can be generated based on the mask of the item to be grasped, and then the circumscribed rectangle can be generated based on the corner points. Specifically, calculate the X coordinate value and Y coordinate value of each pixel in the mask, respectively select the smallest X value, the smallest Y value, the largest X value, and the largest Y value; then, combine the four values into The coordinates of the point, that is, the minimum X value and the minimum Y value form the coordinates (Xmin, Ymin), the maximum X value and the Y value form the coordinates (Xmax, Ymax), and the minimum X value and the maximum Y value form the coordinates (Xmin , Ymax), and the largest X value and the smallest Y value form the coordinates (Xmax, Ymin). In points (Xmin, Ymin), (Xmax, Ymax), (Xmin, Ymax), (Xmax, Ymin) as the four corners of the circumscrib
  • step S221 the key point of this application is to use the inscribed circle algorithm as a part of calculating the correction mask and grab point information, without making any improvements to the inscribed circle algorithm, so the specific inscribed circle algorithm is not limited, and any inscribed circle algorithm is not limited. All inscribed circle algorithms can be used in this application, as long as the selected inscribed circle algorithm can obtain the inscribed circle of the above-mentioned circumscribed rectangle.
  • the mask part enclosed by the inscribed circle can be calculated, and this part of the mask can be used as a correction mask for the object to be grasped.
  • the contour of the obtained correction mask is the same shape and size as the inscribed circle, and the position of the center of the circle is calculated on the correction mask of the inscribed circle, and the information of the center of the circle is obtained, and the information of the center of the circle is used as the relevant information of the grab point.
  • the two-dimensional position information of the circle center for example, the X-axis and Y-axis information of the circle center, may be used as the X-axis position information and the Y-axis position information of the grabbing point.
  • the correction mask obtained in this way has the shape of the mask consistent with the shape of the bottle mouth.
  • the second treatment method includes:
  • Step S230 using the circle detection algorithm to process the mask of the item to be grabbed
  • step S231 the correction mask of the object to be grasped is obtained based on the processing result of the circle detection algorithm, and the relevant information of the grasping point is obtained based on the correction mask.
  • the circle detection algorithm is also called a circle finding algorithm, which can be used to detect circular features in irregular graphics and find circles contained in the graphics.
  • Commonly used algorithms include circle Hough transform algorithm, random Hough transform algorithm, random circle detection algorithm, etc.
  • the focus of this embodiment is to use the circle detection algorithm to find circles from the morphologically processed mask, and does not limit which circle detection algorithm to use specifically. Since the mouth of the bottle itself is circular, and the collected mask contains some features of the shape of the mouth of the bottle, the circle found in the mask area is roughly the position of the mouth of the bottle.
  • step S231 after the circle in the mask is found by the circle detection algorithm, the mask part surrounded by the circle can be used as a correction mask for the object to be grasped. Then calculate the center of the circle, and use the information of the center of the circle as the associated information of the grabbing point. Similar to method 1, the associated information of the grabbing point may be the two-dimensional position information of the center of the circle, and this information is used as the X-axis position information and the Y-axis position information of the grabbing point.
  • the second method does not need to calculate the circumscribed rectangle and inscribed circle that do not exist originally, but only needs to find the circular part from the existing mask area, and the calculation accuracy is higher.
  • the first and second methods have better applicability in the industrial scene where the area to be captured is circular, and can significantly improve the accuracy of the determined associated information of the captured points.
  • the third treatment method includes:
  • Step S240 based on the pre-saved template of the item to be captured, use a template matching algorithm to process the mask of the item to be captured;
  • step S241 the correction mask of the object to be grasped is obtained based on the processing result of the template matching algorithm, and the relevant information of the grasping point is obtained based on the correction mask.
  • the template of the item to be grabbed can be the template of the whole item to be grabbed, or the template of the grabbing area of the item to be grabbed, for example, when the item to be grabbed is a black glass cosmetic bottle, and the clamp
  • the template for the item to be grabbed can be a three-dimensional template for the entire black glass cosmetic bottle, or a template only for the graspable area of the bottle mouth.
  • a matching algorithm is used to match the template in the mask area.
  • the template is equivalent to a known small image
  • the template matching algorithm is equivalent to searching for a target in a large image including a small image. It is known that there is a target in the image, and the target is the same
  • the templates have the same size, orientation and image elements, through which the template matching algorithm can find the target, that is, the small image, and determine its pose.
  • This embodiment does not limit the specific matching algorithm. Since the mask itself will lose color information, the focus is on shape matching rather than color matching. Therefore, this application preferably uses a shape-based matching algorithm for matching.
  • the matching when performing template matching, the matching can be considered successful when the shape similarity reaches 70-95%.
  • Which numerical value is specifically selected can be selected and adjusted according to the needs of actual application scenarios. Certainly, those skilled in the art may also set a specific value or range of values for the above-mentioned shape similarity according to actual matching accuracy requirements.
  • step S241 after the shape matching the pre-stored template is found from the mask, the mask surrounded by the shape can be used as a correction mask, and the relevant information of the grasping point is further calculated.
  • the template Due to the use of the template, no matter what shape the item to be grabbed is in the area to be grabbed, it can be matched and grabbed, not limited to the case where the grabbing area of the item to be grabbed is circular Scenes. Correspondingly, when grabbing different items, the positions of the grabbing points are also different.
  • the grabbing point may be the center point of the correction mask, and the associated information of the grabbing point may be the two-dimensional position information of the center point, that is, the X-axis position information and the Y-axis position information of the grabbing point.
  • the third method can reach a high standard in terms of accuracy and calculation speed, and can be used for grabbing any item.
  • the grasping point association information may be incomplete compared with the grasping point information, so it cannot be directly used by the robot to perform grasping.
  • the grasping point association information is preset, and then the grasping point information for controlling the robot to grasp the object to be grasped is obtained based on the grasping point association information and the preset dimension information missing from the grasping point association information.
  • the associated information of the grasping point can be two-dimensional information of the grasping point (such as X-axis information and Y-axis information), and the grasping point information required by the robot can be three-dimensional information or more dimensional information, such as , the grasping point information is three-dimensional information that also includes Z-axis information, or four-dimensional information that also includes rotation angle/clamping depth, or, the grasping point associated information can be the three-dimensional information of the grasping point, the grasping point required by the robot Point information is four-dimensional information or more dimensional information.
  • the grabbing point information in order to convert the grabbing point associated information into grabbing point information, the grabbing point information can be obtained by pre-setting the missing information of the grabbing point associated information.
  • the grabbing point associated information is two-dimensional Information (or three-dimensional information)
  • the third-dimensional information (or fourth-dimensional information) that is missing in the relevant information of the grab point is preset by pre-detection or manual input.
  • the specific height information can be entered in advance through manual entry, or when the item to be grabbed is a fragile product placed at a specific angle, you can Manually enter its corresponding grabbing angle.
  • the grasping point association information can be further converted into grasping point information based on the preset information. Robot use.
  • This method needs to manually input the missing dimension information before grabbing.
  • the processing efficiency can be significantly improved by manually inputting the unified dimension information.
  • the present application also proposes a method for converting grabbing point association information into grabbing point information without manual intervention and applicable to occasions where the structure and placement of items are not uniform.
  • Fig. 3 shows a schematic flowchart of a method for converting capture point association information into capture point information according to an embodiment of the present application. As shown in Figure 3, the methods include:
  • Step S300 acquiring reference object information of the item to be grabbed
  • Step S310 processing the reference object information of the item to be captured, and obtaining the reference information of the item to be captured, where the reference information is information determined according to the reference object information of the item to be captured;
  • Step S320 based on the reference information of the item to be captured and the associated information of the grabbing point, the information of the grabbing point of the item to be grabbed is generated.
  • the item with a qualified point cloud can be used as a reference object to obtain the missing information in the capture point information of the item to be captured.
  • an item with a point cloud that can recognize the Z-axis information can be used as a reference object.
  • the grasping point association information may be two-dimensional information of the grasping point, for example, the grasping point association information may include X-axis information and Y-axis information of the grasping point.
  • the corresponding grabbing point information may be three-dimensional information of the grabbing point, or four-dimensional or more dimensional information.
  • the current control system needs to determine the information of the X-axis, Y-axis, and Z-axis of the grasping point in order to control the position of the robot’s grasping, after obtaining the two-dimensional information of the grasping point, it cannot To perform grasping based on this information, it is necessary to convert the two-dimensional information of the grasping point into three-dimensional information of the grasping point in a subsequent step.
  • the information that can be obtained through non-point cloud images is usually two-dimensional information (rather than one-dimensional information). Combining the relevant information of the grasping point with the missing height dimension, the corresponding three-dimensional information of the grasping point can be obtained, that is, the grasping point Get some info.
  • the reference object may include other items to be grasped and/or material frames.
  • the reference object may be an item that is closer to the item to be grabbed, or may be other similar items to be grabbed that are placed together with the item to be grabbed.
  • the point cloud of the box is usually complete, and as long as the entire box does not have strong deformation, its height at each position is Consistent, that is, at each position, the height of the material frame is the same as the height of the item to be grabbed or has a fixed height difference from the item to be grabbed, so it is suitable as a reference object. Therefore, when the point cloud of the item to be captured cannot be obtained in this scenario, the 2D color image data of the item and the point cloud data of the recognition frame can be collected for subsequent steps.
  • the reference object should have a qualified point cloud.
  • the quality of the point cloud of each object to be captured is different in the overall point cloud data captured at a certain position. Specifically, it may not be possible to obtain suitable point cloud data of some items to be grasped, but suitable point cloud data of some other items to be grasped may be obtained. In this case, the object to be grasped with a better point cloud can be selected as a reference for other objects to be grasped.
  • the point cloud information can be obtained through a 3D industrial camera.
  • a 3D industrial camera is generally equipped with two lenses to capture the group of objects to be captured from different angles. After processing, the three-dimensional image of the object can be displayed. Place the group of items to be captured under the visual sensor, and shoot the two lenses at the same time.
  • the relative attitude parameters of the two images obtained use the general binocular stereo vision algorithm to calculate the X, The Y and Z coordinate values and the coordinate orientation of each point are converted into point cloud data of the item group to be captured.
  • components such as laser detectors, visible light detectors such as LEDs, infrared detectors, and radar detectors may also be used to generate point clouds, and this application does not limit the specific implementation methods.
  • the two-dimensional color image corresponding to the three-dimensional object area and the depth image corresponding to the two-dimensional color image may also be acquired along a depth direction perpendicular to the object.
  • the two-dimensional color map corresponds to the image of the plane area perpendicular to the preset depth direction; each pixel in the depth map corresponding to the two-dimensional color map corresponds to each pixel in the two-dimensional color map, and each The value of a pixel is the depth value of the pixel.
  • the obtained reference object information may be a point cloud of the reference object or a depth map of the reference object.
  • the reference information is information determined according to the reference object information of the item to be grabbed. Take an industrial scene in which multiple black glass cosmetic bottles to be captured are arranged in a material frame as an example. After obtaining the overall point cloud of the item group, you can further identify the obtained overall point cloud. A clearer reference For example, from the overall point cloud, the point cloud of the material frame or the point cloud of some bottle mouths with clearer point clouds can be identified. Afterwards, the identified point cloud is processed to extract the height information or Z-axis information therein as reference information. Similarly, depth information (corresponding to z-axis information) can also be extracted from the depth map of the reference object as reference information. Although in this embodiment, the missing information in the two-dimensional information of the grabbing point is height information as an example, those skilled in the art can understand that when the missing information is not height information, the reference information may not be height information.
  • the specific method of generating the grabbing point information of the item to be grabbed can be adjusted by preset reference information, and the reference information is adjusted using the reference information adjustment value, and then based on the adjusted reference information and the item to be grabbed Generate the grabbing point information of the item to be grabbed based on the associated information of the grabbing point.
  • the reference information adjustment value may be set according to the original difference between the reference information and the missing information of the grab point association information.
  • the original difference between the reference information and the missing information of the grasping point association information is the height difference between the point cloud of the reference object and the point cloud of the object to be grasped.
  • the height of the point cloud is the same as the height of the mouth of all bottles. At this time, there is no need to set the reference information adjustment value, or the reference information adjustment value is 0 .
  • the gripper After obtaining the height information of the bottle mouth, combine it with the two-dimensional grasping point information to obtain three-dimensional grasping point information, and then the gripper can perform grasping based on the three-dimensional grasping point information.
  • the height obtained through the point cloud may or may not be the same as the bottle. If the heights are different, that is, there is a height difference between the point cloud of the reference object and the point cloud of the object to be grasped, an adjustment value can be preset according to the height difference between the two.
  • the robots or fixtures appearing in the above embodiments can include various general fixtures.
  • the general fixtures refer to fixtures whose structure has been standardized and have a large scope of application, for example, three-jaw chucks and four-jaw chucks for lathes, Flat pliers and indexing heads for milling machines.
  • the fixture can be divided into manual clamping fixture, pneumatic clamping fixture, hydraulic clamping fixture, gas-hydraulic linkage clamping fixture, electromagnetic fixture, vacuum fixture, etc., or other items that can be picked up bionic devices.
  • the present application does not limit the specific type of the gripper, as long as it can realize the grabbing operation of the item.
  • the application can obtain the grasping point information of the object to be grasped with the help of other image data such as color pictures, so that the robot or the fixture can directly rely on The capture point information can realize the capture of the item to be captured without the help of the point cloud of the item, which effectively solves the problem of item capture in the absence of point cloud;
  • this application proposes three methods for correcting the mask And the method of obtaining the relevant information of the grabbing point makes it possible to correct the inaccurate mask when the accurate item mask cannot be obtained, so as to obtain the accurate mask as much as possible and obtain the grabbing point information, effectively avoiding The inaccurate grasping point caused by the inaccurate mask leads to the problem of inaccurate grasping or falling when grasping;
  • the application proposes a method for the robot to automatically convert the inputted grasping point related information into a grasping point
  • a method for obtaining point information which can automatically obtain reference information that can convert the associated information of
  • FIG. 5 shows an apparatus for obtaining grabbing point information according to yet another embodiment of the present application, which includes:
  • An image acquisition module 400 configured to acquire a non-point cloud image containing an item to be captured
  • a mask generating module 410 configured to process the non-point cloud image to obtain the mask of the item to be grabbed
  • a mask processing module 420 configured to process the mask of the item to be grabbed, so as to obtain the information associated with the grabbing point;
  • the grasping point information generating module 430 is configured to obtain grasping point information for controlling the robot to grasp the item to be grasped based on the grasping point association information, wherein the grasping point association information is used to determine the grasping point information
  • the parameter information and the grasping point information are used to indicate the parameter information required by the robot to grasp the object to be grasped.
  • the image acquisition module 400 specifically includes that the non-point cloud images include color images.
  • the mask generation module 410 is specifically configured to process non-point cloud images based on deep learning to generate a mask of the object to be captured.
  • the mask generation module 410 is specifically configured to pre-build a deep learning network, and input non-point cloud images into the deep learning network for processing.
  • the mask processing module 420 specifically includes that the grabbing point includes a center point of the grabable area of the item.
  • the mask processing module 420 is specifically configured to perform morphological processing on the mask of the object to be grabbed.
  • the mask processing module 420 specifically includes that the morphological processing includes morphological dilation processing.
  • the mask processing module 420 is specifically configured to: obtain the circumscribed rectangle of the mask of the item to be captured; generate an inscribed circle of the circumscribed rectangle based on the circumscribed rectangle of the mask of the item to be captured; Obtain the correction mask of the item to be grasped and/or the associated information of the grasping point.
  • the mask processing module 420 is specifically configured to: generate four corner points of a circumscribed rectangle based on the mask of the item to be grasped; generate a circumscribed rectangle based on the corner points.
  • the mask processing module 420 is specifically configured to: use a circle detection algorithm to process the mask of the item to be captured; obtain a corrected mask and/or capture point association of the item to be captured based on the processing result of the circle detection algorithm information.
  • the mask processing module 420 specifically includes that the circle detection algorithm includes a circle Hough transform algorithm, a random Hough transform algorithm and/or a random circle detection algorithm.
  • the mask processing module 420 is specifically configured to: use a template matching algorithm to process the mask of the item to be captured based on the pre-saved template of the item to be captured; obtain the item to be captured based on the processing result of the template matching algorithm Rectification mask and/or grab point association information for .
  • the mask processing module 420 specifically includes that the matching algorithm includes a shape-based matching algorithm.
  • the grabbing point information generation module 430 is specifically configured to: preset dimension information missing in grabbing point associated information; obtain information for Control the grabbing point information of the robot to grab the item to be grabbed.
  • the grasping point information generation module 430 is specifically configured to: acquire reference object information of the item to be grasped; process the reference object information of the item to be grasped, and obtain reference information of the item to be grasped, and the reference information is based on The information determined by the reference object information of the captured item; according to the reference information of the item to be captured and the related information of the captured point, the information of the captured point of the item to be captured is generated.
  • the grabbing point information generating module 430 specifically includes that the grabbing point associated information includes two-dimensional information of the grabbing point.
  • the grabbing point information generation module 430 specifically includes that the grabbing point associated information includes X-axis information and Y-axis information of the grabbing point.
  • the grabbing point information generation module 430 specifically includes that the grabbing point information includes three-dimensional information of the grabbing point.
  • the grabbing point information generating module 430 specifically includes that the reference information includes Z-axis information.
  • the grabbing point information generation module 430 is specifically configured to: preset reference information adjustment values; use reference information adjustment values to adjust reference information; The grab point information of the grabbed item.
  • the capture point information generating module 430 specifically includes that the reference object information includes a point cloud and/or a depth map of the reference object.
  • the capture point information generation module 430 specifically includes that the reference object has a qualified point cloud.
  • the grabbing point information generating module 430 specifically includes that the reference objects include other items to be grabbed and/or material boxes.
  • each module corresponds to the corresponding steps in the method embodiment.
  • the working principle of each module can also refer to the description of the corresponding steps in the method embodiment. , which will not be repeated here.
  • the present application also provides a computer-readable storage medium, on which a computer program is stored, and when the program is executed by a processor, the method in any one of the above-mentioned implementation modes is implemented.
  • the computer program stored in the computer-readable storage medium in the embodiments of the present application can be executed by the processor of the electronic device.
  • the computer-readable storage medium can be a storage medium built in the electronic device, or can be The storage medium of the electronic device is pluggable and pluggable. Therefore, the computer-readable storage medium in the embodiments of the present application has high flexibility and reliability.
  • the electronic device may be a control system/electronic system configured in a car, a mobile terminal (for example, a smart mobile phone, etc.), a personal computer (PC, such as , desktop computer or notebook computer, etc.), tablet computer, server, etc., the specific embodiments of the present application do not limit the specific implementation of the electronic device.
  • a mobile terminal for example, a smart mobile phone, etc.
  • a personal computer PC, such as , desktop computer or notebook computer, etc.
  • tablet computer server, etc.
  • the electronic device may include: a processor (processor) 1202 , a communication interface (Communications Interface) 1204 , a memory (memory) 1206 , and a communication bus 1208 .
  • processor processor
  • Communication interface Communication interface
  • memory memory
  • communication bus 1208 a communication bus
  • the processor 1202 , the communication interface 1204 , and the memory 1206 communicate with each other through the communication bus 1208 .
  • the communication interface 1204 is used to communicate with network elements of other devices such as clients or other servers.
  • the processor 1202 is configured to execute the program 1210, and specifically, may execute relevant steps in the foregoing method embodiments.
  • the program 1210 may include program codes including computer operation instructions.
  • the processor 1202 may be a central processing unit CPU, or a specific integrated circuit ASIC (Application Specific Integrated Circuit), or one or more integrated circuits configured to implement the embodiments of the present application.
  • the one or more processors included in the electronic device may be of the same type, such as one or more CPUs, or may be different types of processors, such as one or more CPUs and one or more ASICs.
  • the memory 1206 is used to store the program 1210 .
  • the memory 1206 may include a high-speed RAM memory, and may also include a non-volatile memory (non-volatile memory), such as at least one disk memory.
  • Program 1210 may be downloaded and installed from a network via communication interface 1204, and/or installed from removable media.
  • the processor 1202 may be made to perform various operations in the foregoing method embodiments.
  • a "computer-readable medium” may be any device that can contain, store, communicate, propagate or transmit a program for use in or in conjunction with an instruction execution system, device or device.
  • computer-readable media include the following: electrical connection with one or more wires (electronic device), portable computer disk case (magnetic device), random access memory (RAM), Read Only Memory (ROM), Erasable and Editable Read Only Memory (EPROM or Flash Memory), Fiber Optic Devices, and Portable Compact Disc Read Only Memory (CDROM).
  • the computer-readable medium may even be paper or other suitable medium on which the program may be printed, as it may be possible, for example, by optically scanning the paper or other medium, followed by editing, interpretation or other suitable means if necessary. Processing to obtain programs electronically and store them in computer memory.
  • the processor can be a central processing unit (Central Processing Unit, CPU), and can also be other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuit (Application Specific Integrated Circuit, ASIC), off-the-shelf programmable gate array (Field- Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • a general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
  • each part of the embodiments of the present application may be realized by hardware, software, firmware or a combination thereof.
  • various steps or methods may be implemented by software or firmware stored in memory and executed by a suitable instruction execution system.
  • a suitable instruction execution system For example, if implemented in hardware, as in another embodiment, it can be implemented by any one or combination of the following techniques known in the art: Discrete logic circuits, ASICs with suitable combinational logic gates, programmable gate arrays (PGAs), field programmable gate arrays (FPGAs), etc.
  • each functional unit in each embodiment of the present application may be integrated into one processing module, each unit may exist separately physically, or two or more units may be integrated into one module.
  • the above-mentioned integrated modules can be implemented in the form of hardware or in the form of software function modules. If the integrated modules are realized in the form of software function modules and sold or used as independent products, they can also be stored in a computer-readable storage medium.
  • the storage medium mentioned above may be a read-only memory, a magnetic disk or an optical disk, and the like.

Abstract

Disclosed in the present application are a grip point information acquisition method and apparatus, an electronic device, and a storage medium. The grip point information acquisition method comprises: obtaining a non-point cloud image containing an object to be gripped; processing the non-point cloud image to obtain a mask of the object; processing the mask of the object to obtain grip point association information; and on the basis of the grip point association information, obtaining grip point information for controlling a robot to grip the object. According to the present application, under the condition that a point cloud of an object to be gripped cannot be obtained, grip point information of the object can be obtained by means of other image data such as a color picture, so that a robot or a gripper can grip the object directly depending on the grip point information without the aid of the point cloud of the object, and the object gripping problem in a point cloud missing environment is effectively solved.

Description

抓取点信息获取方法、装置、电子设备和存储介质Capture point information acquisition method, device, electronic device and storage medium
本申请要求于2021年11月10日提交中国专利局、申请号为CN 202111329097.7、申请名称为“抓取点信息获取方法、装置、电子设备和存储介质”的中国专利申请的优先权,以及2021年11月10日提交中国专利局、申请号为CN 202111329085.4、申请名称为“图像数据处理方法、装置、电子设备和存储介质”的中国专利申请的优先权,以及2021年11月10日提交中国专利局、申请号为CN 202111338191.9、申请名称为“图像数据处理方法、装置、电子设备和存储介质”的中国专利申请的优先权,以及2021年11月10日提交中国专利局、申请号为CN 202111329083.5、申请名称为“抓取点信息获取方法、装置、电子设备和存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application is required to be submitted to the China Patent Office on November 10, 2021, and the application number is CN 202111329097.7, the priority of the Chinese patent application titled "grabbing point information acquisition method, device, electronic equipment and storage medium", and submitted to the China Patent Office on November 10, 2021, the application number is CN 202111329085.4, the application title is The priority of the Chinese patent application for "image data processing method, device, electronic equipment and storage medium", and submitted to the Chinese Patent Office on November 10, 2021, the application number is CN 202111338191.9, and the application name is "image data processing method, device , electronic equipment and storage medium" and the priority of the Chinese patent application for ", and the application number CN 202111329083.5 submitted to the Chinese Patent Office on November 10, 2021, and the title of the application is "Catch point information acquisition method, device, electronic equipment and storage Medium” Chinese patent application priority, the entire content of which is incorporated in this application by reference.
技术领域technical field
本申请实施例涉及机械手臂或夹具的自动控制、程序控制领域,尤其涉及一种订单处理方法、装置、设备、系统、介质及产品。The embodiments of the present application relate to the field of automatic control and program control of a robot arm or a fixture, and in particular to an order processing method, device, equipment, system, medium and product.
背景技术Background technique
近年来电商、快递等行业快速崛起,为物流行业创造了良好的发展机遇。随着人工智能、机器视觉等技术的成熟,在仓储、搬运、分拣这些物流最基础的部分,自动化设备越来越多,工业机器人代替人工已是大势所趋。而夹具作为工业机器人的重要组成部分,越来越多的被应用于物流行业、分拣、3C行业及食品上料中。In recent years, the rapid rise of e-commerce, express delivery and other industries has created good development opportunities for the logistics industry. With the maturity of technologies such as artificial intelligence and machine vision, there are more and more automation equipment in the most basic parts of logistics such as warehousing, handling, and sorting. It is the general trend for industrial robots to replace labor. As an important part of industrial robots, fixtures are more and more used in logistics industry, sorting, 3C industry and food feeding.
目前,机械手臂或夹具的自动控制,需要基于待抓取物品的点云数据。通常的做法是先使用3D相机采集点云数据,再根据点云数据确定机械手臂运动的轨迹点,或者抓取点等信息,之后基于轨迹点或抓取点等信息控制机械手臂执行在适当的位置以适当的轨迹执行抓取操作。机械手臂在立体空间进行操作,其运动轨迹是在长,宽,高三个方向上的立体轨迹,因而轨迹点或者抓取点也应当具有X轴,Y轴和Z轴三个维度的信息,3D相机采集的点云数据即包括三个维度的信息。然而,在一些工业场景中,很难获得清晰的抓取对象的点云信息,例如在化妆品行业中,使用玻璃,特别是黑色玻璃瓶子,盛放液体。这种玻璃瓶子,自身光信号弱,并且反光材质容易被周围物体的多次反射干扰,而且,玻璃透明,会有很多漫反射和多次反射的数据,很难采集到合适的点云数据。具体地,在这种工业场景下,可能无法获取待抓取物品的点云,也可能采集到的待抓取物品的点云有缺失,或者其它点云不佳的情况,这导致机器人要么识别不出要抓取的物品,要么基于错误的点云计算出错误的抓取点,并使用错误的抓取点执行抓取,导致抓取不到,甚至瓶子掉落。目前现有技术中还没有针对这种点云缺失的工业场景下,基于机器人视觉自动控制机器人执行抓取的方案。At present, the automatic control of the robotic arm or fixture needs to be based on the point cloud data of the object to be grasped. The usual practice is to first use a 3D camera to collect point cloud data, and then determine the trajectory point or grasping point of the robot arm based on the point cloud data, and then control the robot arm to execute at an appropriate position based on the trajectory point or grasping point and other information. position to perform the grab operation with the appropriate trajectory. The robot arm operates in a three-dimensional space, and its motion track is a three-dimensional track in the three directions of length, width, and height. Therefore, the track point or grab point should also have three-dimensional information of the X-axis, Y-axis, and Z-axis. 3D The point cloud data collected by the camera includes information in three dimensions. However, in some industrial scenarios, it is difficult to obtain clear point cloud information of grasped objects, such as in the cosmetics industry, where glass, especially black glass bottles, are used to hold liquids. This kind of glass bottle has a weak light signal, and the reflective material is easily interfered by multiple reflections from surrounding objects. Moreover, the glass is transparent, so there will be a lot of diffuse reflection and multiple reflection data, making it difficult to collect suitable point cloud data. Specifically, in this industrial scenario, it may not be possible to obtain the point cloud of the object to be grasped, or the collected point cloud of the object to be grasped may be missing, or other poor point clouds may cause the robot to either recognize If the item to be grasped is not found, or the wrong grasping point is calculated based on the wrong point cloud, and the wrong grasping point is used to perform the grasping, the bottle cannot be grasped or even the bottle falls. At present, there is no solution in the prior art to automatically control the robot to perform grasping based on the robot vision in the industrial scene where the point cloud is missing.
技术解决方案technical solution
鉴于上述问题,提出了本申请以便克服上述问题或者至少部分地解决上述问题。具体地,首先,本申请能够在无法获得待抓取物品的点云的情况下,借助彩色图片等其他图像数据,获取待抓取物品的抓取点信息,使得机器人或者夹具能够直接依赖该抓取点信息而无需借助物品的点云即可实现对待抓取物品的抓取,有效解决了点云缺失环境下的物品抓取问题;其次,本申请提出了对掩膜进行校正并求取抓取点关联信息的方法,使得在无法获取准确物品掩膜的情况下,能够对不准确的掩膜进行校正,以尽可能获得准确的掩膜并进一步获得准确的抓取点信息,有效避免了掩膜不准确导致的抓取点不准确,进而导致抓不准或抓取时掉落的问题;第三,本申请提出了一种由机器人自动将输入的抓取点关联信息转换为抓取点信息的方法,该方法能够根据待抓取物品的环境特征,自动获取能够将抓取点关联信息转换为抓取点信息的参考信息,并基于参考信息获取抓取点信息供机器人抓取,该方案使得在抓取点信息缺失的情形下,仍能够基于现有信息补充获得完整的抓取点信息,并且减少了人工干预。In view of the above-mentioned problems, the present application is proposed in order to overcome the above-mentioned problems or at least partially solve the above-mentioned problems. Specifically, firstly, when the point cloud of the object to be grasped cannot be obtained, the application can obtain the grasping point information of the object to be grasped with the help of other image data such as color pictures, so that the robot or the gripper can directly rely on the grasping point The grabbing of the item to be grabbed can be achieved by obtaining point information without the help of the point cloud of the item, which effectively solves the problem of item grabbing in the absence of point cloud; secondly, this application proposes to correct the mask and obtain the The method of obtaining point-related information makes it possible to correct the inaccurate mask when the accurate item mask cannot be obtained, so as to obtain an accurate mask as much as possible and further obtain accurate grab point information, effectively avoiding the The inaccurate grasping point caused by the inaccurate mask leads to the problem of inaccurate grasping or falling when grasping; thirdly, this application proposes a method for the robot to automatically convert the inputted grasping point related information into a grasping point The method of point information, which can automatically obtain the reference information that can convert the relevant information of the grab point into the grab point information according to the environmental characteristics of the item to be grabbed, and obtain the grab point information based on the reference information for the robot to grab, This solution makes it possible to supplement and obtain complete grabbing point information based on existing information when the grabbing point information is missing, and reduces manual intervention.
本申请权利要求和说明书所披露的所有方案均具有上述一个或多个创新之处,相应地,能够解决上述一个或多个技术问题。具体地,本申请提供一种抓取点信息获取方法、装置、电子设备和存储介质。All the solutions disclosed in the claims and description of the present application have the above-mentioned one or more innovations, and correspondingly, can solve the above-mentioned one or more technical problems. Specifically, the present application provides a method, device, electronic device, and storage medium for acquiring grabbing point information.
本申请的实施方式的抓取点信息获取方法,包括:The capture point information acquisition method of the embodiment of the present application includes:
获取包含待抓取物品的非点云图像;Obtain non-point cloud images containing items to be captured;
对非点云图像进行处理,以获得待抓取物品的掩膜;Process the non-point cloud image to obtain the mask of the object to be grasped;
对待抓取物品的掩膜进行处理,以获取抓取点关联信息;Process the mask of the item to be grabbed to obtain the associated information of the grabbing point;
基于抓取点关联信息,获取用于控制机器人抓取待抓取物品的抓取点信息,其中,抓取点关联信息为用于确定抓取点信息的参数信息,抓取点信息用于表示机器人抓取待抓取物品时所需要的参数信息。Based on the grasping point association information, the grasping point information used to control the robot to grasp the item to be grasped is obtained, wherein the grasping point association information is the parameter information used to determine the grasping point information, and the grasping point information is used to represent The parameter information needed by the robot to grab the item to be grabbed.
在另一个实施例中,提供了一种用于掩膜矫正及抓取点关联信息获取的图像数据处理方法,包括:In another embodiment, an image data processing method for mask correction and capture point related information acquisition is provided, including:
对待抓取物品的掩膜进行形态学处理;Perform morphological processing on the mask of the item to be grabbed;
对待抓取物品的掩膜进行进一步处理,以获取待抓取物品的矫正掩膜,并基于矫正掩膜得到抓取点关联信息。The mask of the item to be grasped is further processed to obtain the correction mask of the object to be grasped, and the grasping point association information is obtained based on the correction mask.
在又一个实施例中,提供了一种获得物品的矫正掩膜并求取抓取点关联信息的方法,包括:In yet another embodiment, a method for obtaining a correction mask of an item and obtaining associated information of grabbing points is provided, including:
获取待抓取物品的掩膜的外接矩形;Get the circumscribed rectangle of the mask of the item to be grabbed;
基于待抓取物品的掩膜的外接矩形,生成该外接矩形的内接圆;Based on the circumscribed rectangle of the mask of the item to be grabbed, an inscribed circle of the circumscribed rectangle is generated;
基于内接圆获取待抓取物品的矫正掩膜,并基于矫正掩膜得到抓取点关联信息。The correction mask of the item to be grasped is obtained based on the inscribed circle, and the relevant information of the grasping point is obtained based on the correction mask.
在又一个实施例中,提供了另一种获得物品的矫正掩膜并求取抓取点关联信息的方法,包括:In yet another embodiment, another method for obtaining a correction mask of an item and obtaining associated information of grabbing points is provided, including:
用圆检测算法对待抓取物品的掩膜进行处理;Use the circle detection algorithm to process the mask of the item to be grabbed;
基于圆检测算法的处理结果获取待抓取物品的矫正掩膜,并基于矫正掩膜得到抓取点关联信息。The correction mask of the object to be grasped is obtained based on the processing result of the circle detection algorithm, and the relevant information of the grasping point is obtained based on the correction mask.
在又一个实施例中,提供了又一种获得物品的矫正掩膜并求取抓取点关联信息的方法,包括:In yet another embodiment, there is provided yet another method for obtaining a correction mask of an item and obtaining associated information of grabbing points, including:
基于预先保存的待抓取物品的模板,使用模板匹配算法对待抓取物品的掩膜进行处理;Based on the pre-saved template of the item to be grabbed, the template matching algorithm is used to process the mask of the item to be grabbed;
基于模板匹配算法的处理结果获取待抓取物品的矫正掩膜,并基于矫正掩膜得到抓取点关联信息。The correction mask of the object to be grasped is obtained based on the processing result of the template matching algorithm, and the relevant information of the grasping point is obtained based on the correction mask.
在又一个实施例中,提供了一种将抓取点关联信息转换为抓取点信息的方法,包括:In yet another embodiment, a method for converting grabbing point association information into grabbing point information is provided, including:
获取待抓取物品的参照物信息;Obtain the reference object information of the item to be grabbed;
处理待抓取物品的参照物信息,获取待抓取物品的参考信息,参考信息是根据待抓取物品的参照物信息确定的信息;Process the reference object information of the item to be grabbed, and obtain the reference information of the item to be grabbed, and the reference information is information determined according to the reference object information of the item to be grabbed;
基于待抓取物品的参考信息以及抓取点关联信息,生成待抓取物品的抓取点信息。Based on the reference information of the item to be grabbed and the associated information of the grabbing point, the grabbing point information of the item to be grabbed is generated.
本申请的实施方式的抓取点信息获取装置,包括:The capture point information acquisition device in the embodiment of the present application includes:
图像获取模块,用于获取包含待抓取物品的非点云图像;An image acquisition module, configured to acquire non-point cloud images containing items to be captured;
掩膜生成模块,用于对非点云图像进行处理,以获得待抓取物品的掩膜;A mask generation module is used to process the non-point cloud image to obtain the mask of the object to be grabbed;
掩膜处理模块,用于对待抓取物品的掩膜进行处理,以获取抓取点关联信息;The mask processing module is used to process the mask of the item to be grabbed, so as to obtain the associated information of the grabbing point;
抓取点信息生成模块,用于基于抓取点关联信息,获取用于控制机器人抓取待抓取物品的抓取点信息,其中,抓取点关联信息为用于确定抓取点信息的参数信息,抓取点信息用于表示机器人抓取待抓取物品时所需要的参数信息。The grasping point information generation module is used to obtain the grasping point information for controlling the robot to grasp the item to be grasped based on the grasping point association information, wherein the grasping point association information is a parameter for determining the grasping point information Information, the grasping point information is used to indicate the parameter information required by the robot to grasp the object to be grasped.
本申请的实施方式的电子设备包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,处理器执行计算机程序时实现上述任一实施方式的抓取点信息获取方法。The electronic device according to the embodiment of the present application includes a memory, a processor, and a computer program stored on the memory and operable on the processor. When the processor executes the computer program, the method for obtaining grabbing point information in any of the above embodiments is implemented.
本申请的实施方式的计算机可读存储介质其上存储有计算机程序,计算机程序被处理器执行时实现上述任一实施方式的抓取点信息获取方法。The computer-readable storage medium in the embodiments of the present application stores a computer program thereon, and when the computer program is executed by a processor, the method for acquiring grabbing point information in any of the above-mentioned embodiments is implemented.
本申请的附加方面和优点将在下面的描述中部分给出,部分将从下面的描述中变得明显,或通过本申请的实践了解到。Additional aspects and advantages of the application will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the application.
附图说明Description of drawings
本申请的上述和/或附加的方面和优点从结合下面附图对实施方式的描述中将变得明显和容易理解,其中:The above and/or additional aspects and advantages of the present application will become apparent and understandable from the description of the embodiments in conjunction with the following drawings, wherein:
图1是本申请某些实施方式的在点云欠佳场景下的抓取点信息获取方法的流程示意图;FIG. 1 is a schematic flow diagram of a method for acquiring information on grabbing points in a scene where the point cloud is not good in some embodiments of the present application;
图2是本申请某些实施方式的获取矫正掩膜及抓取点关联信息的方法的流程示意图;Fig. 2 is a schematic flowchart of a method for obtaining correction mask and information related to grabbing points in some embodiments of the present application;
图3是本申请某些实施方式的将抓取点关联信息转换为抓取点信息的方法的流程示意图;Fig. 3 is a schematic flowchart of a method for converting capture point associated information into capture point information in some embodiments of the present application;
图4是本申请某些实施方式的待抓取物品及使用非专用的深度学习网络获取的物品掩膜的示意图;Fig. 4 is a schematic diagram of an item to be grabbed and an item mask acquired using a non-dedicated deep learning network in some embodiments of the present application;
图5是本申请某些实施方式的抓取点信息获取装置的结构示意图;Fig. 5 is a schematic structural diagram of a capture point information acquisition device in some embodiments of the present application;
图6是本申请某些实施方式的电子设备的结构示意图。Fig. 6 is a schematic structural diagram of an electronic device according to some embodiments of the present application.
本发明的实施方式Embodiments of the present invention
下面将参照附图更详细地描述本申请的示例性实施例。虽然附图中显示了本申请的示例性实施例,然而应当理解,可以以各种形式实现本申请而不应被这里阐述的实施例所限制。相反,提供这些实施例是为了能够更透彻地理解本申请,并且能够将本申请的范围完整的传达给本领域的技术人员。Exemplary embodiments of the present application will be described in more detail below with reference to the accompanying drawings. Although exemplary embodiments of the present application are shown in the drawings, it should be understood that the present application may be embodied in various forms and should not be limited to the embodiments set forth herein. Rather, these embodiments are provided so that the present application can be more thoroughly understood, and the scope of the present application can be fully conveyed to those skilled in the art.
图1示出了根据本申请一个实施例的物品抓取点信息获取方法的流程示意图,如图1所示,该方法包括: Fig. 1 shows a schematic flow chart of a method for acquiring information on an item grabbing point according to an embodiment of the present application. As shown in Fig. 1, the method includes:
步骤S100,获取包含待抓取物品的非点云图像;Step S100, acquiring a non-point cloud image containing an item to be captured;
步骤S110,对非点云图像进行处理,以获得待抓取物品的掩膜;Step S110, processing the non-point cloud image to obtain a mask of the object to be captured;
步骤S120,对待抓取物品的掩膜进行处理,以获取抓取点关联信息,抓取点关联信息为用于确定抓取点信息的参数信息;Step S120, process the mask of the item to be captured to obtain the information related to the captured point, where the associated information of the captured point is parameter information for determining the information of the captured point;
步骤S130,基于抓取点关联信息,获取用于控制机器人抓取待抓取物品的抓取点信息。Step S130, based on the grasping point associated information, acquiring grasping point information for controlling the robot to grasp the object to be grasped.
对于步骤S100,工业场景中,通常需要使用机器人抓取数量众多的物品,因而待抓取物品可能有多个,而不止一个。For step S100, in an industrial scene, it is usually necessary to use a robot to grab a large number of items, so there may be multiple items to be grabbed instead of one.
作为一个较佳的实施例,待抓取的物品,可以是一组放置在料框中的黑色的化妆瓶,机器人需要从料框中抓取化妆瓶,并将其搬运至其它位置。由于黑色玻璃自身光信号弱,并且容易被周围物体的多次反射干扰,有很多漫反射和多次反射的数据,这样使用工业相机拍摄时,很可能无法获得点云,难以采集到通常进行机器人控制所需要的3D点云数据。因此本申请的方法特别适合在这样的场景下使用,以借助点云数据之外的图像数据识别待抓取物品,计算出抓取点信息,并控制机器人执行抓取。As a preferred embodiment, the items to be grabbed can be a group of black cosmetic bottles placed in a material frame, and the robot needs to grab the cosmetic bottles from the material frame and transport them to other locations. Because the light signal of black glass itself is weak, and it is easy to be interfered by multiple reflections of surrounding objects, there are a lot of diffuse reflection and multiple reflection data, so when shooting with an industrial camera, it is likely that it is impossible to obtain a point cloud, which is difficult to collect. Control the required 3D point cloud data. Therefore, the method of the present application is particularly suitable for use in such a scenario, to identify objects to be grasped by means of image data other than point cloud data, calculate grasping point information, and control the robot to perform grasping.
本实施方式中的非点云图像数据优选地可以是物品的彩色图片,如2D彩色图片。与点云不同,2D彩色图能够较为清晰地识别其中所包含的物品,即便该物品是类似黑色透明玻璃瓶这类无法生成点云的物品,也能够清晰地拍摄并识别。可以通过工业相机进行拍摄,将待抓取物品组置于视觉传感器的下方进行拍摄,获取待抓取物品的图像数据。The non-point cloud image data in this embodiment may preferably be a color picture of the item, such as a 2D color picture. Different from the point cloud, the 2D color image can clearly identify the items contained in it, even if the item is an item such as a black transparent glass bottle that cannot generate a point cloud, it can be clearly photographed and identified. It can be photographed by an industrial camera, and the group of objects to be captured is placed under the visual sensor to obtain the image data of the objects to be captured.
对于步骤S110,可以使用现有的任意方式来计算物品的掩膜。For step S110, any existing method can be used to calculate the mask of the item.
优选地,可以基于深度学习网络来获取物品的掩膜,生成待抓取物品的掩膜。图像识别是深度学习网络的一种常规的应用,现有技术中已经存在通用的能够执行图像识别的深度学习网络。作为一种具体的实施方式,可以使用现有的用于识别物品或提取物品掩膜的预先构建的深度学习网络,并将非点云图像输入深度学习网络进行处理,以处理得到掩膜,为了提高识别的精度,可以预先对网络进行训练。Preferably, the mask of the item can be obtained based on the deep learning network to generate the mask of the item to be grabbed. Image recognition is a conventional application of a deep learning network, and a general deep learning network capable of performing image recognition already exists in the prior art. As a specific implementation, an existing pre-built deep learning network for identifying items or extracting item masks can be used, and non-point cloud images can be input into the deep learning network for processing to obtain masks. To improve the accuracy of recognition, the network can be trained in advance.
将获取的2D彩色图像输入该深度学习网络,通过该深度学习网络处理该彩色图像,识别出欲生成掩膜的图像兴趣区域,在该区域生成图像掩膜,并使用图像掩膜遮掩原图像区域。如此,即可在彩色图像的基础上获取了待抓取物品的掩膜。Input the acquired 2D color image into the deep learning network, process the color image through the deep learning network, identify the area of interest in the image to generate a mask, generate an image mask in this area, and use the image mask to cover the original image area . In this way, the mask of the object to be grasped can be obtained on the basis of the color image.
对于步骤S120,在获取物品的掩膜后,可以在掩膜区域中计算物品的抓取点关联信息。其中,抓取点关联信息为相较于抓取点信息缺少部分信息而难以实现机器人抓取目的的信息,或者不能直接供机器人使用,但是能够用于计算抓取点信息的信息。For step S120, after the mask of the item is acquired, the associated information of the grabbing point of the item may be calculated in the mask area. Among them, the grasping point related information is information that lacks some information compared with the grasping point information and is difficult to achieve the purpose of robot grasping, or information that cannot be directly used by the robot but can be used to calculate the grasping point information.
其中,抓取点的具体位置与所使用的夹具以及待抓取的物品有关,如抓取点可以包括物品可抓取区域的中心点。例如,当待抓取物品是黑色的化妆瓶时,可以选择化妆瓶瓶口的中心点作为抓取点。Wherein, the specific position of the grabbing point is related to the clamp used and the item to be grabbed, for example, the grabbing point may include the center point of the grabable area of the item. For example, when the object to be grabbed is a black cosmetic bottle, the center point of the mouth of the cosmetic bottle can be selected as the grabbing point.
非定制的能够在各种情况下使用并识别物品的深度学习网络,在处理拍摄的照片,识别照片中的物品并提取物品掩膜时,生成的掩膜通常并不会与拍摄的物品严丝合缝的对在一起,有时可能稍稍超过物品的可抓取区域,有时可能略小于物品的可抓取区域,并且掩膜的形状通常与物品可抓取区域不一致。A non-customized deep learning network that can use and identify items in various situations. When processing captured photos, identifying items in the photos and extracting item masks, the generated masks usually do not match the captured items. Together, they may sometimes be slightly beyond the item's graspable area, sometimes may be slightly smaller than the item's graspable area, and the shape of the mask often does not coincide with the item's graspable area.
图4示出了待抓取物品以及通过深度学习网络生成的该待抓取物品的典型的掩膜,具体地,图4示出的是待抓取物品为上述化妆瓶的场景,其中的圆形部分是待抓取玻璃瓶的瓶口区域,即机械臂的抓取区域。阴影部分是通过通用的深度学习网络提取的掩膜,可以看出在这种情形下,物品的掩膜不准确,由此计算出的抓取点自然也不准确,不准确的抓取点则可能导致抓不到瓶子或者抓取时瓶子掉落等问题。Figure 4 shows the item to be grabbed and the typical mask of the item to be grabbed generated by the deep learning network, specifically, Figure 4 shows the scene where the item to be grabbed is the above-mentioned cosmetic bottle, where the The shaped part is the bottleneck area of the glass bottle to be grasped, that is, the grasping area of the robotic arm. The shaded part is the mask extracted by the general deep learning network. It can be seen that in this case, the mask of the item is not accurate, and the grasping point calculated from this is naturally inaccurate. The inaccurate grasping point is It may cause problems such as not being able to catch the bottle or dropping the bottle when grabbing.
一种解决方法是设计专用于此场景下的深度学习网络,并通过反复训练以提高将图像输入网络后,处理结果的准确性,如此将图像输入该深度学习网络后就能得到准确的掩膜和抓取点。One solution is to design a deep learning network dedicated to this scenario, and through repeated training to improve the accuracy of the processing results after the image is input into the network, so that an accurate mask can be obtained after the image is input into the deep learning network and grab points.
另一种解决方法是对不准确的掩膜进行处理,从处理后的掩膜中获取抓取点信息。Another solution is to process the inaccurate mask and obtain grab point information from the processed mask.
本申请优选使用后一种解决方案,具体地,本申请提供了一种成本较低,且较为通用的掩膜矫正及抓取点获取方法,这也是本申请的重点之一。This application preferably uses the latter solution. Specifically, this application provides a lower-cost and more general method for mask correction and grabbing point acquisition, which is also one of the key points of this application.
图2示出了根据本申请一个实施例的用于掩膜矫正及抓取点关联信息获取的图像数据处理方法的流程示意图。如图2所示,方法包括:Fig. 2 shows a schematic flowchart of an image data processing method for mask correction and capture point related information acquisition according to an embodiment of the present application. As shown in Figure 2, the methods include:
步骤S200,对待抓取物品的掩膜进行形态学处理;Step S200, performing morphological processing on the mask of the object to be grabbed;
步骤S210,对待抓取物品的掩膜进行进一步处理,以获取待抓取物品的矫正掩膜,并基于矫正掩膜得到抓取点关联信息。In step S210, the mask of the item to be grasped is further processed to obtain a correction mask of the object to be grasped, and information related to the grasping point is obtained based on the correction mask.
对于步骤S200,如图4所示,以一般的深度学习网络获取的物品掩膜,掩膜区域通常不能完美贴合物体轮廓(图4中的抓取区域是瓶口所在区域,不难看出掩膜与实际的瓶口差距较大),呈现歪歪斜斜的状态,内部也可能存在很多孔洞,这对于一般的物品识别应用来说,影响并不大,但是本申请应用于抓取物品的工业场景中,精度要求较高,这样的误差是不可容忍的。因此需要对获取的掩膜进行处理。首先要进行的是形态学处理,改变掩膜区域的图形形态。For step S200, as shown in Figure 4, the item mask obtained by a general deep learning network usually cannot perfectly fit the object outline (the grasping area in Figure 4 is the area where the bottle mouth is located, it is not difficult to see that the mask There is a large gap between the film and the actual bottle mouth), showing a crooked state, and there may be many holes inside, which has little impact on general object recognition applications, but this application is applied to the industry of grabbing objects In the scene, the precision requirement is high, and such errors are intolerable. Therefore, it is necessary to process the obtained mask. The first thing to do is morphological processing, which changes the graphic form of the mask area.
在一种实施方式中,形态学处理可以为膨胀处理。在获取非点云图像信息后,对图像执行膨胀处理,以填补图像的缺失、不规则等缺陷。例如,对于掩膜上的每一个像素点,可以把该点周围一定数量的点,例如8-25个点,设为与该点颜色相同。该步骤相当于把每个像素点的周围都进行填充,因此假如物品掩膜存在缺失,该操作会将缺失部分全部填充,如此处理之后,物品掩膜就会变得完整,不存在缺失,同时掩膜整体上也会因为膨胀而略为变“胖”,适当的膨胀有助于后续进一步的图像处理操作。In one embodiment, the morphological treatment may be a swelling treatment. After obtaining the non-point cloud image information, the image is expanded to fill in the defects such as lack and irregularity of the image. For example, for each pixel point on the mask, a certain number of points around the point, such as 8-25 points, can be set to have the same color as the point. This step is equivalent to filling the surroundings of each pixel, so if there is a missing part in the item mask, this operation will fill in all the missing parts. After this process, the item mask will become complete without missing, and at the same time The overall mask will also become slightly "fat" due to expansion, and proper expansion will help subsequent further image processing operations.
在一种实施方式中,形态学处理也可以为开运算处理或闭运算处理,开运算处理和闭运算处理均为将膨胀处理与腐蚀处理结合的处理方式,通过膨胀处理填补缺失,并通过腐蚀处理将过度填充的部分去除,从而更好的提高掩膜区域的图形精准度。In one embodiment, the morphological processing can also be opening operation processing or closing operation processing. Both the opening operation processing and the closing operation processing are processing methods that combine expansion processing and erosion processing. The processing removes the overfilled part, so as to better improve the graphic accuracy of the mask area.
对于步骤S210,本申请披露了三种处理方式,以获得物品的矫正掩膜并求取抓取点关联信息。For step S210, the present application discloses three processing methods to obtain the correction mask of the item and obtain the relevant information of the grabbing point.
第一种处理方式包括:The first treatment method includes:
步骤S220,获取待抓取物品的掩膜的外接矩形;Step S220, obtaining the circumscribed rectangle of the mask of the item to be grabbed;
步骤S221,基于待抓取物品的掩膜的外接矩形,生成该外接矩形的内接圆;Step S221, based on the circumscribed rectangle of the mask of the item to be captured, an inscribed circle of the circumscribed rectangle is generated;
步骤S222,基于内接圆获取待抓取物品的矫正掩膜,并基于矫正掩膜得到抓取点关联信息。In step S222, the correction mask of the object to be grasped is obtained based on the inscribed circle, and the relevant information of the grasping point is obtained based on the correction mask.
对于步骤S220,可以使用任意的外接矩形算法对掩膜求取外接矩形。For step S220, any circumscribing rectangle algorithm may be used to obtain a circumscribing rectangle for the mask.
作为一种具体的实施方式,可以基于待抓取物品的掩膜,生成外接矩形的4个角点,然后基于角点生成外接矩形。具体的,计算掩膜中每个像素点的X坐标值和Y坐标值,分别选取最小的X值,最小的Y值,最大的X值以及最大的Y值;接着,将4个值组合为点的坐标,即最小的X值和最小的Y值组成坐标(Xmin, Ymin),最大的X值和Y值组成坐标(Xmax, Ymax),最小的X值和最大的Y值组成坐标(Xmin, Ymax),以及最大的X值和最小的Y值组成坐标(Xmax, Ymin)。以点(Xmin, Ymin),(Xmax, Ymax),(Xmin, Ymax),(Xmax, Ymin)分别作为外接矩形的4个角点并连线后,即获得了该外接矩形。As a specific implementation, the four corner points of the circumscribed rectangle can be generated based on the mask of the item to be grasped, and then the circumscribed rectangle can be generated based on the corner points. Specifically, calculate the X coordinate value and Y coordinate value of each pixel in the mask, respectively select the smallest X value, the smallest Y value, the largest X value, and the largest Y value; then, combine the four values into The coordinates of the point, that is, the minimum X value and the minimum Y value form the coordinates (Xmin, Ymin), the maximum X value and the Y value form the coordinates (Xmax, Ymax), and the minimum X value and the maximum Y value form the coordinates (Xmin , Ymax), and the largest X value and the smallest Y value form the coordinates (Xmax, Ymin). In points (Xmin, Ymin), (Xmax, Ymax), (Xmin, Ymax), (Xmax, Ymin) as the four corners of the circumscribing rectangle and connecting them, the circumscribing rectangle is obtained.
对于步骤S221,本申请的关键在于将内接圆算法作为计算矫正掩膜以及抓取点信息的一环,不对内接圆算法做任何改进,因而不限定具体的内接圆算法,任意的内接圆算法都可以用来本申请中,只要所选用的内接圆算法能够获取上述外接矩形的内接圆即可。For step S221, the key point of this application is to use the inscribed circle algorithm as a part of calculating the correction mask and grab point information, without making any improvements to the inscribed circle algorithm, so the specific inscribed circle algorithm is not limited, and any inscribed circle algorithm is not limited. All inscribed circle algorithms can be used in this application, as long as the selected inscribed circle algorithm can obtain the inscribed circle of the above-mentioned circumscribed rectangle.
对于步骤S222,在获取内接圆后,可以计算内接圆所围住的掩膜部分,将这部分掩膜作为待抓取物品的矫正掩膜。获取的矫正掩膜的轮廓与内接圆形状和大小相同,在内接圆矫正掩膜上计算其圆心的位置并获取圆心的信息,该圆心的信息作为抓取点关联信息。具体地,可以将圆心的二维位置信息,例如,圆心的X轴和Y轴信息,作为抓取点的X轴位置信息和Y轴位置信息。For step S222, after the inscribed circle is obtained, the mask part enclosed by the inscribed circle can be calculated, and this part of the mask can be used as a correction mask for the object to be grasped. The contour of the obtained correction mask is the same shape and size as the inscribed circle, and the position of the center of the circle is calculated on the correction mask of the inscribed circle, and the information of the center of the circle is obtained, and the information of the center of the circle is used as the relevant information of the grab point. Specifically, the two-dimensional position information of the circle center, for example, the X-axis and Y-axis information of the circle center, may be used as the X-axis position information and the Y-axis position information of the grabbing point.
以此方式获取的矫正掩膜,掩膜形状已与瓶口形状保持一致。The correction mask obtained in this way has the shape of the mask consistent with the shape of the bottle mouth.
第二种处理方式包括:The second treatment method includes:
步骤S230,使用圆检测算法对待抓取物品的掩膜进行处理;Step S230, using the circle detection algorithm to process the mask of the item to be grabbed;
步骤S231,基于圆检测算法的处理结果获取待抓取物品的矫正掩膜,并基于矫正掩膜得到抓取点关联信息。In step S231, the correction mask of the object to be grasped is obtained based on the processing result of the circle detection algorithm, and the relevant information of the grasping point is obtained based on the correction mask.
对于步骤S230,圆检测算法也叫找圆算法,可以用于在不规则图形中检测圆形特征,并找出图形中包含的圆形。常用的算法包括圆霍夫变换算法,随机霍夫变换算法,随机圆检测算法等。本实施例的重点在于使用圆检测算法从经形态学处理后的掩膜中寻找圆形,而并不限定具体使用何种圆检测算法。由于瓶口本身为圆形,且采集的掩膜中包含了该瓶口形状的部分特征,因而在掩膜区域中找到的圆形,大致就是瓶口的位置。For step S230, the circle detection algorithm is also called a circle finding algorithm, which can be used to detect circular features in irregular graphics and find circles contained in the graphics. Commonly used algorithms include circle Hough transform algorithm, random Hough transform algorithm, random circle detection algorithm, etc. The focus of this embodiment is to use the circle detection algorithm to find circles from the morphologically processed mask, and does not limit which circle detection algorithm to use specifically. Since the mouth of the bottle itself is circular, and the collected mask contains some features of the shape of the mouth of the bottle, the circle found in the mask area is roughly the position of the mouth of the bottle.
对于步骤S231,通过圆检测算法找到掩膜中的圆形后,即可将该圆形所围住的掩膜部分,作为待抓取物品的矫正掩膜。接着计算其圆心,将该圆心的信息作为抓取点关联信息。与方式一类似,抓取点关联信息可以是圆心的二维位置信息,将该信息作为抓取点的X轴位置信息和Y轴位置信息。For step S231, after the circle in the mask is found by the circle detection algorithm, the mask part surrounded by the circle can be used as a correction mask for the object to be grasped. Then calculate the center of the circle, and use the information of the center of the circle as the associated information of the grabbing point. Similar to method 1, the associated information of the grabbing point may be the two-dimensional position information of the center of the circle, and this information is used as the X-axis position information and the Y-axis position information of the grabbing point.
第二种方式不需要计算原本不存在的外接矩形和内接圆,仅需要从已经存在的掩膜区域中寻找圆形部分,计算的精度更高。另外,第一种和第二种方法在待抓取区域为圆形的工业场景具有较好的适用性,能够显著提高确定的抓取点关联信息的精确性。The second method does not need to calculate the circumscribed rectangle and inscribed circle that do not exist originally, but only needs to find the circular part from the existing mask area, and the calculation accuracy is higher. In addition, the first and second methods have better applicability in the industrial scene where the area to be captured is circular, and can significantly improve the accuracy of the determined associated information of the captured points.
第三种处理方式包括:The third treatment method includes:
步骤S240,基于预先保存的待抓取物品的模板,使用模板匹配算法对待抓取物品的掩膜进行处理;Step S240, based on the pre-saved template of the item to be captured, use a template matching algorithm to process the mask of the item to be captured;
步骤S241,基于模板匹配算法的处理结果获取待抓取物品的矫正掩膜,并基于矫正掩膜得到抓取点关联信息。In step S241, the correction mask of the object to be grasped is obtained based on the processing result of the template matching algorithm, and the relevant information of the grasping point is obtained based on the correction mask.
对于步骤S240,待抓取物品的模板可以是待抓取物品整体的模板,也可以是待抓取物品的抓取区域的模板,例如,在待抓取的物品是黑色玻璃化妆瓶,而夹具需要抓取化妆瓶的瓶口这样的场景下,待抓取物品的模板可以是为整个黑色玻璃化妆瓶建立三维的模板,或者仅为瓶口可抓取区域建立模板。For step S240, the template of the item to be grabbed can be the template of the whole item to be grabbed, or the template of the grabbing area of the item to be grabbed, for example, when the item to be grabbed is a black glass cosmetic bottle, and the clamp In a scene where the mouth of a cosmetic bottle needs to be grabbed, the template for the item to be grabbed can be a three-dimensional template for the entire black glass cosmetic bottle, or a template only for the graspable area of the bottle mouth.
在获得了未经矫正的掩膜区域后,基于预存的模板,使用匹配算法将模板在掩膜区域内进行匹配。简单来说,模板相当于一副已知的小图像,而模板匹配算法则相当于在一副包括小图像的大图像中搜寻目标,已知该图中有要找的目标,且该目标同模板有相同的尺寸、方向和图像元素,通过该模板匹配算法可以在图中找到目标,即小图像,并确定其位姿。After obtaining the uncorrected mask area, based on the pre-stored template, a matching algorithm is used to match the template in the mask area. In simple terms, the template is equivalent to a known small image, and the template matching algorithm is equivalent to searching for a target in a large image including a small image. It is known that there is a target in the image, and the target is the same The templates have the same size, orientation and image elements, through which the template matching algorithm can find the target, that is, the small image, and determine its pose.
本实施例对具体的匹配算法不做限制,由于掩膜本身会损失颜色信息,因此重点是形状上的匹配而非颜色上的匹配,因此本申请优选使用基于形状的匹配算法进行匹配。This embodiment does not limit the specific matching algorithm. Since the mask itself will lose color information, the focus is on shape matching rather than color matching. Therefore, this application preferably uses a shape-based matching algorithm for matching.
此外,综合考虑匹配的效率和准确度,在进行模板匹配时,形状相似度达到70-95%时即可认为匹配成功。具体选取哪个数值,可以根据实际应用场景的需要选取及调整。当然,本领域技术人员也可以根据实际的匹配精度需求,设置上述形状相似度的具体数值或数值范围。In addition, considering the efficiency and accuracy of matching, when performing template matching, the matching can be considered successful when the shape similarity reaches 70-95%. Which numerical value is specifically selected can be selected and adjusted according to the needs of actual application scenarios. Certainly, those skilled in the art may also set a specific value or range of values for the above-mentioned shape similarity according to actual matching accuracy requirements.
对于步骤S241,从掩膜中找到与预存的模板相匹配的形状后,即可将该形状围成的掩膜作为矫正掩膜,并进一步计算抓取点关联信息。For step S241, after the shape matching the pre-stored template is found from the mask, the mask surrounded by the shape can be used as a correction mask, and the relevant information of the grasping point is further calculated.
由于采用了模板,因此无论待抓取区域内,待抓取物品以怎样的形状呈现,都可以匹配到该物品并抓取,并不限于待抓取物品的抓取区域为圆形的这种场景。相应地,在抓取不同的物品时,抓取点的位置也不同。Due to the use of the template, no matter what shape the item to be grabbed is in the area to be grabbed, it can be matched and grabbed, not limited to the case where the grabbing area of the item to be grabbed is circular Scenes. Correspondingly, when grabbing different items, the positions of the grabbing points are also different.
在一个实施例中,抓取点可以为矫正掩膜的中心点,抓取点关联信息可以为该中心点的二维位置信息,即抓取点的X轴位置信息和Y轴位置信息。In one embodiment, the grabbing point may be the center point of the correction mask, and the associated information of the grabbing point may be the two-dimensional position information of the center point, that is, the X-axis position information and the Y-axis position information of the grabbing point.
在工程实践中,第三种方法在精度和运算速度上均能达到较高的标准,并可以用于任意物品的抓取。In engineering practice, the third method can reach a high standard in terms of accuracy and calculation speed, and can be used for grabbing any item.
本领域技术人员可以理解的是,虽然本申请的上述优选实施方式是结合基于经形态学处理的掩膜获取矫正掩膜的方式来描述的,但这并不是限定的,在掩膜无缺陷或缺陷程度较小的情形下,还可以不将掩膜进行形态学处理、直接通过非点云图像获得的掩膜获取矫正掩膜并确定抓取点关联信息,本领域技术人员可根据实际应用时的掩膜缺陷程度,在矫正掩膜前选择性地对掩膜进行形态学处理。Those skilled in the art can understand that although the above-mentioned preferred embodiments of the present application are described in conjunction with the method of obtaining a correction mask based on a morphologically processed mask, this is not a limitation. In the case of a small defect, the mask can not be processed morphologically, and the correction mask can be obtained directly through the mask obtained from the non-point cloud image and the relevant information of the grabbing point can be determined. Those skilled in the art can according to the actual application time The degree of mask defect, the mask is selectively morphologically processed before the mask is corrected.
在步骤S130中,如前,抓取点关联信息可能相较于抓取点信息不完整,因而不能直接供机器人使用以执行抓取,针对此场景,在一种可能的实施方式中,可以通过预先设置抓取点关联信息缺失的维度信息,然后基于抓取点关联信息以及预先设置的抓取点关联信息缺失的维度信息,获取用于控制机器人抓取待抓取物品的抓取点信息。In step S130, as before, the grasping point association information may be incomplete compared with the grasping point information, so it cannot be directly used by the robot to perform grasping. For this scenario, in a possible implementation manner, it can be implemented by The missing dimension information of the grasping point association information is preset, and then the grasping point information for controlling the robot to grasp the object to be grasped is obtained based on the grasping point association information and the preset dimension information missing from the grasping point association information.
作为一种具体示例,抓取点关联信息可以是抓取点的二维信息(如X轴信息和Y轴信息),机器人需要的抓取点信息可以是三维信息或更多维的信息,例如,抓取点信息为还包括Z轴信息的三维信息,或者为还包括旋转角度/夹持深度的四维信息,或者,抓取点关联信息可以是抓取点的三维信息,机器人需要的抓取点信息是四维信息或更多维的信息。As a specific example, the associated information of the grasping point can be two-dimensional information of the grasping point (such as X-axis information and Y-axis information), and the grasping point information required by the robot can be three-dimensional information or more dimensional information, such as , the grasping point information is three-dimensional information that also includes Z-axis information, or four-dimensional information that also includes rotation angle/clamping depth, or, the grasping point associated information can be the three-dimensional information of the grasping point, the grasping point required by the robot Point information is four-dimensional information or more dimensional information.
在此情形下,为了将抓取点关联信息转换为抓取点信息,可通过预先设置抓取点关联信息缺失的信息的方式,得到抓取点信息,如在抓取点关联信息为二维信息(或三维信息)时,通过预先检测或人工输入的方式预先设定抓取点关联信息缺失的第三维信息(或第四维信息)。例如,待抓取物品是无法获得抓取点信息的黑色玻璃化妆瓶时,可以通过人工录入的方式预先输入具体的高度信息,或者待抓取物品是特定角度摆放的易碎品时,可以人工录入其对应的抓取角度。In this case, in order to convert the grabbing point associated information into grabbing point information, the grabbing point information can be obtained by pre-setting the missing information of the grabbing point associated information. For example, the grabbing point associated information is two-dimensional Information (or three-dimensional information), the third-dimensional information (or fourth-dimensional information) that is missing in the relevant information of the grab point is preset by pre-detection or manual input. For example, when the item to be grabbed is a black glass cosmetic bottle for which the information of the grabbing point cannot be obtained, the specific height information can be entered in advance through manual entry, or when the item to be grabbed is a fragile product placed at a specific angle, you can Manually enter its corresponding grabbing angle.
如此,在通过前述方案对非点云图像处理后并获得抓取点的抓取点关联信息后,可以进一步基于预设的信息,将抓取点关联信息补全转换为抓取点信息以供机器人使用。In this way, after the non-point cloud image is processed through the aforementioned scheme and the grasping point association information of the grasping point is obtained, the grasping point association information can be further converted into grasping point information based on the preset information. Robot use.
这种方式需要在抓取前,人工输入缺失的维度信息,在物品的结构和摆放较为统一的场合,通过人工输入统一的维度信息,能够显著提高处理效率。This method needs to manually input the missing dimension information before grabbing. In the case where the structure and arrangement of the items are relatively uniform, the processing efficiency can be significantly improved by manually inputting the unified dimension information.
进一步地,本申请还提出了一种无须人工干预,且还能适用于物品结构和摆放较为不统一的场合的将抓取点关联信息转换为抓取点信息的方法。Furthermore, the present application also proposes a method for converting grabbing point association information into grabbing point information without manual intervention and applicable to occasions where the structure and placement of items are not uniform.
图3示出了根据本申请一个实施例的将抓取点关联信息转换为抓取点信息的方法的流程示意图。如图3所示,方法包括:Fig. 3 shows a schematic flowchart of a method for converting capture point association information into capture point information according to an embodiment of the present application. As shown in Figure 3, the methods include:
步骤S300,获取待抓取物品的参照物信息;Step S300, acquiring reference object information of the item to be grabbed;
步骤S310,处理待抓取物品的参照物信息,获取待抓取物品的参考信息,参考信息是根据待抓取物品的参照物信息确定的信息;Step S310, processing the reference object information of the item to be captured, and obtaining the reference information of the item to be captured, where the reference information is information determined according to the reference object information of the item to be captured;
步骤S320,基于待抓取物品的参考信息以及抓取点关联信息,生成待抓取物品的抓取点信息。Step S320, based on the reference information of the item to be captured and the associated information of the grabbing point, the information of the grabbing point of the item to be grabbed is generated.
对于步骤S300,在待抓取物品点云情况不佳的场景,可以将具有合格点云的物品作为参照物,以获得待抓取物品的抓取点信息中所缺失的信息,例如,在能够获得抓取点的X轴信息和Y轴信息的情况下,具有可识别Z轴信息的点云的物品即可以作为参照物。For step S300, in the scene where the point cloud of the item to be captured is not good, the item with a qualified point cloud can be used as a reference object to obtain the missing information in the capture point information of the item to be captured. In the case of obtaining the X-axis information and Y-axis information of the grasping point, an item with a point cloud that can recognize the Z-axis information can be used as a reference object.
作为一种具体的实施方式,抓取点关联信息可以是抓取点的二维信息,例如抓取点关联系可以包括抓取点的X轴信息和Y轴信息。此时,对应的抓取点信息可以是抓取点的三维信息,也可以是四维或更多维的信息。假设目前所用的控制系统,需要确定抓取点的X轴,Y轴,Z轴三个维度的信息后,才能控制机器人抓取的位置,则在获取抓取点的二维信息后,并不能依据该信息执行抓取,还需要在后续的步骤中将抓取点二维的信息转换为抓取点的三维信息。通过非点云图像能得到的信息,通常都是二维信息(而非一维信息),将抓取点关联信息与缺失的高度维度结合,就可以得到对应的抓取点三维信息,即抓取点信息。As a specific implementation manner, the grasping point association information may be two-dimensional information of the grasping point, for example, the grasping point association information may include X-axis information and Y-axis information of the grasping point. At this time, the corresponding grabbing point information may be three-dimensional information of the grabbing point, or four-dimensional or more dimensional information. Assuming that the current control system needs to determine the information of the X-axis, Y-axis, and Z-axis of the grasping point in order to control the position of the robot’s grasping, after obtaining the two-dimensional information of the grasping point, it cannot To perform grasping based on this information, it is necessary to convert the two-dimensional information of the grasping point into three-dimensional information of the grasping point in a subsequent step. The information that can be obtained through non-point cloud images is usually two-dimensional information (rather than one-dimensional information). Combining the relevant information of the grasping point with the missing height dimension, the corresponding three-dimensional information of the grasping point can be obtained, that is, the grasping point Get some info.
具体的,参照物可以包括其它待抓取物品和/或料框。参照物可以是距离待抓取物品较近的物品,也可以是与待抓取物品放在一起的其它同类待抓取物品。具体地,对于将大量待抓取物品,例如化妆瓶,放置在物料框中的工业场景,由于框的点云通常是完整的,并且只要整个框没有强烈的形变,其在各个位置的高度是一致的,即,在各个位置,物料框的高度与待抓取物品高度相同或者与待抓取物品具有固定的高度差,因此适合作为参照物。因此,当在此场景下无法获取待抓取物品的点云时,可以采集物品2D彩色图像数据以及识别框的点云数据进行后续的步骤。Specifically, the reference object may include other items to be grasped and/or material frames. The reference object may be an item that is closer to the item to be grabbed, or may be other similar items to be grabbed that are placed together with the item to be grabbed. Specifically, for an industrial scene where a large number of items to be grasped, such as cosmetic bottles, are placed in a material box, since the point cloud of the box is usually complete, and as long as the entire box does not have strong deformation, its height at each position is Consistent, that is, at each position, the height of the material frame is the same as the height of the item to be grabbed or has a fixed height difference from the item to be grabbed, so it is suitable as a reference object. Therefore, when the point cloud of the item to be captured cannot be obtained in this scenario, the 2D color image data of the item and the point cloud data of the recognition frame can be collected for subsequent steps.
在其它的实施方式中,参照物应具有合格的点云。使用相机进行拍摄时,由于多个待抓取物品处于不同位置,因此在某一个位置拍摄得到的整体的点云数据中,每个待抓取物品的点云质量是有差异的。具体地说,可能无法获得部分待抓取物品的合适的点云数据,但是可以获得另外一些待抓取物品的合适的点云数据。在此情形下,可以选取其中点云较好的待抓取物品作为其它待抓取物品的参照物。In other embodiments, the reference object should have a qualified point cloud. When shooting with a camera, since multiple objects to be captured are in different positions, the quality of the point cloud of each object to be captured is different in the overall point cloud data captured at a certain position. Specifically, it may not be possible to obtain suitable point cloud data of some items to be grasped, but suitable point cloud data of some other items to be grasped may be obtained. In this case, the object to be grasped with a better point cloud can be selected as a reference for other objects to be grasped.
可以通过3D工业相机获取点云信息,3D工业相机一般装配有两个镜头,分别从不同的角度捕捉待抓取物品组,经过处理后能够实现物体的三维图像的展示。将待抓取物品组置于视觉传感器的下方,两个镜头同时拍摄,根据所得到的两个图像的相对姿态参数,使用通用的双目立体视觉算法计算出待填充物体的各点的X、Y、Z坐标值及各点的坐标朝向,进而转变为待抓取物品组的点云数据。具体实施时,也可以使用激光探测器、LED等可见光探测器、红外探测器以及雷达探测器等元件生成点云,本申请对具体实现方式不作限定。The point cloud information can be obtained through a 3D industrial camera. A 3D industrial camera is generally equipped with two lenses to capture the group of objects to be captured from different angles. After processing, the three-dimensional image of the object can be displayed. Place the group of items to be captured under the visual sensor, and shoot the two lenses at the same time. According to the relative attitude parameters of the two images obtained, use the general binocular stereo vision algorithm to calculate the X, The Y and Z coordinate values and the coordinate orientation of each point are converted into point cloud data of the item group to be captured. During specific implementation, components such as laser detectors, visible light detectors such as LEDs, infrared detectors, and radar detectors may also be used to generate point clouds, and this application does not limit the specific implementation methods.
作为一个示例,也可以沿垂直于物品的深度方向获取与三维物品区域相对应的二维彩色图以及对应于二维彩色图的深度图。其中,二维彩色图对应于与预设深度方向垂直的平面区域的图像;对应于二维彩色图的深度图中的各个像素点与二维彩色图中的各个像素点一一对应,且各个像素点的取值为该像素点的深度值。在一个实施方式中,获取的参照物信息可以是参照物的点云或者参照物的深度图。As an example, the two-dimensional color image corresponding to the three-dimensional object area and the depth image corresponding to the two-dimensional color image may also be acquired along a depth direction perpendicular to the object. Among them, the two-dimensional color map corresponds to the image of the plane area perpendicular to the preset depth direction; each pixel in the depth map corresponding to the two-dimensional color map corresponds to each pixel in the two-dimensional color map, and each The value of a pixel is the depth value of the pixel. In one embodiment, the obtained reference object information may be a point cloud of the reference object or a depth map of the reference object.
对于步骤S310,参考信息是根据待抓取物品的参照物信息确定的信息。以多个待抓取的黑色玻璃化妆瓶排列放置在物料框中这样的工业场景为例,在得到物品组整体的点云之后,可以进一步识别所获取的整体点云之中,较为清晰的参照物的点云,例如,可以从整体点云中,识别出物料框的点云或者某些点云较为清晰的瓶口的点云。之后对识别出的点云进行处理,以提取其中的高度信息或Z轴信息作为参考信息。类似的,也可以通过参照物的深度图提取深度信息(对应z轴信息)作为参考信息。虽然本实施例中以抓取点的二维信息中缺少的信息是高度信息为例,然而本领域技术人员能够理解,在缺少的信息非高度信息时,参考信息也可以不是高度信息。For step S310, the reference information is information determined according to the reference object information of the item to be grabbed. Take an industrial scene in which multiple black glass cosmetic bottles to be captured are arranged in a material frame as an example. After obtaining the overall point cloud of the item group, you can further identify the obtained overall point cloud. A clearer reference For example, from the overall point cloud, the point cloud of the material frame or the point cloud of some bottle mouths with clearer point clouds can be identified. Afterwards, the identified point cloud is processed to extract the height information or Z-axis information therein as reference information. Similarly, depth information (corresponding to z-axis information) can also be extracted from the depth map of the reference object as reference information. Although in this embodiment, the missing information in the two-dimensional information of the grabbing point is height information as an example, those skilled in the art can understand that when the missing information is not height information, the reference information may not be height information.
对于步骤S320,生成待抓取物品的抓取点信息的具体的方法,可以通过预设参考信息调整值,并使用参考信息调整值调整参考信息,然后基于调整后的参考信息以及待抓取物品的抓取点关联信息生成待抓取物品的抓取点信息。其中,参考信息调整值可以根据参考信息与抓取点关联信息的缺失信息之间的原始差异进行设置。例如,在高度维度下,参考信息与抓取点关联信息的缺失信息之间的原始差异即为参照物点云与待抓取物品点云的高度差,下面结合两个示例作进一步说明。For step S320, the specific method of generating the grabbing point information of the item to be grabbed can be adjusted by preset reference information, and the reference information is adjusted using the reference information adjustment value, and then based on the adjusted reference information and the item to be grabbed Generate the grabbing point information of the item to be grabbed based on the associated information of the grabbing point. Wherein, the reference information adjustment value may be set according to the original difference between the reference information and the missing information of the grab point association information. For example, in the height dimension, the original difference between the reference information and the missing information of the grasping point association information is the height difference between the point cloud of the reference object and the point cloud of the object to be grasped. The following two examples are used for further illustration.
如果使用的是瓶口的点云,因为物料框中的瓶子种类相同,因此该点云的高度与所有瓶子的瓶口高度相同,此时无需设置参考信息调整值,或参考信息调整值为0。在获得瓶口的高度信息后,将其与二维的抓取点关联信息向结合,获得三维的抓取点信息,之后夹具就可以基于该三维的抓取点信息执行抓取。If you are using the point cloud of the bottle mouth, because the bottles in the material box are of the same type, the height of the point cloud is the same as the height of the mouth of all bottles. At this time, there is no need to set the reference information adjustment value, or the reference information adjustment value is 0 . After obtaining the height information of the bottle mouth, combine it with the two-dimensional grasping point information to obtain three-dimensional grasping point information, and then the gripper can perform grasping based on the three-dimensional grasping point information.
如果使用的是物料框的点云,则通过点云获取的高度可能与瓶子相同,也可能与瓶子不同。如果高度不同,即存在参照物点云与待抓取物品点云的高度差,此时可以根据两者的高度差预设一调整值。If you are using the point cloud of the material box, the height obtained through the point cloud may or may not be the same as the bottle. If the heights are different, that is, there is a height difference between the point cloud of the reference object and the point cloud of the object to be grasped, an adjustment value can be preset according to the height difference between the two.
在获取物料框的高度信息,以及二维的抓取点关联信息后,可以结合该调整值确定三维的抓取点信息。例如,如果框的高度是10cm,调整值为-2cm,则可以计算获得瓶口高度为10-2=8cm,之后与抓取点的X轴和Y轴信息结合后,获得三维的抓取点信息。After obtaining the height information of the material frame and the associated information of the two-dimensional grasping point, the three-dimensional grasping point information can be determined in combination with the adjustment value. For example, if the height of the box is 10cm, and the adjustment value is -2cm, then the height of the bottle mouth can be calculated as 10-2=8cm, and then combined with the X-axis and Y-axis information of the grabbing point, a three-dimensional grabbing point can be obtained information.
在上述实施例中出现的机器人或者夹具,可以包括各类通用夹具,通用夹具是指结构已经标准化,且有较大适用范围的夹具,例如,车床用的三爪卡盘和四爪卡盘,铣床用的平口钳及分度头等。又如,按夹具所用夹紧动力源,可将夹具分为手动夹紧夹具、气动夹紧夹具、液压夹紧夹具、气液联动夹紧夹具、电磁夹具、真空夹具等,或者其它能够拾取物品的仿生器械。本申请不限定夹具的具体类型,只要能够实现物品抓取操作即可。The robots or fixtures appearing in the above embodiments can include various general fixtures. The general fixtures refer to fixtures whose structure has been standardized and have a large scope of application, for example, three-jaw chucks and four-jaw chucks for lathes, Flat pliers and indexing heads for milling machines. As another example, according to the clamping power source used by the fixture, the fixture can be divided into manual clamping fixture, pneumatic clamping fixture, hydraulic clamping fixture, gas-hydraulic linkage clamping fixture, electromagnetic fixture, vacuum fixture, etc., or other items that can be picked up bionic devices. The present application does not limit the specific type of the gripper, as long as it can realize the grabbing operation of the item.
另外,需要说明的是,虽然本申请的每个实施例都具有特定的特征组合,然而,这些特征在实施例之间的进一步组合和交叉组合也是可行的。In addition, it should be noted that although each embodiment of the present application has a specific combination of features, further combinations and cross-combinations of these features among the embodiments are also feasible.
根据上述实施例,首先,本申请能够在无法获得待抓取物品的点云的情况下,借助彩色图片等其他图像数据,获取待抓取物品的抓取点信息,使得机器人或者夹具能够直接依赖该抓取点信息而无需借助物品的点云即可实现对待抓取物品的抓取,有效解决了点云缺失环境下的物品抓取问题;其次,本申请提出了三种对掩膜进行校正并求取抓取点关联信息的方法,使得在无法获取准确物品掩膜的情况下,能够对不准确的掩膜进行校正,以尽可能获得准确的掩膜并获取抓取点信息,有效避免了掩膜不准确导致的抓取点不准确,进而导致抓不准或抓取时掉落的问题;第三,本申请提出了一种由机器人自动将输入的抓取点关联信息转换为抓取点信息的方法,该方法能够根据待抓取物品的环境特征,自动获取能够将抓取点关联信息转换为抓取点信息的参考信息,并基于参考信息确定抓取点信息供机器人抓取,该方案使得在抓取点信息缺失的情形下,仍能够基于现有信息补充获得完整的抓取点信息,并且减少了人工干预。According to the above-mentioned embodiments, firstly, when the point cloud of the object to be grasped cannot be obtained, the application can obtain the grasping point information of the object to be grasped with the help of other image data such as color pictures, so that the robot or the fixture can directly rely on The capture point information can realize the capture of the item to be captured without the help of the point cloud of the item, which effectively solves the problem of item capture in the absence of point cloud; secondly, this application proposes three methods for correcting the mask And the method of obtaining the relevant information of the grabbing point makes it possible to correct the inaccurate mask when the accurate item mask cannot be obtained, so as to obtain the accurate mask as much as possible and obtain the grabbing point information, effectively avoiding The inaccurate grasping point caused by the inaccurate mask leads to the problem of inaccurate grasping or falling when grasping; thirdly, the application proposes a method for the robot to automatically convert the inputted grasping point related information into a grasping point A method for obtaining point information, which can automatically obtain reference information that can convert the associated information of the grab point into grab point information according to the environmental characteristics of the item to be grabbed, and determine the grab point information based on the reference information for the robot to grab , this scheme makes it possible to obtain complete grabbing point information based on existing information in the absence of grabbing point information, and reduces manual intervention.
图5示出了根据本申请又一个实施例的抓取点信息获取装置,该装置包括:FIG. 5 shows an apparatus for obtaining grabbing point information according to yet another embodiment of the present application, which includes:
图像获取模块400,用于获取包含待抓取物品的非点云图像;An image acquisition module 400, configured to acquire a non-point cloud image containing an item to be captured;
掩膜生成模块410,用于对非点云图像进行处理,以获得待抓取物品的掩膜;A mask generating module 410, configured to process the non-point cloud image to obtain the mask of the item to be grabbed;
掩膜处理模块420,用于对待抓取物品的掩膜进行处理,以获取抓取点关联信息;A mask processing module 420, configured to process the mask of the item to be grabbed, so as to obtain the information associated with the grabbing point;
抓取点信息生成模块430,用于基于抓取点关联信息,获取用于控制机器人抓取待抓取物品的抓取点信息,其中,抓取点关联信息为用于确定抓取点信息的参数信息,抓取点信息用于表示机器人抓取待抓取物品时所需要的参数信息。The grasping point information generating module 430 is configured to obtain grasping point information for controlling the robot to grasp the item to be grasped based on the grasping point association information, wherein the grasping point association information is used to determine the grasping point information The parameter information and the grasping point information are used to indicate the parameter information required by the robot to grasp the object to be grasped.
可选地,图像获取模块400具体包括,非点云图像包括彩色图像。Optionally, the image acquisition module 400 specifically includes that the non-point cloud images include color images.
可选地,掩膜生成模块410具体用于,基于深度学习对非点云图像进行处理,生成待抓取物品的掩膜。Optionally, the mask generation module 410 is specifically configured to process non-point cloud images based on deep learning to generate a mask of the object to be captured.
可选地,掩膜生成模块410具体用于,预先构建深度学习网络,将非点云图像输入深度学习网络进行处理。Optionally, the mask generation module 410 is specifically configured to pre-build a deep learning network, and input non-point cloud images into the deep learning network for processing.
可选地,掩膜处理模块420具体包括,抓取点包括物品可抓取区域的中心点。Optionally, the mask processing module 420 specifically includes that the grabbing point includes a center point of the grabable area of the item.
可选地,掩膜处理模块420具体用于,对待抓取物品的掩膜进行形态学处理。Optionally, the mask processing module 420 is specifically configured to perform morphological processing on the mask of the object to be grabbed.
可选地,掩膜处理模块420具体包括,形态学处理包括形态学的膨胀处理。Optionally, the mask processing module 420 specifically includes that the morphological processing includes morphological dilation processing.
可选地,掩膜处理模块420具体用于:获取待抓取物品的掩膜的外接矩形;基于待抓取物品的掩膜的外接矩形,生成该外接矩形的内接圆;基于内接圆获取待抓取物品的矫正掩膜和/或抓取点关联信息。Optionally, the mask processing module 420 is specifically configured to: obtain the circumscribed rectangle of the mask of the item to be captured; generate an inscribed circle of the circumscribed rectangle based on the circumscribed rectangle of the mask of the item to be captured; Obtain the correction mask of the item to be grasped and/or the associated information of the grasping point.
可选地,掩膜处理模块420具体用于:基于待抓取物品的掩膜,生成外接矩形的4个角点;基于角点生成外接矩形。Optionally, the mask processing module 420 is specifically configured to: generate four corner points of a circumscribed rectangle based on the mask of the item to be grasped; generate a circumscribed rectangle based on the corner points.
可选地,掩膜处理模块420具体用于:使用圆检测算法对待抓取物品的掩膜进行处理;基于圆检测算法的处理结果获取待抓取物品的矫正掩膜和/或抓取点关联信息。Optionally, the mask processing module 420 is specifically configured to: use a circle detection algorithm to process the mask of the item to be captured; obtain a corrected mask and/or capture point association of the item to be captured based on the processing result of the circle detection algorithm information.
可选地,掩膜处理模块420具体包括,圆检测算法包括圆霍夫变换算法、随机霍夫变换算法和/或随机圆检测算法。Optionally, the mask processing module 420 specifically includes that the circle detection algorithm includes a circle Hough transform algorithm, a random Hough transform algorithm and/or a random circle detection algorithm.
可选地,掩膜处理模块420具体用于:基于预先保存的待抓取物品的模板,使用模板匹配算法对待抓取物品的掩膜进行处理;基于模板匹配算法的处理结果获取待抓取物品的矫正掩膜和/或抓取点关联信息。Optionally, the mask processing module 420 is specifically configured to: use a template matching algorithm to process the mask of the item to be captured based on the pre-saved template of the item to be captured; obtain the item to be captured based on the processing result of the template matching algorithm Rectification mask and/or grab point association information for .
可选地,掩膜处理模块420具体包括,匹配算法包括基于形状的匹配算法。Optionally, the mask processing module 420 specifically includes that the matching algorithm includes a shape-based matching algorithm.
可选地,抓取点信息生成模块430具体用于:预先设置抓取点关联信息缺失的维度信息;基于抓取点关联信息以及预先设置的抓取点关联信息缺失的维度信息,获取用于控制机器人抓取待抓取物品的抓取点信息。Optionally, the grabbing point information generation module 430 is specifically configured to: preset dimension information missing in grabbing point associated information; obtain information for Control the grabbing point information of the robot to grab the item to be grabbed.
可选地,抓取点信息生成模块430具体用于:获取待抓取物品的参照物信息;处理待抓取物品的参照物信息,获取待抓取物品的参考信息,参考信息是根据待抓取物品的参照物信息确定的信息;根据待抓取物品的参考信息和抓取点关联信息,生成待抓取物品的抓取点信息。Optionally, the grasping point information generation module 430 is specifically configured to: acquire reference object information of the item to be grasped; process the reference object information of the item to be grasped, and obtain reference information of the item to be grasped, and the reference information is based on The information determined by the reference object information of the captured item; according to the reference information of the item to be captured and the related information of the captured point, the information of the captured point of the item to be captured is generated.
可选地,抓取点信息生成模块430具体包括,抓取点关联信息包括抓取点的二维信息。Optionally, the grabbing point information generating module 430 specifically includes that the grabbing point associated information includes two-dimensional information of the grabbing point.
可选地,抓取点信息生成模块430具体包括,抓取点关联信息包括抓取点的X轴信息和Y轴信息。Optionally, the grabbing point information generation module 430 specifically includes that the grabbing point associated information includes X-axis information and Y-axis information of the grabbing point.
可选地,抓取点信息生成模块430具体包括,抓取点信息包括抓取点的三维信息。Optionally, the grabbing point information generation module 430 specifically includes that the grabbing point information includes three-dimensional information of the grabbing point.
可选地,抓取点信息生成模块430具体包括,参考信息包括Z轴信息。Optionally, the grabbing point information generating module 430 specifically includes that the reference information includes Z-axis information.
可选地,抓取点信息生成模块430具体用于:预设参考信息调整值;使用参考信息调整值调整参考信息;基于调整后的参考信息以及待抓取物品的抓取点关联信息生成待抓取物品的抓取点信息。Optionally, the grabbing point information generation module 430 is specifically configured to: preset reference information adjustment values; use reference information adjustment values to adjust reference information; The grab point information of the grabbed item.
可选地,抓取点信息生成模块430具体包括,参照物信息包括参照物的点云和/或深度图。Optionally, the capture point information generating module 430 specifically includes that the reference object information includes a point cloud and/or a depth map of the reference object.
可选地,抓取点信息生成模块430具体包括,参照物具有合格的点云。Optionally, the capture point information generation module 430 specifically includes that the reference object has a qualified point cloud.
可选地,抓取点信息生成模块430具体包括,参照物包括其它待抓取物品和/或料框。Optionally, the grabbing point information generating module 430 specifically includes that the reference objects include other items to be grabbed and/or material boxes.
上述图5所示的装置实施例中,仅描述了模块的主要功能,各个模块的全部功能与方法实施例中相应步骤相对应,各个模块的工作原理同样可以参照方法实施例中相应步骤的描述,此处不再赘述。In the device embodiment shown in FIG. 5 above, only the main functions of the modules are described, and all the functions of each module correspond to the corresponding steps in the method embodiment. The working principle of each module can also refer to the description of the corresponding steps in the method embodiment. , which will not be repeated here.
本申请还提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现上述任一实施方式的方法。需要指出的是,本申请实施方式的计算机可读存储介质存储的计算机程序可以被电子设备的处理器执行,此外,计算机可读存储介质可以是内置在电子设备中的存储介质,也可以是能够插拔地插接在电子设备的存储介质,因此,本申请实施方式的计算机可读存储介质具有较高的灵活性和可靠性。The present application also provides a computer-readable storage medium, on which a computer program is stored, and when the program is executed by a processor, the method in any one of the above-mentioned implementation modes is implemented. It should be noted that the computer program stored in the computer-readable storage medium in the embodiments of the present application can be executed by the processor of the electronic device. In addition, the computer-readable storage medium can be a storage medium built in the electronic device, or can be The storage medium of the electronic device is pluggable and pluggable. Therefore, the computer-readable storage medium in the embodiments of the present application has high flexibility and reliability.
图6示出了根据本申请实施例的一种电子设备的结构示意图,电子设备可以是汽车中配置的控制系统/电子系统、移动终端(例如,智能移动电话等)、个人计算机(PC,例如,台式计算机或者笔记型计算机等)、平板电脑以及服务器等,本申请具体实施例并不对电子设备的具体实现做限定。6 shows a schematic structural diagram of an electronic device according to an embodiment of the present application. The electronic device may be a control system/electronic system configured in a car, a mobile terminal (for example, a smart mobile phone, etc.), a personal computer (PC, such as , desktop computer or notebook computer, etc.), tablet computer, server, etc., the specific embodiments of the present application do not limit the specific implementation of the electronic device.
如图6所示,该电子设备可以包括:处理器(processor)1202、通信接口(Communications Interface)1204、存储器(memory)1206、以及通信总线1208。As shown in FIG. 6 , the electronic device may include: a processor (processor) 1202 , a communication interface (Communications Interface) 1204 , a memory (memory) 1206 , and a communication bus 1208 .
其中:in:
处理器1202、通信接口1204、以及存储器1206通过通信总线1208完成相互间的通信。The processor 1202 , the communication interface 1204 , and the memory 1206 communicate with each other through the communication bus 1208 .
通信接口1204,用于与其它设备比如客户端或其它服务器等的网元通信。The communication interface 1204 is used to communicate with network elements of other devices such as clients or other servers.
处理器1202,用于执行程序1210,具体可以执行上述方法实施例中的相关步骤。The processor 1202 is configured to execute the program 1210, and specifically, may execute relevant steps in the foregoing method embodiments.
具体地,程序1210可以包括程序代码,该程序代码包括计算机操作指令。Specifically, the program 1210 may include program codes including computer operation instructions.
处理器1202可能是中央处理器CPU,或者是特定集成电路ASIC(Application Specific Integrated Circuit),或者是被配置成实施本申请实施例的一个或多个集成电路。电子设备包括的一个或多个处理器,可以是同一类型的处理器,如一个或多个CPU;也可以是不同类型的处理器,如一个或多个CPU以及一个或多个ASIC。The processor 1202 may be a central processing unit CPU, or a specific integrated circuit ASIC (Application Specific Integrated Circuit), or one or more integrated circuits configured to implement the embodiments of the present application. The one or more processors included in the electronic device may be of the same type, such as one or more CPUs, or may be different types of processors, such as one or more CPUs and one or more ASICs.
存储器1206,用于存放程序1210。存储器1206可能包含高速RAM存储器,也可能还包括非易失性存储器(non-volatile memory),例如至少一个磁盘存储器。The memory 1206 is used to store the program 1210 . The memory 1206 may include a high-speed RAM memory, and may also include a non-volatile memory (non-volatile memory), such as at least one disk memory.
程序1210可以通过通信接口1204从网络上被下载及安装,和/或从可拆卸介质被安装。在该程序被处理器1202执行时,可以使得处理器1202执行上述方法实施例中的各项操作。Program 1210 may be downloaded and installed from a network via communication interface 1204, and/or installed from removable media. When the program is executed by the processor 1202, the processor 1202 may be made to perform various operations in the foregoing method embodiments.
在本说明书的描述中,参考术语“一个实施方式”、“一些实施方式”、“示意性实施方式”、“示例”、“具体示例”或“一些示例”等的描述意指结合实施方式或示例描述的具体特征、结构、材料或者特点包含于本申请的至少一个实施方式或示例中。在本说明书中,对上述术语的示意性表述不一定指的是相同的实施方式或示例。而且,描述的具体特征、结构、材料或者特点可以在任何的一个或多个实施方式或示例中以合适的方式结合。In the description of this specification, descriptions with reference to the terms "one embodiment", "some embodiments", "exemplary embodiments", "example", "specific examples" or "some examples" mean that a combination of the embodiments or Examples describe specific features, structures, materials, or characteristics that are included in at least one embodiment or example of the present application. In this specification, schematic representations of the above terms do not necessarily refer to the same embodiment or example. Furthermore, the described specific features, structures, materials or characteristics may be combined in any suitable manner in any one or more embodiments or examples.
流程图中或在此以其他方式描述的任何过程或方法描述可以被理解为,表示包括一个或更多个用于实现特定逻辑功能或过程的步骤的可执行指令的代码的模块、片段或部分,并且本申请的优选实施方式的范围包括另外的实现,其中可以不按所示出或讨论的顺序,包括根据所涉及的功能按基本同时的方式或按相反的顺序,来执行功能,这应被本申请的实施例所属技术领域的技术人员所理解。Any process or method descriptions in flowcharts or otherwise described herein may be understood to represent modules, segments or portions of code comprising one or more executable instructions for implementing specific logical functions or steps of the process , and the scope of preferred embodiments of the present application includes additional implementations in which functions may be performed out of the order shown or discussed, including in substantially simultaneous fashion or in reverse order depending on the functions involved, which shall It should be understood by those skilled in the art to which the embodiments of the present application belong.
在流程图中表示或在此以其他方式描述的逻辑和/或步骤,例如,可以被认为是用于实现逻辑功能的可执行指令的定序列表,可以具体实现在任何计算机可读介质中,以供指令执行系统、装置或设备(如基于计算机的系统、包括处理模块的系统或其他可以从指令执行系统、装置或设备取指令并执行指令的系统)使用,或结合这些指令执行系统、装置或设备而使用。就本说明书而言,"计算机可读介质"可以是任何可以包含、存储、通信、传播或传输程序以供指令执行系统、装置或设备或结合这些指令执行系统、装置或设备而使用的装置。计算机可读介质的更具体的示例(非穷尽性列表)包括以下:具有一个或多个布线的电连接部(电子装置),便携式计算机盘盒(磁装置),随机存取存储器(RAM),只读存储器(ROM),可擦除可编辑只读存储器(EPROM或闪速存储器),光纤装置,以及便携式光盘只读存储器(CDROM)。另外,计算机可读介质甚至可以是可在其上打印程序的纸或其他合适的介质,因为可以例如通过对纸或其他介质进行光学扫描,接着进行编辑、解译或必要时以其他合适方式进行处理来以电子方式获得程序,然后将其存储在计算机存储器中。The logic and/or steps represented in the flowcharts or otherwise described herein, for example, can be considered as a sequenced listing of executable instructions for implementing logical functions, can be embodied in any computer-readable medium, For use with instruction execution systems, devices, or devices (such as computer-based systems, systems including processing modules, or other systems that can fetch instructions from instruction execution systems, devices, or devices and execute instructions), or in conjunction with these instruction execution systems, devices or equipment used. For the purposes of this specification, a "computer-readable medium" may be any device that can contain, store, communicate, propagate or transmit a program for use in or in conjunction with an instruction execution system, device or device. More specific examples (non-exhaustive list) of computer-readable media include the following: electrical connection with one or more wires (electronic device), portable computer disk case (magnetic device), random access memory (RAM), Read Only Memory (ROM), Erasable and Editable Read Only Memory (EPROM or Flash Memory), Fiber Optic Devices, and Portable Compact Disc Read Only Memory (CDROM). In addition, the computer-readable medium may even be paper or other suitable medium on which the program may be printed, as it may be possible, for example, by optically scanning the paper or other medium, followed by editing, interpretation or other suitable means if necessary. Processing to obtain programs electronically and store them in computer memory.
处理器可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field- Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The processor can be a central processing unit (Central Processing Unit, CPU), and can also be other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuit (Application Specific Integrated Circuit, ASIC), off-the-shelf programmable gate array (Field- Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. A general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
应当理解,本申请的实施方式的各部分可以用硬件、软件、固件或它们的组合来实现。在上述实施方式中,多个步骤或方法可以用存储在存储器中且由合适的指令执行系统执行的软件或固件来实现。例如,如果用硬件来实现,和在另一实施方式中一样,可用本领域公知的下列技术中的任一项或他们的组合来实现:具有用于对数据信号实现逻辑功能的逻辑门电路的离散逻辑电路,具有合适的组合逻辑门电路的专用集成电路,可编程门阵列(PGA),现场可编程门阵列(FPGA)等。It should be understood that each part of the embodiments of the present application may be realized by hardware, software, firmware or a combination thereof. In the embodiments described above, various steps or methods may be implemented by software or firmware stored in memory and executed by a suitable instruction execution system. For example, if implemented in hardware, as in another embodiment, it can be implemented by any one or combination of the following techniques known in the art: Discrete logic circuits, ASICs with suitable combinational logic gates, programmable gate arrays (PGAs), field programmable gate arrays (FPGAs), etc.
本技术领域的普通技术人员可以理解实现上述实施例方法携带的全部或部分步骤是可以通过程序来指令相关的硬件完成,的程序可以存储于一种计算机可读存储介质中,该程序在执行时,包括方法实施例的步骤之一或其组合。Those of ordinary skill in the art can understand that all or part of the steps carried by the methods of the above embodiments can be completed by instructing related hardware through a program, and the program can be stored in a computer-readable storage medium. When the program is executed , including one or a combination of the steps of the method embodiment.
此外,在本申请的各个实施例中的各功能单元可以集成在一个处理模块中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个模块中。上述集成的模块既可以采用硬件的形式实现,也可以采用软件功能模块的形式实现。集成的模块如果以软件功能模块的形式实现并作为独立的产品销售或使用时,也可以存储在一个计算机可读取存储介质中。In addition, each functional unit in each embodiment of the present application may be integrated into one processing module, each unit may exist separately physically, or two or more units may be integrated into one module. The above-mentioned integrated modules can be implemented in the form of hardware or in the form of software function modules. If the integrated modules are realized in the form of software function modules and sold or used as independent products, they can also be stored in a computer-readable storage medium.
上述提到的存储介质可以是只读存储器,磁盘或光盘等。The storage medium mentioned above may be a read-only memory, a magnetic disk or an optical disk, and the like.
尽管上面已经示出和描述了本申请的实施例,可以理解的是,上述实施例是示例性的,不能理解为对本申请的限制,本领域的普通技术人员在本申请的范围内可以对上述实施方式进行变化、修改、替换和变型。Although the embodiments of the present application have been shown and described above, it can be understood that the above embodiments are exemplary and should not be construed as limitations on the present application, and those skilled in the art can make the above-mentioned The embodiments are subject to changes, modifications, substitutions and variations.

Claims (19)

  1. 一种抓取点信息获取方法,其特征在于,包括:A method for acquiring grabbing point information, comprising:
    获取包含待抓取物品的非点云图像;Obtain non-point cloud images containing items to be captured;
    对所述非点云图像进行处理,以获得所述待抓取物品的掩膜;Processing the non-point cloud image to obtain a mask of the item to be captured;
    对所述待抓取物品的掩膜进行处理,以获取抓取点关联信息;Processing the mask of the item to be grasped to obtain the relevant information of the grasping point;
    基于所述抓取点关联信息,获取用于控制机器人抓取待抓取物品的抓取点信息,其中,所述抓取点关联信息为用于确定抓取点信息的参数信息,所述抓取点信息用于表示机器人抓取待抓取物品时所需要的参数信息。Based on the grasping point associated information, the grasping point information used to control the robot to grasp the item to be grasped is acquired, wherein the grasping point associated information is parameter information used to determine the grasping point information, and the grasping point The fetching point information is used to indicate the parameter information required by the robot to grab the item to be grabbed.
  2. 根据权利要求1所述的抓取点信息获取方法,其特征在于,所述对所述非点云图像进行处理,以获得所述待抓取物品的掩膜,包括:The method for obtaining grabbing point information according to claim 1, wherein the processing the non-point cloud image to obtain the mask of the item to be grabbed includes:
    基于深度学习对所述非点云图像进行处理,生成所述待抓取物品的掩膜。The non-point cloud image is processed based on deep learning to generate a mask of the object to be captured.
  3. 根据权利要求1所述的抓取点信息获取方法,其特征在于,所述抓取点包括物品可抓取区域的中心点。The method for obtaining grabbing point information according to claim 1, wherein the grabbing point includes a center point of a grabable area of an item.
  4. 根据权利要求1所述的抓取点信息获取方法,其特征在于,所述对所述待抓取物品的掩膜进行处理,以获取抓取点关联信息,包括:The method for obtaining grabbing point information according to claim 1, wherein the processing of the mask of the item to be grabbed to obtain the grabbing point associated information includes:
    获取所述待抓取物品的掩膜的外接矩形;Obtain the circumscribed rectangle of the mask of the item to be grabbed;
    基于所述待抓取物品的掩膜的外接矩形,生成该外接矩形的内接圆;Generate an inscribed circle of the circumscribed rectangle based on the circumscribed rectangle of the mask of the item to be grabbed;
    基于所述内接圆获取待抓取物品的矫正掩膜,并基于所述矫正掩膜得到所述抓取点关联信息。The correction mask of the item to be grasped is obtained based on the inscribed circle, and the grasping point association information is obtained based on the correction mask.
  5. 根据权利要求4所述的抓取点信息获取方法,其特征在于,所述获取待抓取物品的掩膜的外接矩形,包括:The grabbing point information acquisition method according to claim 4, wherein said acquiring the circumscribed rectangle of the mask of the item to be grabbed includes:
    基于所述待抓取物品的掩膜,生成外接矩形的4个角点;Based on the mask of the item to be grabbed, generate 4 corner points of the circumscribed rectangle;
    基于所述角点生成外接矩形。A bounding rectangle is generated based on the corner points.
  6. 根据权利要求1所述的抓取点信息获取方法,其特征在于,所述对所述待抓取物品的掩膜进行处理,以获取抓取点关联信息,包括:The method for obtaining grabbing point information according to claim 1, wherein the processing of the mask of the item to be grabbed to obtain the grabbing point associated information includes:
    使用圆检测算法对所述待抓取物品的掩膜进行处理;Using a circle detection algorithm to process the mask of the item to be grabbed;
    基于圆检测算法的处理结果获取待抓取物品的矫正掩膜,并基于所述矫正掩膜得到所述抓取点关联信息。The correction mask of the object to be grasped is obtained based on the processing result of the circle detection algorithm, and the relevant information of the grasping point is obtained based on the correction mask.
  7. 根据权利要求1所述的抓取点信息获取方法,其特征在于,所述对所述待抓取物品的掩膜进行处理,以获取抓取点关联信息,包括:The method for obtaining grabbing point information according to claim 1, wherein the processing of the mask of the item to be grabbed to obtain the grabbing point associated information includes:
    基于预先保存的待抓取物品的模板,使用模板匹配算法对所述待抓取物品的掩膜进行处理;Processing the mask of the item to be captured using a template matching algorithm based on the pre-saved template of the item to be captured;
    基于模板匹配算法的处理结果获取待抓取物品的矫正掩膜,并基于所述矫正掩膜得到所述抓取点关联信息。The correction mask of the item to be grasped is obtained based on the processing result of the template matching algorithm, and the relevant information of the grasping point is obtained based on the correction mask.
  8. 根据权利要求7所述的抓取点信息获取方法,其特征在于,所述匹配算法包括基于形状的匹配算法。The method for acquiring grabbing point information according to claim 7, wherein the matching algorithm includes a shape-based matching algorithm.
  9. 根据权利要求4至8中任一项所述的抓取点信息获取方法,其特征在于,所述待抓取物品的掩膜为经过形态学处理的掩膜。The method for acquiring grabbing point information according to any one of claims 4 to 8, wherein the mask of the item to be grabbed is a morphologically processed mask.
  10. 根据权利要求9所述的抓取点信息获取方法,其特征在于,所述形态学处理包括形态学的膨胀处理。The method for acquiring grabbing point information according to claim 9, wherein the morphological processing includes morphological expansion processing.
  11. 根据权利要求1至8中任一项所述的抓取点信息获取方法,其特征在于,所述基于所述抓取点关联信息,获取用于控制机器人抓取待抓取物品的抓取点信息,包括:The grabbing point information acquisition method according to any one of claims 1 to 8, characterized in that, based on the grabbing point associated information, acquiring the grabbing point for controlling the robot to grab the item to be grabbed information, including:
    预先设置抓取点关联信息缺失的维度信息;Pre-set the missing dimension information of the capture point association information;
    基于所述抓取点关联信息以及所述预先设置的抓取点关联信息缺失的维度信息,获取用于控制机器人抓取待抓取物品的抓取点信息。Based on the grasping point association information and the dimension information missing from the preset grasping point association information, the grasping point information for controlling the robot to grasp the object to be grasped is acquired.
  12. 根据权利要求1至8中任一项所述的抓取点信息获取方法,其特征在于,所述基于所述抓取点关联信息,获取用于控制机器人抓取待抓取物品的抓取点信息,包括:The grabbing point information acquisition method according to any one of claims 1 to 8, characterized in that, based on the grabbing point associated information, acquiring the grabbing point for controlling the robot to grab the item to be grabbed information, including:
    获取待抓取物品的参照物信息;Obtain the reference object information of the item to be grabbed;
    处理所述待抓取物品的参照物信息,获取所述待抓取物品的参考信息,所述参考信息是根据待抓取物品的参照物信息确定的信息;Processing the reference object information of the item to be grabbed, and obtaining the reference information of the item to be grabbed, the reference information is information determined according to the reference object information of the item to be grabbed;
    根据所述待抓取物品的参考信息和所述抓取点关联信息,生成所述待抓取物品的抓取点信息。The grabbing point information of the item to be grabbed is generated according to the reference information of the item to be grabbed and the associated information of the grabbing point.
  13. 根据权利要求12所述的抓取点信息获取方法,其特征在于,所述根据所述待抓取物品的参考信息和所述抓取点关联信息,生成所述待抓取物品的抓取点信息,包括:The method for obtaining grabbing point information according to claim 12, wherein the grabbing point of the item to be grabbed is generated according to the reference information of the item to be grabbed and the associated information of the grabbing point information, including:
    预设参考信息调整值;Preset reference information adjustment value;
    使用所述参考信息调整值调整所述参考信息;adjusting the reference information using the reference information adjustment value;
    基于调整后的参考信息以及所述待抓取物品的抓取点关联信息生成待抓取物品的抓取点信息。The grasping point information of the item to be grasped is generated based on the adjusted reference information and the grasping point associated information of the item to be grasped.
  14. 根据权利要求12所述的抓取点信息获取方法,其特征在于,所述参照物信息包括参照物的点云或深度图。The method for acquiring grabbing point information according to claim 12, wherein the reference object information includes a point cloud or a depth map of the reference object.
  15. 根据权利要求12所述的抓取点信息获取方法,其特征在于,所述参照物具有合格的点云。The method for acquiring grab point information according to claim 12, wherein the reference object has a qualified point cloud.
  16. 根据权利要求12所述的抓取点信息获取方法,其特征在于,所述参照物包括其它待抓取物品和/或料框。The method for acquiring grabbing point information according to claim 12, wherein the reference objects include other items to be grabbed and/or material frames.
  17. 一种抓取点信息获取装置,其特征在于,包括:A grabbing point information acquisition device, characterized in that it comprises:
    图像获取模块,用于获取包含待抓取物品的非点云图像;An image acquisition module, configured to acquire non-point cloud images containing items to be captured;
    掩膜生成模块,用于对所述非点云图像进行处理,以获得所述待抓取物品的掩膜;A mask generating module, configured to process the non-point cloud image to obtain a mask of the item to be captured;
    掩膜处理模块,用于对所述待抓取物品的掩膜进行处理,以获取抓取点关联信息;A mask processing module, configured to process the mask of the item to be grabbed, so as to obtain information related to the grabbing point;
    抓取点信息生成模块,用于基于所述抓取点关联信息,获取用于控制机器人抓取待抓取物品的抓取点信息,其中,所述抓取点关联信息为用于确定抓取点信息的参数信息,所述抓取点信息用于表示机器人抓取待抓取物品时所需要的参数信息。A grasping point information generation module, configured to obtain grasping point information for controlling the robot to grasp the item to be grasped based on the grasping point associated information, wherein the grasping point associated information is used to determine the grasping point Parameter information of the point information, the grasping point information is used to represent the parameter information required when the robot grasps the object to be grasped.
  18. 一种电子设备,其特征在于,包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现权利要求1至16中任一项所述的抓取点信息获取方法。An electronic device, characterized by comprising: a memory, a processor, and a computer program stored on the memory and operable on the processor, when the processor executes the computer program, claims 1 to 1 are realized. The grabbing point information acquisition method described in any one of 16.
  19. 一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现权利要求1至16中任一项所述的抓取点信息获取方法。A computer-readable storage medium, on which a computer program is stored, wherein, when the computer program is executed by a processor, the grabbing point information acquisition method according to any one of claims 1 to 16 is implemented.
PCT/CN2022/131217 2021-11-10 2022-11-10 Grip point information acquisition method and apparatus, electronic device, and storage medium WO2023083273A1 (en)

Applications Claiming Priority (8)

Application Number Priority Date Filing Date Title
CN202111329097.7A CN114022342A (en) 2021-11-10 2021-11-10 Acquisition method and device for acquisition point information, electronic equipment and storage medium
CN202111329083.5 2021-11-10
CN202111329085.4 2021-11-10
CN202111329083.5A CN114022341A (en) 2021-11-10 2021-11-10 Acquisition method and device for acquisition point information, electronic equipment and storage medium
CN202111338191.9A CN114092428A (en) 2021-11-10 2021-11-10 Image data processing method, image data processing device, electronic equipment and storage medium
CN202111338191.9 2021-11-10
CN202111329097.7 2021-11-10
CN202111329085.4A CN114037595A (en) 2021-11-10 2021-11-10 Image data processing method, image data processing device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2023083273A1 true WO2023083273A1 (en) 2023-05-19

Family

ID=86335117

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/131217 WO2023083273A1 (en) 2021-11-10 2022-11-10 Grip point information acquisition method and apparatus, electronic device, and storage medium

Country Status (1)

Country Link
WO (1) WO2023083273A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116958265A (en) * 2023-09-19 2023-10-27 交通运输部天津水运工程科学研究所 Ship pose measurement method and system based on binocular vision

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110889460A (en) * 2019-12-06 2020-03-17 中山大学 Mechanical arm specified object grabbing method based on cooperative attention mechanism
WO2021069084A1 (en) * 2019-10-11 2021-04-15 Toyota Motor Europe Methods and systems for determining the 3d-locations, the local reference frames and the grasping patterns of grasping points of an object
CN112802105A (en) * 2021-02-05 2021-05-14 梅卡曼德(北京)机器人科技有限公司 Object grabbing method and device
CN112802107A (en) * 2021-02-05 2021-05-14 梅卡曼德(北京)机器人科技有限公司 Robot-based control method and device for clamp group
CN113001552A (en) * 2021-03-16 2021-06-22 中国科学院自动化研究所 Robot operation cooperative grabbing method, system and equipment for impurity targets
CN114022341A (en) * 2021-11-10 2022-02-08 梅卡曼德(北京)机器人科技有限公司 Acquisition method and device for acquisition point information, electronic equipment and storage medium
CN114022342A (en) * 2021-11-10 2022-02-08 梅卡曼德(北京)机器人科技有限公司 Acquisition method and device for acquisition point information, electronic equipment and storage medium
CN114037595A (en) * 2021-11-10 2022-02-11 梅卡曼德(北京)机器人科技有限公司 Image data processing method, image data processing device, electronic equipment and storage medium
CN114092428A (en) * 2021-11-10 2022-02-25 梅卡曼德(北京)机器人科技有限公司 Image data processing method, image data processing device, electronic equipment and storage medium

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021069084A1 (en) * 2019-10-11 2021-04-15 Toyota Motor Europe Methods and systems for determining the 3d-locations, the local reference frames and the grasping patterns of grasping points of an object
CN110889460A (en) * 2019-12-06 2020-03-17 中山大学 Mechanical arm specified object grabbing method based on cooperative attention mechanism
CN112802105A (en) * 2021-02-05 2021-05-14 梅卡曼德(北京)机器人科技有限公司 Object grabbing method and device
CN112802107A (en) * 2021-02-05 2021-05-14 梅卡曼德(北京)机器人科技有限公司 Robot-based control method and device for clamp group
CN113001552A (en) * 2021-03-16 2021-06-22 中国科学院自动化研究所 Robot operation cooperative grabbing method, system and equipment for impurity targets
CN114022341A (en) * 2021-11-10 2022-02-08 梅卡曼德(北京)机器人科技有限公司 Acquisition method and device for acquisition point information, electronic equipment and storage medium
CN114022342A (en) * 2021-11-10 2022-02-08 梅卡曼德(北京)机器人科技有限公司 Acquisition method and device for acquisition point information, electronic equipment and storage medium
CN114037595A (en) * 2021-11-10 2022-02-11 梅卡曼德(北京)机器人科技有限公司 Image data processing method, image data processing device, electronic equipment and storage medium
CN114092428A (en) * 2021-11-10 2022-02-25 梅卡曼德(北京)机器人科技有限公司 Image data processing method, image data processing device, electronic equipment and storage medium

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116958265A (en) * 2023-09-19 2023-10-27 交通运输部天津水运工程科学研究所 Ship pose measurement method and system based on binocular vision

Similar Documents

Publication Publication Date Title
US11967113B2 (en) Method and system for performing automatic camera calibration for a scanning system
JP4309439B2 (en) Object take-out device
CN110580725A (en) Box sorting method and system based on RGB-D camera
US10380767B2 (en) System and method for automatic selection of 3D alignment algorithms in a vision system
CN111721259A (en) Underwater robot recovery positioning method based on binocular vision
US11966996B2 (en) Composite three-dimensional blob tool and method for operating the same
CN115609591B (en) Visual positioning method and system based on 2D Marker and compound robot
CN111784655B (en) Underwater robot recycling and positioning method
CN114092428A (en) Image data processing method, image data processing device, electronic equipment and storage medium
WO2023083273A1 (en) Grip point information acquisition method and apparatus, electronic device, and storage medium
CN116213306A (en) Automatic visual identification method and sorting system
WO2023082417A1 (en) Grabbing point information obtaining method and apparatus, electronic device, and storage medium
CN109863365B (en) Method, electronic device and system for picking up objects from container
JP2021033712A (en) Image processing device, imaging apparatus, robot and robot system
CN114022342A (en) Acquisition method and device for acquisition point information, electronic equipment and storage medium
JPS63311485A (en) Automatic calibration device
CN114037595A (en) Image data processing method, image data processing device, electronic equipment and storage medium
US20230297068A1 (en) Information processing device and information processing method
CN116175542A (en) Grabbing control method, grabbing control device, electronic equipment and storage medium
CN113793349A (en) Target detection method and device, computer readable storage medium and electronic equipment
CN113759346A (en) Laser radar calibration method and device, electronic equipment and storage medium
JPH10213420A (en) Formation of reference image used in pattern matching
US20230368414A1 (en) Pick and place systems and methods
EP4332900A1 (en) Automatic bin detection for robotic applications
US11562527B2 (en) Labeling method, labeling device using the same, pick-and-place system using the same, pick-and-place method using the same and non-transitory computer readable medium using the same

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22892066

Country of ref document: EP

Kind code of ref document: A1