CN110910452B - Low-texture industrial part pose estimation method based on deep learning - Google Patents

Low-texture industrial part pose estimation method based on deep learning Download PDF

Info

Publication number
CN110910452B
CN110910452B CN201911172167.5A CN201911172167A CN110910452B CN 110910452 B CN110910452 B CN 110910452B CN 201911172167 A CN201911172167 A CN 201911172167A CN 110910452 B CN110910452 B CN 110910452B
Authority
CN
China
Prior art keywords
pose
industrial part
network
deep learning
texture
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911172167.5A
Other languages
Chinese (zh)
Other versions
CN110910452A (en
Inventor
庄春刚
赵恒�
李少飞
沈逸超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Jiaotong University
Original Assignee
Shanghai Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Jiaotong University filed Critical Shanghai Jiaotong University
Priority to CN201911172167.5A priority Critical patent/CN110910452B/en
Publication of CN110910452A publication Critical patent/CN110910452A/en
Application granted granted Critical
Publication of CN110910452B publication Critical patent/CN110910452B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection
    • G06T2207/30164Workpiece; Machine component
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Abstract

The invention discloses a low-texture industrial part pose estimation method based on deep learning, which relates to the technical field of computer vision, and comprises the following steps of: firstly, carrying out three-dimensional modeling on an industrial part of which the pose is required to be estimated, constructing a physical simulation environment, and generating a data set of the industrial part in different poses in the simulation environment; secondly, carrying out instance segmentation and clipping on the data set; and finally, establishing a pose estimation sub-network and a pose refinement sub-network based on deep learning to obtain the pose of the low-texture industrial part. According to the invention, the three-dimensional modeling is carried out on the industrial parts, the pose estimation sub-network and the pose refinement sub-network based on deep learning are established, and the RGB image, the depth image, the original point cloud and the new point cloud rendered by the original pose are respectively used as inputs, so that the recognition effect on the low-texture industrial parts with reflective surfaces is greatly improved, and the method has important application value on the capture of the industrial scattered parts.

Description

Low-texture industrial part pose estimation method based on deep learning
Technical Field
The invention relates to the technical field of computer vision, in particular to a low-texture industrial part pose estimation method based on deep learning.
Background
Computer vision techniques play an important role in the perception of unstructured scenes in robots. The visual image is an effective means for acquiring real world information, and features of a corresponding task, such as object position, angle, gesture and the like, are extracted through a visual perception algorithm, so that a robot can execute corresponding operations to complete a designated operation task. For sorting of industrial robots, it is now possible to acquire scene data using visual sensors, but how to recognize a target object from a scene and estimate its position and posture, so that calculating the gripping position and the gripping path of the industrial robot becomes a core problem.
In recent years, with the rapid development of deep learning technology, a pose estimation technology based on deep learning has become a mainstream algorithm in the field of pose estimation. However, the existing mainstream pose estimation algorithm based on deep learning mostly depends on information such as color, texture and the like of the object surface, has poor part identification effect on the industrial low-texture and reflective surface, and has certain obstruction to realizing efficient automatic part sorting.
Therefore, a person skilled in the art is dedicated to developing a low-texture industrial part pose estimation method based on deep learning, simulating a real scene by combining a physical engine and a graphic engine, performing three-dimensional modeling on an industrial part by applying a UV mapping technology, establishing a pose estimation sub-network and a pose refinement sub-network based on deep learning, constructing a corresponding data set, and obtaining an object pose estimation algorithm through continuous iteration so as to improve the sorting capability of a robot on scattered part scenes.
Disclosure of Invention
In view of the above-mentioned drawbacks of the prior art, the present invention aims to solve the problem that the existing pose estimation algorithm mostly depends on the information such as color and texture of the object surface, so that the recognition effect on the part with low texture and reflective surface is poor.
In order to achieve the above object, the present invention provides a low texture industrial part pose estimation method based on deep learning, which is characterized in that the method comprises the following steps:
step 1, carrying out three-dimensional modeling on an industrial part of which the pose is required to be estimated, constructing a physical simulation environment, and generating a data set of the industrial part in different poses in the simulation environment;
step 2, performing instance segmentation and clipping on the data set;
and 3, establishing a pose estimation sub-network and a pose refinement sub-network based on deep learning to obtain the pose of the low-texture industrial part.
Further, the three-dimensional modeling in step 1 is based on UV mapping technology, i.e. mapping the surface of the industrial part to the surface of a three-dimensional model in a two-dimensional map.
Further, the physical simulation environment in the step 1 is to simulate a real scene by combining a physical engine and a graphic engine.
Further, the data set in the step 1 includes an RGB map, a depth map, a category of the industrial part, a bounding box of the industrial part, and a mask of the industrial part.
Further, in the step 2, feature extraction is required to be performed on the RGB image and the depth image after clipping; the feature map sizes of the RGB map and the depth map are 64 XH x W; predicting an initial pose according to the feature map, wherein the loss function is as follows:
wherein N is the number of feature points, [ R|t ]]For the true pose of the person,to predict the resulting pose, x i Is the three-dimensional point coordinates on the model.
Further, the pose estimation sub-network in the step 3 takes the RGB map and the depth map as inputs; the RGB map is an RGB image within a minimum bounding box region containing a single industrial part; the depth map is a depth image within a minimum bounding box region that includes a single industrial part.
Further, the pose refinement sub-network in the step 3 takes an original point cloud and a new point cloud rendered by the initial pose as inputs.
Further, the original point cloud is composed ofThe mask area of the single industrial part on the depth map is calculated and is marked as P 0 The calculation formula is as follows:
wherein ,(xw ,y w ,z w ) The coordinates of the feature points in the camera coordinate system, (u, v) the coordinates of the feature points in the pixel coordinate system, and z c For depth values of feature points, u 0 ,v 0 Dx, dy, f are internal references of the camera.
Further, the new point cloud is recorded as [ R ] by calculating the optimal pose in the initial pose set 0 |t 0 ]Performing projection rendering on the model reconstruction to obtain the optimal pose [ R ] 0 |t 0 ]The depth map below is calculated and a new point cloud under the pose is calculated and is recorded as P 1
Further, the pose refinement sub-network in the step 3 is refined through multiple iterations until the pose meeting the precision requirement is obtained, and the iteration formula is as follows:
wherein ,for the final pose, M is the number of iterations, +.>The pose predicted for the (i+1) th iteration;
the loss function is:
wherein N is the number of feature points, [ R|t ]]For the true pose of the person,for the final pose obtained by the current iteration, x i Is the three-dimensional point coordinates on the model.
Compared with the prior art, the implementation of the invention has at least the following beneficial technical effects:
according to the low-texture industrial part pose estimation method based on deep learning, three-dimensional modeling is carried out on industrial parts, a pose estimation sub-network and a pose refinement sub-network based on deep learning are established, an RGB image, a depth image, an original point cloud and a new point cloud obtained by initial pose rendering are respectively used as inputs, the recognition effect on the low-texture industrial parts with reflective surfaces is greatly improved, and the method has important application value on grabbing industrial scattered parts.
The conception, specific structure, and technical effects of the present invention will be further described with reference to the accompanying drawings to fully understand the objects, features, and effects of the present invention.
Drawings
FIG. 1 is a flow chart of a method for estimating pose of a low-texture industrial part based on deep learning according to a preferred embodiment of the present invention;
FIG. 2 is a flow chart of a physical simulation build dataset provided by a preferred embodiment of the present invention;
FIG. 3 is a diagram of part models and simulation results before and after mapping provided by a preferred embodiment of the present invention;
FIG. 4 is a flow chart of a pose estimation sub-network provided by a preferred embodiment of the present invention;
FIG. 5 is a block diagram of a pose estimation sub-network provided by a preferred embodiment of the present invention;
FIG. 6 is a flow chart of a pose refinement sub-network provided by a preferred embodiment of the present invention;
FIG. 7 is a block diagram of a pose refinement sub-network provided by a preferred embodiment of the present invention;
FIG. 8 is a diagram showing the result of the pose estimation process according to a preferred embodiment of the present invention;
FIG. 9 is a diagram of the final result of pose estimation provided by a preferred embodiment of the present invention.
Detailed Description
The following description of the preferred embodiments of the present invention refers to the accompanying drawings, which make the technical contents thereof more clear and easy to understand. The present invention may be embodied in many different forms of embodiments and the scope of the present invention is not limited to only the embodiments described herein.
In the drawings, like structural elements are referred to by like reference numerals and components having similar structure or function are referred to by like reference numerals. The dimensions and thickness of each component shown in the drawings are arbitrarily shown, and the present invention is not limited to the dimensions and thickness of each component. The thickness of the components is exaggerated in some places in the drawings for clarity of illustration.
As shown in fig. 1, a flowchart of a low-texture industrial part pose estimation method based on deep learning according to a preferred embodiment of the present invention is provided, the method includes the following steps:
step 1, carrying out three-dimensional modeling on an industrial part of which the pose is required to be estimated, constructing a physical simulation environment, and generating a data set of the industrial part in different poses in the simulation environment;
when the three-dimensional modeling is carried out on the industrial part, as shown in fig. 2, the simulation of the real scene is realized by combining a physical engine and a graphic engine, and the rendering result is close to the real scene by further using the attributes such as illumination texture and the like; the physical engine can endow the rigid object in the scene with real physical properties, including quality and collision properties, and the scene has the same physical properties as the real environment by matching with a grid-based rapid collision detection algorithm;
in order to bring the industrial part and the modeling environment into close proximity with the real world, the invention applies a UV mapping technology to map the surface of the industrial part to the surface of the three-dimensional model in a two-dimensional mapping manner; spreading the three-dimensional grid model in a reasonable mode, spreading the surface of the three-dimensional grid model on a two-dimensional plane, wherein the two-dimensional space is called a UV texture space, providing a mapping relation between the surface of an object and a texture image, and determining the position of each pixel on the texture image corresponding to the surface of the model by placing a preset texture image in the UV texture space;
in the physical simulation environment, a data set of the industrial part in different positions is generated in a free fall mode, wherein the data set comprises an RGB image, a depth image, the category of the industrial part, a surrounding frame of the industrial part and a mask of the industrial part.
Step 2, performing instance segmentation and clipping on the data set;
in order to acquire bounding boxes and masks of single parts in an image, an image shot by a camera is subjected to example segmentation by using an existing algorithm, an RGB image and a depth image are cut according to the result of the example segmentation, and the cut result is input into a pose estimation sub-network: firstly, respectively extracting features of a cut RGB image and a cut depth image, wherein the original RGB image is 3 XH×W, and the original depth image is 1 XH×W, wherein H and W respectively represent the height and width of the cut image; secondly, after feature extraction, the sizes of the feature images are 32 XH multiplied by W, and the two feature images are spliced to obtain a feature image with the size of 64 XH multiplied by W; finally, predicting an initial pose for each feature point through a plurality of fully connected layers, wherein the loss function is defined as follows:
wherein N is the number of feature points, [ R|t ]]For the true pose of the person,to predict the resulting pose, x i Is the three-dimensional point coordinates on the model.
Step 3, establishing a pose estimation sub-network and a pose refinement sub-network based on deep learning to obtain the pose of the low-texture industrial part;
aiming at the characteristic of low texture of the industrial part, the invention provides a pose estimation sub-network and a pose refinement sub-network, wherein the pose estimation sub-network takes an RGB image and a depth image as input, the RGB image is an RGB image in a minimum bounding box area containing a single industrial part, and the depth image is a depth image in the minimum bounding box area containing the single industrial part; the pose refinement sub-network takes an original point cloud and a new point cloud rendered by the initial pose as inputs;
for the pose estimation sub-network, inputting the clipped result into the pose estimation sub-network: firstly, respectively carrying out feature extraction on the cut RGB image and the depth image, obtaining a feature image with the size of 64 XH x W after the feature extraction, and finally predicting an initial pose for each feature point through a plurality of full-connection layers;
for the pose refinement sub-network, an original point cloud is calculated from mask areas of single industrial parts on the depth map and is recorded as P 0 The calculation formula is as follows:
wherein ,(xw ,y w ,z w ) The coordinates of the feature points in the camera coordinate system, (u, v) the coordinates of the feature points in the pixel coordinate system, and z c For depth values of feature points, u 0 ,v 0 Dx, dy, f are internal references of the camera;
the pose estimation sub-network predicts an initial pose for each feature point, calculates the optimal pose in the initial pose set through a pose clustering algorithm, and marks the optimal pose as R 0 |t 0 ]Performing projection rendering on the model reconstruction to obtain the optimal pose [ R ] 0 |t 0 ]The depth map below is calculated and a new point cloud under the pose is calculated and is recorded as P 1 The method comprises the steps of carrying out a first treatment on the surface of the Pose refinement sub-network to origin point cloud P 0 And rendering the resulting new point cloud P 1 As input, predicting to obtain more accurate pose, and rendering with the current pose again to obtain new point cloud P 2 Then the original pose P is combined 0 Inputting the data into a pose refinement sub-network, and performing repeated iteration refinement until the accuracy is metThe calculated pose and the iteration formula are as follows:
wherein ,for the final pose, M is the number of iterations, +.>The pose predicted for the (i+1) th iteration;
the loss function is:
wherein N is the number of feature points, [ R|t ]]For the true pose of the person,for the final pose obtained by the current iteration, x i Is the three-dimensional point coordinates on the model.
Examples
In this embodiment, the software platform implemented by the whole deep learning method mainly includes a physical simulation engine Blender and a deep learning framework Pytorch, and the computer hardware is configured as a NVIDIA GeForce GTX 1080TI graphics card.
Step 1, as shown in fig. 1 and fig. 2, three-dimensional modeling is performed on an industrial part with a required pose estimation, a physical simulation environment is constructed by combining a physical engine and a graphic engine, a UV mapping technology is applied to map the surface of the industrial part to the surface of the three-dimensional model in a two-dimensional mapping mode, a data set of the industrial part in different poses is generated in the simulation environment, specifically including an RGB map, a depth map, a category of each industrial part, a bounding box and a mask of the part, about 1000 pictures are generated through physical simulation, and about 3000 examples of each industrial part are generated, wherein the specific generation process is as follows: firstly, randomly initializing the number and the type of parts in a current scene, selecting a corresponding part model initialization pose according to the number and the type of the parts, setting a position random range above a scene stacking area, and superposing random offset generation through a plurality of uniformly distributed position points to ensure that certain distance deviation exists between the parts, wherein the pose is expressed by Euler angles and is an arbitrary value within the range of 0-360 degrees; secondly, calculating a stacking scene formed by the falling of the parts under the action of gravity through a physical simulation model; and finally, randomly placing a light source in a specified range, adjusting the brightness of the light source, and determining the corresponding pose of the camera, thereby forming the rendering configuration of a sample. FIG. 3 shows a three-dimensional model of a part before and after mapping and the RGB and depth maps generated;
step 2, as shown in fig. 4 and fig. 5, are respectively a flow chart and a structure chart of the pose estimation sub-network provided by the present embodiment, an image shot by a camera is subjected to example segmentation by using an existing algorithm, an RGB image and a depth image are cut according to a result of the example segmentation, and a cut result is input into the pose estimation sub-network: firstly, respectively extracting features of a cut RGB image and a cut depth image, wherein the original RGB image is 3 XH×W, and the original depth image is 1 XH×W, wherein H and W respectively represent the height and width of the cut image; secondly, after feature extraction, the sizes of the feature images are 32 XH multiplied by W, and the two feature images are spliced to obtain a feature image with the size of 64 XH multiplied by W; finally, predicting an initial pose for each feature point through a plurality of full-connection layers;
step 3, predicting an initial pose for each feature point by the pose estimation sub-network, calculating to obtain the optimal pose in the initial pose set through a pose clustering algorithm, and recording the optimal pose as [ R ] 0 |t 0 ]Performing projection rendering on the model reconstruction to obtain the optimal pose [ R ] 0 |t 0 ]The depth map below is calculated and a new point cloud under the pose is calculated and is recorded as P 1 The method comprises the steps of carrying out a first treatment on the surface of the As shown in fig. 6 and 7, the pose refinement sub-network provided in the present embodiment is a flowchart and a structure diagram of the pose refinement sub-network using an original point cloud P 0 And rendering the resulting new point cloud P 1 As input, pre-emptMeasuring more accurate pose, and rendering with the current pose again to obtain a new point cloud P 2 Then the original pose P is combined 0 Inputting the pose into a pose refinement sub-network, and performing repeated iteration refinement until the pose meeting the precision requirement is obtained; fig. 8 shows the process result of the pose estimation algorithm of the present embodiment, where the first column is an acquired scene picture, the second column is an image obtained by re-projecting a single three-dimensional model according to the pose predicted by the pose estimation sub-network, the third column is an image obtained by re-projecting a single three-dimensional model according to the pose predicted by the pose refinement sub-network, and the fourth column is an image in which the contours of the projected models of the second column and the third column are displayed on one image; fig. 9 shows the final result of the pose estimation algorithm of the present embodiment, where the first line is an acquired scene picture and the second line is an image obtained by re-projecting the three-dimensional model according to the estimated pose. In this embodiment, the three-dimensional model is re-projected onto the image according to the predicted pose, and the contours of the pose projected models in two stages of pose estimation and pose refinement are displayed on the image at the same time, which indicates that the pose refinement sub-network can improve the final pose accuracy.
The foregoing describes in detail preferred embodiments of the present invention. It should be understood that numerous modifications and variations can be made in accordance with the concepts of the invention without requiring creative effort by one of ordinary skill in the art. Therefore, all technical solutions which can be obtained by logic analysis, reasoning or limited experiments based on the prior art by the person skilled in the art according to the inventive concept shall be within the scope of protection defined by the claims.

Claims (5)

1. A low-texture industrial part pose estimation method based on deep learning, which is characterized by comprising the following steps:
step 1, carrying out three-dimensional modeling on an industrial part of which the pose is required to be estimated, constructing a physical simulation environment, and generating a data set of the industrial part in different poses in the simulation environment; the data set comprises an RGB map, a depth map, a category of the industrial part, a bounding box of the industrial part and a mask of the industrial part;
step 2, performing instance segmentation and clipping on the data set;
step 3, establishing a pose estimation sub-network and a pose refinement sub-network based on deep learning to obtain the pose of the low-texture industrial part; the pose estimation sub-network takes the RGB map and the depth map as input; the pose estimation sub-network performs feature extraction on the cut RGB image and the depth image to obtain a feature image, and predicts an initial pose according to the feature image; the pose refinement sub-network takes an original point cloud and a new point cloud rendered by the initial pose as inputs; the pose refinement sub-network refines through repeated iteration until the pose meeting the precision requirement is obtained;
the original point cloud is obtained by calculating mask areas of single industrial parts on the depth map and is marked as P 0 The calculation formula is as follows:
wherein ,(xw ,y w ,z w ) The coordinates of the feature points in the camera coordinate system, (u, v) the coordinates of the feature points in the pixel coordinate system, and z c For depth values of feature points, u 0 ,v 0 Dx, dy, f are internal references of the camera;
the new point cloud is recorded as [ R ] by calculating the optimal pose in the initial pose set 0 |t 0 ]Performing projection rendering on the model reconstruction to obtain the optimal pose [ R ] 0 |t 0 ]The depth map below is calculated and a new point cloud under the pose is calculated and is recorded as P 1
The iterative formula of the pose refinement sub-network is as follows:
wherein ,for the final pose, M is the number of iterations, +.>The pose predicted for the (i+1) th iteration; the loss function is:
wherein N is the number of feature points, [ R|t ]]For the true pose of the person,for the final pose obtained by the current iteration, x i Is the three-dimensional point coordinates on the model.
2. The method for estimating pose of low texture industrial part based on deep learning according to claim 1, wherein the three-dimensional modeling in step 1 is based on UV mapping technique, i.e. mapping the surface of the industrial part to the surface of three-dimensional model in a two-dimensional map.
3. The method for estimating pose of industrial part with low texture based on deep learning according to claim 1, wherein the physical simulation environment in step 1 is a simulation of real scene by combining a physical engine and a graphic engine.
4. The deep learning based low texture industrial part pose estimation method of claim 1, wherein the initial pose is predicted from the feature map using a loss function of:
wherein N is a featureNumber of dots [ R|t ]]For the true pose of the person,to predict the resulting pose, x i Is the three-dimensional point coordinates on the model.
5. The deep learning based low texture industrial part pose estimation method according to claim 1, wherein the RGB map is an RGB image within a minimum bounding box area containing a single industrial part; the depth map is a depth image within a minimum bounding box region that includes a single industrial part.
CN201911172167.5A 2019-11-26 2019-11-26 Low-texture industrial part pose estimation method based on deep learning Active CN110910452B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911172167.5A CN110910452B (en) 2019-11-26 2019-11-26 Low-texture industrial part pose estimation method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911172167.5A CN110910452B (en) 2019-11-26 2019-11-26 Low-texture industrial part pose estimation method based on deep learning

Publications (2)

Publication Number Publication Date
CN110910452A CN110910452A (en) 2020-03-24
CN110910452B true CN110910452B (en) 2023-08-25

Family

ID=69819648

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911172167.5A Active CN110910452B (en) 2019-11-26 2019-11-26 Low-texture industrial part pose estimation method based on deep learning

Country Status (1)

Country Link
CN (1) CN110910452B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112734727A (en) * 2021-01-11 2021-04-30 安徽理工大学 Apple picking method based on improved deep neural network
CN112801977B (en) * 2021-01-28 2022-11-22 青岛理工大学 Assembly body part relative pose estimation and monitoring method based on deep learning
CN113538569B (en) * 2021-08-11 2023-06-02 广东工业大学 Weak texture object pose estimation method and system
CN113793472B (en) * 2021-09-15 2023-01-20 应急管理部沈阳消防研究所 Image type fire detector pose estimation method based on feature depth aggregation network
CN114332211B (en) * 2022-01-06 2022-12-13 南京航空航天大学 Part pose calculation method based on edge reconstruction and dense fusion network

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106251353A (en) * 2016-08-01 2016-12-21 上海交通大学 Weak texture workpiece and the recognition detection method and system of three-dimensional pose thereof
CN109003325A (en) * 2018-06-01 2018-12-14 网易(杭州)网络有限公司 A kind of method of three-dimensional reconstruction, medium, device and calculate equipment
CN109102547A (en) * 2018-07-20 2018-12-28 上海节卡机器人科技有限公司 Robot based on object identification deep learning model grabs position and orientation estimation method
CN109308737A (en) * 2018-07-11 2019-02-05 重庆邮电大学 A kind of mobile robot V-SLAM method of three stage point cloud registration methods
CN109472820A (en) * 2018-10-19 2019-03-15 清华大学 Monocular RGB-D camera real-time face method for reconstructing and device
CN109523629A (en) * 2018-11-27 2019-03-26 上海交通大学 A kind of object semanteme and pose data set generation method based on physical simulation
CN109685891A (en) * 2018-12-28 2019-04-26 鸿视线科技(北京)有限公司 3 d modeling of building and virtual scene based on depth image generate system
CN109816725A (en) * 2019-01-17 2019-05-28 哈工大机器人(合肥)国际创新研究院 A kind of monocular camera object pose estimation method and device based on deep learning
CN110310362A (en) * 2019-06-24 2019-10-08 中国科学院自动化研究所 High dynamic scene three-dimensional reconstruction method, system based on depth map and IMU
CN110322512A (en) * 2019-06-28 2019-10-11 中国科学院自动化研究所 In conjunction with the segmentation of small sample example and three-dimensional matched object pose estimation method

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106251353A (en) * 2016-08-01 2016-12-21 上海交通大学 Weak texture workpiece and the recognition detection method and system of three-dimensional pose thereof
CN109003325A (en) * 2018-06-01 2018-12-14 网易(杭州)网络有限公司 A kind of method of three-dimensional reconstruction, medium, device and calculate equipment
CN109308737A (en) * 2018-07-11 2019-02-05 重庆邮电大学 A kind of mobile robot V-SLAM method of three stage point cloud registration methods
CN109102547A (en) * 2018-07-20 2018-12-28 上海节卡机器人科技有限公司 Robot based on object identification deep learning model grabs position and orientation estimation method
CN109472820A (en) * 2018-10-19 2019-03-15 清华大学 Monocular RGB-D camera real-time face method for reconstructing and device
CN109523629A (en) * 2018-11-27 2019-03-26 上海交通大学 A kind of object semanteme and pose data set generation method based on physical simulation
CN109685891A (en) * 2018-12-28 2019-04-26 鸿视线科技(北京)有限公司 3 d modeling of building and virtual scene based on depth image generate system
CN109816725A (en) * 2019-01-17 2019-05-28 哈工大机器人(合肥)国际创新研究院 A kind of monocular camera object pose estimation method and device based on deep learning
CN110310362A (en) * 2019-06-24 2019-10-08 中国科学院自动化研究所 High dynamic scene three-dimensional reconstruction method, system based on depth map and IMU
CN110322512A (en) * 2019-06-28 2019-10-11 中国科学院自动化研究所 In conjunction with the segmentation of small sample example and three-dimensional matched object pose estimation method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Dewei Zou 等.An Improved Method for Model-Based Training, Detection and Pose Estimation of Texture-Less 3D Objects in Occlusion Scenes.《ScienceDirect》.2019,第541-546页. *

Also Published As

Publication number Publication date
CN110910452A (en) 2020-03-24

Similar Documents

Publication Publication Date Title
CN110910452B (en) Low-texture industrial part pose estimation method based on deep learning
Park et al. Pix2pose: Pixel-wise coordinate regression of objects for 6d pose estimation
CN109003325B (en) Three-dimensional reconstruction method, medium, device and computing equipment
CN110637305B (en) Learning to reconstruct 3D shapes by rendering many 3D views
JP5778237B2 (en) Backfill points in point cloud
EP2080167B1 (en) System and method for recovering three-dimensional particle systems from two-dimensional images
EP3188033A1 (en) Reconstructing a 3d modeled object
US20200057831A1 (en) Real-time generation of synthetic data from multi-shot structured light sensors for three-dimensional object pose estimation
Riegler et al. Connecting the dots: Learning representations for active monocular depth estimation
JP7294788B2 (en) Classification of 2D images according to the type of 3D placement
JP2021163503A (en) Three-dimensional pose estimation by two-dimensional camera
KR102223484B1 (en) System and method for 3D model generation of cut slopes without vegetation
JP2020160812A (en) Region extraction device and program
Thalhammer et al. SyDPose: Object detection and pose estimation in cluttered real-world depth images trained using only synthetic data
US10885708B2 (en) Automated costume augmentation using shape estimation
EP3846123A1 (en) 3d reconstruction with smooth maps
KR20080058366A (en) Modeling micro-structure for feature extraction
CN116205978A (en) Method, device, equipment and storage medium for determining mapping image of three-dimensional target object
To et al. Bas-relief generation from face photograph based on facial feature enhancement
JP2021176078A (en) Deep layer learning and feature detection through vector field estimation
CN112132845A (en) Three-dimensional model unitization method and device, electronic equipment and readable medium
Akizuki et al. ASM-Net: Category-level Pose and Shape Estimation Using Parametric Deformation.
CN115375847A (en) Material recovery method, three-dimensional model generation method and model training method
US11657506B2 (en) Systems and methods for autonomous robot navigation
Fan et al. Collaborative three-dimensional completion of color and depth in a specified area with superpixels

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant