CN113362467A - Point cloud preprocessing and ShuffleNet-based mobile terminal three-dimensional pose estimation method - Google Patents

Point cloud preprocessing and ShuffleNet-based mobile terminal three-dimensional pose estimation method Download PDF

Info

Publication number
CN113362467A
CN113362467A CN202110634620.0A CN202110634620A CN113362467A CN 113362467 A CN113362467 A CN 113362467A CN 202110634620 A CN202110634620 A CN 202110634620A CN 113362467 A CN113362467 A CN 113362467A
Authority
CN
China
Prior art keywords
dimensional
target
mobile terminal
calculating
pose
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110634620.0A
Other languages
Chinese (zh)
Other versions
CN113362467B (en
Inventor
袁景凌
俞洋
白立华
王梦蝶
李宵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University of Technology WUT
Original Assignee
Wuhan University of Technology WUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University of Technology WUT filed Critical Wuhan University of Technology WUT
Priority to CN202110634620.0A priority Critical patent/CN113362467B/en
Publication of CN113362467A publication Critical patent/CN113362467A/en
Application granted granted Critical
Publication of CN113362467B publication Critical patent/CN113362467B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation

Abstract

The invention discloses a mobile terminal three-dimensional pose estimation method based on point cloud preprocessing and ShuffleNet, which comprises the steps of preprocessing at a PC (personal computer) terminal, performing three-dimensional reconstruction on target point cloud data and importing the data into a three-dimensional rendering engine; obtaining two-dimensional photos of a target under different visual angles by adopting a rotation photographing algorithm in a three-dimensional engine, labeling the photos by using a key voxel block extraction algorithm provided by the invention, and establishing a training data set; the detection model of the key voxel blocks of the target is trained by using ShuffleNetv2-YOLOv3 which has light weight and high performance advantages and is suitable for mobile terminal calculation; reading a video stream from a mobile terminal camera, detecting a target key voxel block through a ShuffleNetv2-YOLOv3 model, and calculating a 2D-3D point pair corresponding to a key voxel block central point through RANSAC and EPNP algorithm to obtain a relative pose of a target. And finally, calculating the pose of the target in the actual three-dimensional world by using the advantages of the mobile terminal through data provided by the built-in IMU and the GPS.

Description

Point cloud preprocessing and ShuffleNet-based mobile terminal three-dimensional pose estimation method
Technical Field
The invention belongs to the technical field of computers, and relates to a mobile terminal three-dimensional pose estimation method based on point cloud preprocessing and ShuffleNet, which can be widely applied to the fields of robot grabbing, vehicle intelligent navigation, augmented reality, medical diagnosis and the like.
Background
The three-dimensional pose estimation plays a very key role in the fields of robot grabbing, vehicle intelligent navigation, augmented reality, medical diagnosis and the like. At present, the mainstream pose estimation methods are divided into two categories, one category is an identification method based on a two-dimensional image, the method predicts 1 central point and 8 angular points of an object for an input RGB or RGB-D image, and then obtains the 6D pose of the object through a PNP or EPNP algorithm. The algorithm has good real-time performance but low accuracy. And the other type of the method is used for positioning based on point cloud data, the method firstly uses a deep network to establish a corresponding relation between 3D point cloud data and a 2D image, and then obtains the 6D posture of the object through a PNP or EPNP algorithm. Due to the use of point cloud data, the accuracy is higher than the first class, but comparatively, the speed is lower.
The mobile terminal of the mobile phone has the advantages of high popularization rate and convenience in carrying, but the hardware configuration is far lower than that of a PC (personal computer), and the recognition speed by adopting a conventional algorithm is difficult to meet the requirement. The portable advantages of the laser radar and the depth camera which need to be externally connected are weakened in configuration, and the mobile end only can adopt an RGB video stream identification scheme, so that the pose analysis accuracy is not high.
The invention mainly aims at the requirement of a mobile terminal on target pose estimation in the field of auxiliary industrial application, and provides a mobile terminal three-dimensional pose estimation method based on point cloud preprocessing and ShuffleNet.
The design scheme adopted by the invention is as follows: a mobile terminal three-dimensional pose estimation method based on point cloud preprocessing and ShuffleNet comprises the following steps:
step 1: performing three-dimensional reconstruction on target point cloud data obtained by laser scanning; importing a three-dimensional model obtained by three-dimensional reconstruction into a rendering engine for photographing;
step 2: respectively acquiring two-dimensional photos and camera poses of the target under different visual angles by adopting a positioning rotation photographing algorithm; extracting two-dimensional photo feature points through SIFT, calculating corresponding three-dimensional feature points, dividing a target model into equal-size voxel blocks, and screening target key voxel blocks according to the number of the three-dimensional feature points; generating two-dimensional projection of key voxel blocks on the photo set and establishing a training data set; training a ShuffLeNet feature detection model aiming at a target through a ShufflLeNet v2-YOLOv3 lightweight network;
and step 3: inputting the video stream into a trained detection model of ShuffleNetv2-YOLOv3 target key voxel blocks, identifying the key voxel blocks to obtain 2D-3D matching point pairs, and calculating the relative pose of the target by combining RANSAC and EPNP algorithms;
and 4, step 4: and calculating the absolute pose of the target in the three-dimensional world by combining the GPS and IMU information of the mobile terminal.
The invention combines the advantages of two types of three-dimensional pose estimation algorithms, and firstly carries out preprocessing at the PC end. And reconstructing a three-dimensional model for the target point cloud data through a Delauay algorithm. The method adopts a positioning rotation photographing and key voxel block extraction algorithm to automatically generate a target voxel characteristic detection data set, and adopts a ShuffleNetv2-YOLOv3 training characteristic detection model which has the advantages of light weight and high performance and is suitable for mobile-end calculation.
The invention fully utilizes the advantages of the mobile terminal hardware equipment in the identification stage and introduces GPS and IMU data to position the pose of the mobile terminal. And detecting key voxel blocks of the target by using a trained ShuffleNetv2-YOLOv3 model, and calculating by using RANSAC and EPNP algorithms to obtain the relative pose between the target and the mobile terminal camera. And finally, calculating the absolute pose of the target in the three-dimensional world. At present, the popularity of the mobile terminal of the mobile phone exceeds 90%, the invention can provide position and pose estimation which does not depend on a depth camera and laser equipment and meets the requirements of industrial auxiliary application in real time and accuracy at the mobile terminal, and has the advantages of portability, practicability and easy popularization.
Drawings
FIG. 1 is a flow chart of an embodiment of the present invention.
Fig. 2 is a schematic block diagram of an embodiment of the present invention.
Detailed Description
In order to facilitate the understanding and implementation of the present invention for those of ordinary skill in the art, the present invention is further described in detail with reference to the accompanying drawings and examples, it is to be understood that the embodiments described herein are merely illustrative and explanatory of the present invention and are not restrictive thereof.
Referring to fig. 1 and fig. 2, the method for estimating the three-dimensional pose of the mobile terminal based on point cloud preprocessing and ShuffleNet provided by the invention comprises the following steps:
step 1: performing three-dimensional reconstruction on target point cloud data obtained by laser scanning; importing a three-dimensional model obtained by three-dimensional reconstruction into a rendering engine for photographing;
the specific implementation of the embodiment includes the following sub-steps:
step 1.1: performing three-dimensional reconstruction on target point cloud data obtained by laser scanning through a Delauay algorithm to obtain a target three-dimensional model;
step 1.2: and importing the target three-dimensional model into a rendering engine, calculating a target three-dimensional model bounding box and a central point, and moving the target three-dimensional model to enable the central point to reach an origin point.
Step 2: respectively acquiring two-dimensional photos and camera poses of the target under different visual angles by adopting a positioning rotation photographing algorithm; extracting two-dimensional photo feature points through SIFT, calculating corresponding three-dimensional feature points, dividing a target model into equal-size voxel blocks, and screening target key voxel blocks according to the number of the three-dimensional feature points; generating two-dimensional projection of key voxel blocks on the photo set and establishing a training data set; training a ShuffLeNet feature detection model aiming at a target through a ShufflLeNet v2-YOLOv3 lightweight network;
the specific implementation of the embodiment includes the following sub-steps:
step 2.1: shooting a target in a three-dimensional engine to obtain a target photo;
step 2.2: detecting two-dimensional feature points in the target picture through an SIFT algorithm to obtain a feature point set K ═ K1,...,kn};
Step 2.3: calculating each characteristic point k by a screen ray projection algorithmiCorresponding three-dimensional coordinate point piAnd recording the three-dimensional feature point set corresponding to the two-dimensional feature point as P ═ P1,...,pn};
Step 2.4: the camera rotates around the target and continues to take pictures of the target, and the step 2.2 and the step 2.3 are repeated until a multi-view picture of the target is obtained and a three-dimensional feature point set PS is obtained through calculation1,...,PNN is the number of photos, and P is a three-dimensional feature point set of each photo;
step 2.5: dividing a target three-dimensional voxel (Volume Pixel) into M blocks B ═ B of the same voxel size1,...,bM}; setting the frequency of the three-dimensional feature points appearing in each voxel block as a voxel block weight q, and screening m blocks KB with the maximum weight as { b }1,...,bmAs key voxel block, where m<M;
Step 2.6: taking the key voxel blocks as categories, calculating the areas of the key voxel blocks on the two-dimensional photo set according to a projection transformation formula, and generating marking information to obtain a training data set;
step 2.7: and training ShuffleNetv2-YOLOv3 through the generated data set to obtain a detection model aiming at the target key voxel block.
And step 3: inputting the video stream into a trained detection model of ShuffleNetv2-YOLOv3 target key voxel blocks, identifying the key voxel blocks to obtain 2D-3D matching point pairs, and calculating the relative pose of the target by combining an EPNP algorithm;
the specific implementation of the embodiment includes the following sub-steps:
step 3.1: reading a video stream, inputting a trained detection model of ShuffleNetv2-YOLOv3 key voxel blocks, and outputting the detection model into two-dimensional areas corresponding to a plurality of key voxel blocks;
step 3.2: calculating the detected central point of the two-dimensional region, and forming a 2D-3D matching point pair with the central point of the corresponding key voxel block;
step 3.3: and calculating the relative pose between the target and the mobile terminal camera through RANSAC and EPNP algorithms.
And 4, step 4: calculating the pose of the mobile end camera in the three-dimensional world through the GPS and IMU information of the mobile end, and calculating the absolute pose of the target in the three-dimensional world by combining the relative position between the target and the mobile end camera;
the specific implementation of the embodiment includes the following sub-steps:
step 4.1: reading GPS and IMU data of a mobile terminal;
step 4.2: calculating the positioning of the mobile terminal in the three-dimensional world through the data acquired in the step 4.1;
step 4.3: and (3) calculating the absolute pose of the target in the three-dimensional world according to the moving end pose calculated in the step 4.2 and the relative pose of the target calculated in the step 3.
The method comprises the steps of estimating the real-time pose at a mobile terminal, reading a video stream from a camera at the mobile terminal, obtaining feature points of a target in an RGB image by using a ShuffleNet model, and obtaining the relative pose through RANSAC and EPNP algorithms. The invention fully utilizes the advantages of the mobile terminal, and calculates the pose of the target in the actual three-dimensional world by combining the relative poses through the positioning information provided by the GPS and IMU of the mobile terminal.
It should be understood that parts of the specification not set forth in detail are well within the prior art.
It should be understood that the above description of the preferred embodiments is given for clarity and not for any purpose of limitation, and that various changes, substitutions and alterations can be made herein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (5)

1. A mobile terminal three-dimensional pose estimation method based on point cloud preprocessing and ShuffleNet comprises the following steps:
step 1: performing three-dimensional reconstruction on target point cloud data obtained by laser scanning; importing a three-dimensional model obtained by three-dimensional reconstruction into a rendering engine for photographing;
step 2: respectively acquiring two-dimensional photos and camera poses of the target under different visual angles by adopting a positioning rotation photographing algorithm; extracting two-dimensional photo feature points through SIFT, calculating corresponding three-dimensional feature points, dividing a target model into equal-size voxel blocks, and screening target key voxel blocks according to the number of the three-dimensional feature points; generating two-dimensional projection of key voxel blocks on the photo set and establishing a training data set; training a ShuffLeNet feature detection model aiming at a target through a ShufflLeNet v2-YOLOv3 lightweight network;
and step 3: inputting the video stream into a trained detection model of ShuffleNetv2-YOLOv3 target key voxel blocks, identifying the key voxel blocks to obtain 2D-3D matching point pairs, and calculating the relative pose of the target through RANSAC and EPNP algorithms;
and 4, step 4: and calculating the absolute pose of the target in the three-dimensional world by combining the GPS and IMU information of the mobile terminal.
2. The point cloud preprocessing and ShuffleNet-based mobile terminal three-dimensional pose estimation method according to claim 1, characterized in that: in the step 1, performing three-dimensional reconstruction on target point cloud data obtained by laser scanning through a Delauay algorithm to obtain a target three-dimensional model; and importing the target three-dimensional model into a rendering engine, calculating a target three-dimensional model bounding box and a central point, and moving the target three-dimensional model to enable the central point to reach an origin point.
3. The method for estimating the three-dimensional pose of the moving end based on the point cloud preprocessing and the ShuffleNet according to claim 1, wherein the concrete implementation of the step 2 comprises the following substeps:
step 2.1: shooting a target in a three-dimensional engine to obtain a target photo;
step 2.2: detecting two-dimensional feature points in the target picture through an SIFT algorithm to obtain a feature point set K ═ K1,...,kn};
Step 2.3: calculating each characteristic point k by a screen ray projection algorithmiCorresponding three-dimensional coordinate point piAnd recording the three-dimensional feature point set corresponding to the two-dimensional feature point as P ═ P1,...,pn};
Step 2.4: the camera rotates around the target and continues to take pictures of the target, and the step 2.2 and the step 2.3 are repeated until a multi-view picture of the target is obtained and a three-dimensional feature point set PS is obtained through calculation1,...,PNN is the number of photos, and P is a three-dimensional feature point set of each photo;
step 2.5: dividing the target three-dimensional voxel into M blocks B ═ B with the same voxel size1,...,bM}; setting the frequency of the three-dimensional feature points appearing in each voxel block as a voxel block weight q, and screening m blocks KB with the maximum weight as { b }1,...,bmAs key voxel block, where m<M;
Step 2.6: taking the key voxel blocks as categories, calculating the areas of the key voxel blocks on the two-dimensional photo set according to a projection transformation formula, and generating marking information to obtain a training data set;
step 2.7: and training ShuffleNetv2-YOLOv3 through the generated data set to obtain a detection model aiming at the target key voxel block.
4. The method for estimating the three-dimensional pose of the moving end based on the point cloud preprocessing and the ShuffleNet according to claim 1, wherein the concrete implementation of the step 3 comprises the following substeps:
step 3.1: reading a video stream, inputting a trained detection model of ShuffleNetv2-YOLOv3 key voxel blocks, and outputting the detection model into two-dimensional areas corresponding to a plurality of key voxel blocks;
step 3.2: calculating the detected central point of the two-dimensional region, and forming a 2D-3D matching point pair with the central point of the corresponding key voxel block;
step 3.3: and calculating the relative pose between the target and the mobile terminal camera through RANSAC and EPNP algorithms.
5. The method for estimating the three-dimensional pose of the moving end based on the point cloud preprocessing and the ShuffleNet according to any one of claims 1 to 4, wherein the specific implementation of the step 4 comprises the following substeps:
step 4.1: reading GPS and IMU data of a mobile terminal;
step 4.2: calculating the positioning of the mobile terminal in the three-dimensional world through the data acquired in the step 4.1;
step 4.3: and (3) calculating the absolute pose of the target in the three-dimensional world according to the moving end pose calculated in the step 4.2 and the relative pose of the target calculated in the step 3.
CN202110634620.0A 2021-06-08 2021-06-08 Point cloud preprocessing and ShuffleNet-based mobile terminal three-dimensional pose estimation method Active CN113362467B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110634620.0A CN113362467B (en) 2021-06-08 2021-06-08 Point cloud preprocessing and ShuffleNet-based mobile terminal three-dimensional pose estimation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110634620.0A CN113362467B (en) 2021-06-08 2021-06-08 Point cloud preprocessing and ShuffleNet-based mobile terminal three-dimensional pose estimation method

Publications (2)

Publication Number Publication Date
CN113362467A true CN113362467A (en) 2021-09-07
CN113362467B CN113362467B (en) 2023-04-07

Family

ID=77532983

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110634620.0A Active CN113362467B (en) 2021-06-08 2021-06-08 Point cloud preprocessing and ShuffleNet-based mobile terminal three-dimensional pose estimation method

Country Status (1)

Country Link
CN (1) CN113362467B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115222731A (en) * 2022-09-07 2022-10-21 西南交通大学 Train fastener abnormity detection method based on two-dimensional image-point cloud mapping
CN115578265A (en) * 2022-12-06 2023-01-06 中汽智联技术有限公司 Point cloud enhancement method, system and storage medium

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090232353A1 (en) * 2006-11-10 2009-09-17 University Of Maryland Method and system for markerless motion capture using multiple cameras
US20150269438A1 (en) * 2014-03-18 2015-09-24 Sri International Real-time system for multi-modal 3d geospatial mapping, object recognition, scene annotation and analytics
EP3276575A1 (en) * 2016-07-25 2018-01-31 Nuctech Company Limited Method, apparatus and system for reconstructing images of 3d surface
WO2018047687A1 (en) * 2016-09-12 2018-03-15 パナソニックIpマネジメント株式会社 Three-dimensional model generating device and three-dimensional model generating method
WO2018148924A1 (en) * 2017-02-17 2018-08-23 深圳市大疆创新科技有限公司 Method and device for reconstructing three-dimensional point cloud
CN109241990A (en) * 2018-07-19 2019-01-18 杭州电子科技大学 A kind of threedimensional model mask method propagated based on multi-tag
CN109387204A (en) * 2018-09-26 2019-02-26 东北大学 The synchronous positioning of the mobile robot of dynamic environment and patterning process in faced chamber
CN109961504A (en) * 2017-12-25 2019-07-02 宏达国际电子股份有限公司 Method for reconstructing three-dimensional model, electronic device and non-transient computer-readable recording medium
US20190213402A1 (en) * 2018-01-09 2019-07-11 Futurewei Technologies, Inc. Head pose and distraction estimation
CN110517209A (en) * 2018-05-21 2019-11-29 北京京东尚科信息技术有限公司 Data processing method, device, system and computer readable storage medium
CN111814683A (en) * 2020-07-09 2020-10-23 北京航空航天大学 Robust visual SLAM method based on semantic prior and deep learning features
CN111899301A (en) * 2020-06-02 2020-11-06 广州中国科学院先进技术研究所 Workpiece 6D pose estimation method based on deep learning
US20210073641A1 (en) * 2019-09-05 2021-03-11 Kabushiki Kaisha Toshiba Learning device, learning system, and learning method
CN112489129A (en) * 2020-12-18 2021-03-12 深圳市优必选科技股份有限公司 Pose recognition model training method and device, pose recognition method and terminal equipment
US20210158023A1 (en) * 2018-05-04 2021-05-27 Northeastern University System and Method for Generating Image Landmarks
US20210407039A1 (en) * 2020-06-30 2021-12-30 Intel Corporation Apparatus and method for approximate trilinear interpolation for scene reconstruction

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090232353A1 (en) * 2006-11-10 2009-09-17 University Of Maryland Method and system for markerless motion capture using multiple cameras
US20150269438A1 (en) * 2014-03-18 2015-09-24 Sri International Real-time system for multi-modal 3d geospatial mapping, object recognition, scene annotation and analytics
EP3276575A1 (en) * 2016-07-25 2018-01-31 Nuctech Company Limited Method, apparatus and system for reconstructing images of 3d surface
WO2018047687A1 (en) * 2016-09-12 2018-03-15 パナソニックIpマネジメント株式会社 Three-dimensional model generating device and three-dimensional model generating method
WO2018148924A1 (en) * 2017-02-17 2018-08-23 深圳市大疆创新科技有限公司 Method and device for reconstructing three-dimensional point cloud
CN109961504A (en) * 2017-12-25 2019-07-02 宏达国际电子股份有限公司 Method for reconstructing three-dimensional model, electronic device and non-transient computer-readable recording medium
US20190213402A1 (en) * 2018-01-09 2019-07-11 Futurewei Technologies, Inc. Head pose and distraction estimation
US20210158023A1 (en) * 2018-05-04 2021-05-27 Northeastern University System and Method for Generating Image Landmarks
CN110517209A (en) * 2018-05-21 2019-11-29 北京京东尚科信息技术有限公司 Data processing method, device, system and computer readable storage medium
CN109241990A (en) * 2018-07-19 2019-01-18 杭州电子科技大学 A kind of threedimensional model mask method propagated based on multi-tag
CN109387204A (en) * 2018-09-26 2019-02-26 东北大学 The synchronous positioning of the mobile robot of dynamic environment and patterning process in faced chamber
US20210073641A1 (en) * 2019-09-05 2021-03-11 Kabushiki Kaisha Toshiba Learning device, learning system, and learning method
CN111899301A (en) * 2020-06-02 2020-11-06 广州中国科学院先进技术研究所 Workpiece 6D pose estimation method based on deep learning
US20210407039A1 (en) * 2020-06-30 2021-12-30 Intel Corporation Apparatus and method for approximate trilinear interpolation for scene reconstruction
CN111814683A (en) * 2020-07-09 2020-10-23 北京航空航天大学 Robust visual SLAM method based on semantic prior and deep learning features
CN112489129A (en) * 2020-12-18 2021-03-12 深圳市优必选科技股份有限公司 Pose recognition model training method and device, pose recognition method and terminal equipment

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
MENGDAN FENG 等: "2D3D-MatchNet: Learning to Match Keypoints Across 2D Image and 3D Point Cloud", 《2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA)》 *
付泰等: "一种具有精确位姿的飞机CAD模型投影图像库设计方法", 《计算机科学》 *
詹川 著: "《反垃圾邮件信息过滤技术研究》", 31 May 2016, 成都:电子科技大学出版社出版 *
马小秋等: "基于透视式增强现实的BIM建筑三维重建仿真", 《计算机仿真》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115222731A (en) * 2022-09-07 2022-10-21 西南交通大学 Train fastener abnormity detection method based on two-dimensional image-point cloud mapping
CN115222731B (en) * 2022-09-07 2022-12-02 西南交通大学 Train fastener abnormity detection method based on two-dimensional image-point cloud mapping
CN115578265A (en) * 2022-12-06 2023-01-06 中汽智联技术有限公司 Point cloud enhancement method, system and storage medium
CN115578265B (en) * 2022-12-06 2023-04-07 中汽智联技术有限公司 Point cloud enhancement method, system and storage medium

Also Published As

Publication number Publication date
CN113362467B (en) 2023-04-07

Similar Documents

Publication Publication Date Title
US11393173B2 (en) Mobile augmented reality system
CN107330439B (en) Method for determining posture of object in image, client and server
EP2915140B1 (en) Fast initialization for monocular visual slam
US20180189974A1 (en) Machine learning based model localization system
KR102097016B1 (en) Apparatus and methdo for analayzing motion
US10043097B2 (en) Image abstraction system
Wei et al. Applications of structure from motion: a survey
CN110458865B (en) In situ formation of planar natural feature targets
US20100194679A1 (en) Gesture recognition system and method thereof
CN108958469B (en) Method for adding hyperlinks in virtual world based on augmented reality
CN113362467B (en) Point cloud preprocessing and ShuffleNet-based mobile terminal three-dimensional pose estimation method
Pan et al. Rapid scene reconstruction on mobile phones from panoramic images
CN103345736A (en) Virtual viewpoint rendering method
TW201222288A (en) Image retrieving system and method and computer program product thereof
KR102464271B1 (en) Pose acquisition method, apparatus, electronic device, storage medium and program
CN115008454A (en) Robot online hand-eye calibration method based on multi-frame pseudo label data enhancement
Zhu et al. Large-scale architectural asset extraction from panoramic imagery
CN111914790A (en) Real-time human body rotation angle identification method based on double cameras under different scenes
CN109118576A (en) Large scene three-dimensional reconstruction system and method for reconstructing based on BDS location-based service
CN104680520A (en) Field three-dimensional information investigation method and system
Kikuchi et al. Automatic diminished reality-based virtual demolition method using semantic segmentation and generative adversarial network for landscape assessment
Kim et al. Vision-based all-in-one solution for augmented reality and its storytelling applications
CN112258435A (en) Image processing method and related product
CN105631938A (en) Image processing method and electronic equipment
WO2022253043A1 (en) Facial deformation compensation method for facial depth image, and imaging apparatus and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant