WO2021115071A1 - Three-dimensional reconstruction method and apparatus for monocular endoscope image, and terminal device - Google Patents

Three-dimensional reconstruction method and apparatus for monocular endoscope image, and terminal device Download PDF

Info

Publication number
WO2021115071A1
WO2021115071A1 PCT/CN2020/129546 CN2020129546W WO2021115071A1 WO 2021115071 A1 WO2021115071 A1 WO 2021115071A1 CN 2020129546 W CN2020129546 W CN 2020129546W WO 2021115071 A1 WO2021115071 A1 WO 2021115071A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
key frame
pixel
coordinates
distortion
Prior art date
Application number
PCT/CN2020/129546
Other languages
French (fr)
Chinese (zh)
Inventor
廖祥云
孙寅紫
王琼
王平安
Original Assignee
中国科学院深圳先进技术研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中国科学院深圳先进技术研究院 filed Critical 中国科学院深圳先进技术研究院
Publication of WO2021115071A1 publication Critical patent/WO2021115071A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/80Geometric correction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10068Endoscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30204Marker
    • G06T2207/30208Marker matrix
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Definitions

  • This application belongs to the field of image processing technology, and in particular relates to a method, device and terminal device for three-dimensional reconstruction of monocular endoscopic images.
  • Three-dimensional reconstruction is one of the research hotspots in computer vision. Its main purpose is to restore the three-dimensional structure of objects from two-dimensional images. It is widely used in augmented reality, virtual navigation and medical fields.
  • the three-dimensional information of the image mainly depends on the visual real-time positioning and map construction (Simultaneous Localization and Mapping, SLAM) technology.
  • the distortion of the monocular endoscope imaging will cause the increase of the pose error, and the endoscope is usually used with a cold light source, and its imaging will also be disturbed by the light, which may affect the feature matching result in the SLAM process. It is usually difficult to use a monocular endoscope to provide accurate samples for training. Combining the SLAM scheme and the depth prediction scheme can the two-dimensional image sequence be densely reconstructed. However, due to the above-mentioned pose and depth map errors and other factors, The accuracy and effect of 3D reconstruction are deteriorated.
  • the embodiments of the present application provide a method and device for three-dimensional reconstruction of a monocular endoscope, which can solve the problem of reducing errors caused by imaging distortion caused by the inherent parameters of the monocular endoscope, and the accuracy of the three-dimensional reconstruction of a two-dimensional image sequence is not high and The problem of poor results.
  • an embodiment of the present application provides a three-dimensional reconstruction method of a monocular endoscopic image, including:
  • the image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud includes:
  • the acquiring the distortion images of a plurality of checkerboard calibration boards taken by a monocular endoscope, and correcting the distortion images of the checkerboard calibration boards to obtain an image sequence includes:
  • the performing distortion correction on the image to be corrected based on the camera coordinate system to obtain the image sequence includes:
  • the pixel coordinates of the preset coordinates in the pixel coordinate system are mapped to the camera coordinate system to obtain the image sequence.
  • the obtaining the pixel coordinates of the key frame includes:
  • the determining a key frame from the image sequence includes:
  • the first image is used as a key frame, where the first image and the second image are Any two adjacent frames of images in the image sequence.
  • the obtaining the pose parameters of the key frame includes:
  • the estimating the depth map of the key frame includes:
  • a reference frame image from the key frames, where the reference frame image is any frame image or multiple frames of images in the key frame;
  • an embodiment of the present application provides a three-dimensional reconstruction device for monocular endoscopic images, including:
  • the acquisition module is used to acquire the distortion images of a plurality of checkerboard calibration boards taken by a monocular endoscope, and perform distortion correction on the distortion images of the checkerboard calibration boards to obtain an image sequence;
  • a calculation module for obtaining the pose parameters of the key frame and estimating the depth map of the key frame
  • the generating module is used for image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud.
  • an embodiment of the present application provides a terminal device, including a memory, a processor, and a computer program stored in the memory and running on the processor.
  • the processor executes the computer program, Realize the above-mentioned three-dimensional reconstruction method.
  • an embodiment of the present application provides a computer-readable storage medium that stores a computer program that implements the above-mentioned three-dimensional reconstruction method when the computer program is executed by a processor.
  • the embodiment of the present application has the following beneficial effects: by obtaining the distortion images of multiple checkerboard calibration boards taken by a monocular endoscope, and performing distortion correction on the distortion images of the multiple checkerboard calibration boards to obtain an image sequence, Determine the key frame from the image sequence, obtain the pose parameters of the key frame, estimate the depth map of the key frame, and reconstruct the image based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud.
  • the above method uses the checkerboard calibration board image to achieve the calibration and distortion correction of the monocular endoscope to obtain the image sequence, which effectively reduces the imaging distortion error caused by the monocular endoscope itself, and determines multiple images that meet the requirements from the image sequence
  • a key frame and determine the pose parameters of the key frame, it can avoid the interference of external factors such as light changes, can accurately estimate the pose parameters and depth map, and perform image reconstruction according to the position parameters and depth map of the key frame to get more
  • the fine three-dimensional point cloud also improves the display effect of the image.
  • FIG. 1 is a schematic flowchart of a method for three-dimensional reconstruction of monocular endoscopic images provided by an embodiment of the present application
  • FIG. 2 is a schematic diagram of the image distortion correction process provided by an embodiment of the present application.
  • FIG. 3 is a schematic structural diagram of a three-dimensional reconstruction device for monocular endoscopic images provided by an embodiment of the present application
  • Fig. 4 is a schematic structural diagram of a terminal device provided by an embodiment of the present application.
  • the term “if” can be construed as “when” or “once” or “in response to determination” or “in response to detecting “.
  • the phrase “if determined” or “if detected [described condition or event]” can be interpreted as meaning “once determined” or “in response to determination” or “once detected [described condition or event]” depending on the context ]” or “in response to detection of [condition or event described]”.
  • FIG. 1 shows a schematic flow chart of a three-dimensional reconstruction method of a monocular endoscopic image provided by the first application of the present application.
  • the three-dimensional reconstruction method of a monocular endoscope image provided by the present application includes S101 To S104, the details are as follows:
  • S101 Obtain distortion images of multiple checkerboard calibration boards taken by a monocular endoscope, and perform distortion correction on the distortion images of the multiple checkerboard calibration boards to obtain an image sequence;
  • the distortion image of the checkerboard calibration board can be used for distortion correction of the monocular endoscope.
  • the checkerboard calibration board is a binary image with two rows of black and white stripe intervals.
  • the monocular endoscope can observe the calibration from different angles.
  • the board obtains multiple monocular endoscope distortion images.
  • the imaging process of the camera mainly involves the transformation between the image pixel coordinate system, the image physical coordinate system, the camera coordinate system and the world coordinate system. Due to the lens imaging principle, the camera imaging distortion occurs, and the distortion correction is to find the corresponding relationship between the point positions before and after the distortion. .
  • the imaging model of the monocular endoscope is different from the small hole imaging model, but is closer to the fisheye camera model.
  • the checkerboard calibration board is a black and white grid arranged at intervals, also called a checkerboard calibration board.
  • the calibration target is used in machine vision, image measurement, photogrammetry, three-dimensional reconstruction and other applications to correct lens distortion and determine the physical size and pixels.
  • the conversion relationship between the two, and to determine the relationship between the three-dimensional geometric position of a point on the surface of a space object and its corresponding point in the image requires the establishment of a geometric model of camera imaging.
  • the camera's geometric model can be obtained by the camera with a fixed-pitch pattern array plate and calculation by the calibration algorithm, thereby obtaining high-precision measurement and reconstruction results.
  • the flat plate with a fixed-pitch pattern array is the calibration plate.
  • the camera calibration of the monocular endoscope can be realized, and the distorted image can be corrected to obtain the image according to the calibration monocular endoscope.
  • Sequence that is, real image, can reduce the error caused by image distortion to image recognition.
  • Figure 2 shows a flow chart of the implementation of distortion correction provided by the present application. As shown in Figure 2, the acquisition of the distortion images of the multiple checkerboard calibration boards taken by the monocular endoscope, and the distortion images of the multiple checkerboard calibration boards Perform correction to obtain an image sequence, including S1011 to S1013:
  • 20 images with a checkerboard calibration board taken with a monocular endoscope at about different angles are acquired, the corner points of the checkerboard in the image are extracted, and the distorted image that meets the fitting conditions is selected.
  • the Canny corner operator can be used to detect the distortion images obtained from all monocular endoscopes observing the checkerboard calibration board, and count the number of corner points in all the distortion images.
  • the distortion image that meets the fitting conditions is preferably the number of corner points detected in the image No less than 6 meshes. Among them, the number of corner points can be selected according to actual conditions, and is not specifically limited here.
  • the parameters of the ellipse equation are obtained by fitting the selected distortion image and the detected corner points.
  • the ellipse equation can be a standard equation including 6 parameters. According to the corner points in the detected distortion image, the least square method is used to obtain the parameters.
  • the parameters of the ellipse equation are obtained from the curved surface projection parameters, and the parameter fitting results of the ellipse equations of multiple distorted images are obtained by means of filtering.
  • fx and fy are the focal lengths of the endoscope in pixels
  • cx and cy are the principal point positions in pixels (that is, the center pixel position of the imaging).
  • the chessboard is a calibration board composed of black and white squares as the calibration object for camera calibration (mapped from the real world to the object in the digital image). Compared with a three-dimensional object, a two-dimensional object lacks some information.
  • the checkerboard is used as a calibration object because the plane checkerboard mode is easier to handle. After changing the position of the checkerboard many times to capture the image, it can obtain richer coordinate information.
  • S1012 Determine an image to be corrected from the distorted image according to the camera parameter and the distortion parameter;
  • calibrating the monocular endoscope can determine the pose of the camera to obtain the camera parameters and distortion parameters of the monocular endoscope.
  • the camera parameters and the distortion parameters are calculated Whether the image has not been distorted to the distorted image to be corrected, that is, it can judge whether the captured multiple images are distorted, or set a preset threshold, and compare the calculation result with the preset threshold to obtain a comparison result, where the comparison result is The ones with greater difference in the comparison result are regarded as distortion and those with little difference in the comparison result are regarded as no distortion, and vice versa.
  • various distortions are often produced in the process of image acquisition or display.
  • the common ones are geometric shape distortion, grayscale distortion, and color distortion.
  • the causes of image distortion are the aberration and distortion of the imaging system. , Limited bandwidth, shooting status, scanning nonlinearity, relative motion, etc., non-uniform lighting conditions or point light source lighting, etc.
  • the image to be corrected is determined from the multiple images taken, which is convenient to eliminate the error of distortion in image recognition and processing, and improves the accuracy of image processing to a certain extent.
  • S1013 Perform distortion correction on the image to be corrected based on the camera coordinate system to obtain the image sequence.
  • the straight line in the distortion space is generally no longer a straight line in the image space, but only the straight line passing through the center of symmetry is an exception.
  • the center of symmetry can be found, and then the general geometric distortion correction process can be performed. .
  • the general steps of distortion correction are to first find the symmetry center of the distortion map, convert the address space relationship represented by the distortion map into a space coordinate system with the center of symmetry as the origin, and then space transformation, rearrange the pixels on the input image, that is, the distortion map to restore The original spatial relationship, that is, use the address mapping relationship to find their corresponding point in the distortion map space for each point in the correction map space, and finally the grayscale difference is to assign the corresponding grayscale value to the pixel after the space transformation to restore the original The gray value of the location.
  • the correction of geometric distortion requires the use of coordinate transformation, including simple transformations such as parallel movement, rotation, enlargement and reduction.
  • the process of distortion correction can be understood as processing a distorted image into an undistorted image, that is, a real image.
  • Different camera models display different images when taking pictures, and may or may not be distorted.
  • the process of using distortion correction can be the same or different.
  • Image distortion mainly includes radial distortion and tangential distortion.
  • Radial distortion refers to the smallest distortion at the center position. As the radius increases, the distortion increases. Radial distortion can be divided into pincushion distortion and barrel distortion. Tangential distortion refers to when the lens is not parallel to the imaging plane, similar to perspective transformation.
  • the performing distortion correction on the image to be corrected based on the camera coordinate system to obtain the image sequence includes steps A1 to A3:
  • Step A1 Obtain the preset coordinates of each pixel of the image to be corrected in the camera coordinates
  • the camera coordinate system can be obtained by calibrating the monocular endoscope.
  • the world coordinate system and the camera coordinate system, the camera coordinate system and the image coordinate system, and the image coordinates can be realized.
  • System to pixel coordinate system conversion The conversion between the world coordinate system and the camera coordinate system is from one three-dimensional coordinate to another three-dimensional coordinate system.
  • the pose parameters of the camera can be obtained through the rotation matrix and the translation vector, that is, the camera coordinate system.
  • From the camera coordinate system to the image coordinate system a three-dimensional coordinate is projected on a two-dimensional plane, and it is estimated based on the distance between the two coordinate systems, that is, the focal length of the camera.
  • the preset coordinates in the camera coordinate system are corrected to obtain the coordinates in the undistorted camera coordinate system, and the coordinates in the undistorted camera coordinate system are mapped to the pixel coordinate system to obtain an undistorted image sequence.
  • Step A2 Project the camera coordinate system onto the plane where each pixel of the image to be corrected is located, and obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
  • the coordinates of the pixel (u', v') of the image taken by the monocular endoscope in the camera coordinate system are (x, y, z), and the pixel point in the camera coordinate system is The coordinates are projected to the plane where the image is located, that is, the image coordinate system.
  • the positional relationship of the origin of the image coordinate system relative to the origin of the pixel coordinate system it can be regarded as projecting the coordinates of the pixel point in the camera coordinate system to the pixel coordinate system, which can be expressed as follows :
  • ⁇ ′ ⁇ (1+k 1 ⁇ 2 +k 2 ⁇ 4 +k 3 ⁇ 6 +k 4 ⁇ 8 )
  • (x', y') are the coordinates projected on the plane
  • r represents the distance between the point and the center on the projection plane (projection radius)
  • represents the angle of incidence.
  • Step A3 Map the pixel coordinates of the preset coordinates in the pixel coordinate system to the camera coordinate system to obtain the image sequence.
  • N undistorted images there are a total of 4 internal parameters + 6N external parameters to calibrate.
  • 10 or 20 images can be generally used to obtain a more accurate solution by using the least square method.
  • the distortion-related parameters can be obtained according to the remaining point coordinates.
  • a remapping process can be used to convert the pixel coordinates of the distorted endoscopic image into the coordinates of the distorted camera coordinate system.
  • the distorted camera coordinate system coordinates are transformed into the undistorted camera coordinate system coordinates, and finally the undistorted camera coordinate system coordinates are transformed into the pixel coordinates of the undistorted image, so that the corrected image sequence can be obtained, and the corresponding image can also be obtained.
  • the pixel coordinates of it is convenient to determine the key frame and the pose parameters of the key frame later.
  • ORB_SLAM2 is an embedded position recognition model, which has the characteristics of relocation, preventing tracking failure (such as occlusion), reinitialization of the mapped scene, loop detection, etc., using the same ORB features for tracking, mapping, and location recognition tasks.
  • Features have good robustness in rotation and scale, and have good invariance to the camera's automatic gain, automatic exposure, and illumination changes. It can also quickly extract features and match features to meet the needs of real-time operation.
  • This application uses ORB_SLAM2 to determine the key frame and pose estimation of monocular endoscopic images.
  • ORB You can use ORB to extract features of the image sequence, estimate the initial pose of the camera through the previous image frame, and initialize the position through global relocation.
  • the four processes of posture, tracking the local map and the judgment standard of the new key frame are used to determine the key frame and the posture parameter of the key frame more accurately.
  • the key frame can be used as a mark of the image sequence and has a guiding effect.
  • the distortion-corrected images in the image sequence are arranged in a preset order, and they can be arranged in sequence according to the shooting time order, which is convenient for feature extraction of each image. Processing to improve the efficiency of monocular endoscope image processing.
  • the determining the key frame from the image sequence includes steps B1 to B2, which are specifically as follows:
  • Step B1 Obtain local features of each image in the image sequence, and perform feature point matching on each image in the image sequence based on the local feature of each image to obtain a matching result;
  • extract the local features of each image in the image sequence and perform feature point matching on each image in the image sequence with the local features of each image to extract the regions corresponding to the coordinates of each image for feature matching, or Extract all pixels in the rich area of the image, and match the feature points of the two images before and after in the preset order, that is, the number of feature points that are successfully matched with the same ORB feature in the two frames of the image is used as the matching result, and set the number of feature points that are successfully matched
  • the threshold is between 50-100.
  • the peripheral edge area of the image imaged by the monocular endoscope is a black area with no information, and useful feature information cannot be extracted. Therefore, the information-rich area in the image is selected and the area can be defined as a region of interest.
  • ORB Oriented FAST and Rotated Brief
  • the ORB algorithm includes feature point extraction and feature point description.
  • the ORB algorithm has the characteristics of fast calculation speed and uses FAST detection
  • the feature point again, is to use the BRIEF algorithm to calculate the descriptor.
  • the unique binary string representation of the descriptor not only saves storage space, but also greatly shortens the matching time.
  • the key frame can be used as a marker to quickly process the image sequence, which can improve the efficiency of monocular endoscope image processing.
  • Step B2 When the matching result is that the number of feature points matched by the first image and the second image is greater than or equal to a preset threshold, use the first image as a key frame, wherein the first image and the first image The two images are any two adjacent frames of images in the image sequence.
  • the threshold for the number of feature points that are successfully matched is set to be between 50 and 100. When the number of feature points matched by the first image and the second image exceeds the threshold, it is determined that the first and the next two frames of images are matched successfully .
  • the constant motion rate model can be used to predict the current camera position (that is, the camera is considered to be moving at a constant speed), and then the corresponding cloud of the feature points in the previous frame image can be searched for in the map
  • the matching point between the point and the current frame image is finally used to further optimize the pose of the current camera by using the searched matching point, so as to obtain the image in the image sequence that meets the requirements to improve the accuracy of determining the key frame.
  • ORB_SLAM2 based on the feature point method can obtain the pose parameters.
  • the pose parameters describe the two images corresponding to the camera.
  • the depth map refers to the number of bits used to store each pixel, which is used to measure the color resolution of the image.
  • obtaining the pose parameters of the key frame includes:
  • the pose of the first image that is, the image of the previous frame
  • the key frame contains
  • extract the ORB feature of each frame of the key frame according to the pose initialization of the first image perform feature matching with the previous frame, and estimate its pose parameters (rotation matrix Ri, translation vector ti )
  • take the image of successful pose estimation as the key frame obtain the pose parameter corresponding to the key frame, and store the key frame and the pose parameter corresponding to the key frame together, so as to perform depth estimation on all the key frames later.
  • the estimating the depth map of the key frame includes:
  • a reference frame image from the key frames, where the reference frame image is any frame image or multiple frames of images in the key frame;
  • the photometric error is minimized according to the first depth map of the key image frame in the monocular video, and the current camera pose between the reference frame image and the key frame in the monocular endoscopic image is determined.
  • the current camera pose triangulates the high gradient image points in the reference frame image and the key frame, determines the second depth map of the key frame, performs Gaussian fusion of the first depth map and the second depth map, and updates the first depth map of the key frame.
  • a depth map If the next camera pose between the next image frame and the key frame of the reference frame image exceeds the preset camera pose, the updated first depth map is determined as the dense depth map of the key frame.
  • depth map estimation one frame of image or multiple frames of images can be selected for estimation.
  • each pixel of each image in the key frame is triangulated.
  • Bayesian probability estimation strategy to get dense depth map.
  • multiple images in the key frame are selected for iterative calculation to obtain the depth value corresponding to each pixel, then the depth map is smoothed and filtered to eliminate some noise in the depth map, which can improve the efficiency and accuracy of depth estimation.
  • the first depth map of the key frame can be a dense depth map obeying the Gaussian distribution obtained by initializing the depth values of the high gradient points in the key frame, or it can be the depth of the previous key frame of the key frame.
  • the value is a dense depth map projected according to the camera pose. For example, if the key frame to be depth estimated is the first key frame in the image sequence, the first depth map of the key frame is the dense depth map obtained by initialization; if the key frame with depth estimation is the first key frame in the image sequence For key frames other than one key frame, the first depth map of the key frame is a dense depth map obtained by projecting the depth value of the previous key frame.
  • Luminosity error refers to the measurement difference between the high gradient point in the projected image and the corresponding high gradient point in the reference frame image.
  • the projected image is based on the initial camera pose between the reference frame and the key frame in the image sequence.
  • the high gradient points corresponding to the pixels in the frame are projected to the reference frame image.
  • the current camera pose includes the rotation and translation between the reference frame and the key frame.
  • the second depth map of the key frame refers to the image sequence according to the The new dense depth map obtained by triangulating the current camera pose between the reference frame image and the key frame; the next frame image of the reference frame image refers to the next frame image adjacent to the pre-reference frame image in the image sequence ,
  • the posture of the latter camera includes the maximum threshold of the posture of the latter camera, which can be preset according to actual conditions and requirements, and there is no specific limitation here.
  • a dense depth map refers to an image that includes depth values corresponding to a large number of feature points, or an image that includes both high gradient points and depth values corresponding to low gradient points.
  • the depth estimation obtains the depth map and the depth value, which is convenient for subsequent restoration of the spatial coordinates of the pixel.
  • S104 Perform image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud.
  • 3D reconstruction refers to the establishment of a 3D model from the input data.
  • Each frame of data scanned by the depth camera not only contains the color RGB image of the point in the scene, but also includes each point to the vertical plane where the depth camera is located. This distance value is called the depth value, and these depth values together constitute the depth map of this frame.
  • the depth map can be regarded as a grayscale image.
  • the grayscale value of each point in the image represents the true distance from the position of the point in reality to the vertical plane where the camera is located.
  • Each point in the RGB image corresponds to a point on the camera. A three-dimensional point in the local coordinate system.
  • the process of 3D reconstruction can be image acquisition, camera calibration, feature extraction, stereo matching, 3D reconstruction, etc., where stereo matching refers to establishing a correspondence between image pairs based on the extracted features. That is, the imaging points of the same physical space point in two different images are mapped one by one.
  • stereo matching refers to establishing a correspondence between image pairs based on the extracted features. That is, the imaging points of the same physical space point in two different images are mapped one by one.
  • stereo matching pay attention to the interference of some factors in the scene, such as lighting conditions, noise interference, distortion of the geometric shape of the scene, surface physical characteristics, and camera characteristics, etc., in order to obtain a high-precision three-dimensional point cloud, and also enhance the vision effect.
  • S104 may include steps C1 to C3, which are specifically as follows:
  • Step C1 Obtain the pixel coordinates of the key frame
  • the pixel coordinate system and the pixel coordinates of each image in the key frame can be determined.
  • the pixel coordinates indicate the position of the pixel in the image, and the key frame can be determined.
  • the pixel position of each image is convenient for subsequent three-dimensional reconstruction of the image.
  • Step C2 calculating the target space coordinates according to the depth map, the pose parameters of the key frame, and the pixel coordinates of the key frame;
  • the depth value corresponding to the depth map of each image in the key is obtained, and the depth value and the pose parameter of the key frame and the pixel coordinates of each image of the key frame are calculated to obtain the spatial coordinates of each image, namely
  • the conversion from two-dimensional coordinates to three-dimensional coordinates, according to the depth value obtained by accurate depth estimation, also improves the accuracy of the calculated target space coordinates.
  • Step C3 Obtain the color information of each pixel in the key frame, and perform point cloud fusion on the key frame according to the color information of each pixel in the key frame and the target space coordinates to obtain the Describe the three-dimensional point cloud.
  • the pixel coordinates [u, v] in the two-dimensional image, the corresponding point cloud contains color information and spatial position information, and the color information is represented by the RGB value of the pixel.
  • the pose parameters of the key frame and the pixel coordinates of the key frame are calculated to obtain the target space coordinates as [x, y, z], and the space coordinates are restored from the pixel coordinates [u, v] and its depth value d by the following formula Means:
  • d represents the depth of the pixel, which is derived from the depth estimation of the REMODE scheme
  • (x',y',z') is the coordinate value in the camera coordinate system
  • (Ri,ti) is the bit corresponding to the frame Pose parameters.
  • the point cloud is a set of discrete points.
  • the point cloud stores the spatial coordinates and color information corresponding to the pixels of the frame.
  • the multi-frame point cloud Stored in a container, and then the repeated point cloud is removed by a filter, and a three-dimensional point cloud of multiple frames of images can be obtained.
  • the above-mentioned three-dimensional reconstruction method may draw a point cloud of multiple frames of images during fusion to obtain finer three-dimensional information.
  • obtaining the pixel coordinates of the key frame in step C1 includes steps C11 to C13:
  • Step C11 Project the camera coordinate system onto the plane where each pixel of the image to be corrected is located, and obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
  • the coordinates of the pixel points in the camera coordinate system are defined, and the correspondence between the camera coordinate system and the image coordinate system is calculated by projection, and then the pixel coordinate system is obtained through the correspondence between the image coordinate system and the pixel coordinate system.
  • the pixel coordinates here are the same as the process of the pixel coordinates obtained in the above-mentioned distortion correction, and will not be repeated here.
  • Step C12 Map the pixel coordinates of the preset coordinates in the pixel coordinate system to the camera coordinate system to obtain the image sequence and the pixel coordinates corresponding to the image sequence;
  • the corrected image sequence and the pixel coordinates corresponding to the image sequence can be obtained through the coordinate system transformation method of the distortion correction.
  • the specific processing process here is the same as the above-mentioned distortion correction process, and will not be repeated here.
  • Step C13 Obtain the pixel coordinates of the key frame based on the pixel coordinates corresponding to the image sequence.
  • the key frame is determined from the image sequence, and the pixel coordinates of the key frame can be obtained. According to the pixel coordinates of each image of the key frame, the moving position relationship of each image relative to the camera can be determined, so as to improve the single The processing efficiency of the endoscopic image.
  • FIG. 3 shows a three-dimensional reconstruction device 300 for monocular endoscopic images provided by an embodiment of the present application.
  • the three-dimensional reconstruction device 300 for monocular endoscope images provided by the present application includes:
  • the acquiring module 310 is configured to acquire the distortion images of a plurality of checkerboard calibration boards taken by a monocular endoscope, and perform distortion correction on the distortion images of the checkerboard calibration boards to obtain an image sequence;
  • the determining module 320 is configured to determine a key frame from the image sequence
  • the calculation module 330 is configured to obtain the pose parameters of the key frame and estimate the depth map of the key frame;
  • the generating module 340 is configured to perform image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud.
  • the device for 3D reconstruction of monocular endoscopic images may be a terminal device, a server, or a device capable of human-computer interaction.
  • the obtaining module 310 specifically includes:
  • the first acquisition unit is configured to acquire the corner points of the chessboard in the distortion images of the multiple chessboard calibration boards, and calibrate the monocular endoscope based on the corner points of the chessboard to obtain the monocular endoscope Camera parameters and distortion parameters of the mirror;
  • a first determining unit configured to determine an image to be corrected from the distorted image according to the camera parameter and the distortion parameter
  • the first processing unit is configured to perform distortion correction on the image to be corrected based on the camera coordinate system to obtain the image sequence.
  • the obtaining module 310 further includes:
  • a second acquiring unit configured to acquire the preset coordinates of each pixel of the image to be corrected in the camera coordinates
  • a second processing unit configured to project the camera coordinate system onto the plane where each pixel of the image to be corrected is located, to obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
  • the third processing unit is configured to map the pixel coordinates of the preset coordinates in the pixel coordinate system to the camera coordinate system to obtain the image sequence.
  • the determining module 320 specifically includes:
  • the third acquiring unit is configured to acquire the local features of each image in the image sequence, and perform feature point matching on each image in the image sequence based on the local features of each image to obtain a matching result;
  • the second determining unit is configured to use the first image as a key frame when the matching result is that the number of feature points matched by the first image and the second image is greater than or equal to a preset threshold, wherein the first image And the second image are any two adjacent frames of images in the image sequence.
  • the determining module 320 further includes:
  • a third determining unit configured to use the first image as a key frame when the number of feature points matched by the first image and the second image is greater than or equal to a preset threshold
  • the fourth processing unit is used to initialize the pose of the first image
  • the first estimation unit is used to estimate the pose parameters of the key frames in the image sequence.
  • the determining module 320 further includes:
  • a fourth determining unit configured to determine a reference frame image from the key frame, wherein the reference frame image is any frame image or multiple frames of images in the key frame;
  • the second estimation unit is configured to perform depth estimation processing on each pixel of the reference frame image based on the pose parameter to obtain the depth map of the key frame.
  • the generating module 340 includes:
  • the fourth acquiring unit is used to acquire the pixel coordinates of the key frame
  • the third estimation unit is configured to calculate the target space coordinates according to the depth map, the pose parameters of the key frame, and the pixel coordinates of the key frame;
  • the first generating unit is used to obtain the color information of each pixel in the key frame, and perform a point cloud on the key frame according to the color information of each pixel in the key frame and the target space coordinates Fusion to obtain the three-dimensional point cloud.
  • the generating module 340 further includes:
  • a first projection unit projecting the camera coordinate system onto a plane where each pixel point of the image to be corrected is located, to obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
  • the second projection unit is configured to map the pixel coordinates of the preset coordinates in the pixel coordinate system to the camera coordinate system to obtain the image sequence and the pixel coordinates corresponding to the image sequence;
  • the second generating unit is configured to obtain the pixel coordinates of the key frame based on the pixel coordinates corresponding to the image sequence.
  • FIG. 4 is a schematic structural diagram of a terminal device 400 provided by an embodiment of the present application.
  • the terminal device 400 includes a memory 410, at least one processor 420, and is stored in the memory 410 and can be stored in the processor 420.
  • the processor 420 executes the computer program 430, the above-mentioned three-dimensional reconstruction method is implemented.
  • the terminal device 400 may be a desktop computer, a mobile phone, a tablet computer, a wearable device, a vehicle-mounted device, an augmented reality (AR)/virtual reality (VR) device, a notebook computer, an ultra mobile personal computer (ultra -Mobile personal computer (UMPC), netbook, personal digital assistant (personal digital assistant, PDA) and other terminal devices, the embodiment of this application does not impose any restrictions on the specific types of terminal devices.
  • AR augmented reality
  • VR virtual reality
  • UMPC ultra mobile personal computer
  • PDA personal digital assistant
  • the terminal device 400 may include but is not limited to a processor 420 and a memory 410. Those skilled in the art can understand that FIG. 4 is only an example of the terminal device 400, and does not constitute a limitation on the terminal device 400. It may include more or less components than those shown in the figure, or a combination of certain components, or different components. , For example, can also include input and output devices.
  • the so-called processor 420 may be a central processing unit (CPU), and the processor 420 may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSP), and application specific integrated circuits (Application Specific Integrated Circuits). , ASIC), ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
  • the memory 410 may be an internal storage unit of the terminal device 400, such as a hard disk or a memory of the terminal device 400. In other embodiments, the memory 410 may also be an external storage device of the terminal device 400, such as a plug-in hard disk equipped on the terminal device 400, a smart media card (SMC), and a secure digital (Secure Digital). Digital, SD) card, flash card (Flash Card), etc. Further, the memory 410 may also include both an internal storage unit of the terminal device 400 and an external storage device. The memory 410 is used to store an operating system, an application program, a boot loader (Boot Loader), data, and other programs, such as the program code of the computer program. The memory 410 may also be used to temporarily store data that has been output or will be output.
  • a boot loader Boot Loader
  • the embodiments of the present application also provide a computer-readable storage medium, where the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the steps in each of the foregoing method embodiments can be realized.
  • the embodiments of the present application provide a computer program product.
  • the steps in the foregoing method embodiments can be realized when the mobile terminal is executed.
  • the integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
  • the computer program can be stored in a computer-readable storage medium.
  • the computer program can be stored in a computer-readable storage medium.
  • the steps of the foregoing method embodiments can be implemented.
  • the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate forms.
  • the computer-readable medium may at least include: any entity or device capable of carrying computer program code to a terminal device, a recording medium, a computer memory, a read-only memory (Read-Only Memory, ROM), and a random access memory (Random Access).
  • Memory RAM
  • electrical carrier signals telecommunications signals
  • software distribution media For example, U disk, mobile hard disk, floppy disk or CD-ROM, etc.
  • computer-readable media cannot be electrical carrier signals and telecommunication signals.
  • the disclosed apparatus/network equipment and method may be implemented in other ways.
  • the device/network device embodiments described above are only illustrative.
  • the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, such as multiple units.
  • components can be combined or integrated into another system, or some features can be omitted or not implemented.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Endoscopes (AREA)
  • Image Processing (AREA)

Abstract

A three-dimensional reconstruction method for a monocular endoscope image. The method comprises: acquiring a plurality of distorted images, photographed by a monocular endoscope, of a checkerboard calibration target, and performing distortion correction on the plurality of distorted images of the checkerboard calibration target to obtain an image sequence (S101); determining a key frame from the image sequence (S102); acquiring a pose parameter of the key frame, and estimating a depth map of the key frame (S103); and performing image reconstruction on the basis of the pose parameter of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud (S104). Further provided are a three-dimensional reconstruction apparatus (300) for a monocular endoscope image, and a terminal device (400). An error caused by imaging distortion of a monocular endoscope is reduced, and the display effect of an image is also improved.

Description

单目内窥镜图像的三维重建方法、装置及终端设备Method, device and terminal equipment for three-dimensional reconstruction of monocular endoscope image 技术领域Technical field
本申请属于图像处理技术领域,尤其涉及一种单目内窥镜图像的三维重建方法、装置及终端设备。This application belongs to the field of image processing technology, and in particular relates to a method, device and terminal device for three-dimensional reconstruction of monocular endoscopic images.
背景技术Background technique
三维重建是计算机视觉中的研究热点之一,其主要目的是从二维图像恢复物体三维结构,在增强现实,虚拟导航和医疗领域都有广泛运用。图像的三维信息主要依赖与视觉即时定位与地图构建(Simultaneous Localization and Mapping,SLAM)技术得出。Three-dimensional reconstruction is one of the research hotspots in computer vision. Its main purpose is to restore the three-dimensional structure of objects from two-dimensional images. It is widely used in augmented reality, virtual navigation and medical fields. The three-dimensional information of the image mainly depends on the visual real-time positioning and map construction (Simultaneous Localization and Mapping, SLAM) technology.
目前,单目内窥镜成像畸变会造成位姿误差的增大,而且内窥镜通常伴随着冷光源一起使用,其成像也会受光线的干扰,可能会影响SLAM过程中的特征匹配结果。通常很难用单目内窥镜提供准确的样本进行训练,结合SLAM方案和深度预测方案,才能够对二维图像序列进行稠密三维重建,但由于上述的位姿和深度图的误差等因素会造成三维重建的精度和效果变差。At present, the distortion of the monocular endoscope imaging will cause the increase of the pose error, and the endoscope is usually used with a cold light source, and its imaging will also be disturbed by the light, which may affect the feature matching result in the SLAM process. It is usually difficult to use a monocular endoscope to provide accurate samples for training. Combining the SLAM scheme and the depth prediction scheme can the two-dimensional image sequence be densely reconstructed. However, due to the above-mentioned pose and depth map errors and other factors, The accuracy and effect of 3D reconstruction are deteriorated.
发明内容Summary of the invention
本申请实施例提供了单目内窥镜的三维重建方法及装置,可以解决减小单目内窥镜固有参数造成成像畸变带来的误差、将二维图像序列进行三维重建出现精度不高和效果差的问题。The embodiments of the present application provide a method and device for three-dimensional reconstruction of a monocular endoscope, which can solve the problem of reducing errors caused by imaging distortion caused by the inherent parameters of the monocular endoscope, and the accuracy of the three-dimensional reconstruction of a two-dimensional image sequence is not high and The problem of poor results.
第一方面,本申请实施例提供了一种单目内窥镜图像的三维重建方法,包括:In the first aspect, an embodiment of the present application provides a three-dimensional reconstruction method of a monocular endoscopic image, including:
获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,对所述多张棋盘标定板的畸变图像进行畸变校正得到图像序列;Acquiring the distortion images of a plurality of checkerboard calibration boards taken by a monocular endoscope, and performing distortion correction on the distortion images of the checkerboard calibration boards to obtain an image sequence;
从所述图像序列中确定关键帧;Determining key frames from the image sequence;
获取所述关键帧的位姿参数,估算所述关键帧的深度图;Acquiring the pose parameters of the key frame, and estimating the depth map of the key frame;
基于所述关键帧的位姿参数以及所述关键帧的深度图进行图像重建,得到三维点云。Perform image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud.
可选地,所述基于所述关键帧的位姿参数以及所述关键帧的深度图进行图像重建,得到三维点云,包括:Optionally, the image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud includes:
获取所述关键帧的像素坐标;Acquiring the pixel coordinates of the key frame;
根据所述深度图、所述关键帧的位姿参数和所述关键帧的像素坐标计算得到目标空间坐标;Calculating the target space coordinates according to the depth map, the pose parameters of the key frame, and the pixel coordinates of the key frame;
获取所述关键帧中的每个像素点的颜色信息,根据所述关键帧中的每个像素点的颜色信息和所述目标空间坐标对所述关键帧进行点云融合,得到所述三维点云。Obtain the color information of each pixel in the key frame, and perform point cloud fusion on the key frame according to the color information of each pixel in the key frame and the target space coordinates to obtain the three-dimensional point cloud.
可选地,所述获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,对所述多张棋盘标定板的畸变图像进行校正得到图像序列,包括:Optionally, the acquiring the distortion images of a plurality of checkerboard calibration boards taken by a monocular endoscope, and correcting the distortion images of the checkerboard calibration boards to obtain an image sequence includes:
获取所述多张棋盘标定板的畸变图像中的棋盘的角点,基于所述棋盘的角点对所述单目内窥镜进行标定,得到所述单目内窥镜的相机参数和畸变参数;Acquire the corner points of the chessboard in the distortion images of the checkerboard calibration boards, calibrate the monocular endoscope based on the corner points of the chessboard, and obtain the camera parameters and distortion parameters of the monocular endoscope ;
根据所述相机参数和所述畸变参数从所述畸变图像中确定待校正图像;Determining an image to be corrected from the distorted image according to the camera parameter and the distortion parameter;
基于相机坐标系对所述待校正图像进行畸变校正得到所述图像序列。Performing distortion correction on the image to be corrected based on the camera coordinate system to obtain the image sequence.
可选地,所述基于相机坐标系对所述待校正图像进行畸变校正得到所述图像序列,包括:Optionally, the performing distortion correction on the image to be corrected based on the camera coordinate system to obtain the image sequence includes:
获取所述待校正图像的每个像素点在所述相机坐标中的预设坐标;Acquiring the preset coordinates of each pixel of the image to be corrected in the camera coordinates;
将所述相机坐标系投影至所述待校正图像的每个像素点所在的平面上,得到所述预设坐标在像素坐标系中的像素坐标;Projecting the camera coordinate system onto the plane where each pixel of the image to be corrected is located to obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
将所述预设坐标在像素坐标系中的像素坐标映射至所述相机坐标系得到所述图像序列。The pixel coordinates of the preset coordinates in the pixel coordinate system are mapped to the camera coordinate system to obtain the image sequence.
可选地,所述获取所述关键帧的像素坐标,包括:Optionally, the obtaining the pixel coordinates of the key frame includes:
将所述相机坐标系投影至所述待校正图像的每个像素点所在的平面上,得到所述预设坐标在像素坐标系中的像素坐标;Projecting the camera coordinate system onto the plane where each pixel of the image to be corrected is located to obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
将所述预设坐标在像素坐标系中的像素坐标映射至所述相机坐标系得到所述图像序列,以及所述图像序列对应的像素坐标;Mapping the pixel coordinates of the preset coordinates in the pixel coordinate system to the camera coordinate system to obtain the image sequence and the pixel coordinates corresponding to the image sequence;
基于所述图像序列对应的像素坐标得到所述关键帧的像素坐标。Obtain the pixel coordinates of the key frame based on the pixel coordinates corresponding to the image sequence.
可选地,所述从所述图像序列中确定关键帧,包括:Optionally, the determining a key frame from the image sequence includes:
获取所述图像序列中的各图像的局部特征,并基于所述各图像的局部特征对所述图像序列中的各图像进行特征点匹配,得到匹配结果;Acquiring local features of each image in the image sequence, and performing feature point matching on each image in the image sequence based on the local feature of each image, to obtain a matching result;
当所述匹配结果为第一图像和第二图像匹配的特征点数量大于或等于预设阈值时,将所述第一图像作为关键帧,其中,所述第一图像和所述第二图像为所述图像序列中相邻的任意两帧图像。When the matching result is that the number of feature points matched by the first image and the second image is greater than or equal to a preset threshold, the first image is used as a key frame, where the first image and the second image are Any two adjacent frames of images in the image sequence.
可选地,所述获取所述关键帧的位姿参数,包括:Optionally, the obtaining the pose parameters of the key frame includes:
将所述第一图像进行位姿初始化;Performing pose initialization on the first image;
估算所述图像序列中的关键帧的位姿参数。Estimate the pose parameters of the key frames in the image sequence.
可选地,所述估算所述关键帧的深度图,包括:Optionally, the estimating the depth map of the key frame includes:
从所述关键帧中确定参考帧图像,其中,所述参考帧图像为所述关键帧中的任一帧图像或多帧图像;Determining a reference frame image from the key frames, where the reference frame image is any frame image or multiple frames of images in the key frame;
基于所述位姿参数对所述参考帧图像的每个像素点进行深度估计处理,得到所述关键帧的深度图。Performing depth estimation processing on each pixel of the reference frame image based on the pose parameter to obtain the depth map of the key frame.
第二方面,本申请实施例提供了一种单目内窥镜图像的三维重建装置,包括:In the second aspect, an embodiment of the present application provides a three-dimensional reconstruction device for monocular endoscopic images, including:
获取模块,用于获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,对所述多张棋盘标定板的畸变图像进行畸变校正得到图像序列;The acquisition module is used to acquire the distortion images of a plurality of checkerboard calibration boards taken by a monocular endoscope, and perform distortion correction on the distortion images of the checkerboard calibration boards to obtain an image sequence;
确定模块,用于从所述图像序列中确定关键帧;A determining module for determining key frames from the image sequence;
计算模块,用于获取所述关键帧的位姿参数,估算所述关键帧的深度图;A calculation module for obtaining the pose parameters of the key frame and estimating the depth map of the key frame;
生成模块,用于基于所述关键帧的位姿参数以及所述关键帧的深度图进行图像重建,得到三维点云。The generating module is used for image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud.
第三方面,本申请实施例提供了一种终端设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现上述的三维重建方法。In the third aspect, an embodiment of the present application provides a terminal device, including a memory, a processor, and a computer program stored in the memory and running on the processor. When the processor executes the computer program, Realize the above-mentioned three-dimensional reconstruction method.
第四方面,本申请实施例提供了一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现上述的三维重建方法。In a fourth aspect, an embodiment of the present application provides a computer-readable storage medium that stores a computer program that implements the above-mentioned three-dimensional reconstruction method when the computer program is executed by a processor.
本申请实施例与现有技术相比存在的有益效果是:通过获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,对多张棋盘标定板的畸变图像进行畸变校正得到图像序列,从图像序列中确定关键帧,获取关键帧的位姿参数,估算关键帧的深度图,基于关键帧的位姿参数和关键帧的深度图进行图像重建得到三维点云。上述方法采用棋盘标定板图像可以实现对单目内窥镜进行标定和畸变校正得到图像序列,有效降低了单目内窥镜自身造成的成像畸变误差,从图像序列中确定符合要求的多张图像作为关键帧,并确定关键帧的位姿参数,可以避免光线变化等外界因素的干扰,可以精确地估计位姿参数和深度图,根据关键帧的位置参数和深度图进行图像重建,可以得到更精细的三维点云,也提高了图像的显示效果。Compared with the prior art, the embodiment of the present application has the following beneficial effects: by obtaining the distortion images of multiple checkerboard calibration boards taken by a monocular endoscope, and performing distortion correction on the distortion images of the multiple checkerboard calibration boards to obtain an image sequence, Determine the key frame from the image sequence, obtain the pose parameters of the key frame, estimate the depth map of the key frame, and reconstruct the image based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud. The above method uses the checkerboard calibration board image to achieve the calibration and distortion correction of the monocular endoscope to obtain the image sequence, which effectively reduces the imaging distortion error caused by the monocular endoscope itself, and determines multiple images that meet the requirements from the image sequence As a key frame, and determine the pose parameters of the key frame, it can avoid the interference of external factors such as light changes, can accurately estimate the pose parameters and depth map, and perform image reconstruction according to the position parameters and depth map of the key frame to get more The fine three-dimensional point cloud also improves the display effect of the image.
附图说明Description of the drawings
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly describe the technical solutions in the embodiments of the present application, the following will briefly introduce the drawings that need to be used in the description of the embodiments or the prior art. Obviously, the drawings in the following description are only of the present application. For some embodiments, for those of ordinary skill in the art, other drawings may be obtained based on these drawings without creative labor.
图1是本申请实施例提供的单目内窥镜图像的三维重建方法的流程示意图;FIG. 1 is a schematic flowchart of a method for three-dimensional reconstruction of monocular endoscopic images provided by an embodiment of the present application;
图2是本申请实施例提供的图像畸变校正的流程示意图;FIG. 2 is a schematic diagram of the image distortion correction process provided by an embodiment of the present application;
图3是本申请实施例提供的单目内窥镜图像的三维重建装置的结构示意图;3 is a schematic structural diagram of a three-dimensional reconstruction device for monocular endoscopic images provided by an embodiment of the present application;
图4是本申请实施例提供的终端设备的结构示意图。Fig. 4 is a schematic structural diagram of a terminal device provided by an embodiment of the present application.
具体实施方式Detailed ways
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本申请实施例。然而,本领域的技术人员应当清楚,在没有这些具体细节的其它实施例中也可以实现本申请。在其它情况中,省略对众所周知的系统、装置、电路以及方法的详细说明,以免不必要的细节妨碍本申请的描述。In the following description, for the purpose of illustration rather than limitation, specific details such as a specific system structure and technology are proposed for a thorough understanding of the embodiments of the present application. However, it should be clear to those skilled in the art that the present application can also be implemented in other embodiments without these specific details. In other cases, detailed descriptions of well-known systems, devices, circuits, and methods are omitted to avoid unnecessary details from obstructing the description of this application.
应当理解,当在本申请说明书和所附权利要求书中使用时,术语“包括”指示所描述特征、整体、步骤、操作、元素和/或组件的存在,但并不排除一个或多个其它特征、整体、步骤、操作、元素、组件和/或其集合的存在或添加。It should be understood that when used in the specification and appended claims of this application, the term "comprising" indicates the existence of the described features, wholes, steps, operations, elements and/or components, but does not exclude one or more other The existence or addition of features, wholes, steps, operations, elements, components, and/or collections thereof.
还应当理解,在本申请说明书和所附权利要求书中使用的术语“和/或”是指相关联列出的项中的一个或多个的任何组合以及所有可能组合,并且包括这些组合。It should also be understood that the term "and/or" used in the specification and appended claims of this application refers to any combination of one or more of the items listed in the associated and all possible combinations, and includes these combinations.
如在本申请说明书和所附权利要求书中所使用的那样,术语“如果”可以依据上下文被解释为“当...时”或“一旦”或“响应于确定”或“响应于检测到”。类似地,短语“如果确定”或“如果检测到[所描述条件或事件]”可以依据上下文被解释为意指“一旦确定”或“响应于确定”或“一旦检测到[所描述条件或事件]”或“响应于检测到[所描述条件或事件]”。As used in the description of this application and the appended claims, the term "if" can be construed as "when" or "once" or "in response to determination" or "in response to detecting ". Similarly, the phrase "if determined" or "if detected [described condition or event]" can be interpreted as meaning "once determined" or "in response to determination" or "once detected [described condition or event]" depending on the context ]" or "in response to detection of [condition or event described]".
另外,在本申请说明书和所附权利要求书的描述中,术语“第一”、“第二”、“第三”等仅用于区分描述,而不能理解为指示或暗示相对重要性。In addition, in the description of the specification of this application and the appended claims, the terms "first", "second", "third", etc. are only used to distinguish the description, and cannot be understood as indicating or implying relative importance.
在本申请说明书中描述的参考“一个实施例”或“一些实施例”等意味着在本申请的一个或多个实施例中包括结合该实施例描述的特定特征、结构或特点。由此,在本说明书中的不同之处出现的语句“在一个实施例中”、“在一些实施例中”、“在其他一些实施例中”、“在另外一些实施例中”等不是必然都参考相同的实施例,而是意味着“一个或多个但不是所有的实施例”,除非是以其他方式另外特别强调。术语“包括”、“包含”、“具有”及它们的变形都意味着“包括但不限于”,除非是以其他方式另外特别强调。The reference to "one embodiment" or "some embodiments" described in the specification of this application means that one or more embodiments of this application include a specific feature, structure, or characteristic described in combination with the embodiment. Therefore, the sentences "in one embodiment", "in some embodiments", "in some other embodiments", "in some other embodiments", etc. appearing in different places in this specification are not necessarily All refer to the same embodiment, but mean "one or more but not all embodiments" unless it is specifically emphasized otherwise. The terms "including", "including", "having" and their variations all mean "including but not limited to", unless otherwise specifically emphasized.
图1示出了本申请一提供的单目内窥镜图像的三维重建方法的流程示意图,如图1,本申请提供的单目内窥镜图像的三维重建方法,所述三维重建方法包括S101至S104,具体如下:FIG. 1 shows a schematic flow chart of a three-dimensional reconstruction method of a monocular endoscopic image provided by the first application of the present application. As shown in FIG. 1, the three-dimensional reconstruction method of a monocular endoscope image provided by the present application, the three-dimensional reconstruction method includes S101 To S104, the details are as follows:
S101:获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,对所述多张棋盘标定板的畸 变图像进行畸变校正得到图像序列;S101: Obtain distortion images of multiple checkerboard calibration boards taken by a monocular endoscope, and perform distortion correction on the distortion images of the multiple checkerboard calibration boards to obtain an image sequence;
在本实施例中,棋盘标定板的畸变图像可以用于单目内窥镜畸变校正的棋盘标定板是双排黑白条纹间隔的二值化图片,单目内窥镜可以从不同角度观测该标定板得到多张单目内窥镜畸变图像。摄像机的成像过程主要涉及到图像像素坐标系、图像物理坐标系、摄像机坐标系和世界坐标系之间进行变换,由于透镜成像原理出现摄像机成像畸变,畸变校正是找到畸变前后的点位置的对应关系。In this embodiment, the distortion image of the checkerboard calibration board can be used for distortion correction of the monocular endoscope. The checkerboard calibration board is a binary image with two rows of black and white stripe intervals. The monocular endoscope can observe the calibration from different angles. The board obtains multiple monocular endoscope distortion images. The imaging process of the camera mainly involves the transformation between the image pixel coordinate system, the image physical coordinate system, the camera coordinate system and the world coordinate system. Due to the lens imaging principle, the camera imaging distortion occurs, and the distortion correction is to find the corresponding relationship between the point positions before and after the distortion. .
需要说明的是,单目内窥镜的成像模型不同于小孔成像模型,而是更贴近于鱼眼摄像头模型。棋盘标定板是按照间隔排列的黑白格,也称棋盘格标定板,标定板(Calibration Target)在机器视觉、图像测量、摄影测量、三维重建等应用中,为校正镜头畸变,确定物理尺寸和像素间的换算关系,以及确定空间物体表面某点的三维几何位置与其在图像中对应点之间的相互关系,需要建立相机成像的几何模型。通过相机拍摄带有固定间距图案阵列平板、经过标定算法的计算,可以得出相机的几何模型,从而得到高精度的测量和重建结果。而带有固定间距图案阵列的平板就是标定板。It should be noted that the imaging model of the monocular endoscope is different from the small hole imaging model, but is closer to the fisheye camera model. The checkerboard calibration board is a black and white grid arranged at intervals, also called a checkerboard calibration board. The calibration target is used in machine vision, image measurement, photogrammetry, three-dimensional reconstruction and other applications to correct lens distortion and determine the physical size and pixels The conversion relationship between the two, and to determine the relationship between the three-dimensional geometric position of a point on the surface of a space object and its corresponding point in the image requires the establishment of a geometric model of camera imaging. The camera's geometric model can be obtained by the camera with a fixed-pitch pattern array plate and calculation by the calibration algorithm, thereby obtaining high-precision measurement and reconstruction results. The flat plate with a fixed-pitch pattern array is the calibration plate.
应理解,通过获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,可以实现对单目内窥镜进行相机标定,并根据标定单目内窥镜可以对畸变图像进行畸变校正得到图像序列,即真实图像,可以降低图像畸变给图像识别带来的误差。It should be understood that by acquiring the distortion images of multiple checkerboard calibration boards taken by the monocular endoscope, the camera calibration of the monocular endoscope can be realized, and the distorted image can be corrected to obtain the image according to the calibration monocular endoscope. Sequence, that is, real image, can reduce the error caused by image distortion to image recognition.
图2示出了本申请提供的畸变校正的实现流程图,如图2,所述获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,对所述多张棋盘标定板的畸变图像进行校正得到图像序列,包括S1011至S1013:Figure 2 shows a flow chart of the implementation of distortion correction provided by the present application. As shown in Figure 2, the acquisition of the distortion images of the multiple checkerboard calibration boards taken by the monocular endoscope, and the distortion images of the multiple checkerboard calibration boards Perform correction to obtain an image sequence, including S1011 to S1013:
S1011:获取所述多张棋盘标定板的畸变图像中的棋盘的角点,基于所述棋盘的角点对所述单目内窥镜进行标定,得到所述单目内窥镜的相机参数和畸变参数;S1011: Obtain the corner points of the chessboard in the distortion images of the multiple checkerboard calibration boards, and calibrate the monocular endoscope based on the corner points of the chessboard to obtain the camera parameters of the monocular endoscope and Distortion parameter
在本实施例中,获取20张左右不同角度下用单目内窥镜拍摄的带有棋盘标定板的图像,提取图像中的棋盘的角点,选择满足拟合条件的畸变图像。可以采用Canny角点算子检测所有单目内窥镜观测棋盘标定板得到的畸变图像,统计所有畸变图像中的角点数目,满足拟合条件的畸变图像优选为该图像中检测得到的角点数目不小于6个。其中,角点的数目可以根据实际情况选择,此处不作具体限定。In this embodiment, 20 images with a checkerboard calibration board taken with a monocular endoscope at about different angles are acquired, the corner points of the checkerboard in the image are extracted, and the distorted image that meets the fitting conditions is selected. The Canny corner operator can be used to detect the distortion images obtained from all monocular endoscopes observing the checkerboard calibration board, and count the number of corner points in all the distortion images. The distortion image that meets the fitting conditions is preferably the number of corner points detected in the image No less than 6 meshes. Among them, the number of corner points can be selected according to actual conditions, and is not specifically limited here.
具体的,根据选择的畸变图像和检测到的角点拟合得到椭圆方程的参数,椭圆方程可以为包括6个参数的标准方程,根据检测得到的畸变图像中的角点,采用最小二乘法得到曲面投影参数得到该椭圆方程参数,多张畸变图像的椭圆方程的参数拟合结果采用均值滤波得到平均值。构建曲面投影模型,建立椭圆方程的参数,即便图像像素点坐标和真实图像像素点坐标之间的对应关系,之后构建曲面投影模型,根据曲面投影原则建立曲面模型参数,得到 畸变图像点坐标与真实图像点坐标之间的对应关系,以此对单目内窥镜进行标定,通过对单目内窥镜进行标定可以得到单目内窥镜的相机参数和畸变参数。通过标定可得到单目内窥镜的内参矩阵K与畸变参数矩阵(k 1 k 2 k 3 k 4)其中K可表示为: Specifically, the parameters of the ellipse equation are obtained by fitting the selected distortion image and the detected corner points. The ellipse equation can be a standard equation including 6 parameters. According to the corner points in the detected distortion image, the least square method is used to obtain the parameters. The parameters of the ellipse equation are obtained from the curved surface projection parameters, and the parameter fitting results of the ellipse equations of multiple distorted images are obtained by means of filtering. Construct a curved surface projection model, establish the parameters of the ellipse equation, even if the corresponding relationship between the image pixel coordinates and the real image pixel coordinates, then construct the curved surface projection model, and establish the curved surface model parameters according to the curved surface projection principle to obtain the distorted image point coordinates and the real The corresponding relationship between the coordinates of the image points is used to calibrate the monocular endoscope, and the camera parameters and distortion parameters of the monocular endoscope can be obtained by calibrating the monocular endoscope. Through calibration, the internal parameter matrix K and the distortion parameter matrix (k 1 k 2 k 3 k 4 ) of the monocular endoscope can be obtained, where K can be expressed as:
Figure PCTCN2020129546-appb-000001
其中fx与fy是内窥镜以像素为单位的焦距,cx与cy是以像素为单位的主点位置(即成像的中心像素位置)。
Figure PCTCN2020129546-appb-000001
Where fx and fy are the focal lengths of the endoscope in pixels, and cx and cy are the principal point positions in pixels (that is, the center pixel position of the imaging).
需要说明的是,棋盘是一块由黑白方块间隔组成的标定板作为相机标定的标定物(从真实世界映射到数字图像内的对象)。二维物体相对于三维物体会缺少一部分信息,采用棋盘作为标定物是因为平面棋盘模式更容易处理,经过多次改变棋盘的方位来捕捉图像,以此获得更丰富的坐标信息。It should be noted that the chessboard is a calibration board composed of black and white squares as the calibration object for camera calibration (mapped from the real world to the object in the digital image). Compared with a three-dimensional object, a two-dimensional object lacks some information. The checkerboard is used as a calibration object because the plane checkerboard mode is easier to handle. After changing the position of the checkerboard many times to capture the image, it can obtain richer coordinate information.
S1012:根据所述相机参数和所述畸变参数从所述畸变图像中确定待校正图像;S1012: Determine an image to be corrected from the distorted image according to the camera parameter and the distortion parameter;
在本实施例中,对单目内窥镜进行标定可以确定相机的位姿,以获得单目内窥镜的相机参数和畸变参数,以一副图像为例,通过相机参数和畸变参数计算出该图像是否没有发生畸变得到待校正畸变图像,即可以对拍摄的多张图像进行判断是否发生畸变,或者设定预设阈值,将计算结果与预设阈值比较得到比较结果,其中,将比较结果中相差较大的作为有畸变和比较结果中相差不大的作为无畸变,反之亦然。In this embodiment, calibrating the monocular endoscope can determine the pose of the camera to obtain the camera parameters and distortion parameters of the monocular endoscope. Taking an image as an example, the camera parameters and the distortion parameters are calculated Whether the image has not been distorted to the distorted image to be corrected, that is, it can judge whether the captured multiple images are distorted, or set a preset threshold, and compare the calculation result with the preset threshold to obtain a comparison result, where the comparison result is The ones with greater difference in the comparison result are regarded as distortion and those with little difference in the comparison result are regarded as no distortion, and vice versa.
需要说明的是,在图像的获取或显示过程中往往会产生各种失真(畸变),常见的有几何形状失真、灰度失真、颜色失真,引起图像失真的原因有成像系统的象差、畸变、带宽有限、拍摄状态、扫描非线性、相对运动等,非均匀光照条件或点光源照明等。根据相机参数和畸变参数对拍摄的多张图像中确定待校正图像,便于排除畸变对图像识别和处理的误差,一定程度上提高了图像处理的精度。It should be noted that various distortions (distortion) are often produced in the process of image acquisition or display. The common ones are geometric shape distortion, grayscale distortion, and color distortion. The causes of image distortion are the aberration and distortion of the imaging system. , Limited bandwidth, shooting status, scanning nonlinearity, relative motion, etc., non-uniform lighting conditions or point light source lighting, etc. According to the camera parameters and distortion parameters, the image to be corrected is determined from the multiple images taken, which is convenient to eliminate the error of distortion in image recognition and processing, and improves the accuracy of image processing to a certain extent.
S1013:基于所述相机坐标系对所述待校正图像进行畸变校正得到所述图像序列。S1013: Perform distortion correction on the image to be corrected based on the camera coordinate system to obtain the image sequence.
在本实施例中,畸变空间中的直线在像空间中一般不再是直线,而只有通过对称中心的直线是例外,在进行畸变校正时可以找出对称中心,再进行通用的几何畸变校正过程。畸变校正的一般步骤为先找出畸变图对称中心,将畸变图代表的地址空间关系转换为以对称中心为原点的空间坐标系,接着空间变换,对输入图像即畸变图上像素重新排列以恢复原空间关系,即利用地址映射关系为校正图空间上的每一个点找到它们在畸变图空间上的对应点,最后灰度差值即对空间变换后的像素赋予相应的灰度值以恢复原位置的灰度值。几何畸变的校正需要使用坐标转换,包括平行移动、旋转、扩大缩小等简单的变换。In this embodiment, the straight line in the distortion space is generally no longer a straight line in the image space, but only the straight line passing through the center of symmetry is an exception. When performing distortion correction, the center of symmetry can be found, and then the general geometric distortion correction process can be performed. . The general steps of distortion correction are to first find the symmetry center of the distortion map, convert the address space relationship represented by the distortion map into a space coordinate system with the center of symmetry as the origin, and then space transformation, rearrange the pixels on the input image, that is, the distortion map to restore The original spatial relationship, that is, use the address mapping relationship to find their corresponding point in the distortion map space for each point in the correction map space, and finally the grayscale difference is to assign the corresponding grayscale value to the pixel after the space transformation to restore the original The gray value of the location. The correction of geometric distortion requires the use of coordinate transformation, including simple transformations such as parallel movement, rotation, enlargement and reduction.
需要说明的是,畸变校正的过程可以理解成将有畸变的图像处理成无畸变的图像即真实图像,不同的摄像头模型在拍照时图像的显示是不同的,可能发生畸变或不发生畸变,相应的采用畸变校正过程可以相同,也可以不同。图像的畸变主要有径向畸变和切向畸变,径向畸变指的是正中心位置的畸变最小,随着半径的增大畸变增大,径向畸变可以分为枕形畸变和桶形畸变。切向畸变指的是在透镜与成像平面不平行时产生,类似于透视变换。通过对待校正图像进行畸变校正得到校正后的图像序列,一定程度上确保了图像处理的可靠性。It should be noted that the process of distortion correction can be understood as processing a distorted image into an undistorted image, that is, a real image. Different camera models display different images when taking pictures, and may or may not be distorted. The process of using distortion correction can be the same or different. Image distortion mainly includes radial distortion and tangential distortion. Radial distortion refers to the smallest distortion at the center position. As the radius increases, the distortion increases. Radial distortion can be divided into pincushion distortion and barrel distortion. Tangential distortion refers to when the lens is not parallel to the imaging plane, similar to perspective transformation. By performing distortion correction on the image to be corrected to obtain a corrected image sequence, the reliability of image processing is ensured to a certain extent.
可选地,所述基于所述相机坐标系对所述待校正图像进行畸变校正得到所述图像序列,包括步骤A1~A3:Optionally, the performing distortion correction on the image to be corrected based on the camera coordinate system to obtain the image sequence includes steps A1 to A3:
步骤A1:获取所述待校正图像的每个像素点在所述相机坐标中的预设坐标;Step A1: Obtain the preset coordinates of each pixel of the image to be corrected in the camera coordinates;
在本实施例中,在对单目内窥镜进行标定可以得到相机坐标系,根据相机成像模型和相机坐标系可以实现世界坐标系与相机坐标系、相机坐标系与图像坐标系、以及图像坐标系到像素坐标系的转换。世界坐标系与相机坐标系之间转换是从一个三维坐标到另一个三维坐标系,可以通过旋转矩阵和平移向量得到相机的位姿参数,即相机坐标系。相机坐标系到图像坐标系是将一个三维坐标投影在一个二维平面上,根据两个坐标系距离即相机焦距进行估算。换言之,将相机坐标系中的预设坐标进行校正得到无畸变的相机坐标系下的坐标,将无畸变的相机坐标系下的坐标映射至像素坐标系中,得到无畸变的图像序列。In this embodiment, the camera coordinate system can be obtained by calibrating the monocular endoscope. According to the camera imaging model and the camera coordinate system, the world coordinate system and the camera coordinate system, the camera coordinate system and the image coordinate system, and the image coordinates can be realized. System to pixel coordinate system conversion. The conversion between the world coordinate system and the camera coordinate system is from one three-dimensional coordinate to another three-dimensional coordinate system. The pose parameters of the camera can be obtained through the rotation matrix and the translation vector, that is, the camera coordinate system. From the camera coordinate system to the image coordinate system, a three-dimensional coordinate is projected on a two-dimensional plane, and it is estimated based on the distance between the two coordinate systems, that is, the focal length of the camera. In other words, the preset coordinates in the camera coordinate system are corrected to obtain the coordinates in the undistorted camera coordinate system, and the coordinates in the undistorted camera coordinate system are mapped to the pixel coordinate system to obtain an undistorted image sequence.
步骤A2:将所述相机坐标系投影至所述待校正图像的每个像素点所在的平面上,得到所述预设坐标在像素坐标系中的像素坐标;Step A2: Project the camera coordinate system onto the plane where each pixel of the image to be corrected is located, and obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
在本实施例中,假设单目内窥镜拍摄的图像的像素点(u′,v′)在相机坐标系中的坐标为(x,y,z),将像素点在相机坐标系中的坐标投影到图像所在的平面即图像坐标系,根据图像坐标系原点在相对于像素坐标系原点的位置关系,可以看成将像素点在相机坐标系中的坐标投影至像素坐标系,可以表示如下:In this embodiment, it is assumed that the coordinates of the pixel (u', v') of the image taken by the monocular endoscope in the camera coordinate system are (x, y, z), and the pixel point in the camera coordinate system is The coordinates are projected to the plane where the image is located, that is, the image coordinate system. According to the positional relationship of the origin of the image coordinate system relative to the origin of the pixel coordinate system, it can be regarded as projecting the coordinates of the pixel point in the camera coordinate system to the pixel coordinate system, which can be expressed as follows :
x′=x/z,y′=y/z,r 2=x′ 2+y′ 2 x′=x/z,y′=y/z,r 2 =x′ 2 +y′ 2
θ=arctan(r)θ=arctan(r)
θ′=θ(1+k 1θ 2+k 2θ 4+k 3θ 6+k 4θ 8) θ′=θ(1+k 1 θ 2 +k 2 θ 4 +k 3 θ 6 +k 4 θ 8 )
x′=(θ′/r)x,y′=(θ′/r)yx′=(θ′/r)x,y′=(θ′/r)y
u=f xx′+c x u=f x x′+c x
v=f yy′+c y v=f y y′+c y
其中,(x′,y′)是投影在平面上的坐标,r则表示投影平面上该点与中心的距离(投影半径),θ表示入射角。通过上述公式可以确定相机坐标系和图像坐标系之间的对应关系,这 样便于后续确定像素坐标系及像素坐标。Among them, (x', y') are the coordinates projected on the plane, r represents the distance between the point and the center on the projection plane (projection radius), and θ represents the angle of incidence. The corresponding relationship between the camera coordinate system and the image coordinate system can be determined through the above formula, which facilitates the subsequent determination of the pixel coordinate system and pixel coordinates.
步骤A3:将所述预设坐标在像素坐标系中的像素坐标映射至所述相机坐标系得到所述图像序列。Step A3: Map the pixel coordinates of the preset coordinates in the pixel coordinate system to the camera coordinate system to obtain the image sequence.
在本实施例中,对于N张无畸变的图像来说,共有4个内参+6N个外参来标定,每张棋盘图上有4个有效的角点,可以提供8个约束,则需要8N>=4+6N,则至少需要2张无畸变的图像可以求出相机的内参和外参,实际上一般可以取10张或者20张,从而利用最小二乘法得到更精确的解,在求出了内参和外参后,即可根据剩余的点坐标求出畸变相关参数。In this embodiment, for N undistorted images, there are a total of 4 internal parameters + 6N external parameters to calibrate. There are 4 effective corner points on each chessboard graph, and 8 constraints can be provided, so 8N is required. >=4+6N, at least 2 undistorted images are needed to obtain the internal and external parameters of the camera. In fact, 10 or 20 images can be generally used to obtain a more accurate solution by using the least square method. After the internal and external parameters are added, the distortion-related parameters can be obtained according to the remaining point coordinates.
需要说明的是,在对待校正图像进行畸变校正可能出现至少两张图像呈线性关系,可以采用重映射的过程,将有畸变内窥镜图像像素坐标转化为有畸变的相机坐标系坐标,再将有畸变的相机坐标系坐标转化为无畸变的相机坐标系坐标,最后将无畸变的相机坐标系坐标转化为无畸变图像的像素坐标,以此得到校正后的图像序列,也可以得到相应的图像的像素坐标,便于后续确定关键帧及关键帧的位姿参数。It should be noted that when performing distortion correction on the image to be corrected, at least two images may have a linear relationship. A remapping process can be used to convert the pixel coordinates of the distorted endoscopic image into the coordinates of the distorted camera coordinate system. The distorted camera coordinate system coordinates are transformed into the undistorted camera coordinate system coordinates, and finally the undistorted camera coordinate system coordinates are transformed into the pixel coordinates of the undistorted image, so that the corrected image sequence can be obtained, and the corresponding image can also be obtained The pixel coordinates of, it is convenient to determine the key frame and the pose parameters of the key frame later.
S102:从所述图像序列中确定关键帧;S102: Determine a key frame from the image sequence;
在本实施例中,将畸变校正后的图像作为图像序列,从中确定关键帧。ORB_SLAM2是一种嵌入式位置识别模型,具有重定位,防止跟踪失败(如遮挡)、已建图场景重新初始化、回环检测等特点,使用相同的ORB特征进行跟踪、建图和位置识别任务,这些特征在旋转和尺度上有良好的鲁棒性,对摄像机的自动增益和自动曝光以及光照变化具有良好的不变性,它还能够迅速提取特征和匹配特征,满足实时操作的需求。本申请采用ORB_SLAM2对单目内窥镜图像进行关键帧的判定和位姿估计,可以采用ORB对图像序列进行特征提取、通过前一图像帧估计相机的初始位姿、通过全局重定位来初始化位姿、跟踪局部地图和新关键帧的判断标准这四个过程,以更精确地确定关键帧和关键帧的位姿参数。In this embodiment, the image after distortion correction is used as an image sequence, from which key frames are determined. ORB_SLAM2 is an embedded position recognition model, which has the characteristics of relocation, preventing tracking failure (such as occlusion), reinitialization of the mapped scene, loop detection, etc., using the same ORB features for tracking, mapping, and location recognition tasks. Features have good robustness in rotation and scale, and have good invariance to the camera's automatic gain, automatic exposure, and illumination changes. It can also quickly extract features and match features to meet the needs of real-time operation. This application uses ORB_SLAM2 to determine the key frame and pose estimation of monocular endoscopic images. You can use ORB to extract features of the image sequence, estimate the initial pose of the camera through the previous image frame, and initialize the position through global relocation. The four processes of posture, tracking the local map and the judgment standard of the new key frame are used to determine the key frame and the posture parameter of the key frame more accurately.
需要说明的是,关键帧可以作为图像序列的标记,具有引导作用,将图像序列中经过畸变校正的图像按照预设顺序进行排列,可以按照拍摄时间顺序依次排列,这样便于对各图像进行特征提取处理,提高单目内窥镜图像处理的效率。It should be noted that the key frame can be used as a mark of the image sequence and has a guiding effect. The distortion-corrected images in the image sequence are arranged in a preset order, and they can be arranged in sequence according to the shooting time order, which is convenient for feature extraction of each image. Processing to improve the efficiency of monocular endoscope image processing.
可选地,所述从所述图像序列中确定关键帧,包括步骤B1~B2,具体如下:Optionally, the determining the key frame from the image sequence includes steps B1 to B2, which are specifically as follows:
步骤B1:获取所述图像序列中的各图像的局部特征,并基于所述各图像的局部特征对所述图像序列中的各图像进行特征点匹配,得到匹配结果;Step B1: Obtain local features of each image in the image sequence, and perform feature point matching on each image in the image sequence based on the local feature of each image to obtain a matching result;
在本实施例中,提取图像序列中的各图像的局部特征,并对各图像的局部特征对图像序列中的各图像进行特征点匹配,以提取各图像的坐标对应的区域进行特征匹配,或者提取图像丰富区域的所有像素点,按照预设顺序将前后两帧图像进行特征点匹配,即两帧图像中相同的ORB特征匹配成功的特征点数量作为匹配结果,设定匹配成功的特征点数量的阈值在 50~100之间。In this embodiment, extract the local features of each image in the image sequence, and perform feature point matching on each image in the image sequence with the local features of each image to extract the regions corresponding to the coordinates of each image for feature matching, or Extract all pixels in the rich area of the image, and match the feature points of the two images before and after in the preset order, that is, the number of feature points that are successfully matched with the same ORB feature in the two frames of the image is used as the matching result, and set the number of feature points that are successfully matched The threshold is between 50-100.
需要说明的是,单目内窥镜成像的图像四周边缘区域为黑色无信息的区域,不能提取有用的特征信息,因此选取图像中信息丰富的区域,将该区域可以定义为感兴趣区域,以提取该区域的ORB特征,ORB(Oriented FAST and Rotated BRIEF)是一种快速特征点提取和描述的算法,ORB算法包括特征点提取和特征点描述,ORB算法具有计算速度快的特点,使用FAST检测特征点,再次是使用BRIEF算法计算描述子,该描述子特有的二进制串的表现形式不仅节约了存储空间,而且大大缩短了匹配的时间。It should be noted that the peripheral edge area of the image imaged by the monocular endoscope is a black area with no information, and useful feature information cannot be extracted. Therefore, the information-rich area in the image is selected and the area can be defined as a region of interest. To extract the ORB features of the region, ORB (Oriented FAST and Rotated Brief) is a fast feature point extraction and description algorithm. The ORB algorithm includes feature point extraction and feature point description. The ORB algorithm has the characteristics of fast calculation speed and uses FAST detection The feature point, again, is to use the BRIEF algorithm to calculate the descriptor. The unique binary string representation of the descriptor not only saves storage space, but also greatly shortens the matching time.
应理解,通过从渐变校正后的图像序列中确定关键帧,可以关键帧作为标记对该图像序列进行快速处理,可以提高单目内窥镜图像处理的效率。It should be understood that by determining the key frame from the gradually corrected image sequence, the key frame can be used as a marker to quickly process the image sequence, which can improve the efficiency of monocular endoscope image processing.
步骤B2:当所述匹配结果为第一图像和第二图像匹配的特征点数量大于或等于预设阈值时,将所述第一图像作为关键帧,其中,所述第一图像和所述第二图像为所述图像序列中相邻的任意两帧图像。Step B2: When the matching result is that the number of feature points matched by the first image and the second image is greater than or equal to a preset threshold, use the first image as a key frame, wherein the first image and the first image The two images are any two adjacent frames of images in the image sequence.
在本实施例中,设定匹配成功的特征点数量的阈值在50~100之间,当第一图像和第二图像匹配的特征点数量超过该阈值,则确定首次有前后两帧图像匹配成功。In this embodiment, the threshold for the number of feature points that are successfully matched is set to be between 50 and 100. When the number of feature points matched by the first image and the second image exceeds the threshold, it is determined that the first and the next two frames of images are matched successfully .
需要说明的是,如果上一帧图像跟踪成功,可以用运动速率恒定模型来预测当前相机的位置(即认为摄像头处于匀速运动),然后搜索上一帧图像中的特征点在地图中的对应云点与当前帧图像的匹配点,最后利用搜索到的匹配点对当前相机的位姿进一步优化,从而获得符合要求的图像序列中的图像,以提高判定关键帧的准确性。It should be noted that if the previous frame image is successfully tracked, the constant motion rate model can be used to predict the current camera position (that is, the camera is considered to be moving at a constant speed), and then the corresponding cloud of the feature points in the previous frame image can be searched for in the map The matching point between the point and the current frame image is finally used to further optimize the pose of the current camera by using the searched matching point, so as to obtain the image in the image sequence that meets the requirements to improve the accuracy of determining the key frame.
S103:获取所述关键帧的位姿参数,估算所述关键帧的深度图;S103: Obtain the pose parameters of the key frame, and estimate the depth map of the key frame;
在本实施例中,基于特征点法的ORB_SLAM2可以得到位姿参数,对于有相对位姿参数的图像序列,即两张图像之间存在线性关系,位姿参数描述的是两张图像对应相机的相对移动关系。深度图是指存储每个像素使用的位数,用于量度图像的色彩分辨率,In this embodiment, ORB_SLAM2 based on the feature point method can obtain the pose parameters. For image sequences with relative pose parameters, that is, there is a linear relationship between the two images, the pose parameters describe the two images corresponding to the camera. Relative movement relationship. The depth map refers to the number of bits used to store each pixel, which is used to measure the color resolution of the image.
可选地,获取所述关键帧的位姿参数包括:Optionally, obtaining the pose parameters of the key frame includes:
将所述第一图像进行位姿初始化;估算所述图像序列中的关键帧的位姿参数。Performing pose initialization on the first image; estimating pose parameters of key frames in the image sequence.
在本实施例中,当第一图像和第二图像匹配的特征点数量超过设定的阈值时,将第一图像即前一帧图像的位姿初始化为(R0,t0),该关键帧包含多张特征点匹配成功的图像,根据第一图像的位姿初始化对关键帧中的每一帧图像提取ORB特征与前一帧进行特征匹配并估算其位姿参数(旋转矩阵Ri,平移向量ti),将位姿估计成功的图像作为关键帧,得到关键帧对应的位姿参数,将关键帧和关键帧对应的位姿参数一起存储,以便后续对所有关键帧进行深度估计。In this embodiment, when the number of feature points matched by the first image and the second image exceeds the set threshold, the pose of the first image, that is, the image of the previous frame, is initialized to (R0, t0), and the key frame contains For multiple images with successful feature point matching, extract the ORB feature of each frame of the key frame according to the pose initialization of the first image, perform feature matching with the previous frame, and estimate its pose parameters (rotation matrix Ri, translation vector ti ), take the image of successful pose estimation as the key frame, obtain the pose parameter corresponding to the key frame, and store the key frame and the pose parameter corresponding to the key frame together, so as to perform depth estimation on all the key frames later.
应理解,根据第一图像的位姿初始化对关键帧中的其他图像也采用上述特征点匹配的过 程进行判断,并估算当前图像的位姿参数,将特征点匹配成功的图像作为关键帧,并根据初始化的位姿参数得到图像序列中的各图像的位姿参数即关键帧的位姿参数,这样提高了位姿参数估计的精确度。It should be understood that, according to the pose initialization of the first image, other images in the key frame are also judged by the above-mentioned feature point matching process, and the pose parameters of the current image are estimated, and the image whose feature points are successfully matched is used as the key frame, and According to the initialized pose parameters, the pose parameters of each image in the image sequence, that is, the pose parameters of the key frame are obtained, which improves the accuracy of the estimation of the pose parameters.
可选地,所述估算所述关键帧的深度图,包括:Optionally, the estimating the depth map of the key frame includes:
从所述关键帧中确定参考帧图像,其中,所述参考帧图像为所述关键帧中的任一帧图像或多帧图像;Determining a reference frame image from the key frames, where the reference frame image is any frame image or multiple frames of images in the key frame;
在本实施例中,根据单目视频中关键图像帧的第一深度图将光度误差进行最小化,确定单目内窥镜图像中的参考帧图像与关键帧之间的当前相机位姿.根据当前相机位姿对参考帧图像和关键帧中的高梯度像索点进行三角测量,确定关键帧的第二深度图,将第一深度图和第二深度图进行高斯融合,更新关键帧的第一深度图,若参考帧图像的后一图像帧与关键帧之间的后一相机位姿超过预设相机位姿,则将更新后的第一深度图确定为关键帧的稠密深度图。In this embodiment, the photometric error is minimized according to the first depth map of the key image frame in the monocular video, and the current camera pose between the reference frame image and the key frame in the monocular endoscopic image is determined. According to The current camera pose triangulates the high gradient image points in the reference frame image and the key frame, determines the second depth map of the key frame, performs Gaussian fusion of the first depth map and the second depth map, and updates the first depth map of the key frame. A depth map. If the next camera pose between the next image frame and the key frame of the reference frame image exceeds the preset camera pose, the updated first depth map is determined as the dense depth map of the key frame.
需要说明的是,进行深度图估算可以选取一帧图像或者多帧图像进行估算,当选取关键帧中的一帧图像作为参考帧时,对关键帧中的各图像的每个像素点进行三角测量和贝叶斯概率估计策略得到稠密深度图。当选取关键帧中的多帧图像进行迭代计算得到每个像素点对应的深度值,之后对深度图进行平滑滤波处理,以消除一些深度图中的噪声,可以提高深度估计的效率和精确度。It should be noted that for depth map estimation, one frame of image or multiple frames of images can be selected for estimation. When one frame of image in the key frame is selected as the reference frame, each pixel of each image in the key frame is triangulated. And Bayesian probability estimation strategy to get dense depth map. When multiple images in the key frame are selected for iterative calculation to obtain the depth value corresponding to each pixel, then the depth map is smoothed and filtered to eliminate some noise in the depth map, which can improve the efficiency and accuracy of depth estimation.
基于所述位姿参数对所述参考帧图像的每个像素点进行深度估计处理,得到所述关键帧的深度图。Performing depth estimation processing on each pixel of the reference frame image based on the pose parameter to obtain the depth map of the key frame.
在本实施例中,关键帧的第一深度图可以是通过初始化关键帧中高梯度点的深度值得到的服从于高斯分布的稠密深度图,也可以是将该关键帧的前一关键帧的深度值根据相机位姿进行投影后的稠密深度图。例如,若待深度估计的关键帧为图像序列中的第一关键帧,则该关键帧的第一深度图为通过初始化得到的稠密深度图;若带深度估计的关键帧为图像序列中除第一关键帧之外的其他关键帧,则该关键帧的第一深度图为将前一关键帧的深度值进行投影后的稠密深度图。光度误差是指投影图像中的高梯度点与参考帧图像中对应的高梯度点之间的量度差,投影图像是按照图像序列中的参考帧与关键帧之间的初始相机位姿,将关键帧中的像素点对应的高梯度点投影至参考帧图像后得到的图像,当前相机位姿包括参考帧与关键帧之间的旋转和平移,关键帧的第二深度图是指根据图像序列中的参考帧图像与关键帧之间的当前相机位姿进行三角测量得到的新的稠密深度图;参考帧图像的后一帧图像是指在图像序列中预参考帧图像相邻的后一帧图像,后一相机位姿包括后一相机位姿的最大阈值,可以根据实际情况和需求预先设定,此处不作具体限定。In this embodiment, the first depth map of the key frame can be a dense depth map obeying the Gaussian distribution obtained by initializing the depth values of the high gradient points in the key frame, or it can be the depth of the previous key frame of the key frame. The value is a dense depth map projected according to the camera pose. For example, if the key frame to be depth estimated is the first key frame in the image sequence, the first depth map of the key frame is the dense depth map obtained by initialization; if the key frame with depth estimation is the first key frame in the image sequence For key frames other than one key frame, the first depth map of the key frame is a dense depth map obtained by projecting the depth value of the previous key frame. Luminosity error refers to the measurement difference between the high gradient point in the projected image and the corresponding high gradient point in the reference frame image. The projected image is based on the initial camera pose between the reference frame and the key frame in the image sequence. The high gradient points corresponding to the pixels in the frame are projected to the reference frame image. The current camera pose includes the rotation and translation between the reference frame and the key frame. The second depth map of the key frame refers to the image sequence according to the The new dense depth map obtained by triangulating the current camera pose between the reference frame image and the key frame; the next frame image of the reference frame image refers to the next frame image adjacent to the pre-reference frame image in the image sequence , The posture of the latter camera includes the maximum threshold of the posture of the latter camera, which can be preset according to actual conditions and requirements, and there is no specific limitation here.
需要说明的是,稠密深度图是指包括大量特征点对应的深度值的图像,或者包括高梯度点也包括低梯度点对应的深度值的图像,通过对参考帧图像中的每个像素点进行深度估计得到深度图和深度值,便于后续恢复该像素点的空间坐标。It should be noted that a dense depth map refers to an image that includes depth values corresponding to a large number of feature points, or an image that includes both high gradient points and depth values corresponding to low gradient points. The depth estimation obtains the depth map and the depth value, which is convenient for subsequent restoration of the spatial coordinates of the pixel.
S104:基于所述关键帧的位姿参数以及所述关键帧的深度图进行图像重建,得到三维点云。S104: Perform image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud.
在本实施例中,三维重建是指从输入数据中建立3D模型,深度相机扫描得到的每一帧数据不仅包含场景中的点的彩色RGB图像,还包含每个点到深度相机所在的垂直平面的距离值,这个距离值称为深度值,这些深度值共同组成了这一帧的深度图。深度图可以看成一副灰度图,图像中的每一个点的灰度值代表了这个点在现实中的位置到相机所在垂直平面的真实距离,RGB图像中的每个点都会对应一个在相机的局部坐标系中的三维点。In this embodiment, 3D reconstruction refers to the establishment of a 3D model from the input data. Each frame of data scanned by the depth camera not only contains the color RGB image of the point in the scene, but also includes each point to the vertical plane where the depth camera is located. This distance value is called the depth value, and these depth values together constitute the depth map of this frame. The depth map can be regarded as a grayscale image. The grayscale value of each point in the image represents the true distance from the position of the point in reality to the vertical plane where the camera is located. Each point in the RGB image corresponds to a point on the camera. A three-dimensional point in the local coordinate system.
需要说明的是,三维重建的过程可以是图像获取、摄像机标定、特征提取、立体匹配和三维重建等,其中,立体匹配是指根据所提取的特征来建立图像对之间的一种对应关系,也就是将同一物理空间点在两幅不同图像中的成像点进行一一对应起来。在进行匹配时要注意场景中一些因素的干扰,比如光照条件、噪声干扰、景物几何形状畸变、表面物理特性以及摄像机机特性等诸多变化因素,以得到高精度的三维点云,也增强了视觉效果。It should be noted that the process of 3D reconstruction can be image acquisition, camera calibration, feature extraction, stereo matching, 3D reconstruction, etc., where stereo matching refers to establishing a correspondence between image pairs based on the extracted features. That is, the imaging points of the same physical space point in two different images are mapped one by one. When matching, pay attention to the interference of some factors in the scene, such as lighting conditions, noise interference, distortion of the geometric shape of the scene, surface physical characteristics, and camera characteristics, etc., in order to obtain a high-precision three-dimensional point cloud, and also enhance the vision effect.
可选地,S104可以包括步骤C1~C3,具体如下:Optionally, S104 may include steps C1 to C3, which are specifically as follows:
步骤C1:获取所述关键帧的像素坐标;Step C1: Obtain the pixel coordinates of the key frame;
在本实施例中,根据上述对单目内窥镜进行相机标定,可以确定像素坐标系和关键帧中的各图像的像素坐标,该像素坐标表示像素在图像中的位置,可以确定关键帧中的各图像的像素位置,便于后续对图像进行三维重建。In this embodiment, by performing camera calibration on the monocular endoscope described above, the pixel coordinate system and the pixel coordinates of each image in the key frame can be determined. The pixel coordinates indicate the position of the pixel in the image, and the key frame can be determined The pixel position of each image is convenient for subsequent three-dimensional reconstruction of the image.
步骤C2:根据所述深度图、所述关键帧的位姿参数和所述关键帧的像素坐标计算得到目标空间坐标;Step C2: calculating the target space coordinates according to the depth map, the pose parameters of the key frame, and the pixel coordinates of the key frame;
在本实施例中,获取关键中的各图像的深度图对应的深度值,将深度值和关键帧的位姿参数和关键帧的各图像的像素坐标进行计算,得到各图像的空间坐标,即由二维坐标到三维坐标的转化,根据精确地深度估计得到的深度值,计算所得的目标空间坐标的精度也提高。In this embodiment, the depth value corresponding to the depth map of each image in the key is obtained, and the depth value and the pose parameter of the key frame and the pixel coordinates of each image of the key frame are calculated to obtain the spatial coordinates of each image, namely The conversion from two-dimensional coordinates to three-dimensional coordinates, according to the depth value obtained by accurate depth estimation, also improves the accuracy of the calculated target space coordinates.
步骤C3:获取所述关键帧中的每个像素点的颜色信息,根据所述关键帧中的每个像素点的颜色信息和所述目标空间坐标对所述关键帧进行点云融合,得到所述三维点云。Step C3: Obtain the color information of each pixel in the key frame, and perform point cloud fusion on the key frame according to the color information of each pixel in the key frame and the target space coordinates to obtain the Describe the three-dimensional point cloud.
在本实施例中,二维图像中的像素点坐标[u,v],对应的点云包含了颜色信息和空间位置信息,颜色信息由该像素点的RGB值表示,根据所述深度图、所述关键帧的位姿参数和所述关键帧的像素坐标计算得到目标空间坐标为[x,y,z],从像素坐标[u,v]及其深度值d恢复空间坐标,由如下公式表示:In this embodiment, the pixel coordinates [u, v] in the two-dimensional image, the corresponding point cloud contains color information and spatial position information, and the color information is represented by the RGB value of the pixel. According to the depth map, The pose parameters of the key frame and the pixel coordinates of the key frame are calculated to obtain the target space coordinates as [x, y, z], and the space coordinates are restored from the pixel coordinates [u, v] and its depth value d by the following formula Means:
z′=dz′=d
x′=z(u-c x)/f x x′=z(uc x )/f x
y′=z(v-c y)/f y y′=z(vc y )/f y
(x,y,z) T=(R i,t i)(x′,y′,z′) T (x,y,z) T = (R i ,t i )(x′,y′,z′) T
其中,d表示该像素点的深度,由REMODE方案的深度估计得出,(x’,y’,z’)是像相机坐标系下的坐标值,(Ri,ti)是该帧对应的位姿参数。Among them, d represents the depth of the pixel, which is derived from the depth estimation of the REMODE scheme, (x',y',z') is the coordinate value in the camera coordinate system, (Ri,ti) is the bit corresponding to the frame Pose parameters.
需要说明的是,点云是一组离散的点表示的图,点云中存储了该帧像素点对应的空间坐标和颜色信息,在进行多帧的点云融合时,将多帧的点云存储在一个容器中,然后通过滤波器将重复的点云去除,可以得出多帧图像融合在一起的三维点云。上述三维重建方法可以是在融合时绘制了多帧图像的点云,以得到更精细的三维信息。It should be noted that the point cloud is a set of discrete points. The point cloud stores the spatial coordinates and color information corresponding to the pixels of the frame. When multi-frame point cloud fusion is performed, the multi-frame point cloud Stored in a container, and then the repeated point cloud is removed by a filter, and a three-dimensional point cloud of multiple frames of images can be obtained. The above-mentioned three-dimensional reconstruction method may draw a point cloud of multiple frames of images during fusion to obtain finer three-dimensional information.
可选地,当S101包括S1011至S1013时,步骤C1所述获取所述关键帧的像素坐标,包括步骤C11~C13:Optionally, when S101 includes S1011 to S1013, obtaining the pixel coordinates of the key frame in step C1 includes steps C11 to C13:
步骤C11:将所述相机坐标系投影至所述待校正图像的每个像素点所在的平面上,得到所述预设坐标在像素坐标系中的像素坐标;Step C11: Project the camera coordinate system onto the plane where each pixel of the image to be corrected is located, and obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
在本实施方式中,定义像素点在相机坐标系中的坐标,利用投影计算相机坐标系和图像坐标系之间的对应关系,之后通过图像坐标系与像素坐标系的对应关系得到像素坐标系,此处的像素坐标与上述的畸变校正中得到的像素坐标的过程相同,此处不再赘述。In this embodiment, the coordinates of the pixel points in the camera coordinate system are defined, and the correspondence between the camera coordinate system and the image coordinate system is calculated by projection, and then the pixel coordinate system is obtained through the correspondence between the image coordinate system and the pixel coordinate system. The pixel coordinates here are the same as the process of the pixel coordinates obtained in the above-mentioned distortion correction, and will not be repeated here.
步骤C12:将所述预设坐标在像素坐标系中的像素坐标映射至所述相机坐标系得到所述图像序列,以及所述图像序列对应的像素坐标;Step C12: Map the pixel coordinates of the preset coordinates in the pixel coordinate system to the camera coordinate system to obtain the image sequence and the pixel coordinates corresponding to the image sequence;
在本实施例中,通过畸变校正的坐标系转化方式可以得到校正后的图像序列和图像序列对应的像素坐标,此处的具体处理过程与上述畸变校正过程相同,此处不再赘述。In this embodiment, the corrected image sequence and the pixel coordinates corresponding to the image sequence can be obtained through the coordinate system transformation method of the distortion correction. The specific processing process here is the same as the above-mentioned distortion correction process, and will not be repeated here.
步骤C13:基于所述图像序列对应的像素坐标得到所述关键帧的像素坐标。Step C13: Obtain the pixel coordinates of the key frame based on the pixel coordinates corresponding to the image sequence.
在本实施方式中,从图像序列中确定了关键帧,就可以得到关键帧的像素坐标,根据关键帧的各图像的像素坐标就可以确定各图像相对于相机的移动位置关系,以提高了单目内窥镜图像的处理效率。In this embodiment, the key frame is determined from the image sequence, and the pixel coordinates of the key frame can be obtained. According to the pixel coordinates of each image of the key frame, the moving position relationship of each image relative to the camera can be determined, so as to improve the single The processing efficiency of the endoscopic image.
图3示出了本申请实施例提供的单目内窥镜图像的三维重建装置300,如图3,本申请提供的单目内窥镜图像的三维重建装置300,包括:FIG. 3 shows a three-dimensional reconstruction device 300 for monocular endoscopic images provided by an embodiment of the present application. As shown in FIG. 3, the three-dimensional reconstruction device 300 for monocular endoscope images provided by the present application includes:
获取模块310,用于获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,对所述多张棋盘标定板的畸变图像进行畸变校正得到图像序列;The acquiring module 310 is configured to acquire the distortion images of a plurality of checkerboard calibration boards taken by a monocular endoscope, and perform distortion correction on the distortion images of the checkerboard calibration boards to obtain an image sequence;
确定模块320,用于从所述图像序列中确定关键帧;The determining module 320 is configured to determine a key frame from the image sequence;
计算模块330,用于获取所述关键帧的位姿参数,估算所述关键帧的深度图;The calculation module 330 is configured to obtain the pose parameters of the key frame and estimate the depth map of the key frame;
生成模块340,用于基于所述关键帧的位姿参数以及所述关键帧的深度图进行图像重建,得到三维点云。The generating module 340 is configured to perform image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud.
在本实施例中,单目内窥镜图像的三维重建装置可以是终端设备,也可以是服务器,也可以是能进行人机交互的设备。In this embodiment, the device for 3D reconstruction of monocular endoscopic images may be a terminal device, a server, or a device capable of human-computer interaction.
可选地,所述获取模块310具体包括:Optionally, the obtaining module 310 specifically includes:
第一获取单元,用于获取所述多张棋盘标定板的畸变图像中的棋盘的角点,基于所述棋盘的角点对所述单目内窥镜进行标定,得到所述单目内窥镜的相机参数和畸变参数;The first acquisition unit is configured to acquire the corner points of the chessboard in the distortion images of the multiple chessboard calibration boards, and calibrate the monocular endoscope based on the corner points of the chessboard to obtain the monocular endoscope Camera parameters and distortion parameters of the mirror;
第一确定单元,用于根据所述相机参数和所述畸变参数从所述畸变图像中确定待校正图像;A first determining unit, configured to determine an image to be corrected from the distorted image according to the camera parameter and the distortion parameter;
第一处理单元,用于基于所述相机坐标系对所述待校正图像进行畸变校正得到所述图像序列。The first processing unit is configured to perform distortion correction on the image to be corrected based on the camera coordinate system to obtain the image sequence.
可选地,所述获取模块310还包括:Optionally, the obtaining module 310 further includes:
第二获取单元,用于获取所述待校正图像的每个像素点在所述相机坐标中的预设坐标;A second acquiring unit, configured to acquire the preset coordinates of each pixel of the image to be corrected in the camera coordinates;
第二处理单元,用于将所述相机坐标系投影至所述待校正图像的每个像素点所在的平面上,得到所述预设坐标在像素坐标系中的像素坐标;A second processing unit, configured to project the camera coordinate system onto the plane where each pixel of the image to be corrected is located, to obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
第三处理单元,用于将所述预设坐标在像素坐标系中的像素坐标映射至所述相机坐标系得到所述图像序列。The third processing unit is configured to map the pixel coordinates of the preset coordinates in the pixel coordinate system to the camera coordinate system to obtain the image sequence.
可选地,所述确定模块320具体包括:Optionally, the determining module 320 specifically includes:
第三获取单元,用于获取所述图像序列中的各图像的局部特征,并基于所述各图像的局部特征对所述图像序列中的各图像进行特征点匹配,得到匹配结果;The third acquiring unit is configured to acquire the local features of each image in the image sequence, and perform feature point matching on each image in the image sequence based on the local features of each image to obtain a matching result;
第二确定单元,用于当所述匹配结果为第一图像和第二图像匹配的特征点数量大于或等于预设阈值时,将所述第一图像作为关键帧,其中,所述第一图像和所述第二图像为所述图像序列中相邻的任意两帧图像。The second determining unit is configured to use the first image as a key frame when the matching result is that the number of feature points matched by the first image and the second image is greater than or equal to a preset threshold, wherein the first image And the second image are any two adjacent frames of images in the image sequence.
可选地,所述确定模块320还包括:Optionally, the determining module 320 further includes:
第三确定单元,用于当第一图像和第二图像匹配的特征点数量大于或等于预设阈值时,将所述第一图像作为关键帧;A third determining unit, configured to use the first image as a key frame when the number of feature points matched by the first image and the second image is greater than or equal to a preset threshold;
第四处理单元,用于将所述第一图像进行位姿初始化;The fourth processing unit is used to initialize the pose of the first image;
第一估算单元,用于估算所述图像序列中的关键帧的位姿参数。The first estimation unit is used to estimate the pose parameters of the key frames in the image sequence.
可选地,所述确定模块320还包括:Optionally, the determining module 320 further includes:
第四确定单元,用于从所述关键帧中确定参考帧图像,其中,所述参考帧图像为所述 关键帧中的任一帧图像或多帧图像;A fourth determining unit, configured to determine a reference frame image from the key frame, wherein the reference frame image is any frame image or multiple frames of images in the key frame;
第二估算单元,用于基于所述位姿参数对所述参考帧图像的每个像素点进行深度估计处理,得到所述关键帧的深度图。The second estimation unit is configured to perform depth estimation processing on each pixel of the reference frame image based on the pose parameter to obtain the depth map of the key frame.
可选地,所述生成模块340包括:Optionally, the generating module 340 includes:
第四获取单元,用于获取所述关键帧的像素坐标;The fourth acquiring unit is used to acquire the pixel coordinates of the key frame;
第三估算单元,用于根据所述深度图、所述关键帧的位姿参数和所述关键帧的像素坐标计算得到目标空间坐标;The third estimation unit is configured to calculate the target space coordinates according to the depth map, the pose parameters of the key frame, and the pixel coordinates of the key frame;
第一生成单元,用于获取所述关键帧中的每个像素点的颜色信息,根据所述关键帧中的每个像素点的颜色信息和所述目标空间坐标对所述关键帧进行点云融合,得到所述三维点云。The first generating unit is used to obtain the color information of each pixel in the key frame, and perform a point cloud on the key frame according to the color information of each pixel in the key frame and the target space coordinates Fusion to obtain the three-dimensional point cloud.
可选地,所述生成模块340还包括:Optionally, the generating module 340 further includes:
第一投影单元,将所述相机坐标系投影至所述待校正图像的每个像素点所在的平面上,得到所述预设坐标在像素坐标系中的像素坐标;A first projection unit, projecting the camera coordinate system onto a plane where each pixel point of the image to be corrected is located, to obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
第二投影单元,用于将所述预设坐标在像素坐标系中的像素坐标映射至所述相机坐标系得到所述图像序列,以及所述图像序列对应的像素坐标;The second projection unit is configured to map the pixel coordinates of the preset coordinates in the pixel coordinate system to the camera coordinate system to obtain the image sequence and the pixel coordinates corresponding to the image sequence;
第二生成单元,用于基于所述图像序列对应的像素坐标得到所述关键帧的像素坐标。The second generating unit is configured to obtain the pixel coordinates of the key frame based on the pixel coordinates corresponding to the image sequence.
请参阅图4,图4是本申请实施例还提供的终端设备400的结构示意图,终端设备400包括存储器410、至少一个处理器420以及存储在所述存储器410中并可在所述处理器420上运行的计算机程序430,所述处理器420执行所述计算机程序430时实现上述的三维重建方法。Please refer to FIG. 4, which is a schematic structural diagram of a terminal device 400 provided by an embodiment of the present application. The terminal device 400 includes a memory 410, at least one processor 420, and is stored in the memory 410 and can be stored in the processor 420. When the processor 420 executes the computer program 430, the above-mentioned three-dimensional reconstruction method is implemented.
终端设备400可以是桌上型计算机、手机、平板电脑、可穿戴设备、车载设备、增强现实(augmented reality,AR)/虚拟现实(virtual reality,VR)设备、笔记本电脑、超级移动个人计算机(ultra-mobile personal computer,UMPC)、上网本、个人数字助理(personal digital assistant,PDA)等终端设备上,本申请实施例对终端设备的具体类型不作任何限制。The terminal device 400 may be a desktop computer, a mobile phone, a tablet computer, a wearable device, a vehicle-mounted device, an augmented reality (AR)/virtual reality (VR) device, a notebook computer, an ultra mobile personal computer (ultra -Mobile personal computer (UMPC), netbook, personal digital assistant (personal digital assistant, PDA) and other terminal devices, the embodiment of this application does not impose any restrictions on the specific types of terminal devices.
该终端设备400可包括但不仅限于处理器420、存储器410。本领域技术人员可以理解,图4仅仅是终端设备400的举例,并不构成对终端设备400的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如还可以包括输入输出设备等。The terminal device 400 may include but is not limited to a processor 420 and a memory 410. Those skilled in the art can understand that FIG. 4 is only an example of the terminal device 400, and does not constitute a limitation on the terminal device 400. It may include more or less components than those shown in the figure, or a combination of certain components, or different components. , For example, can also include input and output devices.
所称处理器420可以是中央处理单元(Central Processing Unit,CPU),该处理器420还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。 通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The so-called processor 420 may be a central processing unit (CPU), and the processor 420 may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSP), and application specific integrated circuits (Application Specific Integrated Circuits). , ASIC), ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc. The general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
所述存储器410在一些实施例中可以是终端设备400的内部存储单元,例如终端设备400的硬盘或内存。所述存储器410在另一些实施例中也可以是所述终端设备400的外部存储设备,例如终端设备400上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器410还可以既包括终端设备400的内部存储单元也包括外部存储设备。所述存储器410用于存储操作系统、应用程序、引导装载程序(Boot Loader)、数据以及其他程序等,例如所述计算机程序的程序代码等。所述存储器410还可以用于暂时地存储已经输出或者将要输出的数据。In some embodiments, the memory 410 may be an internal storage unit of the terminal device 400, such as a hard disk or a memory of the terminal device 400. In other embodiments, the memory 410 may also be an external storage device of the terminal device 400, such as a plug-in hard disk equipped on the terminal device 400, a smart media card (SMC), and a secure digital (Secure Digital). Digital, SD) card, flash card (Flash Card), etc. Further, the memory 410 may also include both an internal storage unit of the terminal device 400 and an external storage device. The memory 410 is used to store an operating system, an application program, a boot loader (Boot Loader), data, and other programs, such as the program code of the computer program. The memory 410 may also be used to temporarily store data that has been output or will be output.
需要说明的是,上述生成装置/单元之间的信息交互、执行过程等内容,由于与本申请方法实施例基于同一构思,其具体功能及带来的技术效果,具体可参见方法实施例部分,此处不再赘述。It should be noted that the information interaction and execution process between the above-mentioned generating devices/units are based on the same concept as the method embodiment of this application, and its specific functions and technical effects can be found in the method embodiment section. I won't repeat them here.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。实施例中的各功能单元、模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中,上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。另外,各功能单元、模块的具体名称也只是为了便于相互区分,并不用于限制本申请的保护范围。上述辅助拍摄装置中单元、模块的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that, for the convenience and conciseness of description, only the division of the above functional units and modules is used as an example. In practical applications, the above functions can be allocated to different functional units and modules as needed. Module completion, that is, the internal structure of the device is divided into different functional units or modules to complete all or part of the functions described above. The functional units and modules in the embodiments can be integrated into one processing unit, or each unit can exist alone physically, or two or more units can be integrated into one unit. The above-mentioned integrated units can be hardware-based Formal realization can also be realized in the form of a software functional unit. In addition, the specific names of the functional units and modules are only for the convenience of distinguishing each other, and are not used to limit the protection scope of the present application. For the specific working process of the units and modules in the above-mentioned auxiliary photographing device, reference may be made to the corresponding process in the foregoing method embodiment, which will not be repeated here.
本申请实施例还提供了一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现可实现上述各个方法实施例中的步骤。The embodiments of the present application also provide a computer-readable storage medium, where the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the steps in each of the foregoing method embodiments can be realized.
本申请实施例提供了一种计算机程序产品,当计算机程序产品在移动终端上运行时,使得移动终端执行时实现可实现上述各个方法实施例中的步骤。The embodiments of the present application provide a computer program product. When the computer program product runs on a mobile terminal, the steps in the foregoing method embodiments can be realized when the mobile terminal is executed.
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请实现上述实施例方法中的全部或部分流程,可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一计算机可读存储介质中,该计算机程序在被处理器执行时,可实现上述各个方法实施例的步骤。其中,所述计算机程序包括计算机程序代码,所述计算机程序代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。所述计算机可读介质至少可以包括:能够将计算机程序代码携带到终端设备的任何实体或装置、记录介质、计算机存储器、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、 电载波信号、电信信号以及软件分发介质。例如U盘、移动硬盘、磁碟或者光盘等。在某些司法管辖区,根据立法和专利实践,计算机可读介质不可以是电载波信号和电信信号。If the integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium. Based on this understanding, the implementation of all or part of the processes in the above-mentioned embodiment methods in the present application can be accomplished by instructing relevant hardware through a computer program. The computer program can be stored in a computer-readable storage medium. The computer program can be stored in a computer-readable storage medium. When executed by the processor, the steps of the foregoing method embodiments can be implemented. Wherein, the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate forms. The computer-readable medium may at least include: any entity or device capable of carrying computer program code to a terminal device, a recording medium, a computer memory, a read-only memory (Read-Only Memory, ROM), and a random access memory (Random Access). Memory, RAM), electrical carrier signals, telecommunications signals, and software distribution media. For example, U disk, mobile hard disk, floppy disk or CD-ROM, etc. In some jurisdictions, according to legislation and patent practices, computer-readable media cannot be electrical carrier signals and telecommunication signals.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。In the above-mentioned embodiments, the description of each embodiment has its own focus. For parts that are not described in detail or recorded in an embodiment, reference may be made to related descriptions of other embodiments.
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。A person of ordinary skill in the art may realize that the units and algorithm steps of the examples described in combination with the embodiments disclosed herein can be implemented by electronic hardware or a combination of computer software and electronic hardware. Whether these functions are executed by hardware or software depends on the specific application and design constraint conditions of the technical solution. Professionals and technicians can use different methods for each specific application to implement the described functions, but such implementation should not be considered beyond the scope of this application.
在本申请说明书中描述的参考“一个实施例”或“一些实施例”等意味着在本申请的一个或多个实施例中包括结合该实施例描述的特定特征、结构或特点。由此,在本说明书中的不同之处出现的语句“在一个实施例中”、“在一些实施例中”、“在其他一些实施例中”、“在另外一些实施例中”等不是必然都参考相同的实施例,而是意味着“一个或多个但不是所有的实施例”,除非是以其他方式另外特别强调。术语“包括”、“包含”、“具有”及它们的变形都意味着“包括但不限于”,除非是以其他方式另外特别强调。The reference to "one embodiment" or "some embodiments" described in the specification of this application means that one or more embodiments of this application include a specific feature, structure, or characteristic described in combination with the embodiment. Therefore, the sentences "in one embodiment", "in some embodiments", "in some other embodiments", "in some other embodiments", etc. appearing in different places in this specification are not necessarily All refer to the same embodiment, but mean "one or more but not all embodiments" unless it is specifically emphasized otherwise. The terms "including", "including", "having" and their variations all mean "including but not limited to", unless otherwise specifically emphasized.
在本申请所提供的实施例中,应该理解到,所揭露的装置/网络设备和方法,可以通过其它的方式实现。例如,以上所描述的装置/网络设备实施例仅仅是示意性的,例如,所述模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通讯连接可以是通过一些接口,装置或单元的间接耦合或通讯连接,可以是电性,机械或其它的形式。In the embodiments provided in this application, it should be understood that the disclosed apparatus/network equipment and method may be implemented in other ways. For example, the device/network device embodiments described above are only illustrative. For example, the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, such as multiple units. Or components can be combined or integrated into another system, or some features can be omitted or not implemented. In addition, the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-mentioned embodiments are only used to illustrate the technical solutions of the present application, not to limit them; although the present application has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that they can still implement the foregoing The technical solutions recorded in the examples are modified, or some of the technical features are equivalently replaced; these modifications or replacements do not cause the essence of the corresponding technical solutions to deviate from the spirit and scope of the technical solutions of the embodiments of the application, and should be included in Within the scope of protection of this application.

Claims (10)

  1. 一种单目内窥镜图像的三维重建方法,其特征在于,所述三维重建方法包括:A method for three-dimensional reconstruction of monocular endoscopic images, characterized in that the three-dimensional reconstruction method includes:
    获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,对所述多张棋盘标定板的畸变图像进行畸变校正得到图像序列;Acquiring the distortion images of a plurality of checkerboard calibration boards taken by a monocular endoscope, and performing distortion correction on the distortion images of the checkerboard calibration boards to obtain an image sequence;
    从所述图像序列中确定关键帧;Determining key frames from the image sequence;
    获取所述关键帧的位姿参数,估算所述关键帧的深度图;Acquiring the pose parameters of the key frame, and estimating the depth map of the key frame;
    基于所述关键帧的位姿参数以及所述关键帧的深度图进行图像重建,得到三维点云。Perform image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud.
  2. 根据权利要求1所述的三维重建方法,其特征在于,所述基于所述关键帧的位姿参数以及所述关键帧的深度图进行图像重建,得到三维点云,包括:The 3D reconstruction method according to claim 1, wherein the image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a 3D point cloud comprises:
    获取所述关键帧的像素坐标;Acquiring the pixel coordinates of the key frame;
    根据所述深度图、所述关键帧的位姿参数和所述关键帧的像素坐标计算得到目标空间坐标;Calculating the target space coordinates according to the depth map, the pose parameters of the key frame, and the pixel coordinates of the key frame;
    获取所述关键帧中的每个像素点的颜色信息,根据所述关键帧中的每个像素点的颜色信息和所述目标空间坐标对所述关键帧进行点云融合,得到所述三维点云。Obtain the color information of each pixel in the key frame, and perform point cloud fusion on the key frame according to the color information of each pixel in the key frame and the target space coordinates to obtain the three-dimensional point cloud.
  3. 根据权利要求1或2所述的三维重建方法,其特征在于,所述获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,对所述多张棋盘标定板的畸变图像进行校正得到图像序列,包括:The three-dimensional reconstruction method according to claim 1 or 2, characterized in that said acquiring the distortion images of a plurality of checkerboard calibration boards taken by a monocular endoscope, and correcting the distortion images of the plurality of checkerboard calibration boards to obtain Image sequence, including:
    获取所述多张棋盘标定板的畸变图像中的棋盘的角点,基于所述棋盘的角点对所述单目内窥镜进行标定,得到所述单目内窥镜的相机参数和畸变参数;Acquire the corner points of the chessboard in the distortion images of the checkerboard calibration boards, calibrate the monocular endoscope based on the corner points of the chessboard, and obtain the camera parameters and distortion parameters of the monocular endoscope ;
    根据所述相机参数和所述畸变参数从所述畸变图像中确定待校正图像;Determining an image to be corrected from the distorted image according to the camera parameter and the distortion parameter;
    基于相机坐标系对所述待校正图像进行畸变校正得到所述图像序列。Performing distortion correction on the image to be corrected based on the camera coordinate system to obtain the image sequence.
  4. 根据权利要求3所述的三维重建方法,其特征在于,所述基于相机坐标系对所述待校正图像进行畸变校正得到所述图像序列,包括:The three-dimensional reconstruction method according to claim 3, wherein said performing distortion correction on said image to be corrected based on a camera coordinate system to obtain said image sequence comprises:
    获取所述待校正图像的每个像素点在所述相机坐标中的预设坐标;Acquiring the preset coordinates of each pixel of the image to be corrected in the camera coordinates;
    将所述相机坐标系投影至所述待校正图像的每个像素点所在的平面上,得到所述预设坐标在像素坐标系中的像素坐标;Projecting the camera coordinate system onto the plane where each pixel of the image to be corrected is located to obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
    将所述预设坐标在像素坐标系中的像素坐标映射至所述相机坐标系得到所述图像序列。The pixel coordinates of the preset coordinates in the pixel coordinate system are mapped to the camera coordinate system to obtain the image sequence.
  5. 根据权利要求3所述的三维重建方法,其特征在于,所述获取所述关键帧的像素坐标,包括:The three-dimensional reconstruction method according to claim 3, wherein said obtaining the pixel coordinates of the key frame comprises:
    将所述相机坐标系投影至所述待校正图像的每个像素点所在的平面上,得到所述预设坐标在像素坐标系中的像素坐标;Projecting the camera coordinate system onto the plane where each pixel of the image to be corrected is located to obtain the pixel coordinates of the preset coordinates in the pixel coordinate system;
    将所述预设坐标在像素坐标系中的像素坐标映射至所述相机坐标系得到所述图像序列,以及所述图像序列对应的像素坐标;Mapping the pixel coordinates of the preset coordinates in the pixel coordinate system to the camera coordinate system to obtain the image sequence and the pixel coordinates corresponding to the image sequence;
    基于所述图像序列对应的像素坐标得到所述关键帧的像素坐标。Obtain the pixel coordinates of the key frame based on the pixel coordinates corresponding to the image sequence.
  6. 根据权利要求1-2、4-5任一项所述的三维重建方法,其特征在于,所述从所述图像序列中确定关键帧,包括:The three-dimensional reconstruction method according to any one of claims 1-2 and 4-5, wherein said determining a key frame from the image sequence comprises:
    获取所述图像序列中的各图像的局部特征,并基于所述各图像的局部特征对所述图像序列中的各图像进行特征点匹配,得到匹配结果;Acquiring local features of each image in the image sequence, and performing feature point matching on each image in the image sequence based on the local feature of each image, to obtain a matching result;
    当所述匹配结果为第一图像和第二图像匹配的特征点数量大于或等于预设阈值时,将所述第一图像作为关键帧,其中,所述第一图像和所述第二图像为所述图像序列中相邻的任意两帧图像。When the matching result is that the number of feature points matched by the first image and the second image is greater than or equal to a preset threshold, the first image is used as a key frame, where the first image and the second image are Any two adjacent frames of images in the image sequence.
  7. 根据权利要求6所述的三维重建方法,其特征在于,所述获取所述关键帧的位姿参数,包括:The 3D reconstruction method according to claim 6, wherein said obtaining the pose parameters of the key frame comprises:
    将所述第一图像进行位姿初始化;Performing pose initialization on the first image;
    估算所述图像序列中的关键帧的位姿参数。Estimate the pose parameters of the key frames in the image sequence.
  8. 根据权利要求1或7所述的三维重建方法,其特征在于,所述估算所述关键帧的深度图,包括:The 3D reconstruction method according to claim 1 or 7, wherein the estimating the depth map of the key frame comprises:
    从所述关键帧中确定参考帧图像,其中,所述参考帧图像为所述关键帧中的任一帧图像或多帧图像;Determining a reference frame image from the key frames, where the reference frame image is any frame image or multiple frames of images in the key frame;
    基于所述位姿参数对所述参考帧图像的每个像素点进行深度估计处理,得到所述关键帧的深度图。Performing depth estimation processing on each pixel of the reference frame image based on the pose parameter to obtain the depth map of the key frame.
  9. 一种单目内窥镜图像的三维重建装置,其特征在于,包括:A three-dimensional reconstruction device for monocular endoscopic images, which is characterized in that it comprises:
    获取模块,用于获取单目内窥镜拍摄的多张棋盘标定板的畸变图像,对所述多张棋盘标定板的畸变图像进行畸变校正得到图像序列;The acquisition module is used to acquire the distortion images of a plurality of checkerboard calibration boards taken by a monocular endoscope, and perform distortion correction on the distortion images of the checkerboard calibration boards to obtain an image sequence;
    确定模块,用于从所述图像序列中确定关键帧;A determining module for determining key frames from the image sequence;
    计算模块,用于获取所述关键帧的位姿参数,估算所述关键帧的深度图;A calculation module for obtaining the pose parameters of the key frame and estimating the depth map of the key frame;
    生成模块,用于基于所述关键帧的位姿参数以及所述关键帧的深度图进行图像重建,得到三维点云。The generating module is used for image reconstruction based on the pose parameters of the key frame and the depth map of the key frame to obtain a three-dimensional point cloud.
  10. 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1至8任一项所述的三维重建方法。A computer-readable storage medium storing a computer program, wherein the computer program implements the three-dimensional reconstruction method according to any one of claims 1 to 8 when the computer program is executed by a processor.
PCT/CN2020/129546 2019-12-12 2020-11-17 Three-dimensional reconstruction method and apparatus for monocular endoscope image, and terminal device WO2021115071A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201911275140.9A CN111145238B (en) 2019-12-12 2019-12-12 Three-dimensional reconstruction method and device for monocular endoscopic image and terminal equipment
CN201911275140.9 2019-12-12

Publications (1)

Publication Number Publication Date
WO2021115071A1 true WO2021115071A1 (en) 2021-06-17

Family

ID=70518247

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/129546 WO2021115071A1 (en) 2019-12-12 2020-11-17 Three-dimensional reconstruction method and apparatus for monocular endoscope image, and terminal device

Country Status (2)

Country Link
CN (1) CN111145238B (en)
WO (1) WO2021115071A1 (en)

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113596432A (en) * 2021-07-30 2021-11-02 成都市谛视科技有限公司 3D video production method, device and equipment with variable visual angle and storage medium
CN113793413A (en) * 2021-08-13 2021-12-14 北京迈格威科技有限公司 Three-dimensional reconstruction method and device, electronic equipment and storage medium
CN113838138A (en) * 2021-08-06 2021-12-24 杭州灵西机器人智能科技有限公司 System calibration method, system, device and medium for optimizing feature extraction
CN113884025A (en) * 2021-09-16 2022-01-04 河南垂天智能制造有限公司 Additive manufacturing structure optical loopback detection method and device, electronic equipment and storage medium
CN113902846A (en) * 2021-10-11 2022-01-07 岱悟智能科技(上海)有限公司 Indoor three-dimensional modeling method based on monocular depth camera and mileage sensor
CN113925441A (en) * 2021-12-17 2022-01-14 极限人工智能有限公司 Imaging method and imaging system based on endoscope
CN114155349A (en) * 2021-12-14 2022-03-08 杭州联吉技术有限公司 Three-dimensional mapping method, three-dimensional mapping device and robot
CN114529613A (en) * 2021-12-15 2022-05-24 深圳市华汉伟业科技有限公司 Method for extracting characteristic point high-precision coordinates of circular array calibration plate
CN114820787A (en) * 2022-04-22 2022-07-29 聊城大学 Image correction method and system for large-view-field planar vision measurement
CN114862961A (en) * 2022-04-13 2022-08-05 上海人工智能创新中心 Position detection method and device for calibration plate, electronic equipment and readable storage medium
CN114882058A (en) * 2022-04-26 2022-08-09 上海人工智能创新中心 Angular point detection method and device and calibration board
CN114926515A (en) * 2022-06-08 2022-08-19 北京化工大学 Infrared and visible light image registration method based on time-space domain depth information completion
CN114972531A (en) * 2022-05-17 2022-08-30 上海人工智能创新中心 Calibration board, corner detection method, equipment and readable storage medium
CN117173342A (en) * 2023-11-02 2023-12-05 中国海洋大学 Underwater monocular and binocular camera-based natural light moving three-dimensional reconstruction device and method
CN117218089A (en) * 2023-09-18 2023-12-12 中南大学 Asphalt pavement structure depth detection method
CN117237553A (en) * 2023-09-14 2023-12-15 广东省核工业地质局测绘院 Three-dimensional map mapping system based on point cloud image fusion
CN117392329A (en) * 2023-12-08 2024-01-12 中国海洋大学 Use method based on mobile multispectral luminosity stereoscopic device
WO2024022062A1 (en) * 2022-07-28 2024-02-01 杭州堃博生物科技有限公司 Endoscope pose estimation method and apparatus, and storage medium
CN117557660A (en) * 2024-01-09 2024-02-13 北京集度科技有限公司 Data processing method and device, electronic equipment and vehicle
CN117893693A (en) * 2024-03-15 2024-04-16 南昌航空大学 Dense SLAM three-dimensional scene reconstruction method and device
CN114862961B (en) * 2022-04-13 2024-06-07 上海人工智能创新中心 Position detection method and device for calibration plate, electronic equipment and readable storage medium

Families Citing this family (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111145238B (en) * 2019-12-12 2023-09-22 中国科学院深圳先进技术研究院 Three-dimensional reconstruction method and device for monocular endoscopic image and terminal equipment
CN111798387A (en) * 2020-06-24 2020-10-20 海南大学 Image processing method and system for confocal endoscope
CN111882657B (en) * 2020-06-29 2024-01-26 杭州易现先进科技有限公司 Three-dimensional reconstruction scale recovery method, device, system and computer equipment
CN111784754B (en) * 2020-07-06 2024-01-12 浙江得图网络有限公司 Tooth orthodontic method, device, equipment and storage medium based on computer vision
CN111862120B (en) * 2020-07-22 2023-07-11 苏州大学 Monocular SLAM scale recovery method
CN111899345B (en) * 2020-08-03 2023-09-01 成都圭目机器人有限公司 Three-dimensional reconstruction method based on 2D visual image
CN112330729B (en) * 2020-11-27 2024-01-12 中国科学院深圳先进技术研究院 Image depth prediction method, device, terminal equipment and readable storage medium
CN112261399B (en) * 2020-12-18 2021-03-16 安翰科技(武汉)股份有限公司 Capsule endoscope image three-dimensional reconstruction method, electronic device and readable storage medium
CN112802123B (en) * 2021-01-21 2023-10-27 北京科技大学设计研究院有限公司 Binocular linear array camera static calibration method based on stripe virtual target
CN112907620A (en) * 2021-01-25 2021-06-04 北京地平线机器人技术研发有限公司 Camera pose estimation method and device, readable storage medium and electronic equipment
CN112767489B (en) * 2021-01-29 2024-05-14 北京达佳互联信息技术有限公司 Three-dimensional pose determining method and device, electronic equipment and storage medium
CN112927340B (en) * 2021-04-06 2023-12-01 中国科学院自动化研究所 Three-dimensional reconstruction acceleration method, system and equipment independent of mechanical placement
CN113569843A (en) * 2021-06-21 2021-10-29 影石创新科技股份有限公司 Corner point detection method and device, computer equipment and storage medium
CN113724194B (en) * 2021-06-22 2024-01-30 山东交通学院 Engine endoscopic flame measurement system and image processing method
WO2023007641A1 (en) * 2021-07-29 2023-02-02 株式会社エビデント Three-dimensional reconstruction device, three-dimensional reconstruction method, and program
CN116168143A (en) * 2021-11-25 2023-05-26 华为技术有限公司 Multi-view three-dimensional reconstruction method
CN114332028A (en) * 2021-12-30 2022-04-12 小荷医疗器械(海南)有限公司 Endoscope image processing method and device, readable medium and electronic equipment
CN114677572B (en) * 2022-04-08 2023-04-18 北京百度网讯科技有限公司 Object description parameter generation method and deep learning model training method
CN114782470B (en) * 2022-06-22 2022-09-13 浙江鸿禾医疗科技有限责任公司 Three-dimensional panoramic recognition positioning method of alimentary canal, storage medium and equipment
CN115471556B (en) * 2022-09-22 2023-11-14 南京博视医疗科技有限公司 Monocular camera image target point three-dimensional positioning method and device
CN116704152B (en) * 2022-12-09 2024-04-19 荣耀终端有限公司 Image processing method and electronic device
CN116958147B (en) * 2023-09-21 2023-12-22 青岛美迪康数字工程有限公司 Target area determining method, device and equipment based on depth image characteristics
CN117115358A (en) * 2023-10-11 2023-11-24 世优(北京)科技有限公司 Automatic digital person modeling method and device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103247075A (en) * 2013-05-13 2013-08-14 北京工业大学 Variational mechanism-based indoor scene three-dimensional reconstruction method
US20140111507A1 (en) * 2012-10-23 2014-04-24 Electronics And Telecommunications Research Institute 3-dimensional shape reconstruction device using depth image and color image and the method
CN108898630A (en) * 2018-06-27 2018-11-27 清华-伯克利深圳学院筹备办公室 A kind of three-dimensional rebuilding method, device, equipment and storage medium
CN109448041A (en) * 2018-10-29 2019-03-08 重庆金山医疗器械有限公司 A kind of capsule endoscope 3-dimensional reconstruction method and system
CN109544677A (en) * 2018-10-30 2019-03-29 山东大学 Indoor scene main structure method for reconstructing and system based on depth image key frame
CN111145238A (en) * 2019-12-12 2020-05-12 中国科学院深圳先进技术研究院 Three-dimensional reconstruction method and device of monocular endoscope image and terminal equipment

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019169540A1 (en) * 2018-03-06 2019-09-12 斯坦德机器人(深圳)有限公司 Method for tightly-coupling visual slam, terminal and computer readable storage medium
CN108416840B (en) * 2018-03-14 2020-02-18 大连理工大学 Three-dimensional scene dense reconstruction method based on monocular camera
CN109087349B (en) * 2018-07-18 2021-01-26 亮风台(上海)信息科技有限公司 Monocular depth estimation method, device, terminal and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140111507A1 (en) * 2012-10-23 2014-04-24 Electronics And Telecommunications Research Institute 3-dimensional shape reconstruction device using depth image and color image and the method
CN103247075A (en) * 2013-05-13 2013-08-14 北京工业大学 Variational mechanism-based indoor scene three-dimensional reconstruction method
CN108898630A (en) * 2018-06-27 2018-11-27 清华-伯克利深圳学院筹备办公室 A kind of three-dimensional rebuilding method, device, equipment and storage medium
CN109448041A (en) * 2018-10-29 2019-03-08 重庆金山医疗器械有限公司 A kind of capsule endoscope 3-dimensional reconstruction method and system
CN109544677A (en) * 2018-10-30 2019-03-29 山东大学 Indoor scene main structure method for reconstructing and system based on depth image key frame
CN111145238A (en) * 2019-12-12 2020-05-12 中国科学院深圳先进技术研究院 Three-dimensional reconstruction method and device of monocular endoscope image and terminal equipment

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113596432A (en) * 2021-07-30 2021-11-02 成都市谛视科技有限公司 3D video production method, device and equipment with variable visual angle and storage medium
CN113596432B (en) * 2021-07-30 2024-04-30 成都市谛视科技有限公司 Visual angle variable 3D video production method, visual angle variable 3D video production device, visual angle variable 3D video production equipment and storage medium
CN113838138A (en) * 2021-08-06 2021-12-24 杭州灵西机器人智能科技有限公司 System calibration method, system, device and medium for optimizing feature extraction
CN113793413A (en) * 2021-08-13 2021-12-14 北京迈格威科技有限公司 Three-dimensional reconstruction method and device, electronic equipment and storage medium
CN113884025A (en) * 2021-09-16 2022-01-04 河南垂天智能制造有限公司 Additive manufacturing structure optical loopback detection method and device, electronic equipment and storage medium
CN113884025B (en) * 2021-09-16 2024-05-03 河南垂天智能制造有限公司 Method and device for detecting optical loop of additive manufacturing structure, electronic equipment and storage medium
CN113902846B (en) * 2021-10-11 2024-04-12 岱悟智能科技(上海)有限公司 Indoor three-dimensional modeling method based on monocular depth camera and mileage sensor
CN113902846A (en) * 2021-10-11 2022-01-07 岱悟智能科技(上海)有限公司 Indoor three-dimensional modeling method based on monocular depth camera and mileage sensor
CN114155349A (en) * 2021-12-14 2022-03-08 杭州联吉技术有限公司 Three-dimensional mapping method, three-dimensional mapping device and robot
CN114155349B (en) * 2021-12-14 2024-03-22 杭州联吉技术有限公司 Three-dimensional image construction method, three-dimensional image construction device and robot
CN114529613A (en) * 2021-12-15 2022-05-24 深圳市华汉伟业科技有限公司 Method for extracting characteristic point high-precision coordinates of circular array calibration plate
CN113925441B (en) * 2021-12-17 2022-05-03 极限人工智能有限公司 Imaging method and imaging system based on endoscope
CN113925441A (en) * 2021-12-17 2022-01-14 极限人工智能有限公司 Imaging method and imaging system based on endoscope
CN114862961A (en) * 2022-04-13 2022-08-05 上海人工智能创新中心 Position detection method and device for calibration plate, electronic equipment and readable storage medium
CN114862961B (en) * 2022-04-13 2024-06-07 上海人工智能创新中心 Position detection method and device for calibration plate, electronic equipment and readable storage medium
CN114820787B (en) * 2022-04-22 2024-05-28 聊城大学 Image correction method and system for large-view-field plane vision measurement
CN114820787A (en) * 2022-04-22 2022-07-29 聊城大学 Image correction method and system for large-view-field planar vision measurement
CN114882058B (en) * 2022-04-26 2024-06-07 上海人工智能创新中心 Corner detection method, corner detection device and calibration plate
CN114882058A (en) * 2022-04-26 2022-08-09 上海人工智能创新中心 Angular point detection method and device and calibration board
CN114972531A (en) * 2022-05-17 2022-08-30 上海人工智能创新中心 Calibration board, corner detection method, equipment and readable storage medium
CN114926515B (en) * 2022-06-08 2024-05-14 北京化工大学 Infrared and visible light image registration method based on time-space domain depth information complementation
CN114926515A (en) * 2022-06-08 2022-08-19 北京化工大学 Infrared and visible light image registration method based on time-space domain depth information completion
WO2024022062A1 (en) * 2022-07-28 2024-02-01 杭州堃博生物科技有限公司 Endoscope pose estimation method and apparatus, and storage medium
CN117237553A (en) * 2023-09-14 2023-12-15 广东省核工业地质局测绘院 Three-dimensional map mapping system based on point cloud image fusion
CN117218089B (en) * 2023-09-18 2024-04-19 中南大学 Asphalt pavement structure depth detection method
CN117218089A (en) * 2023-09-18 2023-12-12 中南大学 Asphalt pavement structure depth detection method
CN117173342A (en) * 2023-11-02 2023-12-05 中国海洋大学 Underwater monocular and binocular camera-based natural light moving three-dimensional reconstruction device and method
CN117392329B (en) * 2023-12-08 2024-02-06 中国海洋大学 Use method based on mobile multispectral luminosity stereoscopic device
CN117392329A (en) * 2023-12-08 2024-01-12 中国海洋大学 Use method based on mobile multispectral luminosity stereoscopic device
CN117557660B (en) * 2024-01-09 2024-04-12 北京集度科技有限公司 Data processing method and device, electronic equipment and vehicle
CN117557660A (en) * 2024-01-09 2024-02-13 北京集度科技有限公司 Data processing method and device, electronic equipment and vehicle
CN117893693A (en) * 2024-03-15 2024-04-16 南昌航空大学 Dense SLAM three-dimensional scene reconstruction method and device
CN117893693B (en) * 2024-03-15 2024-05-28 南昌航空大学 Dense SLAM three-dimensional scene reconstruction method and device

Also Published As

Publication number Publication date
CN111145238A (en) 2020-05-12
CN111145238B (en) 2023-09-22

Similar Documents

Publication Publication Date Title
WO2021115071A1 (en) Three-dimensional reconstruction method and apparatus for monocular endoscope image, and terminal device
US7733404B2 (en) Fast imaging system calibration
WO2020259271A1 (en) Image distortion correction method and apparatus
US8447140B1 (en) Method and apparatus for estimating rotation, focal lengths and radial distortion in panoramic image stitching
US20190385285A1 (en) Image Processing Method and Device
CN110070598B (en) Mobile terminal for 3D scanning reconstruction and 3D scanning reconstruction method thereof
WO2021136386A1 (en) Data processing method, terminal, and server
WO2022095596A1 (en) Image alignment method, image alignment apparatus and terminal device
WO2021139176A1 (en) Pedestrian trajectory tracking method and apparatus based on binocular camera calibration, computer device, and storage medium
US20210295467A1 (en) Method for merging multiple images and post-processing of panorama
WO2019232793A1 (en) Two-camera calibration method, electronic device and computer-readable storage medium
Daftry et al. Flexible and User-Centric Camera Calibration using Planar Fiducial Markers.
CN112085771A (en) Image registration method and device, terminal equipment and computer readable storage medium
US20220405968A1 (en) Method, apparatus and system for image processing
CN111383264B (en) Positioning method, positioning device, terminal and computer storage medium
CN113516719B (en) Camera calibration method, system and storage medium based on multiple homography matrixes
JP6086491B2 (en) Image processing apparatus and database construction apparatus thereof
CN111260574B (en) Seal photo correction method, terminal and computer readable storage medium
CN112907657A (en) Robot repositioning method, device, equipment and storage medium
JP7312026B2 (en) Image processing device, image processing method and program
CN112102378A (en) Image registration method and device, terminal equipment and computer readable storage medium
JP2018036884A (en) Light source estimation device and program
WO2022174603A1 (en) Pose prediction method, pose prediction apparatus, and robot
Paudel et al. Localization of 2D cameras in a known environment using direct 2D-3D registration
JP3452188B2 (en) Tracking method of feature points in 2D video

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20897788

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20897788

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 20.01.2023)

122 Ep: pct application non-entry in european phase

Ref document number: 20897788

Country of ref document: EP

Kind code of ref document: A1