WO2020048484A1 - Super-resolution image reconstruction method and apparatus, and terminal and storage medium - Google Patents

Super-resolution image reconstruction method and apparatus, and terminal and storage medium Download PDF

Info

Publication number
WO2020048484A1
WO2020048484A1 PCT/CN2019/104388 CN2019104388W WO2020048484A1 WO 2020048484 A1 WO2020048484 A1 WO 2020048484A1 CN 2019104388 W CN2019104388 W CN 2019104388W WO 2020048484 A1 WO2020048484 A1 WO 2020048484A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
dimensional
dimensional model
target object
target
Prior art date
Application number
PCT/CN2019/104388
Other languages
French (fr)
Chinese (zh)
Inventor
方璐
戴琼海
李广涵
Original Assignee
清华-伯克利深圳学院筹备办公室
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 清华-伯克利深圳学院筹备办公室 filed Critical 清华-伯克利深圳学院筹备办公室
Publication of WO2020048484A1 publication Critical patent/WO2020048484A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects
    • G06T15/20Perspective computation
    • G06T15/205Image-based rendering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2215/00Indexing scheme for image rendering
    • G06T2215/06Curved planar reformation of 3D line structures

Definitions

  • the embodiments of the present application relate to the field of computational vision technology, for example, to a method, a device, a terminal, and a storage medium for super-resolution image reconstruction.
  • the accuracy of computer vision algorithms depends on the imaging quality of the input image or video. Therefore, the resolution of the input image or video needs to be improved.
  • the scene corresponding to an image or video contains two parts, static and dynamic, and the dynamic part contains rigid deformed objects and non-rigid deformed objects.
  • the shape and attitude of a rigidly deformed object will not change over time, you can directly use any frame of high-definition images to improve its resolution; for non-rigidly deformed objects, because their own shape and attitude will change over time, they cannot Increase the resolution of any frame of high-resolution images. Therefore, the difficulty in improving the accuracy of computer vision algorithms is to improve the resolution of non-rigidly deformed objects.
  • the related art there are mainly two methods for improving the resolution of a specific target object (ie, super-resolution reconstruction), one is a single-image super-resolution algorithm, and the other is a super-resolution algorithm based on a reference image.
  • super-resolution reconstruction when the input image and the training set are not similar, the single-image super-resolution algorithm cannot do a good super-resolution reconstruction of low-resolution input images with severe loss of detail, and all high-frequency details generated by this method are all low-frequency The information is generated and the authenticity is not high.
  • the reference image-based super-resolution algorithm needs to input the depth map of high-definition images. Although it can better complement high-frequency details, in practical applications, it is difficult to obtain high-definition depth images, and the algorithm is not universal.
  • the present application provides a super-resolution image reconstruction method, device, terminal, and storage medium to improve the resolution of non-rigid target objects in a low-resolution global image sequence.
  • an embodiment of the present application provides a super-resolution image reconstruction method.
  • the method includes: acquiring a first image of a target region in a current region at a first moment, and generating and the first image according to the first image.
  • At least one three-dimensional model corresponding to at least one first target object in the target region, wherein the first target object is a first non-rigid target object; obtaining a second image of the current region at a second time after the first time Extracting a third image corresponding to the target area from the second image, and updating the at least one three-dimensional model based on the third image; mapping the updated at least one three-dimensional model to at least one two Dimensional image, and stitching the at least one two-dimensional image into the second image to obtain a target super-resolution image.
  • an embodiment of the present application further provides a super-resolution image reconstruction apparatus, the apparatus includes: a three-dimensional model generating module configured to acquire a first image of a target region in a current region at a first moment, and The first image generates at least one three-dimensional model corresponding to at least one first target object in the target area, wherein the first target object is a first non-rigid target object; a three-dimensional model update module is configured to obtain the first target object; A second image of the current area at a second time after the first time, extracting a third image corresponding to the target area from the second image, and updating the at least one three-dimensional model based on the third image;
  • the super-resolution image acquisition module is configured to map the updated at least one three-dimensional model into at least one two-dimensional image, and stitch the at least one two-dimensional image into the second image to obtain a target super-resolution image.
  • an embodiment of the present application further provides a super-resolution image reconstruction terminal.
  • the terminal includes: one or more processors; and a storage device configured to store one or more programs.
  • the program is executed by the one or more processors, so that the one or more processors implement the super-resolution image reconstruction method according to any embodiment of the present application.
  • an embodiment of the present application further provides a computer-readable storage medium on which a computer program is stored.
  • the program is executed by a processor, the super-resolution image reconstruction method according to any one of the embodiments of the present application is implemented.
  • FIG. 1 is a flowchart of a super-resolution image reconstruction method according to an embodiment of the present application
  • FIG. 2 is a schematic structural diagram of a super-resolution image reconstruction device according to an embodiment of the present application
  • FIG. 3 is a schematic structural diagram of a super-resolution image reconstruction terminal in another embodiment of the present application.
  • FIG. 1 is a flowchart of a super-resolution image reconstruction method according to an embodiment of the present application. This embodiment is applicable to a case where the resolution of a non-rigid target object in a low-resolution global image sequence needs to be improved.
  • the image reconstruction apparatus performs the operation. As shown in FIG. 1, the method in this embodiment includes steps S110 to S130.
  • step S110 a first image of a target region in the current region at a first moment is obtained, and at least one three-dimensional model corresponding to at least one first target object in the target region is generated according to the first image, where the first target object Is the first non-rigid target object.
  • the target area may be an area containing at least one first target object.
  • the first target object may be a first non-rigid target object.
  • the non-rigid target object is an object whose shape and attitude can change with time.
  • a non-rigid target object may be a pedestrian.
  • the first image acquired at the first moment is a local image corresponding to the target region of the current region.
  • the first image can be acquired by a camera with a relatively small field of view.
  • the sharpness of the image is relatively high.
  • At least one two-dimensional image corresponding to the at least one first target object may be extracted from the first image. Based on the correspondence between the two-dimensional image and the three-dimensional model, the at least one two-dimensional image may be generated. At least one three-dimensional model corresponding to at least one first target object.
  • step S120 a second image of the current region at a second time after the first time is acquired, a third image corresponding to the target region is extracted from the second image, and at least one three-dimensional model is updated based on the third image.
  • the second image acquired at a second time after the first time is a global image corresponding to the current region, and the second image can use a relatively large field of view (compared to the camera that acquired the first image).
  • the camera acquires, correspondingly, the sharpness of the second image is relatively low.
  • the resolution of the camera that acquires the first image is the same as the resolution of the camera that acquires the second image, that is, the size of the first image and the second image are the same.
  • the second image Since the second image is acquired after the first image, the shape and posture of the first target object corresponding to the second image will be updated relative to the target object corresponding to the first image, so the second image can be used Update at least one three-dimensional model obtained by using the first image.
  • a third image corresponding to the target area may be extracted from the second image, and at least one may be updated with at least one two-dimensional image corresponding to at least one first target object extracted from the third image.
  • Three-dimensional model Three-dimensional model.
  • step S130 the updated at least one three-dimensional model is mapped into at least one two-dimensional image, and the at least one two-dimensional image is stitched into a second image to obtain a target super-resolution image.
  • super-resolution means increasing the resolution of the original image by means of hardware or software
  • super-resolution means the image after increasing the resolution.
  • mapping relationship between the three-dimensional model and the two-dimensional image. After obtaining the updated at least one three-dimensional model, the mapping relationship can be used to map the at least one three-dimensional model into at least one two-dimensional image.
  • the sharpness of the at least one two-dimensional image obtained by the mapping is equivalent to that of the first image, and is higher than the sharpness of the second image.
  • the at least one two-dimensional image obtained by the mapping is stitched into the second image by using an image stitching method , Using the at least one two-dimensional image with high definition to replace the corresponding low-resolution portion of the second image, and finally obtain the target super-resolution image.
  • the corresponding scenes are mainly static scenes and rigidly deformed objects.
  • rigidly deformed objects will move with time, but Since the shape and posture of the rigid deformation object will not change with time, in order to improve the overall resolution of the second image, the static scene of the second image and the corresponding part of the rigid deformation object in the first image can be directly stitched to The corresponding position of the second image to increase its resolution.
  • the at least one three-dimensional model generated using the first image includes both the shape and posture information of the at least one first target object and the texture information of the at least one first target object.
  • the system configured to obtain the first image and the second image may be a rotatable high-definition monitoring PTZ system.
  • the system may include a first-scale camera, a second-scale camera, and a rotatable PTZ There are three parts.
  • the first-scale camera is installed on the rotatable gimbal, and can rotate following the rotation of the gimbal.
  • the first scale camera may be a small field of view camera, configured to acquire a first image of the target area in the current area
  • the second scale camera may be a large field of view camera, configured to monitor the current area in real time and be able to continuously acquire the second area of the current area. image.
  • the resolution of the first-scale camera may be the same as that of the second-scale camera, and the size of the first image may be the same as that of the second image. Accordingly, the sharpness of the first image is higher than that of the second image. Clarity. Based on this, the resolution of the second image acquired at the second time may be improved by using the first image acquired at the first time based on the above scheme.
  • the super-resolution image reconstruction method obtains a first image of a target region in a current region at a first moment, and generates at least one three-dimensional model corresponding to at least one first target object in the target region according to the first image.
  • the first target object is a first non-rigid target object
  • a second image of the current region at a second time after the first time is obtained
  • a third image corresponding to the target region is extracted from the second image, and based on the first Three images update at least one three-dimensional model, map the updated at least one three-dimensional model to at least one two-dimensional image, and stitch at least one two-dimensional image into the second image to obtain the target super-resolution image, which improves the low-resolution global image Resolution of non-rigid target objects in the sequence.
  • generating at least one three-dimensional model corresponding to at least one first target object in the target area according to the first image includes: performing target object detection on the first image based on a preset target object detection method to obtain At least one first partial image corresponding to at least one first target object; using a preset two-dimensional pose point estimation method, two-dimensional pose point estimation is performed on each first partial image separately, and Each first two-dimensional pose point corresponding to the target object; For each first target object, the initial three-dimensional model is optimized using each first two-dimensional pose point to obtain a three-dimensional model corresponding to the first target object; for each three-dimensional model , Using the texture information in the corresponding first partial image to render the three-dimensional model to update the three-dimensional model.
  • At least one two-dimensional image corresponding to at least one target object may be acquired.
  • a first target object in the first image may be detected by using a preset target object detection method to obtain at least one first partial image corresponding to at least one first target object.
  • a first target object corresponds to a first partial image, and each first partial image can be labeled from the first image by using a square area.
  • the preset target object detection method may be a faster-rcnn detection algorithm, which has a high detection accuracy and a fast operation speed.
  • the faster-rcnn detection algorithm uses a deep learning method to propose an RPN network structure.
  • the convolutional neural network outputs two branches, one branch is the parameters corresponding to all candidate regions: the center coordinates x and y of the region, and the length and width of the region w. , H; The other branch is the probability that the candidate area is the first target object. Based on the two branches of the convolutional neural network output, the specific position of the at least one first target object in the first image can be determined to determine the position of the at least one first partial image.
  • the posture information of each first target object may be determined using each first partial image.
  • a preset two-dimensional pose point estimation method may be used to separately perform two-dimensional pose point estimation on each first partial image to obtain each first two-dimensional pose point corresponding to each first target object.
  • the preset two-dimensional pose point estimation method may be Openpose, which uses deep learning methods to predict each first partial image separately to obtain the two-dimensional pose points of all first target objects in each first partial image. , And then divide all the two-dimensional pose points according to the characteristics of the first target object, and finally determine the two-dimensional pose points corresponding to each first target object.
  • the first two-dimensional pose points obtained by predicting each first partial image by using Openpose are corresponding to each first target object. The first two-dimensional pose point.
  • an initial three-dimensional model can be constructed by using initialization parameters.
  • the decibels use each first two-dimensional attitude point to initialize the
  • the three-dimensional model is optimized to obtain a three-dimensional model corresponding to the first target object.
  • the three-dimensional model obtained by using the above method does not include the texture information of the first target object, and the two-dimensional image obtained by mapping the three-dimensional model does not include color information.
  • the three-dimensional model can be rendered using the texture information in the corresponding first partial image to update the three-dimensional model, so that the updated three-dimensional model contains both the shape and posture information of the first target object and the first Color information of the target object.
  • the initial three-dimensional model is optimized by using each first two-dimensional pose point to obtain a three-dimensional model corresponding to the first target object.
  • the method includes: The initial shape factor matrix ⁇ and the initial attitude angle vector ⁇ are used to construct an initial three-dimensional model.
  • the initial camera model parameter matrix K is used to perform a two-dimensional mapping on the initial three-dimensional model to obtain the initial two-dimensional attitude points corresponding to the initial three-dimensional model.
  • a target object Calculate a shape factor matrix ⁇ 1 and a first attitude angle vector ⁇ 1 that satisfy a preset condition, where the preset condition is a pair of matching points of each first two-dimensional pose point and each initial two-dimensional pose point.
  • the initial three-dimensional model is optimized by using the shape factor matrix ⁇ 1 and the first attitude angle vector ⁇ 1 to obtain a three-dimensional model corresponding to the first target object.
  • a 3D model consists of dense point clouds in a 3D space.
  • an initial three-dimensional model may be constructed based on a preset three-dimensional model construction method, an initial shape factor matrix ⁇ , and an initial attitude angle vector ⁇ .
  • an SMPLily algorithm may be used to construct the initial three-dimensional model. Taking the first target object as a human body as an example, the SMPLily algorithm constructs a three-dimensional model using a three-dimensional model of the SMPL human body, a shape factor matrix ⁇ , and an attitude angle vector ⁇ .
  • the three-dimensional human body model includes 6890 three-dimensional points and 24 three-dimensional joint points, of which 24 The three-dimensional joint points are used to control the position of the entire three-dimensional model point cloud, and then control the attitude of the three-dimensional model.
  • the shape factor matrix ⁇ controls the characteristics of the three-dimensional model such as height, fat, and thinness. The angle of rotation of the position of this point in the model.
  • Each of the 6,890 three-dimensional points in the three-dimensional model can be represented by a linear weighted average of 24 attitude angle vectors.
  • the initial camera model parameter matrix K can be used to perform two-dimensional mapping on the 24 three-dimensional joint points in the initial three-dimensional model. Each initial two-dimensional pose point corresponding to the three-dimensional model.
  • the shape factor matrix ⁇ 1 and The initial attitude angle vector ⁇ 1 is optimized by using the shape factor matrix ⁇ 1 and the initial attitude angle vector ⁇ 1 to obtain a three-dimensional model corresponding to the first target object.
  • a shape factor matrix ⁇ 1 and a first attitude angle vector ⁇ 1 that satisfy a preset condition may be calculated, where the preset conditions are each first two-dimensional pose point and each The sum of the differences between the pair of matching points of the initial two-dimensional pose points is the smallest, and the shape factor matrix ⁇ 1 is the smallest; the shape factor matrix ⁇ 1 and the first attitude angle vector ⁇ 1 are used to optimize the initial three-dimensional model to obtain A three-dimensional model corresponding to the first target object.
  • the method further includes: calculating a camera model parameter matrix that satisfies a preset condition.
  • Each three-dimensional model uses the texture information in the corresponding first partial image to render the three-dimensional model to update the three-dimensional model, including: for each three-dimensional model: using the camera model parameter matrix K1, converting the texture information in the corresponding first partial image Map to the 3D model to update the 3D model.
  • the camera model parameter matrix K may also be used to render texture information on the three-dimensional models.
  • a camera model parameter matrix K 1 that satisfies a preset condition is calculated, where the preset condition is a difference between each first two-dimensional pose point and each matching point pair of each initial two-dimensional pose point. The sum is minimal, and the form factor matrix ⁇ 1 is minimal.
  • the model using a camera parameter matrix K1 mapping texture information corresponding to a first partial image to the three-dimensional model, in order to update the three-dimensional model.
  • the method further includes: using a preset interpolation algorithm to map the three-dimensional model obtained.
  • the texture information is interpolated to obtain the texture information of the complete 3D model.
  • the first partial image that provides texture information for the three-dimensional model. Because the first partial image is a two-dimensional image, when the texture information in the first partial image is mapped to the three-dimensional model, the three-dimensional model is There must be some 3D points where texture information cannot be obtained, and these 3D points may include 3D points that can enter the field of view; in addition, when mapping a 3D model to a 2D image, it is only necessary to use the 3D model Able to enter three-dimensional points in the field of view. Therefore, texture information interpolation processing can be performed on the three-dimensional points that can enter the field of view and cannot obtain texture information, so that when the three-dimensional model is mapped to a two-dimensional image, complete texture information can be obtained. In an embodiment, a bilinear interpolation algorithm may be used to perform interpolation processing on the texture information of the mapped three-dimensional model to obtain the complete three-dimensional model texture information.
  • updating at least one three-dimensional model based on the third image includes: performing target object detection on the third image based on a preset target object detection method to obtain a one-to-one correspondence with at least one second target object in the target area.
  • At least one second partial image wherein the second target object is a second non-rigid target object; matching each first partial image with each second partial image to obtain a match between the at least one first partial image and the second partial image Pair to determine at least one three-dimensional model corresponding to a second target object in at least one second partial image; using a preset two-dimensional pose point estimation method to perform two-dimensional pose point estimation on each second partial image to obtain Each second two-dimensional pose point corresponding to each second target object; for each second target object, the second three-dimensional model corresponding to the second target object is updated using each second two-dimensional pose point.
  • the second target object may be the first target object after its shape and posture are changed.
  • the method for obtaining the second partial image is the same as the method for obtaining the first partial image, and also uses a faster-rcnn detection algorithm. After using the faster-rcnn detection algorithm to obtain at least one second partial image, the image matching algorithm is used to match each first partial image with each second partial image to obtain each second partial image that matches each first partial image. Since each first partial image corresponds to a three-dimensional model, based on each first partial image, each three-dimensional model corresponding to a second target object in each second partial image can be determined.
  • Each three-dimensional model corresponding to each second partial image determined by using the foregoing steps is determined using each first partial image. Therefore, the pose information of each three-dimensional model corresponds to the pose information of the first target object in each first partial image.
  • the posture information of each three-dimensional model may be updated using the posture information of each second target object.
  • a preset two-dimensional pose point estimation method may be used to separately perform two-dimensional pose point estimation on each second partial image to obtain each second two-dimensional pose point corresponding to each second target object. And for each second target object, using each second two-dimensional pose point to update the three-dimensional model corresponding to the second target object.
  • the preset two-dimensional pose point estimation method may be Openpose. The process of obtaining the second two-dimensional pose point by using Openpose is the same as the process of obtaining the first two-dimensional pose point by using Openpose.
  • updating the three-dimensional model corresponding to the second target object using each second two-dimensional pose point includes: for each second target object: using a preset deep learning algorithm , Each second two-dimensional attitude point is converted into a second attitude angle vector ⁇ 2 ; the shape factor matrix ⁇ 1 and the second attitude angle vector ⁇ 2 are used to update the three-dimensional model corresponding to the second target object to obtain the second target The 3D model corresponding to the object.
  • the three-dimensional model corresponding to the first partial image is obtained by using the shape factor matrix and the attitude angle vector optimization, the three-dimensional model can also be updated by using the updated two parameters, and because the target object is determined, the shape factor matrix There will be no change, so the 3D model can be updated with the updated attitude angle vector.
  • the second three-dimensional model corresponding to the second target object is updated by using each second two-dimensional pose point. After obtaining the second two-dimensional pose points corresponding to each second target object, the Set a deep learning algorithm to convert each second two-dimensional attitude point into a second attitude angle vector ⁇ 2 , and use the shape factor matrix ⁇ 1 and the second attitude angle vector ⁇ 2 to update the three-dimensional model corresponding to the second target object.
  • a three-dimensional model corresponding to the second target object is obtained.
  • the deep learning method is based on a deep residual network, and uses a combination of the most basic linear layer, RELU activation function, and reasonable network parameters to finally obtain the second attitude angle vector ⁇ 2 .
  • matching each first partial image with each second partial image to obtain at least one matching pair of the first partial image and the second partial image includes: determining each of the first partial image and each of the second partial image. The center point of the partial image; for each second partial image: Calculate the Euclidean distance between the center point of the second partial image and the center point of each first partial image; use the first partial image that minimizes the Euclidean distance as the second Matching pairs of local images.
  • the image matching algorithm that matches each first partial image with each second partial image may be a method of determining the center points of each of the first partial images and each of the second partial images, wherein the method of determining the center points is It can be the average value of the horizontal and vertical coordinates of the four vertices in the square area.
  • mapping the updated at least one three-dimensional model into at least one two-dimensional image includes: using the camera model parameter matrix K 1 to map the updated at least one three-dimensional model into at least one two-dimensional image.
  • FIG. 2 is a schematic structural diagram of a super-resolution image reconstruction device according to an embodiment of the present application.
  • the super-resolution image reconstruction device of this embodiment includes a three-dimensional model generation module 210, a three-dimensional model update module 220, and a super-resolution image acquisition module 230.
  • the three-dimensional model generating module 210 is configured to acquire a first image of a target region in the current region at a first moment, and generate at least one three-dimensional model corresponding to at least one first target object in the target region according to the first image.
  • a target object is a first non-rigid target object.
  • the three-dimensional model update module 220 is configured to obtain a second image of the current region at a second time after the first time, extract a third image corresponding to the target region from the second image, and update at least one three-dimensional model based on the third image. .
  • the super-resolution image acquisition module 230 is configured to map the updated at least one three-dimensional model into at least one two-dimensional image, and stitch the at least one two-dimensional image into a second image to obtain a target super-resolution image.
  • the super-resolution image reconstruction device obtains a first image of a target region in a current region at a first moment through a three-dimensional model generation module, and generates a first image corresponding to at least one first target object in the target region according to the first image.
  • At least one three-dimensional model wherein the first target object is a first non-rigid target object, and a three-dimensional model update module is used to obtain a second image of the current region at a second time after the first time, and the target region is extracted from the second image.
  • a corresponding third image and update at least one three-dimensional model based on the third image, and use the super-resolution image acquisition module to map the updated at least one three-dimensional model to at least one two-dimensional image, and stitch the at least one two-dimensional image to the first In the two images, the target super-resolution image is obtained, which improves the resolution of non-rigid target objects in the low-resolution global image sequence.
  • the three-dimensional model generation module 210 may include a first partial image acquisition sub-module, a first two-dimensional pose point acquisition sub-module, a three-dimensional model determination sub-module, and a texture information rendering sub-module.
  • the first partial image acquisition sub-module is configured to perform target object detection on the first image based on a preset target object detection method to obtain at least one first partial image corresponding to at least one first target object.
  • the first two-dimensional pose point acquisition sub-module is configured to use a preset two-dimensional pose point estimation method to perform two-dimensional pose point estimation on each first partial image respectively, and obtain each first partial object corresponding to each first target object. A two-dimensional pose point.
  • the three-dimensional model determination sub-module is set to optimize the initial three-dimensional model with each first two-dimensional pose point for each first target object to obtain a three-dimensional model corresponding to the first target object.
  • the texture information rendering sub-module is configured to, for each three-dimensional model, use the texture information in the corresponding first partial image to render the three-dimensional model to update the three-dimensional model.
  • the three-dimensional model determination sub-module may include an initial three-dimensional model construction unit, an initial two-dimensional pose point acquisition unit, a parameter acquisition unit, and a three-dimensional model acquisition unit.
  • the initial three-dimensional model construction unit is configured to construct an initial three-dimensional model based on a preset three-dimensional model construction method, an initial shape factor matrix ⁇ , and an initial attitude angle vector ⁇ .
  • the initial two-dimensional pose point acquisition unit is set to perform two-dimensional mapping on the initial three-dimensional model by using the initial camera model parameter matrix K to obtain each initial two-dimensional pose point corresponding to the initial three-dimensional model.
  • a parameter acquisition unit is set for each first target object: calculating a form factor matrix ⁇ 1 and a first attitude angle vector ⁇ 1 that satisfy a preset condition, where the preset conditions are each first two-dimensional pose point and each initial The sum of the differences between the pairs of matching points of the two-dimensional pose points is the smallest, and the form factor matrix ⁇ 1 is the smallest.
  • the three-dimensional model acquisition unit is configured to optimize the initial three-dimensional model by using the shape factor matrix ⁇ 1 and the first attitude angle vector ⁇ 1 to obtain a three-dimensional model corresponding to the first target object.
  • the parameter obtaining unit may be further configured to calculate a camera model parameter matrix K 1 that satisfies a preset condition, where the preset condition is each matching of each first two-dimensional pose point and each initial two-dimensional pose point. The sum of the differences between the point pairs is minimal, and the form factor matrix ⁇ 1 is minimal.
  • the texture information rendering sub-module may be configured to: for each three-dimensional model: use the camera model parameter matrix K1 to map the texture information in the corresponding first partial image to the three-dimensional model to update the three-dimensional model.
  • the texture information rendering sub-module may be further configured to: for each three-dimensional model: after using the camera model parameter matrix K1 to map the texture information in the corresponding first partial image onto the three-dimensional model, a preset is adopted The interpolation algorithm interpolates the texture information of the mapped 3D model to obtain the texture coordinates of the complete 3D model.
  • the three-dimensional model update module 220 may include a second local image acquisition sub-module, a local image matching sub-module, a second two-dimensional pose point acquisition sub-module, and a three-dimensional model update sub-module.
  • the second partial image acquisition sub-module is configured to perform target object detection on the third image based on a preset target object detection method to obtain at least one second partial image corresponding to at least one second target object in the target area.
  • the second target object is a second non-rigid target object.
  • the partial image matching sub-module is configured to match each first partial image with each second partial image to obtain a matching pair of at least one first partial image and a second partial image, so as to determine a first pair of at least one second partial image.
  • the second two-dimensional pose point acquisition sub-module is configured to use a preset two-dimensional pose point estimation method to perform two-dimensional pose point estimation on each second partial image respectively, and obtain each first partial object corresponding to each second target object. Two-dimensional and two-dimensional pose points.
  • the three-dimensional model update sub-module is configured to update the three-dimensional model corresponding to the second target object with each second two-dimensional pose point for each second target object.
  • the local image matching sub-module may include an image center point determination unit, a European-style distance calculation unit, and a local image matching pair determination unit.
  • the image center point determination unit is configured to determine a center point of each of the first partial images and each of the second partial images.
  • the Euclidean distance calculation unit is set for each second partial image: respectively calculate the Euclidean distance between the center point of the second partial image and the center point of each first partial image.
  • the local image matching pair determination unit is configured to use the first partial image with the smallest Euclidean distance as the matching pair of the second partial image.
  • the three-dimensional model updating sub-module may include: for each second target object: a second attitude angle vector determination unit, configured to use a preset deep learning algorithm to convert each second two-dimensional attitude point into a first Two attitude angle vectors ⁇ 2 ; a three-dimensional model update unit configured to update the three-dimensional model corresponding to the second target object by using the shape factor matrix ⁇ 1 and the second attitude angle vector ⁇ 2 to obtain the three-dimensional model corresponding to the second target object .
  • the super-resolution image acquisition module 230 is configured to: use the camera model parameter matrix K 1 to map the updated at least one three-dimensional model into at least one two-dimensional image.
  • the super-resolution image reconstruction apparatus provided in the embodiment of the present application can execute the super-resolution image reconstruction method provided in any embodiment of the present application, and has corresponding function modules for executing the method.
  • FIG. 3 is a schematic structural diagram of a super-resolution image reconstruction terminal according to another embodiment of the present application.
  • FIG. 3 shows a block diagram of an exemplary super-resolution image reconstruction terminal 312 suitable for implementing the embodiments of the present application.
  • the super-resolution image reconstruction terminal 312 shown in FIG. 3 is merely an example, and should not impose any limitation on the functions and scope of use of the embodiments of the present application.
  • the super-resolution image reconstruction terminal 312 is expressed in the form of a general-purpose computing device.
  • the components of the super-resolution image reconstruction terminal 312 may include, but are not limited to, one or more processors 316, a memory 328, and a bus 318 connecting different system components (including the memory 328 and the processor 316).
  • the bus 318 represents one or more of several types of bus structures, including a memory bus or a memory controller, a peripheral bus, a graphics acceleration port, a processor, or a local bus using any of a variety of bus structures.
  • these architectures include, but are not limited to, the Industry Standard Architecture (ISA) bus, the Micro Channel Architecture (MAC) bus, the enhanced ISA bus, and the Video Electronics Standards Association (Vedio Electronics Standard). Association (VESA) local area bus and Peripheral Component Interconnect (PCI) bus.
  • the super-resolution image reconstruction terminal 312 typically includes a variety of computer system-readable media. These media can be any available media that can be accessed by the super-resolution image reconstruction terminal 312, including volatile and non-volatile media, removable and non-removable media.
  • the memory 328 may include a computer system readable medium in the form of volatile memory, such as Random Access Memory (RAM) 330 and / or cache memory 332.
  • the super-resolution image reconstruction terminal 312 may include other removable / non-removable, volatile / nonvolatile computer system storage media.
  • the storage device 334 may be configured to read and write non-removable, non-volatile magnetic media (not shown in FIG. 3 and is commonly referred to as a “hard drive”).
  • a disk drive for reading and writing to a removable non-volatile disk such as a “floppy disk”
  • a read-only storage for a removable non-volatile optical disk such as a compact optical disk
  • each drive may be connected to the bus 318 through one or more data medium interfaces.
  • the memory 328 may include at least one program product having a set (eg, at least one) of program modules configured to perform the functions of the embodiments of the present application.
  • a program / utility tool 340 having a set (at least one) of program modules 342 may be stored in, for example, the memory 328.
  • Such program modules 342 include, but are not limited to, an operating system, one or more application programs, other program modules, and program data Each of these examples, or some combination, may include an implementation of a network environment.
  • the program module 342 generally performs functions and / or methods in the embodiments described in this application.
  • the super-resolution image reconstruction terminal 312 may also communicate with one or more external devices 314 (such as a keyboard, a pointing device, a display 324, etc., where the display 324 can decide whether to configure it according to actual needs), and may also communicate with one or more users
  • external devices 314 such as a keyboard, a pointing device, a display 324, etc., where the display 324 can decide whether to configure it according to actual needs
  • This communication can be performed through an input / output (I / O) interface 322.
  • the super-resolution image reconstruction terminal 312 may also communicate with one or more networks (such as a local area network (LAN), a wide area network (WAN), and / or a public network, such as the Internet) through the network adapter 320. As shown, the network adapter 320 communicates with other modules of the super-resolution image reconstruction terminal 312 through the bus 318. It should be understood that although not shown in FIG. 3, other hardware and / or software modules may be used in combination with the super-resolution image reconstruction terminal 312, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID Systems, tape drives, and data backup storage devices.
  • the processor 316 executes various functional applications and data processing by running a program stored in the memory 328, for example, implementing a super-resolution image reconstruction method provided by any embodiment of the present application.
  • An embodiment of the present application further provides a computer-readable storage medium on which a computer program is stored.
  • the program is executed by a processor, the super-resolution image reconstruction method provided by the embodiment of the present application is implemented.
  • the method includes: A first image of a target region in the region at a first moment, and at least one three-dimensional model corresponding to at least one first target object in the target region is generated according to the first image, where the first target object is a first non-rigid target object ; Obtain a second image of the current area at a second time after the first time, extract a third image corresponding to the target area from the second image, and update at least one three-dimensional model based on the third image; The three-dimensional model is mapped into at least one two-dimensional image, and the at least one two-dimensional image is stitched into a second image to obtain a target super-resolution image.
  • the computer-readable storage medium provided in the embodiment of the present application is not limited to the method operations described above, and the computer program stored on the computer program may also be implemented in the super-resolution image reconstruction method provided by any embodiment of the present application. Related operations.
  • the computer storage medium in the embodiments of the present application may adopt any combination of one or more computer-readable media.
  • the computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium.
  • the computer-readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in combination with an instruction execution system, apparatus, or device.
  • the computer-readable signal medium may include a data signal in baseband or propagated as part of a carrier wave, which carries a computer-readable program code. Such a propagated data signal may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • the computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, and the computer-readable medium may send, propagate, or transmit a program for use by or in connection with an instruction execution system, apparatus, or device .
  • the program code contained on the computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wire, optical fiber cable, radio frequency (RF), etc., or any suitable combination of the foregoing.
  • any appropriate medium including but not limited to wireless, wire, optical fiber cable, radio frequency (RF), etc., or any suitable combination of the foregoing.
  • Computer program code for performing the operations of this application may be written in one or more programming languages, or a combination thereof, including programming languages such as Java, Smalltalk, C ++, and also conventional Procedural programming language—such as "C" or similar programming language.
  • the program code can be executed entirely on the user's computer, partly on the user's computer, as an independent software package, partly on the user's computer, partly on a remote computer, or entirely on a remote computer or server.
  • the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (such as through an Internet service provider) Internet connection).
  • LAN local area network
  • WAN wide area network
  • Internet service provider Internet service provider

Abstract

Disclosed in embodiments of the present application are a super-resolution image reconstruction method and device, and a terminal and a storage medium. The super-resolution image reconstruction method comprises: obtaining a first image of a target region in a current region at a first time, and generating at least one three-dimensional model corresponding to at least one first target object in the target region according to the first image, wherein the first target object is a first non-rigid target object; obtaining a second image of the current region at a second time after the first time, extracting a third image corresponding to the target region from the second image, and updating at least one three-dimensional model based on the third image; and mapping at least one updated three-dimensional model into at least one two-dimensional image, and splicing the at least one two-dimensional image into the second image to obtain a target super-resolution image.

Description

超分辨图像重建方法、装置、终端和存储介质Super-resolution image reconstruction method, device, terminal and storage medium
本申请要求在2018年09月04日提交中国专利局、申请号为201811027057.5的中国专利申请的优先权,该申请的全部内容通过引用结合在本申请中。This application claims priority from a Chinese patent application filed with the Chinese Patent Office on September 4, 2018, with application number 201811027057.5, the entire contents of which are incorporated herein by reference.
技术领域Technical field
本申请实施例涉及计算视觉技术领域,例如涉及一种超分辨图像重建方法、装置、终端和存储介质。The embodiments of the present application relate to the field of computational vision technology, for example, to a method, a device, a terminal, and a storage medium for super-resolution image reconstruction.
背景技术Background technique
计算机视觉算法的准确性依赖于输入图像或视频的成像质量,因此,需要提高输入图像或视频的分辨率。通常图像或视频对应的场景包含静态和动态两部分,而动态部分中又包含刚性形变物体和非刚性形变物体。其中,由于刚性形变物体自身形状以及姿态不会随时间发生改变,因此可以直接利用任意帧高清图像提高其分辨率;而非刚性形变物体,由于其自身形状以及姿态会随着时间发生改变,无法利用任意帧高清图像提高其分辨率。因此提高计算机视觉算法的准确性的难点在于提高非刚性形变物体的分辨率。The accuracy of computer vision algorithms depends on the imaging quality of the input image or video. Therefore, the resolution of the input image or video needs to be improved. Usually the scene corresponding to an image or video contains two parts, static and dynamic, and the dynamic part contains rigid deformed objects and non-rigid deformed objects. Among them, since the shape and attitude of a rigidly deformed object will not change over time, you can directly use any frame of high-definition images to improve its resolution; for non-rigidly deformed objects, because their own shape and attitude will change over time, they cannot Increase the resolution of any frame of high-resolution images. Therefore, the difficulty in improving the accuracy of computer vision algorithms is to improve the resolution of non-rigidly deformed objects.
相关技术中的提高特定目标物体分辨率(即超分辨重建)的方法主要有两种,一种是单图像超分辨算法,一种是基于参考图像的超分辨算法。其中,当输入图像和训练集不相似时,单图像超分辨算法对于细节损失严重的低清输入图像无法做到很好的超分辨重建,且这种方法生成的全部高频细节全是由低频信息生成,真实性不高。而基于参考图像的超分辨算法需要输入高清图像的深度图,虽然可以更好地补充高频细节,但是在实际应用中,高清深度图像难以获取,算法的普适性较差。In the related art, there are mainly two methods for improving the resolution of a specific target object (ie, super-resolution reconstruction), one is a single-image super-resolution algorithm, and the other is a super-resolution algorithm based on a reference image. Among them, when the input image and the training set are not similar, the single-image super-resolution algorithm cannot do a good super-resolution reconstruction of low-resolution input images with severe loss of detail, and all high-frequency details generated by this method are all low-frequency The information is generated and the authenticity is not high. The reference image-based super-resolution algorithm needs to input the depth map of high-definition images. Although it can better complement high-frequency details, in practical applications, it is difficult to obtain high-definition depth images, and the algorithm is not universal.
发明内容Summary of the Invention
本申请提供一种超分辨图像重建方法、装置、终端和存储介质,以提高低清全局图像序列中非刚性目标物体的分辨率。The present application provides a super-resolution image reconstruction method, device, terminal, and storage medium to improve the resolution of non-rigid target objects in a low-resolution global image sequence.
第一方面,本申请实施例提供了一种超分辨图像重建方法,所述方法包括:获取当前区域中的目标区域在第一时刻的第一图像,并根据所述第一图像生成 与所述目标区域中至少一个第一目标物体对应的至少一个三维模型,其中,所述第一目标物体为第一非刚性目标物体;获取所述当前区域在第一时刻之后的第二时刻的第二图像,从所述第二图像中提取出所述目标区域对应的第三图像,并基于所述第三图像更新所述至少一个三维模型;将更新后的所述至少一个三维模型映射为至少一个二维图像,并将所述至少一个二维图像拼接至所述第二图像中,得到目标超分辨图像。In a first aspect, an embodiment of the present application provides a super-resolution image reconstruction method. The method includes: acquiring a first image of a target region in a current region at a first moment, and generating and the first image according to the first image. At least one three-dimensional model corresponding to at least one first target object in the target region, wherein the first target object is a first non-rigid target object; obtaining a second image of the current region at a second time after the first time Extracting a third image corresponding to the target area from the second image, and updating the at least one three-dimensional model based on the third image; mapping the updated at least one three-dimensional model to at least one two Dimensional image, and stitching the at least one two-dimensional image into the second image to obtain a target super-resolution image.
第二方面,本申请实施例还提供了一种超分辨图像重建装置,所述装置包括:三维模型生成模块,设置为获取当前区域中的目标区域在第一时刻的第一图像,并根据所述第一图像生成与所述目标区域中至少一个第一目标物体对应的至少一个三维模型,其中,所述第一目标物体为第一非刚性目标物体;三维模型更新模块,设置为获取所述当前区域在第一时刻之后的第二时刻的第二图像,从所述第二图像中提取出所述目标区域对应的第三图像,并基于所述第三图像更新所述至少一个三维模型;超分辨图像获取模块,设置为将更新后的所述至少一个三维模型映射为至少一个二维图像,并将所述至少一个二维图像拼接至所述第二图像中,得到目标超分辨图像。In a second aspect, an embodiment of the present application further provides a super-resolution image reconstruction apparatus, the apparatus includes: a three-dimensional model generating module configured to acquire a first image of a target region in a current region at a first moment, and The first image generates at least one three-dimensional model corresponding to at least one first target object in the target area, wherein the first target object is a first non-rigid target object; a three-dimensional model update module is configured to obtain the first target object; A second image of the current area at a second time after the first time, extracting a third image corresponding to the target area from the second image, and updating the at least one three-dimensional model based on the third image; The super-resolution image acquisition module is configured to map the updated at least one three-dimensional model into at least one two-dimensional image, and stitch the at least one two-dimensional image into the second image to obtain a target super-resolution image.
第三方面,本申请实施例还提供了一种超分辨图像重建终端,所述终端包括:一个或多个处理器;存储装置,设置为存储一个或多个程序,当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如本申请任一实施例所述的超分辨图像重建方法。According to a third aspect, an embodiment of the present application further provides a super-resolution image reconstruction terminal. The terminal includes: one or more processors; and a storage device configured to store one or more programs. The program is executed by the one or more processors, so that the one or more processors implement the super-resolution image reconstruction method according to any embodiment of the present application.
第四方面,本申请实施例还提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现如本申请任一实施例所述的超分辨图像重建方法。According to a fourth aspect, an embodiment of the present application further provides a computer-readable storage medium on which a computer program is stored. When the program is executed by a processor, the super-resolution image reconstruction method according to any one of the embodiments of the present application is implemented.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1是本申请一实施例中的超分辨图像重建方法的流程图;1 is a flowchart of a super-resolution image reconstruction method according to an embodiment of the present application;
图2是本申请一实施例中的超分辨图像重建装置的结构示意图;2 is a schematic structural diagram of a super-resolution image reconstruction device according to an embodiment of the present application;
图3是本申请另一实施例中的超分辨图像重建终端的结构示意图。FIG. 3 is a schematic structural diagram of a super-resolution image reconstruction terminal in another embodiment of the present application.
具体实施方式detailed description
下面结合附图和实施例对本申请作详细说明。可以理解的是,此处所描述 的具体实施例仅仅用于解释本申请,而非对本申请的限定。另外还需要说明的是,为了便于描述,附图中仅示出了与本申请相关的部分而非全部结构。The following describes the present application in detail with reference to the drawings and embodiments. It can be understood that the specific embodiments described herein are only used to explain the present application, rather than limiting the present application. In addition, it should be noted that, for convenience of description, only some parts related to the present application are shown in the drawings instead of the entire structure.
图1为本申请一实施例提供的一种超分辨图像重建方法的流程图,本实施例可适用于需要提高低清全局图像序列中非刚性目标物体的分辨率的情况,该方法可以由超分辨图像重建装置来执行,如图1所示,本实施例的方法包括步骤S110至步骤S130。FIG. 1 is a flowchart of a super-resolution image reconstruction method according to an embodiment of the present application. This embodiment is applicable to a case where the resolution of a non-rigid target object in a low-resolution global image sequence needs to be improved. The image reconstruction apparatus performs the operation. As shown in FIG. 1, the method in this embodiment includes steps S110 to S130.
在步骤S110中,获取当前区域中的目标区域在第一时刻的第一图像,并根据第一图像生成与目标区域中至少一个第一目标物体对应的至少一个三维模型,其中,第一目标物体为第一非刚性目标物体。In step S110, a first image of a target region in the current region at a first moment is obtained, and at least one three-dimensional model corresponding to at least one first target object in the target region is generated according to the first image, where the first target object Is the first non-rigid target object.
本实施例中,目标区域可以是包含至少一个第一目标物体的区域,第一目标物体可以为第一非刚性目标物体,非刚性目标物体即为自身形状以及姿态可以随时间发生改变的物体,例如非刚性目标物体可以是行人。在第一时刻获取到的第一图像是当前区域的目标区域所对应的局部图像,该第一图像可以利用视场角相对较小的相机获取,相对应的,该图像的清晰度也相对较高。In this embodiment, the target area may be an area containing at least one first target object. The first target object may be a first non-rigid target object. The non-rigid target object is an object whose shape and attitude can change with time. For example, a non-rigid target object may be a pedestrian. The first image acquired at the first moment is a local image corresponding to the target region of the current region. The first image can be acquired by a camera with a relatively small field of view. Correspondingly, the sharpness of the image is relatively high.
在一实施例中,可以从第一图像中提取出与至少一个第一目标物体对应的至少一个二维图像,基于二维图像与三维模型之间的对应关系,可以利用至少一个二维图像生成与至少一个第一目标物体对应的至少一个三维模型。In an embodiment, at least one two-dimensional image corresponding to the at least one first target object may be extracted from the first image. Based on the correspondence between the two-dimensional image and the three-dimensional model, the at least one two-dimensional image may be generated. At least one three-dimensional model corresponding to at least one first target object.
在步骤S120中,获取当前区域在第一时刻之后的第二时刻的第二图像,从第二图像中提取出目标区域对应的第三图像,并基于第三图像更新至少一个三维模型。In step S120, a second image of the current region at a second time after the first time is acquired, a third image corresponding to the target region is extracted from the second image, and at least one three-dimensional model is updated based on the third image.
其中,在第一时刻之后的第二时刻获取到的第二图像是当前区域所对应的全局图像,该第二图像可以利用视场角相对较大(与获取第一图像的相机相比)的相机获取,相对应的,该第二图像的清晰度也相对较低。本实施例中,在一实施例中,获取第一图像的相机的分辨率与获取第二图像的相机的分辨率相同,即第一图像与第二图像的尺寸相同。The second image acquired at a second time after the first time is a global image corresponding to the current region, and the second image can use a relatively large field of view (compared to the camera that acquired the first image). The camera acquires, correspondingly, the sharpness of the second image is relatively low. In this embodiment, in an embodiment, the resolution of the camera that acquires the first image is the same as the resolution of the camera that acquires the second image, that is, the size of the first image and the second image are the same.
由于第二图像是在第一图像之后获取到的,因此,第二图像对应的第一目标物体的自身形状以及姿态相对于第一图像对应的目标物体会有所更新,因此可以利用第二图像更新上述利用第一图像获取到的至少一个三维模型。在一实施例中,可以从第二图像中提取出与目标区域对应的第三图像,并利用从第三 图像中提取出的与至少一个第一目标物体对应的至少一个二维图像更新至少一个三维模型。Since the second image is acquired after the first image, the shape and posture of the first target object corresponding to the second image will be updated relative to the target object corresponding to the first image, so the second image can be used Update at least one three-dimensional model obtained by using the first image. In an embodiment, a third image corresponding to the target area may be extracted from the second image, and at least one may be updated with at least one two-dimensional image corresponding to at least one first target object extracted from the third image. Three-dimensional model.
在步骤S130中,将更新后的至少一个三维模型映射为至少一个二维图像,并将至少一个二维图像拼接至第二图像中,得到目标超分辨图像。In step S130, the updated at least one three-dimensional model is mapped into at least one two-dimensional image, and the at least one two-dimensional image is stitched into a second image to obtain a target super-resolution image.
其中,超分辨即通过硬件或软件的方法提高原有图像的分辨率,超分辨图像即提高分辨率之后的图像。三维模型与二维图像之间存在对应的映射关系,在获取到更新后的至少一个三维模型之后,可以利用该映射关系将至少一个三维模型映射为至少一个二维图像。上述映射得到的至少一个二维图像的清晰度与第一图像的清晰度相当,并高于第二图像的清晰度,利用图像拼接方法将映射得到的至少一个二维图像拼接至第二图像中,以利用该清晰度高的至少一个二维图像替代第二图像中相应的清晰度低的部分,最终得到目标超分辨图像。Among them, super-resolution means increasing the resolution of the original image by means of hardware or software, and super-resolution means the image after increasing the resolution. There is a corresponding mapping relationship between the three-dimensional model and the two-dimensional image. After obtaining the updated at least one three-dimensional model, the mapping relationship can be used to map the at least one three-dimensional model into at least one two-dimensional image. The sharpness of the at least one two-dimensional image obtained by the mapping is equivalent to that of the first image, and is higher than the sharpness of the second image. The at least one two-dimensional image obtained by the mapping is stitched into the second image by using an image stitching method , Using the at least one two-dimensional image with high definition to replace the corresponding low-resolution portion of the second image, and finally obtain the target super-resolution image.
在此需要说明的是,对于第二图像中没有利用至少一个二维图像进行图像拼接的其他部分,其对应的场景主要是静态场景以及刚性形变物体,虽然刚性形变物体会随时间发生移动,但是由于刚性形变物体自身形状以及姿态不会随时间发生改变,因此,为了提高第二图像整体的分辨率,可以直接将第一图像中与第二图像的静态场景以及刚性形变物体对应的部分拼接至第二图像相应的位置,以提高其分辨率。此外,利用第一图像生成的至少一个三维模型既包括至少一个第一目标物体的形状姿态信息,又包括至少一个第一目标物体的纹理信息。It should be noted here that for other parts of the second image that do not use at least one two-dimensional image for image stitching, the corresponding scenes are mainly static scenes and rigidly deformed objects. Although rigidly deformed objects will move with time, but Since the shape and posture of the rigid deformation object will not change with time, in order to improve the overall resolution of the second image, the static scene of the second image and the corresponding part of the rigid deformation object in the first image can be directly stitched to The corresponding position of the second image to increase its resolution. In addition, the at least one three-dimensional model generated using the first image includes both the shape and posture information of the at least one first target object and the texture information of the at least one first target object.
示例性的,设置为获取第一图像和第二图像的系统可以是可转动高清监控云台系统,在一实施例中,该系统可以包括第一尺度相机、第二尺度相机、可转动云台三部分,其中,第一尺度相机安装在可转动云台上,能够跟随云台的转动而转动。第一尺度相机可以为小视场相机,设置为获取当前区域中的目标区域的第一图像,第二尺度相机可以为大视场相机,设置为实时监测当前区域并能够连续获取当前区域的第二图像。且第一尺度相机的分辨率可以与第二尺度相机的分辨率相同,第一图像的尺寸大小可以与第二图像的尺寸大小相同,相应的,第一图像的清晰度高于第二图像的清晰度。基于此,可以基于上述方案,利用在第一时刻获取的第一图像提高在第二时刻获取到的第二图像的分辨率。Exemplarily, the system configured to obtain the first image and the second image may be a rotatable high-definition monitoring PTZ system. In an embodiment, the system may include a first-scale camera, a second-scale camera, and a rotatable PTZ There are three parts. Among them, the first-scale camera is installed on the rotatable gimbal, and can rotate following the rotation of the gimbal. The first scale camera may be a small field of view camera, configured to acquire a first image of the target area in the current area, and the second scale camera may be a large field of view camera, configured to monitor the current area in real time and be able to continuously acquire the second area of the current area. image. And the resolution of the first-scale camera may be the same as that of the second-scale camera, and the size of the first image may be the same as that of the second image. Accordingly, the sharpness of the first image is higher than that of the second image. Clarity. Based on this, the resolution of the second image acquired at the second time may be improved by using the first image acquired at the first time based on the above scheme.
本实施例提供的超分辨图像重建方法,通过获取当前区域中的目标区域在 第一时刻的第一图像,并根据第一图像生成与目标区域中至少一个第一目标物体对应的至少一个三维模型,其中,第一目标物体为第一非刚性目标物体,获取当前区域在第一时刻之后的第二时刻的第二图像,从第二图像中提取出目标区域对应的第三图像,并基于第三图像更新至少一个三维模型,将更新后的至少一个三维模型映射为至少一个二维图像,并将至少一个二维图像拼接至第二图像中,得到目标超分辨图像,提高了低清全局图像序列中非刚性目标物体的分辨率。The super-resolution image reconstruction method provided by this embodiment obtains a first image of a target region in a current region at a first moment, and generates at least one three-dimensional model corresponding to at least one first target object in the target region according to the first image. Where the first target object is a first non-rigid target object, a second image of the current region at a second time after the first time is obtained, a third image corresponding to the target region is extracted from the second image, and based on the first Three images update at least one three-dimensional model, map the updated at least one three-dimensional model to at least one two-dimensional image, and stitch at least one two-dimensional image into the second image to obtain the target super-resolution image, which improves the low-resolution global image Resolution of non-rigid target objects in the sequence.
在上述实施例的基础上,根据第一图像生成与目标区域中至少一个第一目标物体对应的至少一个三维模型,包括:基于预设目标物体检测方法,对第一图像进行目标物体检测,得到与至少一个第一目标物体一一对应的至少一个第一局部图像;利用预设二维姿态点预估方法,分别对各第一局部图像进行二维姿态点预估,得到与每个第一目标物体对应的各第一二维姿态点;针对每个第一目标物体,利用各第一二维姿态点对初始三维模型进行优化,得到与第一目标物体对应的三维模型;针对各三维模型,分别利用相应的第一局部图像中的纹理信息渲染三维模型,以更新三维模型。Based on the above embodiment, generating at least one three-dimensional model corresponding to at least one first target object in the target area according to the first image includes: performing target object detection on the first image based on a preset target object detection method to obtain At least one first partial image corresponding to at least one first target object; using a preset two-dimensional pose point estimation method, two-dimensional pose point estimation is performed on each first partial image separately, and Each first two-dimensional pose point corresponding to the target object; For each first target object, the initial three-dimensional model is optimized using each first two-dimensional pose point to obtain a three-dimensional model corresponding to the first target object; for each three-dimensional model , Using the texture information in the corresponding first partial image to render the three-dimensional model to update the three-dimensional model.
本实施例中,在生成至少一个三维模型之前,可以先获取与至少一个目标物体对应的至少一个二维图像。在一实施例中,可以利用预设的目标物体检测方法对第一图像中的第一目标物体进行检测,得到与至少一个第一目标物体一一对应的至少一个第一局部图像。其中,一个第一目标物体对应一个第一局部图像,且每个第一局部图像可以利用方形区域从第一图像中进行标示。本实施例中,预设目标物体检测方法可以是faster-rcnn检测算法,该检测算法检测精度高且运算速度快。faster-rcnn检测算法利用深度学习方法,提出了RPN网络结构,在卷积神经网络输出两个分支,一个分支是全部候选区域对应的参数:分别是区域中心坐标x、y,区域的长宽w、h;另一个分支则是候选区域是第一目标物体的概率。基于卷积神经网络输出的两个分支即可确定至少一个第一目标物体在第一图像中的具体位置,以确定至少一个第一局部图像的位置。In this embodiment, before generating at least one three-dimensional model, at least one two-dimensional image corresponding to at least one target object may be acquired. In an embodiment, a first target object in the first image may be detected by using a preset target object detection method to obtain at least one first partial image corresponding to at least one first target object. A first target object corresponds to a first partial image, and each first partial image can be labeled from the first image by using a square area. In this embodiment, the preset target object detection method may be a faster-rcnn detection algorithm, which has a high detection accuracy and a fast operation speed. The faster-rcnn detection algorithm uses a deep learning method to propose an RPN network structure. The convolutional neural network outputs two branches, one branch is the parameters corresponding to all candidate regions: the center coordinates x and y of the region, and the length and width of the region w. , H; The other branch is the probability that the candidate area is the first target object. Based on the two branches of the convolutional neural network output, the specific position of the at least one first target object in the first image can be determined to determine the position of the at least one first partial image.
在得到至少一个第一局部图像之后,可以利用各第一局部图像确定各第一目标物体的姿态信息。在一实施例中,可以利用预设二维姿态点预估方法,分别对各第一局部图像进行二维姿态点预估,得到与每个第一目标物体对应的各第一二维姿态点。其中,预设二维姿态点预估方法可以是Openpose,该方法利 用深度学习方法,分别对各第一局部图像进行预测,得到每个第一局部图像中所有第一目标物体的二维姿态点,然后再根据第一目标物体的特征对所有二维姿态点进行划分,最终确定每个第一目标物体对应的二维姿态点。本实施例中,由于一个第一局部图像只包含一个第一目标物体,因此,利用Openpose分别对各第一局部图像进行预测得到的第一二维姿态点,即为各第一目标物体对应的第一二维姿态点。After obtaining at least one first partial image, the posture information of each first target object may be determined using each first partial image. In an embodiment, a preset two-dimensional pose point estimation method may be used to separately perform two-dimensional pose point estimation on each first partial image to obtain each first two-dimensional pose point corresponding to each first target object. . The preset two-dimensional pose point estimation method may be Openpose, which uses deep learning methods to predict each first partial image separately to obtain the two-dimensional pose points of all first target objects in each first partial image. , And then divide all the two-dimensional pose points according to the characteristics of the first target object, and finally determine the two-dimensional pose points corresponding to each first target object. In this embodiment, since a first partial image includes only a first target object, the first two-dimensional pose points obtained by predicting each first partial image by using Openpose are corresponding to each first target object. The first two-dimensional pose point.
本实施例中,在生成至少一个第一目标物体对应的至少一个三维模型之前,可以利用初始化参数构建一个初始三维模型,针对每个第一目标物体,分贝利用各第一二维姿态点对初始三维模型进行优化,得到与第一目标物体对应的三维模型。利用上述方法得到的三维模型并不包含第一目标物体的纹理信息,由该三维模型映射得到的二维图像也不包含颜色信息。因此,针对各三维模型,可以分别利用相应的第一局部图像中的纹理信息渲染三维模型,以更新三维模型,使得更新后的三维模型既包含第一目标物体的形状姿态信息,又包含第一目标物体的颜色信息。In this embodiment, before generating at least one three-dimensional model corresponding to at least one first target object, an initial three-dimensional model can be constructed by using initialization parameters. For each first target object, the decibels use each first two-dimensional attitude point to initialize the The three-dimensional model is optimized to obtain a three-dimensional model corresponding to the first target object. The three-dimensional model obtained by using the above method does not include the texture information of the first target object, and the two-dimensional image obtained by mapping the three-dimensional model does not include color information. Therefore, for each three-dimensional model, the three-dimensional model can be rendered using the texture information in the corresponding first partial image to update the three-dimensional model, so that the updated three-dimensional model contains both the shape and posture information of the first target object and the first Color information of the target object.
在一实施例中,针对每个第一目标物体,利用各第一二维姿态点对初始三维模型进行优化,得到与第一目标物体对应的三维模型,包括:基于预设三维模型构建方法、初始形状因子矩阵β和初始姿态角向量θ构建初始三维模型;利用初始相机模型参数矩阵K对初始三维模型进行二维映射,得到与初始三维模型对应的各初始二维姿态点;针对每个第一目标物体:计算满足预设条件的形状因子矩阵β 1和第一姿态角向量θ 1,其中,预设条件为各第一二维姿态点与各初始二维姿态点的各匹配点对之间的差值的加和最小,且形状因子矩阵β 1最小;利用形状因子矩阵β 1和第一姿态角向量θ 1对初始三维模型进行优化,得到与第一目标物体对应的三维模型。 In one embodiment, for each first target object, the initial three-dimensional model is optimized by using each first two-dimensional pose point to obtain a three-dimensional model corresponding to the first target object. The method includes: The initial shape factor matrix β and the initial attitude angle vector θ are used to construct an initial three-dimensional model. The initial camera model parameter matrix K is used to perform a two-dimensional mapping on the initial three-dimensional model to obtain the initial two-dimensional attitude points corresponding to the initial three-dimensional model. A target object: Calculate a shape factor matrix β 1 and a first attitude angle vector θ 1 that satisfy a preset condition, where the preset condition is a pair of matching points of each first two-dimensional pose point and each initial two-dimensional pose point. The sum of the differences between them is the smallest, and the shape factor matrix β 1 is the smallest; the initial three-dimensional model is optimized by using the shape factor matrix β 1 and the first attitude angle vector θ 1 to obtain a three-dimensional model corresponding to the first target object.
一般的,三维模型由三维空间上的密集点云组成。本实施例中,可以基于预设三维模型构建方法、初始形状因子矩阵β和初始姿态角向量θ构建初始三维模型,在一实施例中,可以利用SMPLily算法来构建初始三维模型。以第一目标物体为人体为例,SMPLily算法利用SMPL人体三维模型、形状因子矩阵β以及姿态角向量θ构建三维模型,该人体三维模型包括6890个三维点以及24个三维关节点,其中,24个三维关节点用于控制整个三维模型点云的位置,进而控制三维模型的姿态,形状因子矩阵β控制三维模型的高矮、胖瘦等特征结构,姿 态角向量θ由三维关节点相对于初始三维模型中此点的位置所转动的角度来表示。三维模型中的6890个三维点中的每个点,都可以用24个姿态角向量进行线性加权平均表示。在利用SMPLily算法、初始形状因子矩阵β和初始姿态角向量θ得到初始三维模型之后,可以利用初始相机模型参数矩阵K,对初始三维模型中的24个三维关节点进行二维映射,得到与初始三维模型对应的各初始二维姿态点。Generally, a 3D model consists of dense point clouds in a 3D space. In this embodiment, an initial three-dimensional model may be constructed based on a preset three-dimensional model construction method, an initial shape factor matrix β, and an initial attitude angle vector θ. In one embodiment, an SMPLily algorithm may be used to construct the initial three-dimensional model. Taking the first target object as a human body as an example, the SMPLily algorithm constructs a three-dimensional model using a three-dimensional model of the SMPL human body, a shape factor matrix β, and an attitude angle vector θ. The three-dimensional human body model includes 6890 three-dimensional points and 24 three-dimensional joint points, of which 24 The three-dimensional joint points are used to control the position of the entire three-dimensional model point cloud, and then control the attitude of the three-dimensional model. The shape factor matrix β controls the characteristics of the three-dimensional model such as height, fat, and thinness. The angle of rotation of the position of this point in the model. Each of the 6,890 three-dimensional points in the three-dimensional model can be represented by a linear weighted average of 24 attitude angle vectors. After using the SMPLily algorithm, the initial form factor matrix β, and the initial attitude angle vector θ to obtain the initial three-dimensional model, the initial camera model parameter matrix K can be used to perform two-dimensional mapping on the 24 three-dimensional joint points in the initial three-dimensional model. Each initial two-dimensional pose point corresponding to the three-dimensional model.
由于初始三维模型是利用初始形状因子矩阵β和初始姿态角向量θ确定的,因此,为了得到与第一目标物体对应的三维模型,可以先确定与第一目标物体对应的形状因子矩阵β 1和初始姿态角向量θ 1,利用形状因子矩阵β 1和初始姿态角向量θ 1对初始三维模型进行优化,以得到与第一目标物体对应的三维模型。在一实施例中,针对每个第一目标物体:可以计算满足预设条件的形状因子矩阵β 1和第一姿态角向量θ 1,其中,预设条件为各第一二维姿态点与各初始二维姿态点的各匹配点对之间的差值的加和最小,且形状因子矩阵β 1最小;利用形状因子矩阵β 1和第一姿态角向量θ 1对初始三维模型进行优化,得到与第一目标物体对应的三维模型。 Since the initial three-dimensional model is determined using the initial shape factor matrix β and the initial attitude angle vector θ, in order to obtain the three-dimensional model corresponding to the first target object, the shape factor matrix β 1 and The initial attitude angle vector θ 1 is optimized by using the shape factor matrix β 1 and the initial attitude angle vector θ 1 to obtain a three-dimensional model corresponding to the first target object. In an embodiment, for each first target object, a shape factor matrix β 1 and a first attitude angle vector θ 1 that satisfy a preset condition may be calculated, where the preset conditions are each first two-dimensional pose point and each The sum of the differences between the pair of matching points of the initial two-dimensional pose points is the smallest, and the shape factor matrix β 1 is the smallest; the shape factor matrix β 1 and the first attitude angle vector θ 1 are used to optimize the initial three-dimensional model to obtain A three-dimensional model corresponding to the first target object.
在一实施例中,针对每个第一目标物体:除了计算满足预设条件的形状因子矩阵β 1和第一姿态角向量θ 1之外,还包括:计算满足预设条件的相机模型参数矩阵K 1,其中,预设条件为各第一二维姿态点与各初始二维姿态点的各匹配点对之间的差值的加和最小,且形状因子矩阵β 1最小;相应的,针对各三维模型,分别利用相应的第一局部图像中的纹理信息渲染三维模型,以更新三维模型,包括:针对各三维模型:利用相机模型参数矩阵K1,将相应的第一局部图像中的纹理信息映射到三维模型上,以更新三维模型。 In an embodiment, for each first target object, in addition to calculating a shape factor matrix β 1 and a first attitude angle vector θ 1 that satisfy a preset condition, the method further includes: calculating a camera model parameter matrix that satisfies a preset condition. K 1 , where the preset condition is that the sum of the differences between each first two-dimensional pose point and each pair of matching points of each initial two-dimensional pose point is the smallest, and the shape factor matrix β 1 is the smallest; accordingly, for Each three-dimensional model uses the texture information in the corresponding first partial image to render the three-dimensional model to update the three-dimensional model, including: for each three-dimensional model: using the camera model parameter matrix K1, converting the texture information in the corresponding first partial image Map to the 3D model to update the 3D model.
在一实施例中,在得到与各第一目标物体对应的各三维模型之后,还可以利用相机模型参数矩阵K对各三维模型进行纹理信息渲染。在一实施例中,计算满足预设条件的相机模型参数矩阵K 1,其中,预设条件为各第一二维姿态点与各初始二维姿态点的各匹配点对之间的差值的加和最小,且形状因子矩阵β 1最小。在得到相机模型参数矩阵K 1之后,针对各三维模型,利用相机模型参数矩阵K1,将相应的第一局部图像中的纹理信息映射到三维模型上,以更新三维模型。 In one embodiment, after obtaining the three-dimensional models corresponding to the first target objects, the camera model parameter matrix K may also be used to render texture information on the three-dimensional models. In an embodiment, a camera model parameter matrix K 1 that satisfies a preset condition is calculated, where the preset condition is a difference between each first two-dimensional pose point and each matching point pair of each initial two-dimensional pose point. The sum is minimal, and the form factor matrix β 1 is minimal. After obtaining the camera model parameter matrix K 1, for each of the three-dimensional model, the model using a camera parameter matrix K1, mapping texture information corresponding to a first partial image to the three-dimensional model, in order to update the three-dimensional model.
在一实施例中,针对各三维模型:利用相机模型参数矩阵K1,将相应的第 一局部图像中的纹理信息映射到三维模型上之后,还包括:采用预设插值算法对映射得到的三维模型的纹理信息进行插值处理,以得到完整的三维模型的纹理信息。In one embodiment, for each three-dimensional model: after using the camera model parameter matrix K1 to map the texture information in the corresponding first partial image onto the three-dimensional model, the method further includes: using a preset interpolation algorithm to map the three-dimensional model obtained. The texture information is interpolated to obtain the texture information of the complete 3D model.
本实施例中,为三维模型提供纹理信息的是第一局部图像,由于第一局部图像为二维图像,因此,在将第一局部图像中的纹理信息映射到三维模型中时,三维模型中必然存在无法获取到纹理信息的部分三维点,而在这些三维点中可能包含能够进入视野范围内的三维点;此外,在将三维模型映射为二维图像时,也只需要用到三维模型中能够进入视野范围内的三维点。因此,可以对能够进入视野范围内且无法获取到纹理信息的三维点进行纹理信息插值处理,以便在将三维模型映射为二维图像时,能够得到完整的纹理信息。在一实施例中,可以采用双线性插值算法对映射得到的三维模型的纹理信息进行插值处理,以得到完整的三维模型的纹理信息。In this embodiment, it is the first partial image that provides texture information for the three-dimensional model. Because the first partial image is a two-dimensional image, when the texture information in the first partial image is mapped to the three-dimensional model, the three-dimensional model is There must be some 3D points where texture information cannot be obtained, and these 3D points may include 3D points that can enter the field of view; in addition, when mapping a 3D model to a 2D image, it is only necessary to use the 3D model Able to enter three-dimensional points in the field of view. Therefore, texture information interpolation processing can be performed on the three-dimensional points that can enter the field of view and cannot obtain texture information, so that when the three-dimensional model is mapped to a two-dimensional image, complete texture information can be obtained. In an embodiment, a bilinear interpolation algorithm may be used to perform interpolation processing on the texture information of the mapped three-dimensional model to obtain the complete three-dimensional model texture information.
在一实施例中,基于第三图像更新至少一个三维模型,包括:基于预设目标物体检测方法,对第三图像进行目标物体检测,得到与目标区域中至少一个第二目标物体一一对应的至少一个第二局部图像,其中,第二目标物体为第二非刚性目标物体;将各第一局部图像与各第二局部图像进行匹配,得到至少一个第一局部图像与第二局部图像的匹配对,以确定至少一个第二局部图像中的第二目标物体对应的至少一个三维模型;利用预设二维姿态点预估方法,分别对各第二局部图像进行二维姿态点预估,得到与每个第二目标物体对应的各第二二维姿态点;针对每个第二目标物体,利用各第二二维姿态点对第二目标物体对应的三维模型进行更新。In an embodiment, updating at least one three-dimensional model based on the third image includes: performing target object detection on the third image based on a preset target object detection method to obtain a one-to-one correspondence with at least one second target object in the target area. At least one second partial image, wherein the second target object is a second non-rigid target object; matching each first partial image with each second partial image to obtain a match between the at least one first partial image and the second partial image Pair to determine at least one three-dimensional model corresponding to a second target object in at least one second partial image; using a preset two-dimensional pose point estimation method to perform two-dimensional pose point estimation on each second partial image to obtain Each second two-dimensional pose point corresponding to each second target object; for each second target object, the second three-dimensional model corresponding to the second target object is updated using each second two-dimensional pose point.
本实施例中,第二目标物体可以是自身形状以及姿态发生变化后的第一目标物体。获取第二局部图像的方法与获取第一局部图像的方法相同,同样采用faster-rcnn检测算法。在利用faster-rcnn检测算法获取到至少一个第二局部图像之后,利用图像匹配算法,将各第一局部图像与各第二局部图像进行匹配,得到与各第一局部图像相匹配的各第二局部图像,由于各第一局部图像都对应一个三维模型,因此,基于各第一局部图像,可以确定与各第二局部图像中的第二目标物体对应的各三维模型。In this embodiment, the second target object may be the first target object after its shape and posture are changed. The method for obtaining the second partial image is the same as the method for obtaining the first partial image, and also uses a faster-rcnn detection algorithm. After using the faster-rcnn detection algorithm to obtain at least one second partial image, the image matching algorithm is used to match each first partial image with each second partial image to obtain each second partial image that matches each first partial image. Since each first partial image corresponds to a three-dimensional model, based on each first partial image, each three-dimensional model corresponding to a second target object in each second partial image can be determined.
利用上述步骤确定的与各第二局部图像对应的各三维模型是利用各第一局部图像确定的,因此,各三维模型的姿态信息对应各第一局部图像中的第一目 标物体的姿态信息,为了使各三维模型与各第二局部图像相匹配,可以利用各第二目标物体的姿态信息更新各三维模型的姿态信息。在一实施例中,可以利用预设二维姿态点预估方法,分别对各第二局部图像进行二维姿态点预估,得到与每个第二目标物体对应的各第二二维姿态点,并针对每个第二目标物体,利用各第二二维姿态点对第二目标物体对应的三维模型进行更新。其中,预设二维姿态点预估方法可以是Openpose,利用Openpose获取第二二维姿态点的过程与利用Openpose获取第一二维姿态点的过程相同。Each three-dimensional model corresponding to each second partial image determined by using the foregoing steps is determined using each first partial image. Therefore, the pose information of each three-dimensional model corresponds to the pose information of the first target object in each first partial image. In order to match each three-dimensional model with each second partial image, the posture information of each three-dimensional model may be updated using the posture information of each second target object. In an embodiment, a preset two-dimensional pose point estimation method may be used to separately perform two-dimensional pose point estimation on each second partial image to obtain each second two-dimensional pose point corresponding to each second target object. And for each second target object, using each second two-dimensional pose point to update the three-dimensional model corresponding to the second target object. The preset two-dimensional pose point estimation method may be Openpose. The process of obtaining the second two-dimensional pose point by using Openpose is the same as the process of obtaining the first two-dimensional pose point by using Openpose.
在一实施例中,针对每个第二目标物体,利用各第二二维姿态点对第二目标物体对应的三维模型进行更新,包括:针对每个第二目标物体:利用预设深度学习算法,将各第二二维姿态点转换为第二姿态角向量θ 2;利用形状因子矩阵β 1和第二姿态角向量θ 2对第二目标物体对应的三维模型进行更新,得到与第二目标物体对应的三维模型。 In an embodiment, for each second target object, updating the three-dimensional model corresponding to the second target object using each second two-dimensional pose point includes: for each second target object: using a preset deep learning algorithm , Each second two-dimensional attitude point is converted into a second attitude angle vector θ 2 ; the shape factor matrix β 1 and the second attitude angle vector θ 2 are used to update the three-dimensional model corresponding to the second target object to obtain the second target The 3D model corresponding to the object.
由于与第一局部图像对应的三维模型是利用形状因子矩阵和姿态角向量优化得到的,因此,同样可以利用更新后的上述两个参数对三维模型进行更新,又由于目标物体确定,形状因子矩阵不会发生变化,因此,可以利用更新后的姿态角向量对三维模型进行更新。在一实施例中,利用各第二二维姿态点对第二目标物体对应的三维模型进行更新,可以是在得到与每个第二目标物体对应的各第二二维姿态点后,利用预设深度学习算法,将各第二二维姿态点转换为第二姿态角向量θ 2,并利用形状因子矩阵β 1和第二姿态角向量θ 2对第二目标物体对应的三维模型进行更新,得到与第二目标物体对应的三维模型。其中,深度学习方法基于深度残差网络,使用最基本的线性层、RELU激活函数以及合理的网络参数的组合,最终实现对第二姿态角向量θ 2的获取。 Since the three-dimensional model corresponding to the first partial image is obtained by using the shape factor matrix and the attitude angle vector optimization, the three-dimensional model can also be updated by using the updated two parameters, and because the target object is determined, the shape factor matrix There will be no change, so the 3D model can be updated with the updated attitude angle vector. In one embodiment, the second three-dimensional model corresponding to the second target object is updated by using each second two-dimensional pose point. After obtaining the second two-dimensional pose points corresponding to each second target object, the Set a deep learning algorithm to convert each second two-dimensional attitude point into a second attitude angle vector θ 2 , and use the shape factor matrix β 1 and the second attitude angle vector θ 2 to update the three-dimensional model corresponding to the second target object. A three-dimensional model corresponding to the second target object is obtained. Among them, the deep learning method is based on a deep residual network, and uses a combination of the most basic linear layer, RELU activation function, and reasonable network parameters to finally obtain the second attitude angle vector θ 2 .
在一实施例中,将各第一局部图像与各第二局部图像进行匹配,得到至少一个第一局部图像与第二局部图像的匹配对,包括:分别确定各第一局部图像与各第二局部图像的中心点;针对各第二局部图像:分别计算第二局部图像的中心点与各第一局部图像的中心点之间的欧式距离;将使得欧式距离最小的第一局部图像作为第二局部图像的匹配对。In one embodiment, matching each first partial image with each second partial image to obtain at least one matching pair of the first partial image and the second partial image includes: determining each of the first partial image and each of the second partial image. The center point of the partial image; for each second partial image: Calculate the Euclidean distance between the center point of the second partial image and the center point of each first partial image; use the first partial image that minimizes the Euclidean distance as the second Matching pairs of local images.
本实施例中,将各第一局部图像与各第二局部图像进行匹配的图像匹配算法,可以是分别确定各第一局部图像与各第二局部图像的中心点,其中,确定中心点的方法可以是取方形区域四个顶点横纵坐标的平均值。在确定中心点之 后,针对各第二局部图像:分别计算第二局部图像的中心点与各第一局部图像的中心点之间的欧式距离,并比较各欧式距离的大小关系,最终将使得欧式距离最小的第一局部图像作为第二局部图像的匹配对。In this embodiment, the image matching algorithm that matches each first partial image with each second partial image may be a method of determining the center points of each of the first partial images and each of the second partial images, wherein the method of determining the center points is It can be the average value of the horizontal and vertical coordinates of the four vertices in the square area. After determining the center point, for each second partial image: Calculate the Euclidean distance between the center point of the second partial image and the center point of each first partial image, and compare the magnitude relationship of each Euclidean distance. The first partial image with the smallest distance is used as a matching pair of the second partial image.
在一实施例中,将更新后的至少一个三维模型映射为至少一个二维图像,包括:利用相机模型参数矩阵K 1,将更新后的至少一个三维模型映射为至少一个二维图像。 In an embodiment, mapping the updated at least one three-dimensional model into at least one two-dimensional image includes: using the camera model parameter matrix K 1 to map the updated at least one three-dimensional model into at least one two-dimensional image.
图2是本申请一实施例中的一种超分辨图像重建装置的结构示意图。如图2所示,本实施例的超分辨图像重建装置包括三维模型生成模块210,三维模型更新模块220和超分辨图像获取模块230。FIG. 2 is a schematic structural diagram of a super-resolution image reconstruction device according to an embodiment of the present application. As shown in FIG. 2, the super-resolution image reconstruction device of this embodiment includes a three-dimensional model generation module 210, a three-dimensional model update module 220, and a super-resolution image acquisition module 230.
三维模型生成模块210,设置为获取当前区域中的目标区域在第一时刻的第一图像,并根据第一图像生成与目标区域中至少一个第一目标物体对应的至少一个三维模型,其中,第一目标物体为第一非刚性目标物体。The three-dimensional model generating module 210 is configured to acquire a first image of a target region in the current region at a first moment, and generate at least one three-dimensional model corresponding to at least one first target object in the target region according to the first image. A target object is a first non-rigid target object.
三维模型更新模块220,设置为获取当前区域在第一时刻之后的第二时刻的第二图像,从第二图像中提取出目标区域对应的第三图像,并基于第三图像更新至少一个三维模型。The three-dimensional model update module 220 is configured to obtain a second image of the current region at a second time after the first time, extract a third image corresponding to the target region from the second image, and update at least one three-dimensional model based on the third image. .
超分辨图像获取模块230,设置为将更新后的至少一个三维模型映射为至少一个二维图像,并将至少一个二维图像拼接至第二图像中,得到目标超分辨图像。The super-resolution image acquisition module 230 is configured to map the updated at least one three-dimensional model into at least one two-dimensional image, and stitch the at least one two-dimensional image into a second image to obtain a target super-resolution image.
本实施例提供的超分辨图像重建装置,通过三维模型生成模块获取当前区域中的目标区域在第一时刻的第一图像,并根据第一图像生成与目标区域中至少一个第一目标物体对应的至少一个三维模型,其中,第一目标物体为第一非刚性目标物体,利用三维模型更新模块获取当前区域在第一时刻之后的第二时刻的第二图像,从第二图像中提取出目标区域对应的第三图像,并基于第三图像更新至少一个三维模型,并利用超分辨图像获取模块将更新后的至少一个三维模型映射为至少一个二维图像,并将至少一个二维图像拼接至第二图像中,得到目标超分辨图像,提高了低清全局图像序列中非刚性目标物体的分辨率。The super-resolution image reconstruction device provided in this embodiment obtains a first image of a target region in a current region at a first moment through a three-dimensional model generation module, and generates a first image corresponding to at least one first target object in the target region according to the first image. At least one three-dimensional model, wherein the first target object is a first non-rigid target object, and a three-dimensional model update module is used to obtain a second image of the current region at a second time after the first time, and the target region is extracted from the second image. A corresponding third image, and update at least one three-dimensional model based on the third image, and use the super-resolution image acquisition module to map the updated at least one three-dimensional model to at least one two-dimensional image, and stitch the at least one two-dimensional image to the first In the two images, the target super-resolution image is obtained, which improves the resolution of non-rigid target objects in the low-resolution global image sequence.
在上述实施例的基础上,三维模型生成模块210可以包括第一局部图像获取子模块,第一二维姿态点获取子模块,三维模型确定子模块以及纹理信息渲染子模块。Based on the above embodiments, the three-dimensional model generation module 210 may include a first partial image acquisition sub-module, a first two-dimensional pose point acquisition sub-module, a three-dimensional model determination sub-module, and a texture information rendering sub-module.
第一局部图像获取子模块,设置为基于预设目标物体检测方法,对第一图像进行目标物体检测,得到与至少一个第一目标物体一一对应的至少一个第一局部图像。The first partial image acquisition sub-module is configured to perform target object detection on the first image based on a preset target object detection method to obtain at least one first partial image corresponding to at least one first target object.
第一二维姿态点获取子模块,设置为利用预设二维姿态点预估方法,分别对各第一局部图像进行二维姿态点预估,得到与每个第一目标物体对应的各第一二维姿态点。The first two-dimensional pose point acquisition sub-module is configured to use a preset two-dimensional pose point estimation method to perform two-dimensional pose point estimation on each first partial image respectively, and obtain each first partial object corresponding to each first target object. A two-dimensional pose point.
三维模型确定子模块,设置为针对每个第一目标物体,利用各第一二维姿态点对初始三维模型进行优化,得到与第一目标物体对应的三维模型。The three-dimensional model determination sub-module is set to optimize the initial three-dimensional model with each first two-dimensional pose point for each first target object to obtain a three-dimensional model corresponding to the first target object.
纹理信息渲染子模块,设置为针对各三维模型,分别利用相应的第一局部图像中的纹理信息渲染三维模型,以更新三维模型。The texture information rendering sub-module is configured to, for each three-dimensional model, use the texture information in the corresponding first partial image to render the three-dimensional model to update the three-dimensional model.
在一实施例中,三维模型确定子模块可以包括初始三维模型构建单元,初始二维姿态点获取单元,参数获取单元以及三维模型获取单元。In an embodiment, the three-dimensional model determination sub-module may include an initial three-dimensional model construction unit, an initial two-dimensional pose point acquisition unit, a parameter acquisition unit, and a three-dimensional model acquisition unit.
初始三维模型构建单元,设置为基于预设三维模型构建方法、初始形状因子矩阵β和初始姿态角向量θ构建初始三维模型。The initial three-dimensional model construction unit is configured to construct an initial three-dimensional model based on a preset three-dimensional model construction method, an initial shape factor matrix β, and an initial attitude angle vector θ.
初始二维姿态点获取单元,设置为利用初始相机模型参数矩阵K对初始三维模型进行二维映射,得到与初始三维模型对应的各初始二维姿态点。The initial two-dimensional pose point acquisition unit is set to perform two-dimensional mapping on the initial three-dimensional model by using the initial camera model parameter matrix K to obtain each initial two-dimensional pose point corresponding to the initial three-dimensional model.
参数获取单元,设置为针对每个第一目标物体:计算满足预设条件的形状因子矩阵β 1和第一姿态角向量θ 1,其中,预设条件为各第一二维姿态点与各初始二维姿态点的各匹配点对之间的差值的加和最小,且形状因子矩阵β 1最小。 A parameter acquisition unit is set for each first target object: calculating a form factor matrix β 1 and a first attitude angle vector θ 1 that satisfy a preset condition, where the preset conditions are each first two-dimensional pose point and each initial The sum of the differences between the pairs of matching points of the two-dimensional pose points is the smallest, and the form factor matrix β 1 is the smallest.
三维模型获取单元,设置为利用形状因子矩阵β 1和第一姿态角向量θ 1对初始三维模型进行优化,得到与第一目标物体对应的三维模型。 The three-dimensional model acquisition unit is configured to optimize the initial three-dimensional model by using the shape factor matrix β 1 and the first attitude angle vector θ 1 to obtain a three-dimensional model corresponding to the first target object.
在一实施例中,参数获取单元还可以设置为:计算满足预设条件的相机模型参数矩阵K 1,其中,预设条件为各第一二维姿态点与各初始二维姿态点的各匹配点对之间的差值的加和最小,且形状因子矩阵β 1最小。相应的,纹理信息渲染子模块可以设置为:针对各三维模型:利用相机模型参数矩阵K1,将相应的第一局部图像中的纹理信息映射到三维模型上,以更新三维模型。 In an embodiment, the parameter obtaining unit may be further configured to calculate a camera model parameter matrix K 1 that satisfies a preset condition, where the preset condition is each matching of each first two-dimensional pose point and each initial two-dimensional pose point. The sum of the differences between the point pairs is minimal, and the form factor matrix β 1 is minimal. Correspondingly, the texture information rendering sub-module may be configured to: for each three-dimensional model: use the camera model parameter matrix K1 to map the texture information in the corresponding first partial image to the three-dimensional model to update the three-dimensional model.
在一实施例中,纹理信息渲染子模块还可以设置为:针对各三维模型:在利用相机模型参数矩阵K1,将相应的第一局部图像中的纹理信息映射到三维模型上之后,采用预设插值算法对映射得到的三维模型的纹理信息进行插值处理,以得到完整的三维模型的纹理坐标。In an embodiment, the texture information rendering sub-module may be further configured to: for each three-dimensional model: after using the camera model parameter matrix K1 to map the texture information in the corresponding first partial image onto the three-dimensional model, a preset is adopted The interpolation algorithm interpolates the texture information of the mapped 3D model to obtain the texture coordinates of the complete 3D model.
在一实施例中,三维模型更新模块220可以包括第二局部图像获取子模块,局部图像匹配子模块,第二二维姿态点获取子模块以及三维模型更新子模块。In one embodiment, the three-dimensional model update module 220 may include a second local image acquisition sub-module, a local image matching sub-module, a second two-dimensional pose point acquisition sub-module, and a three-dimensional model update sub-module.
第二局部图像获取子模块,设置为基于预设目标物体检测方法,对第三图像进行目标物体检测,得到与目标区域中至少一个第二目标物体一一对应的至少一个第二局部图像,其中,第二目标物体为第二非刚性目标物体。The second partial image acquisition sub-module is configured to perform target object detection on the third image based on a preset target object detection method to obtain at least one second partial image corresponding to at least one second target object in the target area. , The second target object is a second non-rigid target object.
局部图像匹配子模块,设置为将各第一局部图像与各第二局部图像进行匹配,得到至少一个第一局部图像与第二局部图像的匹配对,以确定至少一个第二局部图像中的第二目标物体对应的至少一个三维模型。The partial image matching sub-module is configured to match each first partial image with each second partial image to obtain a matching pair of at least one first partial image and a second partial image, so as to determine a first pair of at least one second partial image. At least one three-dimensional model corresponding to two target objects.
第二二维姿态点获取子模块,设置为利用预设二维姿态点预估方法,分别对各第二局部图像进行二维姿态点预估,得到与每个第二目标物体对应的各第二二维姿态点。The second two-dimensional pose point acquisition sub-module is configured to use a preset two-dimensional pose point estimation method to perform two-dimensional pose point estimation on each second partial image respectively, and obtain each first partial object corresponding to each second target object. Two-dimensional and two-dimensional pose points.
三维模型更新子模块,设置为针对每个第二目标物体,利用各第二二维姿态点对第二目标物体对应的三维模型进行更新。The three-dimensional model update sub-module is configured to update the three-dimensional model corresponding to the second target object with each second two-dimensional pose point for each second target object.
在一实施例中,局部图像匹配子模块可以包括图像中心点确定单元,欧式距离计算单元以及局部图像匹配对确定单元。In one embodiment, the local image matching sub-module may include an image center point determination unit, a European-style distance calculation unit, and a local image matching pair determination unit.
图像中心点确定单元,设置为分别确定各第一局部图像与各第二局部图像的中心点。The image center point determination unit is configured to determine a center point of each of the first partial images and each of the second partial images.
欧式距离计算单元,设置为针对各第二局部图像:分别计算第二局部图像的中心点与各第一局部图像的中心点之间的欧式距离。The Euclidean distance calculation unit is set for each second partial image: respectively calculate the Euclidean distance between the center point of the second partial image and the center point of each first partial image.
局部图像匹配对确定单元,设置为将使得欧式距离最小的第一局部图像作为第二局部图像的匹配对。The local image matching pair determination unit is configured to use the first partial image with the smallest Euclidean distance as the matching pair of the second partial image.
在一实施例中,三维模型更新子模块可以包括:针对每个第二目标物体:第二姿态角向量确定单元,设置为利用预设深度学习算法,将各第二二维姿态点转换为第二姿态角向量θ 2;三维模型更新单元,设置为利用形状因子矩阵β 1和第二姿态角向量θ 2对第二目标物体对应的三维模型进行更新,得到与第二目标物体对应的三维模型。 In an embodiment, the three-dimensional model updating sub-module may include: for each second target object: a second attitude angle vector determination unit, configured to use a preset deep learning algorithm to convert each second two-dimensional attitude point into a first Two attitude angle vectors θ 2 ; a three-dimensional model update unit configured to update the three-dimensional model corresponding to the second target object by using the shape factor matrix β 1 and the second attitude angle vector θ 2 to obtain the three-dimensional model corresponding to the second target object .
在一实施例中,超分辨图像获取模块230设置为:利用相机模型参数矩阵K 1,将更新后的至少一个三维模型映射为至少一个二维图像。 In one embodiment, the super-resolution image acquisition module 230 is configured to: use the camera model parameter matrix K 1 to map the updated at least one three-dimensional model into at least one two-dimensional image.
本申请实施例所提供的超分辨图像重建装置可执行本申请任意实施例所提供的超分辨图像重建方法,具备执行方法相应的功能模块。The super-resolution image reconstruction apparatus provided in the embodiment of the present application can execute the super-resolution image reconstruction method provided in any embodiment of the present application, and has corresponding function modules for executing the method.
图3为本申请另一实施例提供的超分辨图像重建终端的结构示意图。图3示出了适于用来实现本申请实施方式的示例性超分辨图像重建终端312的框图。图3显示的超分辨图像重建终端312仅仅是一个示例,不应对本申请实施例的功能和使用范围带来任何限制。FIG. 3 is a schematic structural diagram of a super-resolution image reconstruction terminal according to another embodiment of the present application. FIG. 3 shows a block diagram of an exemplary super-resolution image reconstruction terminal 312 suitable for implementing the embodiments of the present application. The super-resolution image reconstruction terminal 312 shown in FIG. 3 is merely an example, and should not impose any limitation on the functions and scope of use of the embodiments of the present application.
如图3所示,超分辨图像重建终端312以通用计算设备的形式表现。超分辨图像重建终端312的组件可以包括但不限于:一个或者多个处理器316,存储器328,连接不同系统组件(包括存储器328和处理器316)的总线318。As shown in FIG. 3, the super-resolution image reconstruction terminal 312 is expressed in the form of a general-purpose computing device. The components of the super-resolution image reconstruction terminal 312 may include, but are not limited to, one or more processors 316, a memory 328, and a bus 318 connecting different system components (including the memory 328 and the processor 316).
总线318表示几类总线结构中的一种或多种,包括存储器总线或者存储器控制器,外围总线,图形加速端口,处理器或者使用多种总线结构中的任意总线结构的局域总线。举例来说,这些体系结构包括但不限于工业标准体系结构(Industry Standard Architecture,ISA)总线,微通道体系结构(Micro Channel Architecture,MAC)总线,增强型ISA总线、视频电子标准协会(Vedio Electronic Standard Association,VESA)局域总线以及外围组件互连(Peripheral Component Interconnect,PCI)总线。The bus 318 represents one or more of several types of bus structures, including a memory bus or a memory controller, a peripheral bus, a graphics acceleration port, a processor, or a local bus using any of a variety of bus structures. For example, these architectures include, but are not limited to, the Industry Standard Architecture (ISA) bus, the Micro Channel Architecture (MAC) bus, the enhanced ISA bus, and the Video Electronics Standards Association (Vedio Electronics Standard). Association (VESA) local area bus and Peripheral Component Interconnect (PCI) bus.
超分辨图像重建终端312典型地包括多种计算机系统可读介质。这些介质可以是任何能够被超分辨图像重建终端312访问的可用介质,包括易失性和非易失性介质,可移动的和不可移动的介质。The super-resolution image reconstruction terminal 312 typically includes a variety of computer system-readable media. These media can be any available media that can be accessed by the super-resolution image reconstruction terminal 312, including volatile and non-volatile media, removable and non-removable media.
存储器328可以包括易失性存储器形式的计算机系统可读介质,例如随机存取存储器(Random Access Memory,RAM)330和/或高速缓存存储器332。超分辨图像重建终端312可以包括其它可移动/不可移动的、易失性/非易失性计算机系统存储介质。仅作为举例,存储装置334可以设置为读写不可移动的、非易失性磁介质(图3未显示,通常称为“硬盘驱动器”)。尽管图3中未示出,可以提供用于对可移动非易失性磁盘(例如“软盘”)读写的磁盘驱动器,以及对可移动非易失性光盘(例如紧凑型光盘只读储存器(Compact Disc Read-Only Memory,CD-ROM),数字视盘(Digital Video Disc-Read Only Memory,DVD-ROM)或者其它光介质)读写的光盘驱动器。在这些情况下,每个驱动器可以通过一个或者多个数据介质接口与总线318相连。存储器328可以包括至少一个程序产品,该程序产品具有一组(例如至少一个)程序模块,这些程序模块被配置以执行本申请各实施例的功能。The memory 328 may include a computer system readable medium in the form of volatile memory, such as Random Access Memory (RAM) 330 and / or cache memory 332. The super-resolution image reconstruction terminal 312 may include other removable / non-removable, volatile / nonvolatile computer system storage media. For example only, the storage device 334 may be configured to read and write non-removable, non-volatile magnetic media (not shown in FIG. 3 and is commonly referred to as a “hard drive”). Although not shown in FIG. 3, a disk drive for reading and writing to a removable non-volatile disk (such as a “floppy disk”) and a read-only storage for a removable non-volatile optical disk (such as a compact optical disk) may be provided. (Compact Disc-Read-Only Memory (CD-ROM), Digital Video Disc (Read-Only Memory, DVD-ROM) or other optical media). In these cases, each drive may be connected to the bus 318 through one or more data medium interfaces. The memory 328 may include at least one program product having a set (eg, at least one) of program modules configured to perform the functions of the embodiments of the present application.
具有一组(至少一个)程序模块342的程序/实用工具340,可以存储在例如存储器328中,这样的程序模块342包括但不限于操作系统、一个或者多个应用程序、其它程序模块以及程序数据,这些示例中的每一个或某种组合中可能包括网络环境的实现。程序模块342通常执行本申请所描述的实施例中的功能和/或方法。A program / utility tool 340 having a set (at least one) of program modules 342 may be stored in, for example, the memory 328. Such program modules 342 include, but are not limited to, an operating system, one or more application programs, other program modules, and program data Each of these examples, or some combination, may include an implementation of a network environment. The program module 342 generally performs functions and / or methods in the embodiments described in this application.
超分辨图像重建终端312也可以与一个或多个外部设备314(例如键盘、指向设备、显示器324等,其中,显示器324可根据实际需要决定是否配置)通信,还可与一个或者多个使得用户能与该超分辨图像重建终端312交互的设备通信,和/或与使得该超分辨图像重建终端312能与一个或多个其它计算设备进行通信的任何设备(例如网卡,调制解调器等等)通信。这种通信可以通过输入/输出(Input/Output,I/O)接口322进行。并且,超分辨图像重建终端312还可以通过网络适配器320与一个或者多个网络(例如局域网(Local Area Network,LAN),广域网(Wide Area Network,WAN)和/或公共网络,例如因特网)通信。如图所示,网络适配器320通过总线318与超分辨图像重建终端312的其它模块通信。应当明白,尽管图3中未示出,可以结合超分辨图像重建终端312使用其它硬件和/或软件模块,包括但不限于:微代码、设备驱动器、冗余处理单元、外部磁盘驱动阵列、RAID系统、磁带驱动器以及数据备份存储装置等。The super-resolution image reconstruction terminal 312 may also communicate with one or more external devices 314 (such as a keyboard, a pointing device, a display 324, etc., where the display 324 can decide whether to configure it according to actual needs), and may also communicate with one or more users A device capable of interacting with the super-resolution image reconstruction terminal 312 and / or any device (such as a network card, modem, etc.) that enables the super-resolution image reconstruction terminal 312 to communicate with one or more other computing devices. This communication can be performed through an input / output (I / O) interface 322. In addition, the super-resolution image reconstruction terminal 312 may also communicate with one or more networks (such as a local area network (LAN), a wide area network (WAN), and / or a public network, such as the Internet) through the network adapter 320. As shown, the network adapter 320 communicates with other modules of the super-resolution image reconstruction terminal 312 through the bus 318. It should be understood that although not shown in FIG. 3, other hardware and / or software modules may be used in combination with the super-resolution image reconstruction terminal 312, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID Systems, tape drives, and data backup storage devices.
处理器316通过运行存储在存储器328中的程序,从而执行各种功能应用以及数据处理,例如实现本申请任意实施例所提供的超分辨图像重建方法。The processor 316 executes various functional applications and data processing by running a program stored in the memory 328, for example, implementing a super-resolution image reconstruction method provided by any embodiment of the present application.
本申请实施例还提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现如本申请实施例所提供的超分辨图像重建方法,该方法包括:获取当前区域中的目标区域在第一时刻的第一图像,并根据第一图像生成与目标区域中至少一个第一目标物体对应的至少一个三维模型,其中,第一目标物体为第一非刚性目标物体;获取当前区域在第一时刻之后的第二时刻的第二图像,从第二图像中提取出目标区域对应的第三图像,并基于第三图像更新至少一个三维模型;将更新后的至少一个三维模型映射为至少一个二维图像,并将至少一个二维图像拼接至第二图像中,得到目标超分辨图像。An embodiment of the present application further provides a computer-readable storage medium on which a computer program is stored. When the program is executed by a processor, the super-resolution image reconstruction method provided by the embodiment of the present application is implemented. The method includes: A first image of a target region in the region at a first moment, and at least one three-dimensional model corresponding to at least one first target object in the target region is generated according to the first image, where the first target object is a first non-rigid target object ; Obtain a second image of the current area at a second time after the first time, extract a third image corresponding to the target area from the second image, and update at least one three-dimensional model based on the third image; The three-dimensional model is mapped into at least one two-dimensional image, and the at least one two-dimensional image is stitched into a second image to obtain a target super-resolution image.
当然,本申请实施例所提供的一种计算机可读存储介质,其上存储的计算机程序不限于如上所述的方法操作,还可以执行本申请任意实施例所提供的超 分辨图像重建方法中的相关操作。Certainly, the computer-readable storage medium provided in the embodiment of the present application is not limited to the method operations described above, and the computer program stored on the computer program may also be implemented in the super-resolution image reconstruction method provided by any embodiment of the present application. Related operations.
本申请实施例的计算机存储介质,可以采用一个或多个计算机可读的介质的任意组合。计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子(非穷举的列表)包括:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机存取存储器(RAM)、只读存储器(Read Only Memory,ROM)、可擦式可编程只读存储器(Erasable Programmable Read Only Memory,EPROM)或闪存、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本文件中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。The computer storage medium in the embodiments of the present application may adopt any combination of one or more computer-readable media. The computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium. The computer-readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof. More specific examples (non-exhaustive list) of computer-readable storage media include: electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read-only memory (Read Only Memory , ROM), Erasable Programmable Read Only Memory (EPROM) or flash memory, optical fiber, portable compact disk read-only memory (CD-ROM), optical storage device, magnetic storage device, or any of the above The right combination. In this document, a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in combination with an instruction execution system, apparatus, or device.
计算机可读的信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读的信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。The computer-readable signal medium may include a data signal in baseband or propagated as part of a carrier wave, which carries a computer-readable program code. Such a propagated data signal may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing. The computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, and the computer-readable medium may send, propagate, or transmit a program for use by or in connection with an instruction execution system, apparatus, or device .
计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括——但不限于无线、电线、光缆、射频(Radio Frequency,RF)等等,或者上述的任意合适的组合。The program code contained on the computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wire, optical fiber cable, radio frequency (RF), etc., or any suitable combination of the foregoing.
可以以一种或多种程序设计语言或其组合来编写用于执行本申请操作的计算机程序代码,所述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。Computer program code for performing the operations of this application may be written in one or more programming languages, or a combination thereof, including programming languages such as Java, Smalltalk, C ++, and also conventional Procedural programming language—such as "C" or similar programming language. The program code can be executed entirely on the user's computer, partly on the user's computer, as an independent software package, partly on the user's computer, partly on a remote computer, or entirely on a remote computer or server. In the case of a remote computer, the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (such as through an Internet service provider) Internet connection).

Claims (12)

  1. 一种超分辨图像重建方法,包括:A super-resolution image reconstruction method includes:
    获取当前区域中的目标区域在第一时刻的第一图像,并根据所述第一图像生成与所述目标区域中至少一个第一目标物体对应的至少一个三维模型,其中,所述第一目标物体为第一非刚性目标物体;Acquiring a first image of a target region in the current region at a first moment, and generating at least one three-dimensional model corresponding to at least one first target object in the target region according to the first image, wherein the first target The object is a first non-rigid target object;
    获取所述当前区域在第一时刻之后的第二时刻的第二图像,从所述第二图像中提取出所述目标区域对应的第三图像,并基于所述第三图像更新所述至少一个三维模型;Acquiring a second image of the current area at a second time after the first time, extracting a third image corresponding to the target area from the second image, and updating the at least one based on the third image Three-dimensional model
    将更新后的所述至少一个三维模型映射为至少一个二维图像,并将所述至少一个二维图像拼接至所述第二图像中,得到目标超分辨图像。The updated at least one three-dimensional model is mapped into at least one two-dimensional image, and the at least one two-dimensional image is stitched into the second image to obtain a target super-resolution image.
  2. 根据权利要求1所述的方法,其中,所述根据所述第一图像生成与所述目标区域中至少一个第一目标物体对应的至少一个三维模型,包括:The method according to claim 1, wherein generating at least one three-dimensional model corresponding to at least one first target object in the target area based on the first image comprises:
    基于预设目标物体检测方法,对所述第一图像进行目标物体检测,得到与所述至少一个第一目标物体一一对应的至少一个第一局部图像;Performing target object detection on the first image based on a preset target object detection method to obtain at least one first partial image corresponding to the at least one first target object;
    利用预设二维姿态点预估方法,分别对每个第一局部图像进行二维姿态点预估,得到与每个第一目标物体对应的第一二维姿态点;Using a preset two-dimensional pose point estimation method to perform two-dimensional pose point estimation on each first partial image to obtain a first two-dimensional pose point corresponding to each first target object;
    针对所述每个第一目标物体,利用所述第一二维姿态点对初始三维模型进行优化,得到与所述每个第一目标物体对应的三维模型;For each first target object, using the first two-dimensional pose point to optimize an initial three-dimensional model to obtain a three-dimensional model corresponding to each first target object;
    针对所述每个第一目标物体对应的三维模型,分别利用相应的第一局部图像中的纹理信息渲染所述每个第一目标物体对应的三维模型,以更新所述至少一个三维模型。For the three-dimensional model corresponding to each first target object, the three-dimensional model corresponding to each first target object is rendered using the texture information in the corresponding first partial image to update the at least one three-dimensional model.
  3. 根据权利要求2所述的方法,其中,所述针对所述每个第一目标物体,利用所述第一二维姿态点对初始三维模型进行优化,得到与所述每个第一目标物体对应的三维模型,包括:The method according to claim 2, wherein for each of the first target objects, the initial three-dimensional model is optimized by using the first two-dimensional pose points to obtain a correspondence with each of the first target objects 3D models, including:
    基于预设三维模型构建方法、初始形状因子矩阵β和初始姿态角向量θ构建所述初始三维模型;Constructing the initial three-dimensional model based on a preset three-dimensional model construction method, an initial shape factor matrix β, and an initial attitude angle vector θ;
    利用初始相机模型参数矩阵K对所述初始三维模型进行二维映射,得到与所述初始三维模型对应的初始二维姿态点;Performing a two-dimensional mapping on the initial three-dimensional model by using an initial camera model parameter matrix K to obtain an initial two-dimensional pose point corresponding to the initial three-dimensional model;
    针对所述每个第一目标物体:For each first target object:
    计算满足预设条件的形状因子矩阵β 1和第一姿态角向量θ 1,其中,所述预设条件为:所述第一二维姿态点与所述初始二维姿态点的匹配点对之间的差值的 加和最小,且形状因子矩阵β 1最小; Calculate a shape factor matrix β 1 and a first attitude angle vector θ 1 that satisfy a preset condition, where the preset condition is that the matching points of the first two-dimensional pose point and the initial two-dimensional pose point are The sum of the differences between them is the smallest, and the form factor matrix β 1 is the smallest;
    利用所述形状因子矩阵β 1和所述第一姿态角向量θ 1对所述初始三维模型进行优化,得到与所述第一目标物体对应的三维模型。 The initial three-dimensional model is optimized by using the shape factor matrix β 1 and the first attitude angle vector θ 1 to obtain a three-dimensional model corresponding to the first target object.
  4. 根据权利要求3所述的方法,针对所述每个第一目标物体:计算满足预设条件的形状因子矩阵β 1和第一姿态角向量θ 1,还包括: The method according to claim 3, for each of the first target objects: calculating a form factor matrix β 1 and a first attitude angle vector θ 1 satisfying a preset condition, further comprising:
    计算满足预设条件的相机模型参数矩阵K1,其中,所述预设条件为:所述第一二维姿态点与所述所述初始二维姿态点的匹配点对之间的差值的加和最小,且形状因子矩阵β 1最小; Calculate a camera model parameter matrix K1 that satisfies a preset condition, wherein the preset condition is: an addition of a difference between a pair of matching points of the first two-dimensional pose point and the initial two-dimensional pose point And minimum, and the form factor matrix β 1 is the smallest;
    所述针对所述每个第一目标物体对应的三维模型,分别利用相应的第一局部图像中的纹理信息渲染所述每个第一目标物体对应的三维模型,以更新所述至少一个三维模型,包括:For the three-dimensional model corresponding to each first target object, using the texture information in the corresponding first partial image to render the three-dimensional model corresponding to each first target object to update the at least one three-dimensional model. ,include:
    针对所述每个第一目标物体对应的三维模型:利用所述相机模型参数矩阵K1,将相应的第一局部图像中的纹理信息映射到所述每个第一目标物体对应的三维模型上,以更新所述至少一个三维模型。For the three-dimensional model corresponding to each first target object: using the camera model parameter matrix K1 to map the texture information in the corresponding first partial image to the three-dimensional model corresponding to each first target object, To update the at least one three-dimensional model.
  5. 根据权利要求4所述的方法,所述针对所述每个第一目标物体对应的三维模型:在利用所述相机模型参数矩阵K1,将相应的第一局部图像中的纹理信息映射到所述每个第一目标物体对应的三维模型上之后,还包括:The method according to claim 4, wherein the three-dimensional model corresponding to each first target object: using the camera model parameter matrix K1, mapping texture information in a corresponding first partial image to the first partial image After the three-dimensional model corresponding to each first target object, the method further includes:
    采用预设插值算法对映射得到的三维模型的纹理信息进行插值处理,以得到完整的三维模型的纹理坐标。The preset interpolation algorithm is used to interpolate the texture information of the mapped 3D model to obtain the texture coordinates of the complete 3D model.
  6. 根据权利要求3所述的方法,其中,所述基于所述第三图像更新所述至少一个三维模型,包括:The method of claim 3, wherein said updating the at least one three-dimensional model based on the third image comprises:
    基于所述预设目标物体检测方法,对所述第三图像进行目标物体检测,得到与所述目标区域中至少一个第二目标物体一一对应的至少一个第二局部图像,其中,所述第二目标物体为第二非刚性目标物体;Based on the preset target object detection method, target object detection is performed on the third image, and at least one second partial image corresponding to at least one second target object in the target area is obtained, where the first The two target objects are second non-rigid target objects;
    将所述第一局部图像与所述第二局部图像进行匹配,得到至少一个所述第一局部图像与所述第二局部图像的匹配对,以确定所述至少一个第二局部图像中的第二目标物体对应的至少一个三维模型;Matching the first partial image with the second partial image to obtain at least one matching pair of the first partial image and the second partial image, so as to determine the first of the at least one second partial image At least one three-dimensional model corresponding to two target objects;
    利用所述预设二维姿态点预估方法,分别对每个第二局部图像进行二维姿态点预估,得到与每个第二目标物体对应的第二二维姿态点;Using the preset two-dimensional pose point estimation method to separately perform two-dimensional pose point estimation on each second partial image to obtain a second two-dimensional pose point corresponding to each second target object;
    针对所述每个第二目标物体,利用所述第二二维姿态点对所述第二目标物 体对应的三维模型进行更新。For each second target object, the three-dimensional model corresponding to the second target object is updated using the second two-dimensional pose point.
  7. 根据权利要求6所述的方法,案子,所述将所述第一局部图像与所述第二局部图像进行匹配,得到至少一个所述第一局部图像与所述第二局部图像的匹配对,包括:The method according to claim 6, the case, said matching the first partial image with the second partial image to obtain at least one matching pair of the first partial image and the second partial image, include:
    分别确定所述每个第一局部图像与所述每个第二局部图像的中心点;Determining a center point of each of the first partial images and each of the second partial images;
    针对所述每个第二局部图像:For each second partial image:
    分别计算所述每个第二局部图像的中心点与所述每个第一局部图像的中心点之间的欧式距离;Calculating the Euclidean distance between the center point of each second partial image and the center point of each first partial image;
    将使得所述欧式距离最小的第一局部图像作为所述第二局部图像的匹配对。A first partial image that minimizes the Euclidean distance is used as a matching pair for the second partial image.
  8. 根据权利要求6所述的方法,其中,所述针对所述每个第二目标物体,利用所述第二二维姿态点对所述第二目标物体对应的三维模型进行更新,包括:The method according to claim 6, wherein, for each of the second target objects, updating the three-dimensional model corresponding to the second target object by using the second two-dimensional pose point comprises:
    针对所述每个第二目标物体:For each second target object:
    利用预设深度学习算法,将所述第二二维姿态点转换为第二姿态角向量θ 2Using a preset deep learning algorithm to convert the second two-dimensional attitude point into a second attitude angle vector θ 2 ;
    利用所述形状因子矩阵β 1和所述第二姿态角向量θ 2对所述第二目标物体对应的三维模型进行更新,得到与所述第二目标物体对应的三维模型。 The three-dimensional model corresponding to the second target object is updated by using the shape factor matrix β 1 and the second attitude angle vector θ 2 to obtain a three-dimensional model corresponding to the second target object.
  9. 根据权利要求4所述的方法,其中,所述将更新后的所述至少一个三维模型映射为至少一个二维图像,包括:The method according to claim 4, wherein said mapping said updated at least one three-dimensional model to at least one two-dimensional image comprises:
    利用所述相机模型参数矩阵K1,将更新后的所述至少一个三维模型映射为至少一个二维图像。Utilizing the camera model parameter matrix K1 to map the updated at least one three-dimensional model into at least one two-dimensional image.
  10. 一种超分辨图像重建装置,包括:A super-resolution image reconstruction device includes:
    三维模型生成模块,设置为获取当前区域中的目标区域在第一时刻的第一图像,并根据所述第一图像生成与所述目标区域中至少一个第一目标物体对应的至少一个三维模型,其中,所述第一目标物体为第一非刚性目标物体;A three-dimensional model generating module configured to obtain a first image of a target region in the current region at a first moment, and generate at least one three-dimensional model corresponding to at least one first target object in the target region according to the first image, Wherein, the first target object is a first non-rigid target object;
    三维模型更新模块,设置为获取所述当前区域在第一时刻之后的第二时刻的第二图像,从所述第二图像中提取出所述目标区域对应的第三图像,并基于所述第三图像更新所述至少一个三维模型;The three-dimensional model update module is configured to obtain a second image of the current area at a second time after the first time, extract a third image corresponding to the target area from the second image, and based on the first Three images update the at least one three-dimensional model;
    超分辨图像获取模块,设置为将更新后的所述至少一个三维模型映射为至少一个二维图像,并将所述至少一个二维图像拼接至所述第二图像中,得到目标超分辨图像。The super-resolution image acquisition module is configured to map the updated at least one three-dimensional model into at least one two-dimensional image, and stitch the at least one two-dimensional image into the second image to obtain a target super-resolution image.
  11. 一种超分辨图像重建终端,包括:A super-resolution image reconstruction terminal includes:
    至少一个处理器;At least one processor;
    存储装置,设置为存储至少一个程序,A storage device configured to store at least one program,
    当所述至少一个程序被所述至少一个处理器执行,使得所述至少一个处理器实现如权利要求1-9中任一所述的超分辨图像重建方法。When the at least one program is executed by the at least one processor, the at least one processor implements the super-resolution image reconstruction method according to any one of claims 1-9.
  12. 一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现如权利要求1-9中任一所述的超分辨图像重建方法。A computer-readable storage medium having stored thereon a computer program that, when executed by a processor, implements the super-resolution image reconstruction method according to any one of claims 1-9.
PCT/CN2019/104388 2018-09-04 2019-09-04 Super-resolution image reconstruction method and apparatus, and terminal and storage medium WO2020048484A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811027057.5 2018-09-04
CN201811027057.5A CN109191554B (en) 2018-09-04 2018-09-04 Super-resolution image reconstruction method, device, terminal and storage medium

Publications (1)

Publication Number Publication Date
WO2020048484A1 true WO2020048484A1 (en) 2020-03-12

Family

ID=64914431

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/104388 WO2020048484A1 (en) 2018-09-04 2019-09-04 Super-resolution image reconstruction method and apparatus, and terminal and storage medium

Country Status (2)

Country Link
CN (1) CN109191554B (en)
WO (1) WO2020048484A1 (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109191554B (en) * 2018-09-04 2021-01-01 清华-伯克利深圳学院筹备办公室 Super-resolution image reconstruction method, device, terminal and storage medium
CN109859296B (en) * 2019-02-01 2022-11-29 腾讯科技(深圳)有限公司 Training method of SMPL parameter prediction model, server and storage medium
EP4036863A4 (en) * 2019-09-30 2023-02-08 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Human body model reconstruction method and reconstruction system, and storage medium
CN113643433A (en) * 2020-04-27 2021-11-12 成都术通科技有限公司 Form and attitude estimation method, device, equipment and storage medium
CN112634139B (en) * 2021-02-25 2021-05-28 杭州海康威视数字技术股份有限公司 Optical field super-resolution imaging method, device and equipment
CN113112402A (en) * 2021-03-22 2021-07-13 深兰科技(上海)有限公司 Model acquisition method and device, electronic equipment and storage medium
CN113538649B (en) * 2021-07-14 2022-09-16 深圳信息职业技术学院 Super-resolution three-dimensional texture reconstruction method, device and equipment
CN113610713B (en) * 2021-08-13 2023-11-28 北京达佳互联信息技术有限公司 Training method of video super-resolution model, video super-resolution method and device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100124383A1 (en) * 2008-11-19 2010-05-20 Nec Laboratories America, Inc. Systems and methods for resolution-invariant image representation
CN102682442A (en) * 2012-04-28 2012-09-19 东南大学 Motion target super-resolution image reconstruction method based on optical flow field
CN103810685A (en) * 2014-02-25 2014-05-21 清华大学深圳研究生院 Super resolution processing method for depth image
CN105741252A (en) * 2015-11-17 2016-07-06 西安电子科技大学 Sparse representation and dictionary learning-based video image layered reconstruction method
CN109191554A (en) * 2018-09-04 2019-01-11 清华-伯克利深圳学院筹备办公室 A kind of super resolution image reconstruction method, device, terminal and storage medium

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104091364B (en) * 2014-07-10 2017-01-11 西北工业大学 Single-image super-resolution reconstruction method
CN105006018B (en) * 2015-06-30 2017-11-14 四川大学 Three dimensional CT core image super resolution ratio reconstruction method
CN105023275B (en) * 2015-07-14 2018-08-28 清华大学 Super-resolution optical field acquisition device and its three-dimensional rebuilding method
US9918792B1 (en) * 2016-05-31 2018-03-20 American Medical Technologies, Llc Methods and system for atrial fibrillation ablation using a fluoroscopy and/or medical images based cardiac mapping system with optional esophageal temperature monitoring
CN107888707B (en) * 2017-12-08 2021-04-20 北京奇艺世纪科技有限公司 Picture transmission method and device and electronic equipment
CN107944428B (en) * 2017-12-15 2021-07-30 北京工业大学 Indoor scene semantic annotation method based on super-pixel set
CN108038905B (en) * 2017-12-25 2018-12-07 北京航空航天大学 A kind of Object reconstruction method based on super-pixel
CN108416821B (en) * 2018-03-08 2019-08-02 山东财经大学 A kind of CT Image Super-resolution Reconstruction method of deep neural network
CN108447020A (en) * 2018-03-12 2018-08-24 南京信息工程大学 A kind of face super-resolution reconstruction method based on profound convolutional neural networks
CN108376386A (en) * 2018-03-23 2018-08-07 深圳天琴医疗科技有限公司 A kind of construction method and device of the super-resolution model of image

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100124383A1 (en) * 2008-11-19 2010-05-20 Nec Laboratories America, Inc. Systems and methods for resolution-invariant image representation
CN102682442A (en) * 2012-04-28 2012-09-19 东南大学 Motion target super-resolution image reconstruction method based on optical flow field
CN103810685A (en) * 2014-02-25 2014-05-21 清华大学深圳研究生院 Super resolution processing method for depth image
CN105741252A (en) * 2015-11-17 2016-07-06 西安电子科技大学 Sparse representation and dictionary learning-based video image layered reconstruction method
CN109191554A (en) * 2018-09-04 2019-01-11 清华-伯克利深圳学院筹备办公室 A kind of super resolution image reconstruction method, device, terminal and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ZHENG, HAITIAN ET AL.: "Combining Exemplar-based Approach and Learning-Based Approach for Light Field Super-resolution Using a Hybrid Imaging System", 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, 31 December 2017 (2017-12-31), XP033303718 *

Also Published As

Publication number Publication date
CN109191554B (en) 2021-01-01
CN109191554A (en) 2019-01-11

Similar Documents

Publication Publication Date Title
WO2020048484A1 (en) Super-resolution image reconstruction method and apparatus, and terminal and storage medium
EP3786890B1 (en) Method and apparatus for determining pose of image capture device, and storage medium therefor
WO2020001168A1 (en) Three-dimensional reconstruction method, apparatus, and device, and storage medium
CN108335353B (en) Three-dimensional reconstruction method, device and system of dynamic scene, server and medium
JP7373554B2 (en) Cross-domain image transformation
WO2019238114A1 (en) Three-dimensional dynamic model reconstruction method, apparatus and device, and storage medium
CN112927362A (en) Map reconstruction method and device, computer readable medium and electronic device
US20220358675A1 (en) Method for training model, method for processing video, device and storage medium
CN112598780B (en) Instance object model construction method and device, readable medium and electronic equipment
WO2021237743A1 (en) Video frame interpolation method and apparatus, and computer-readable storage medium
US11961266B2 (en) Multiview neural human prediction using implicit differentiable renderer for facial expression, body pose shape and clothes performance capture
CN113313832A (en) Semantic generation method and device of three-dimensional model, storage medium and electronic equipment
WO2022208440A1 (en) Multiview neural human prediction using implicit differentiable renderer for facial expression, body pose shape and clothes performance capture
CN116863078A (en) Three-dimensional human body model reconstruction method, three-dimensional human body model reconstruction device, electronic equipment and readable medium
CN115578515B (en) Training method of three-dimensional reconstruction model, three-dimensional scene rendering method and device
CN112085842A (en) Depth value determination method and device, electronic equipment and storage medium
US11741671B2 (en) Three-dimensional scene recreation using depth fusion
KR20230078502A (en) Apparatus and method for image processing
CN116797713A (en) Three-dimensional reconstruction method and terminal equipment
CN116246026B (en) Training method of three-dimensional reconstruction model, three-dimensional scene rendering method and device
CN115272575B (en) Image generation method and device, storage medium and electronic equipment
CN113920282B (en) Image processing method and device, computer readable storage medium, and electronic device
CN115100360B (en) Image generation method and device, storage medium and electronic equipment
US20240071009A1 (en) Visually coherent lighting for mobile augmented reality
CN117294956A (en) Method for generating image and optical flow based on moving target, related method and related product

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19856744

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19856744

Country of ref document: EP

Kind code of ref document: A1