WO2019161813A1 - Dynamic scene three-dimensional reconstruction method, apparatus and system, server, and medium - Google Patents

Dynamic scene three-dimensional reconstruction method, apparatus and system, server, and medium Download PDF

Info

Publication number
WO2019161813A1
WO2019161813A1 PCT/CN2019/083816 CN2019083816W WO2019161813A1 WO 2019161813 A1 WO2019161813 A1 WO 2019161813A1 CN 2019083816 W CN2019083816 W CN 2019083816W WO 2019161813 A1 WO2019161813 A1 WO 2019161813A1
Authority
WO
WIPO (PCT)
Prior art keywords
array
observation point
uav
dynamic scene
energy term
Prior art date
Application number
PCT/CN2019/083816
Other languages
French (fr)
Chinese (zh)
Inventor
方璐
刘烨斌
许岚
程巍
戴琼海
Original Assignee
清华-伯克利深圳学院筹备办公室
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 清华-伯克利深圳学院筹备办公室 filed Critical 清华-伯克利深圳学院筹备办公室
Priority to US16/975,242 priority Critical patent/US11954870B2/en
Publication of WO2019161813A1 publication Critical patent/WO2019161813A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/292Multi-camera tracking
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64CAEROPLANES; HELICOPTERS
    • B64C39/00Aircraft not otherwise provided for
    • B64C39/02Aircraft not otherwise provided for characterised by special use
    • B64C39/024Aircraft not otherwise provided for characterised by special use of the remote controlled vehicle type, i.e. RPV
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/251Fusion techniques of input or preprocessed data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/593Depth or shape recovery from multiple images from stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/593Depth or shape recovery from multiple images from stereo images
    • G06T7/596Depth or shape recovery from multiple images from stereo images from three or more stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/803Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of input or preprocessed data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/17Terrestrial scenes taken from planes or by drones
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • G06V20/653Three-dimensional objects by matching three-dimensional models, e.g. conformal mapping of Riemann surfaces
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U2101/00UAVs specially adapted for particular uses or applications
    • B64U2101/30UAVs specially adapted for particular uses or applications for imaging, photography or videography
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30244Camera pose

Definitions

  • Embodiments of the present invention relate to the field of computer vision technology, for example, to a three-dimensional reconstruction method of a dynamic scene, an apparatus and system, a server, and a medium.
  • Embodiments of the present invention provide a three-dimensional reconstruction method for a dynamic scene, a device and a system, a server, and a medium, to overcome the defect of how to automatically complete the three-dimensional reconstruction of the dynamic scene without affecting the comfort of the collector and being restricted by the shooting space.
  • an embodiment of the present invention provides a method for three-dimensional reconstruction of a dynamic scene, where the method includes:
  • the UAV array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
  • the embodiment of the present invention further provides a three-dimensional reconstruction device for a dynamic scene, the device comprising:
  • An image sequence acquisition module configured to acquire a plurality of consecutive depth image sequences of the dynamic scene, wherein the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera;
  • An image fusion module is configured to fuse the plurality of consecutive depth image sequences to establish a three-dimensional reconstruction model of the dynamic scene
  • a target observation point calculation module configured to calculate a target observation point of the UAV array according to the three-dimensional reconstruction model and the current pose of the UAV array;
  • a reconstruction model updating module is configured to instruct the UAV array to move to the target observation point for shooting, and update the three-dimensional reconstruction model according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
  • an embodiment of the present invention further provides a three-dimensional reconstruction system for a dynamic scene, where the system includes an unmanned aerial vehicle array and a three-dimensional reconstruction platform;
  • each drone in the UAV array is equipped with a depth camera, which is set to capture a depth image sequence of the dynamic scene;
  • the three-dimensional reconstruction platform includes a three-dimensional reconstruction device for a dynamic scene according to any one of the embodiments of the present application, and is configured to generate a three-dimensional reconstruction model of the dynamic scene according to a plurality of consecutive depth image sequences captured by the UAV array.
  • the embodiment of the present invention further provides a server, including:
  • One or more processors are One or more processors;
  • a storage device configured to store one or more programs
  • the one or more processors When the one or more programs are executed by the one or more processors, the one or more processors implement a three-dimensional reconstruction method of a dynamic scene as described in any of the embodiments of the present application.
  • the embodiment of the present invention further provides a computer readable storage medium, where the computer program is stored, and the program is executed by the processor to implement a three-dimensional reconstruction method of a dynamic scene according to any embodiment of the present application. .
  • Embodiment 1 is a flowchart of a method for reconstructing a dynamic scene according to Embodiment 1 of the present invention
  • FIG. 2 is a flowchart of a method for reconstructing a dynamic scene according to Embodiment 2 of the present invention
  • FIG. 3 is a flowchart of a method for reconstructing a dynamic scene according to Embodiment 3 of the present invention
  • FIG. 4 is a schematic structural diagram of a three-dimensional reconstruction apparatus for a dynamic scene according to Embodiment 4 of the present invention.
  • FIG. 5 is a schematic structural diagram of a three-dimensional reconstruction system for a dynamic scene according to Embodiment 5 of the present invention.
  • FIG. 6 is a schematic structural diagram of a server according to Embodiment 6 of the present invention.
  • Embodiment 1 is a flowchart of a method for three-dimensional reconstruction of a dynamic scene according to Embodiment 1 of the present invention.
  • the present embodiment is applicable to a situation in which a dynamic scene is three-dimensionally reconstructed, such as a scene in which a dancer dances on a stage.
  • the method can be performed by a three-dimensional reconstruction device of a dynamic scene, the device can be implemented in software and/or hardware, and can be integrated in a server. As shown in FIG. 1, the method may include, for example, steps S110 to S140.
  • step S110 a plurality of consecutive depth image sequences of the dynamic scene are acquired.
  • the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera.
  • the UAS array may include multiple UAVs, for example, 3 or 5 U.S. units, for example, may be configured according to the actual needs of the dynamic scenario, which is not limited in this embodiment of the present invention.
  • Different drones in the UAV array can simultaneously capture dynamic scenes at different viewpoints to obtain depth image sequences of dynamic scenes from different angles for better 3D reconstruction.
  • Each drone is equipped with a depth camera for taking deep images of dynamic scenes.
  • a depth image is an image or image channel that contains information about the distance from the surface of the scene object of the viewpoint.
  • Each pixel value in the depth image is the actual distance of the camera from the object, by which a three-dimensional model can be constructed.
  • the shooting by the drone array equipped with the depth camera is not restricted by the shooting space as in the fixed camera array of the related art, and the drone array can be controlled to automatically take the image.
  • the drone array can be controlled to be located at the initial position above the dynamic scene and simultaneously photographed, because the dynamic scene is three-dimensionally reconstructed, and the position or posture of the person or scene in the dynamic scene changes in real time, so each The unmanned person continues to shoot and sends the captured continuous depth image sequence to the 3D reconstruction device for processing in real time.
  • the continuous depth image sequence refers to a sequence of depth images continuously captured in chronological order.
  • the depth camera can continuously capture 30 frames of images per second, and each image is arranged in chronological order to obtain a sequence of images.
  • the acquiring a plurality of consecutive depth image sequences of the dynamic scene includes:
  • the plurality of original depth image sequences are original image sequences captured by different drones, and although the UAV array is simultaneously photographed, there is a certain error in time, therefore, it is required These original depth image sequences are aligned according to the synchronization time stamp, thereby ensuring temporal consistency of different depth image sequences captured by different drones from different perspectives, thereby improving the accuracy of the reconstruction model.
  • step S120 the plurality of consecutive depth image sequences are fused to establish a three-dimensional reconstruction model of the dynamic scene.
  • the continuous depth image sequence may be projected into the three-dimensional space according to the internal reference matrix of the depth camera to obtain a three-dimensional point cloud, and then the three-dimensional point cloud is registered and fused, and finally a three-dimensional reconstruction model is established.
  • it can be performed by using the registration and fusion algorithms in the related art, and will not be repeated here.
  • step S130 the target observation point of the UAV array is determined according to the three-dimensional reconstruction model and the current pose of the UAV array.
  • the target observation point of the UAV array at the next moment is calculated in real time, and the target observation point is the optimal observation point, thereby indicating that the UAV array moves to the target observation in real time.
  • the point is taken to update the reconstructed model to achieve accurate reproduction of the dynamic scene.
  • the pose can be represented by a rotation angle and a translation distance of the drone, and correspondingly, the control of the drone can also include two parameters of rotation and translation.
  • the drone can be controlled to move to the optimal target observation point, and the rotation angle of the drone from the current observation point to the target observation point is controlled, that is, the optimal shooting of the drone is controlled.
  • Viewpoint The target observation point can be determined according to a preset standard, and the shooting points that may exist in the drone are evaluated, and the shooting point whose evaluation result meets the standard is determined as the best target observation point.
  • the possible shooting points may be determined according to the current pose of the UAV array, and the evaluation process may be performed according to the possible existing observation points and the established three-dimensional reconstruction model, and the evaluation may exist in different possible In the observation point, the effect of the three-dimensional reconstruction model established by the depth image sequence captured by which observation point conforms to a preset criterion, for example, by calculating the energy function of the observation point that may exist.
  • step S140 the drone array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
  • the technical solution of the embodiment uses the UAV array to capture the dynamic scene, and performs image fusion according to the captured multiple consecutive depth image sequences to obtain a three-dimensional reconstruction model of the dynamic scene. Therefore, it is not necessary to rely on additional equipment to ensure the The comfort of the collector. Moreover, in the reconstruction process, the calculation of the target observation point in real time indicates that the UAV array moves to the target observation point for shooting, and the model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point, thereby A more accurate 3D reconstruction model is obtained, and the reconstruction process can be completed automatically without being restricted by the shooting space.
  • FIG. 2 is a flowchart of a three-dimensional reconstruction method of a dynamic scene according to Embodiment 2 of the present invention. As shown in FIG. 2, the method may include, for example, steps S210 to S270.
  • step S210 a plurality of consecutive depth image sequences of the dynamic scene are acquired.
  • the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera.
  • step S220 the plurality of consecutive depth image sequences are fused, and the key frame reconstructed body is determined according to a preset period.
  • step S230 in each preset period, determining a deformation parameter of the non-rigid deformation node in the current key frame reconstruction body, and updating the reconstruction model in the current key frame reconstruction body to the current data frame reconstruction according to the deformation parameter In the body.
  • the current data frame reconstructed body refers to a reconstructed body in real time at each moment.
  • step S240 a three-dimensional reconstruction model of the dynamic scene is extracted from the current data frame reconstructed body.
  • step S250 the current key frame reconstructed body is replaced with the current data frame reconstructed body as the key frame reconstructed body in the next preset period.
  • step S220-step S250 the depth image is actually merged by using a key frame strategy, and a real-time reconstruction model of the dynamic scene is obtained.
  • an initial three-dimensional reconstruction model may be established by image fusion according to a plurality of consecutive depth image sequences, and then a key frame reconstruction body of the three-dimensional reconstruction model is determined according to a preset period, for example, 100 frames, and each preset is During the period, the operations of steps S230 to S250 are performed.
  • the non-rigid deformation node may represent a node in which a character or a scene changes in a dynamic scene
  • the reconstruction model in the current key frame reconstruction body is updated to the current data frame reconstruction body by determining the deformation parameter of the non-rigid deformation node, and then from the current
  • the 3D reconstruction model is extracted from the data frame reconstruction body, so that the details of the dynamic scene can be captured, the accuracy of the reconstruction model can be improved, errors and confusion can be avoided, and the jam can be avoided.
  • the current data frame reconstruction body is used to replace the current key frame reconstruction body as the key frame reconstruction body in the next preset period, thereby implementing the iteration of the current data frame reconstruction body and the key frame reconstruction body, thereby realizing each of the dynamic scenes.
  • the reconstructed body can be understood as a hypothesis in the process of 3D reconstruction. It is assumed that the reconstructed body can surround the entire dynamic scene (or reconstructed object). The reconstructed body is composed of many uniform voxels, through registration, fusion and other algorithms. A three-dimensional reconstruction model of the dynamic scene is built on top of this reconstructed body. The nodes in the reconstructed body and the deformation parameters of the nodes represent the characteristics of the dynamic scene. Therefore, the three-dimensional reconstruction model of the dynamic scene can be extracted from the reconstructed body. In this embodiment, by integrating the depth image by the key frame strategy described above, errors caused by data fusion when the point cloud registration is inaccurate can be avoided.
  • the deformation parameter includes rotation and translation parameters of each deformation node, and the calculation process thereof can be obtained, for example, by solving an energy equation of a non-rigid motion constraint, the energy equation being composed of a non-rigid motion constraint term and a local rigid motion constraint term.
  • u i represents the position coordinates of the 3D point cloud in the same matching point pair
  • c i represents the i-th element in the set of matching point pairs.
  • i represents the i-th vertex on the model
  • It represents the positional transformation effect of the non-rigid motion acting on v i and v j simultaneously on v j , that is, to ensure that the non-rigid driving effects of adjacent vertices on the model are as uniform as possible.
  • the non-rigid motion constraint term E n ensures that the model driven by the non-rigid motion is aligned with the three-dimensional point cloud obtained from the depth image as much as possible, and the local rigid motion constraint E g can be guaranteed while the model as a whole is subjected to the local rigid constraint motion. A large amount of reasonable non-rigid motion can also be well solved.
  • the deformed vertices are approximated as follows:
  • the cumulative transformation matrix of the model vertex v i up to the previous frame is a known quantity; I is a four-dimensional unit matrix; make That is, the model vertices after the previous frame transformation are transformed:
  • step S260 the target observation point of the UAV array is determined according to the three-dimensional reconstruction model and the current pose of the UAV array.
  • step S270 the drone array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the drone array at the target observation point.
  • the technical solution of the embodiment utilizes an unmanned aerial vehicle array to capture a dynamic scene, performs image fusion according to a plurality of consecutive continuous depth image sequences, and adopts a key frame strategy to finally obtain a three-dimensional reconstruction model of the dynamic scene, which is not subject to shooting.
  • the accuracy of the reconstruction model is improved, and errors caused by inaccurate registration of point clouds are avoided.
  • FIG. 3 is a flowchart of a three-dimensional reconstruction method of a dynamic scene according to Embodiment 3 of the present invention.
  • the calculation operation of the target observation point is further optimized based on the foregoing embodiment.
  • the method may include, for example, steps S310 to S360.
  • step S310 a plurality of consecutive depth image sequences of the dynamic scene are acquired.
  • the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera.
  • step S320 the plurality of consecutive depth image sequences are fused to establish a three-dimensional reconstruction model of the dynamic scene.
  • step S330 according to the current pose of the UAV array, the spatial neighborhood is rasterized to establish a set of candidate observation points.
  • the current pose of the UAV array characterizes the current observation point of each UAV in the UAV array, including the current coordinates and shooting angle.
  • the spatial neighborhood range is delineated according to the current pose and the preset distance, and the candidate observation points are determined by rasterizing the spatial neighborhood, that is, each node represents a candidate observation point after rasterization.
  • step S340 the total energy value of each of the candidate observation points in the set of candidate observation points is determined using the validity energy function.
  • step S350 candidate observation points whose total energy values meet the preset criteria are taken as the target observation points.
  • the validity energy function includes a depth energy term, a central energy term, and a motion energy term
  • the depth energy term is used to determine the extent to which the average depth value of the candidate observation point is close to the target depth value
  • the central energy term is used to determine how close the reconstructed model observed by the candidate observation point is to the central portion of the captured image frame
  • the kinetic energy term is used to determine the amount of motion occurring in the dynamic scene observed by the candidate observation point.
  • the validity energy function is represented by the following formula:
  • E t is the total energy term
  • Ed is the depth energy term
  • Ec is the central energy term
  • Em is the motion energy term
  • ⁇ d , ⁇ c and ⁇ m correspond to the depth energy term, the central energy term and the motion energy term, respectively.
  • the depth energy term, the center energy term, and the motion energy term are respectively represented by the following formula:
  • T c and T V are the poses of the UAV array and the candidate observation points in the reconstruction model respectively;
  • t v is the translation component of the pose of the candidate observation point;
  • x n is the voxel of the reconstruction model hit by the ray N
  • x is the normal of the voxel;
  • x i is the node where the reconstructed model is non-rigid deformed;
  • x' i is the node after the non-rigid deformation;
  • ⁇ () is the projection perspective transform from the three-dimensional space to the two-dimensional image plane d avg and d o represent the average depth value and the target depth value of the candidate observation point respectively;
  • the ⁇ () function represents the penalty term for the distance;
  • is the
  • the candidate observation points can be comprehensively evaluated to determine the effect of the 3D reconstruction model established by the depth image sequence captured by the UAV at which observation point.
  • the standard is to comprehensively consider the average depth, average center degree and cumulative motion information of the depth image collected at the candidate observation points, so that the depth image acquired at the target observation point is more favorable for the current dynamic scene reconstruction.
  • candidate observation points with the largest total energy value may be selected as the optimal target observation points.
  • step S360 the drone array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the drone array at the target observation point.
  • the technical solution of the embodiment uses a UAV array to capture a dynamic scene, performs image fusion according to the captured multiple consecutive depth image sequences, obtains a three-dimensional reconstruction model of the dynamic scene, and performs candidate observation points through a validity energy function. Calculate and evaluate, determine the optimal target observation point, and instruct the drone array to move to the target observation point for shooting, thereby not only achieving automatic shooting and reconstruction, but also improving the reconstruction effect of the three-dimensional model, and is simple and easy Line, has broad application prospects.
  • FIG. 4 is a schematic structural diagram of a three-dimensional reconstruction apparatus for a dynamic scene according to Embodiment 4 of the present invention.
  • This embodiment can be applied to a case where a dynamic scene is three-dimensionally reconstructed, such as a scene in which a dancer dances on a stage.
  • the three-dimensional reconstruction device of the dynamic scene provided by the embodiment of the present invention can perform the three-dimensional reconstruction method of the dynamic scene provided by any embodiment of the present application, and has the corresponding functional modules and beneficial effects of the execution method.
  • the apparatus includes an image sequence acquisition module 410, an image fusion module 420, a target observation point calculation module 430, and a reconstruction model update module 440.
  • the image sequence acquisition module 410 is configured to acquire a plurality of consecutive depth image sequences of the dynamic scene, wherein the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera.
  • the image fusion module 420 is configured to fuse the plurality of consecutive depth image sequences to establish a three-dimensional reconstruction model of the dynamic scene.
  • the target observation point calculation module 430 is configured to determine a target observation point of the UAV array according to the three-dimensional reconstruction model and the current pose of the UAV array.
  • the reconstruction model update module 440 is configured to instruct the UAV array to move to the target observation point for shooting, and update the three-dimensional reconstruction model according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
  • the image sequence acquisition module 410 includes:
  • An original image sequence acquiring unit configured to acquire a plurality of original depth image sequences of the dynamic scene captured by the UAV array
  • an image sequence alignment unit configured to align the plurality of original depth image sequences according to a synchronization time stamp to obtain the plurality of consecutive depth image sequences.
  • the image fusion module 420 is further configured to:
  • the plurality of consecutive depth image sequences are fused, and the key frame reconstructed body is determined according to a preset period, and in each preset period, the following operations are performed:
  • the current data frame reconstructed body is replaced with the current data frame reconstructed body as a key frame reconstructed body in the next preset period.
  • the target observation point calculation module 430 includes:
  • the candidate observation point establishing unit is configured to rasterize the spatial neighborhood according to the current pose of the UAV array to establish a set of candidate observation points;
  • An energy value calculation unit configured to determine a total energy value of each of the candidate observation points in the set of candidate observation points by using a validity energy function
  • the target observation point determining unit is configured to select a candidate observation point whose total energy value conforms to a preset criterion as the target observation point.
  • the validity energy function includes a depth energy term, a central energy term, and a motion energy term
  • the depth energy term is used to determine the extent to which the average depth value of the candidate observation point is close to the target depth value
  • the central energy term is used to determine how close the reconstructed model observed by the candidate observation point is to the central portion of the captured image frame
  • the kinetic energy term is used to determine the amount of motion occurring in the dynamic scene observed by the candidate observation point.
  • the validity energy function is represented by the following formula:
  • E t is the total energy term
  • Ed is the depth energy term
  • Ec is the central energy term
  • Em is the motion energy term
  • ⁇ d , ⁇ c and ⁇ m correspond to the depth energy term, the central energy term and the motion energy term, respectively.
  • the depth energy term, the center energy term, and the motion energy term are respectively represented by the following formula:
  • T c and T V are the poses of the UAV array and the candidate observation points in the reconstruction model respectively;
  • t v is the translation component of the pose of the candidate observation point;
  • x n is the voxel of the reconstruction model hit by the ray N
  • x is the normal of the voxel;
  • x i is the node where the reconstructed model is non-rigid deformed;
  • x' i is the node after the non-rigid deformation;
  • ⁇ () is the projection perspective transform from the three-dimensional space to the two-dimensional image plane d avg and d o represent the average depth value and the target depth value of the candidate observation point respectively;
  • the ⁇ () function represents the penalty term for the distance;
  • is the
  • the technical solution of the embodiment uses the UAV array to capture the dynamic scene, and performs image fusion according to the captured multiple consecutive depth image sequences to obtain a three-dimensional reconstruction model of the dynamic scene. Therefore, it is not necessary to rely on additional equipment to ensure the The comfort of the collector. Moreover, in the reconstruction process, the calculation of the target observation point in real time indicates that the UAV array moves to the target observation point for shooting, and the model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point, thereby A more accurate 3D reconstruction model is obtained, and the reconstruction process can be completed automatically without being restricted by the shooting space.
  • FIG. 5 is a schematic structural diagram of a three-dimensional reconstruction system for a dynamic scene according to Embodiment 5 of the present invention. As shown in FIG. 5, the UAV array 1 and the three-dimensional reconstruction platform 2 are included.
  • each drone in the UAV array 1 is equipped with a depth camera, which is set to capture a depth image sequence of a dynamic scene.
  • FIG. 5 shows that the UAV array 1 includes three UAVs, namely the UAV 11, the UAV 12 and the UAV 13, but the embodiment of the present invention is in the UAV array.
  • the number of drones is not limited, and can be configured according to the actual situation of the dynamic scene to be reconstructed.
  • the three-dimensional reconstruction platform 2 includes the three-dimensional reconstruction device 21 of the dynamic scene described in any of the above embodiments, and is configured to generate a three-dimensional reconstruction model of the dynamic scene according to the plurality of consecutive depth image sequences captured by the UAV array.
  • the three-dimensional reconstruction platform 2 further includes a wireless communication module 22, and is wirelessly connected to the UAV array 1 and configured to receive a plurality of consecutive depth image sequences captured by the UAV array, and is further configured to set the three-dimensional reconstruction device 22 The calculated position information of the target observation point is sent to the drone array 1.
  • each of the UAV arrays 1 further includes a navigation module configured to control the drone to move to the target observation point to capture the dynamic scene according to the position information.
  • the technical solution of the embodiment uses the UAV array to capture the dynamic scene, and performs image fusion according to the captured multiple consecutive depth image sequences to obtain a three-dimensional reconstruction model of the dynamic scene. Therefore, it is not necessary to rely on additional equipment to ensure the The comfort of the collector. Moreover, in the reconstruction process, the calculation of the target observation point in real time indicates that the UAV array moves to the target observation point for shooting, and the model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point, thereby A more accurate 3D reconstruction model is obtained, and the reconstruction process can be completed automatically without being restricted by the shooting space.
  • FIG. 6 is a schematic structural diagram of a server according to Embodiment 6 of the present invention.
  • FIG. 6 shows a block diagram of an exemplary server 612 suitable for use in implementing embodiments of the present invention.
  • the server 612 shown in FIG. 6 is merely an example and should not impose any limitation on the function and scope of use of the embodiments of the present invention.
  • server 612 is represented in the form of a general purpose server.
  • Components of server 612 may include, but are not limited to, one or more processors 616, storage 628, and bus 618 that connect different system components, including storage 628 and processor 616.
  • Bus 618 represents one or more of several types of bus structures, including a memory device bus or memory device controller, a peripheral bus, a graphics acceleration port, a processor, or a local bus using any of a variety of bus structures.
  • these architectures include, but are not limited to, the Industry Subversive Alliance (ISA) bus, the Micro Channel Architecture (MAC) bus, the Enhanced ISA Bus, and the Video Electronics Standards Association. Association, VESA) Local Bus and Peripheral Component Interconnect (PCI) bus.
  • Server 612 typically includes a variety of computer system readable media. These media can be any available media that can be accessed by server 612, including volatile and non-volatile media, removable and non-removable media.
  • Storage device 628 can include computer system readable media in the form of volatile memory, such as random access memory (RAM) 630 and/or cache memory 632.
  • Server 612 may further include other removable/non-removable, volatile/non-volatile computer system storage media.
  • storage system 634 can be used to read and write non-removable, non-volatile magnetic media (not shown in Figure 6, commonly referred to as "hard disk drives").
  • a disk drive for reading and writing to a removable non-volatile disk for example, a "floppy disk”
  • a removable non-volatile disk such as a read-only disk (Compact Disc Read)
  • Storage device 628 can include at least one program product having a set (eg, at least one) of program modules configured to perform the functions of the various embodiments of the present application.
  • Program module 642 typically performs the functions and/or methods of the embodiments described herein.
  • Server 612 may also be in communication with one or more external devices 614 (eg, a keyboard, pointing device, display 624, etc.), and may also be in communication with one or more devices that enable a user to interact with the server 612, and/or Server 612 can communicate with any device (e.g., network card, modem, etc.) that is in communication with one or more other computing devices. This communication can take place via an input/output (I/O) interface 622. Moreover, the server 612 can also communicate with one or more networks (eg, a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet) through the network adapter 620. As shown in FIG.
  • LAN local area network
  • WAN wide area network
  • public network such as the Internet
  • network adapter 620 communicates with other modules of server 612 via bus 618. It should be understood that although not shown in the figures, other hardware and/or software modules may be utilized in connection with server 612, including but not limited to: microcode, device drivers, redundant processors, external disk drive arrays, disk arrays (Redundant Arrays of Independent Disks, RAID) systems, tape drives, and data backup storage systems.
  • the processor 616 performs a three-dimensional reconstruction method of the dynamic scene provided by the embodiment of the present invention by executing a program stored in the storage device 628, and the method includes:
  • the UAV array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
  • the seventh embodiment of the present invention further provides a computer readable storage medium, where the computer program is stored, and the program is executed by the processor to implement a three-dimensional reconstruction method of a dynamic scene according to the embodiment of the present invention, including:
  • the UAV array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
  • the computer storage medium of the embodiments of the present invention may employ any combination of one or more computer readable mediums.
  • the computer readable medium can be a computer readable signal medium or a computer readable storage medium.
  • the computer readable storage medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the above.
  • a computer readable storage medium can be any tangible medium that can contain or store a program, which can be used by or in connection with an instruction execution system, apparatus or device.
  • the computer readable signal medium may comprise a data signal that is propagated in the baseband or as part of a carrier, carrying computer readable program code. Such propagated data signals can take a variety of forms including, but not limited to, electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • the computer readable signal medium can also be any computer readable medium other than a computer readable storage medium, which can transmit, propagate, or transport a program for use by or in connection with the instruction execution system, apparatus, or device. .
  • Program code embodied on a computer readable medium can be transmitted by any suitable medium, including but not limited to wireless, wire, optical cable, radio frequency (RF), and the like, or any suitable combination of the foregoing.
  • suitable medium including but not limited to wireless, wire, optical cable, radio frequency (RF), and the like, or any suitable combination of the foregoing.
  • Computer program code for performing the operations of the present application may be written in one or more programming languages, or a combination thereof, including an object oriented programming language such as Java, Smalltalk, C++, and conventional Procedural programming language—such as the "C" language or a similar programming language.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer, partly on the remote computer, or entirely on the remote computer or server.
  • the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or can be connected to an external computer (eg, using an Internet service provider) Internet connection).
  • LAN local area network
  • WAN wide area network
  • Internet service provider Internet service provider

Abstract

Disclosed are a dynamic scene three-dimensional reconstruction method, apparatus and system, a server, and a medium. The method comprises: obtaining a plurality of continuous depth image sequences of a dynamic scene, the plurality of continuous depth image sequences being photographed by an unmanned aerial vehicle array equipped with depth cameras; combining the plurality of continuous depth image sequences, and establishing a three-dimensional reconstruction model of the dynamic scene; calculating a target observation point of the unmanned aerial vehicle array according to the three-dimensional reconstruction model and a current position of the unmanned aerial vehicle array; instructing the unmanned aerial vehicle array to move to the target observation point for photographing, and updating the three-dimensional reconstruction model according to a plurality of continuous depth image sequences photographed by the unmanned aerial vehicle array at the target observation point.

Description

动态场景的三维重建方法以及装置和系统、服务器、介质3D reconstruction method for dynamic scene, device and system, server, medium
本申请要求在2018年02月23日提交中国专利局、申请号为201810155616.4的中国专利申请的优先权,该申请的全部内容通过引用结合在本申请中。The present application claims the priority of the Chinese Patent Application No. 201101155616.4, filed on Jan. 23, 2011, the entire disclosure of which is hereby incorporated by reference.
技术领域Technical field
本发明实施例涉及计算机视觉技术领域,例如涉及一种动态场景的三维重建方法以及装置和系统、服务器、介质。Embodiments of the present invention relate to the field of computer vision technology, for example, to a three-dimensional reconstruction method of a dynamic scene, an apparatus and system, a server, and a medium.
背景技术Background technique
随着消费级深度相机的逐渐普及,特别是最新的Iphone X中更是内置了基于结构光的深度相机,使得基于动态三维重建的虚拟现实、混合现实应用成为可能,并具有广泛的应用前景和重要的应用价值。With the gradual popularization of consumer-grade depth cameras, especially the latest Iphone X has a built-in depth camera based on structured light, which makes virtual reality and mixed reality applications based on dynamic 3D reconstruction possible, and has broad application prospects. Important application value.
相关的动态场景三维重建方法通常依赖价格昂贵的激光扫描仪,虽然精度较高,但是扫描过程依赖于额外的穿戴式传感器,损害了采集者的舒适度。此外,还可以利用相机阵列系统来实现动态场景三维重建,但是,这种方法由于受限于固定的相机阵列,拍摄空间非常受限,而且需要额外的人力资源来控制拍摄相机和选择拍摄视点,无法全自动完成重建过程。Related dynamic scene 3D reconstruction methods often rely on expensive laser scanners, although the accuracy is high, but the scanning process relies on additional wearable sensors, which compromises the comfort of the collector. In addition, the camera array system can also be used to realize dynamic scene 3D reconstruction. However, this method is limited by the fixed camera array, the shooting space is very limited, and additional human resources are needed to control the camera and select the shooting viewpoint. The rebuild process cannot be fully automated.
发明内容Summary of the invention
本发明实施例提供一种动态场景的三维重建方法以及装置和系统、服务器、介质,以克服在不影响采集者舒适度并且不受拍摄空间限制的条件下如何自动完成动态场景三维重建的缺陷。Embodiments of the present invention provide a three-dimensional reconstruction method for a dynamic scene, a device and a system, a server, and a medium, to overcome the defect of how to automatically complete the three-dimensional reconstruction of the dynamic scene without affecting the comfort of the collector and being restricted by the shooting space.
第一方面,本发明实施例提供了一种动态场景的三维重建方法,该方法包括:In a first aspect, an embodiment of the present invention provides a method for three-dimensional reconstruction of a dynamic scene, where the method includes:
获取动态场景的多个连续深度图像序列,其中,所述多个连续深度图像序列是由搭载深度相机的无人机阵列拍摄得到;Acquiring a plurality of consecutive depth image sequences of the dynamic scene, wherein the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera;
对所述多个连续深度图像序列进行融合,建立所述动态场景的三维重建模型;Performing fusion on the plurality of consecutive depth image sequences to establish a three-dimensional reconstruction model of the dynamic scene;
根据所述三维重建模型和无人机阵列当前的位姿计算得到无人机阵列的目标观测点;Calculating a target observation point of the UAV array according to the three-dimensional reconstruction model and the current pose of the UAV array;
指示无人机阵列移动至所述目标观测点进行拍摄,并根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新所述三维重建模型。The UAV array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
第二方面,本发明实施例还提供了一种动态场景的三维重建装置,该装置包括:In a second aspect, the embodiment of the present invention further provides a three-dimensional reconstruction device for a dynamic scene, the device comprising:
图像序列获取模块,设置为获取动态场景的多个连续深度图像序列,其中,所述多个连续深度图像序列是由搭载深度相机的无人机阵列拍摄得到;An image sequence acquisition module, configured to acquire a plurality of consecutive depth image sequences of the dynamic scene, wherein the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera;
图像融合模块,设置为对所述多个连续深度图像序列进行融合,建立所述动态场景的三维重建模型;An image fusion module is configured to fuse the plurality of consecutive depth image sequences to establish a three-dimensional reconstruction model of the dynamic scene;
目标观测点计算模块,设置为根据所述三维重建模型和无人机阵列当前的位姿计算得到无人机阵列的目标观测点;a target observation point calculation module, configured to calculate a target observation point of the UAV array according to the three-dimensional reconstruction model and the current pose of the UAV array;
重建模型更新模块,设置为指示无人机阵列移动至所述目标观测点进行拍摄,并根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新所述三维重建模型。And a reconstruction model updating module is configured to instruct the UAV array to move to the target observation point for shooting, and update the three-dimensional reconstruction model according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
第三方面,本发明实施例还提供了一种动态场景的三维重建系统,该系统包括无人机阵列和三维重建平台;In a third aspect, an embodiment of the present invention further provides a three-dimensional reconstruction system for a dynamic scene, where the system includes an unmanned aerial vehicle array and a three-dimensional reconstruction platform;
其中,无人机阵列中的每个无人机搭载有深度相机,设置为拍摄动态场景的深度图像序列;Wherein each drone in the UAV array is equipped with a depth camera, which is set to capture a depth image sequence of the dynamic scene;
三维重建平台包括如本申请任一实施例所述的动态场景的三维重建装置,设置为根据无人机阵列拍摄的多个连续深度图像序列生成动态场景的三维重建模型。The three-dimensional reconstruction platform includes a three-dimensional reconstruction device for a dynamic scene according to any one of the embodiments of the present application, and is configured to generate a three-dimensional reconstruction model of the dynamic scene according to a plurality of consecutive depth image sequences captured by the UAV array.
第四方面,本发明实施例还提供了一种服务器,包括:In a fourth aspect, the embodiment of the present invention further provides a server, including:
一个或多个处理器;One or more processors;
存储装置,设置为存储一个或多个程序,a storage device configured to store one or more programs,
当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如本申请任一实施例所述的动态场景的三维重建方法。When the one or more programs are executed by the one or more processors, the one or more processors implement a three-dimensional reconstruction method of a dynamic scene as described in any of the embodiments of the present application.
第五方面,本发明实施例还提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现如本申请任一实施例所述的动态场景的三维重建方法。In a fifth aspect, the embodiment of the present invention further provides a computer readable storage medium, where the computer program is stored, and the program is executed by the processor to implement a three-dimensional reconstruction method of a dynamic scene according to any embodiment of the present application. .
附图概述BRIEF abstract
图1是本发明实施例一提供的动态场景的三维重建方法的流程图;1 is a flowchart of a method for reconstructing a dynamic scene according to Embodiment 1 of the present invention;
图2是本发明实施例二提供的动态场景的三维重建方法的流程图;2 is a flowchart of a method for reconstructing a dynamic scene according to Embodiment 2 of the present invention;
图3是本发明实施例三提供的动态场景的三维重建方法的流程图;3 is a flowchart of a method for reconstructing a dynamic scene according to Embodiment 3 of the present invention;
图4是本发明实施例四提供的动态场景的三维重建装置的结构示意图;4 is a schematic structural diagram of a three-dimensional reconstruction apparatus for a dynamic scene according to Embodiment 4 of the present invention;
图5是本发明实施例五提供的动态场景的三维重建系统的结构示意图;5 is a schematic structural diagram of a three-dimensional reconstruction system for a dynamic scene according to Embodiment 5 of the present invention;
图6是本发明实施例六提供的一种服务器的结构示意图。FIG. 6 is a schematic structural diagram of a server according to Embodiment 6 of the present invention.
具体实施方式Detailed ways
下面结合附图和实施例对本申请作进一步的详细说明。可以理解的是,此处所描述的具体实施例仅仅用于解释本申请,而非对本申请的限定。另外还需要说明的是,为了便于描述,附图中仅示出了与本申请相关的部分而非全部结构。The present application will be further described in detail below with reference to the accompanying drawings and embodiments. It is understood that the specific embodiments described herein are merely illustrative of the application and are not intended to be limiting. In addition, it should be noted that, for the convenience of description, only some but not all of the structures related to the present application are shown in the drawings.
实施例一 Embodiment 1
图1是本发明实施例一提供的动态场景的三维重建方法的流程图,本实施例可适用于对动态场景进行三维重建的情况,所述动态场景例如舞者在舞台跳舞的场景。该方法可以由动态场景的三维重建装置来执行,该装置可以采用软件和/或硬件的方式实现,并可集成在服务器中。如图1所示,该方法例如可以包括:步骤S110至步骤S140。1 is a flowchart of a method for three-dimensional reconstruction of a dynamic scene according to Embodiment 1 of the present invention. The present embodiment is applicable to a situation in which a dynamic scene is three-dimensionally reconstructed, such as a scene in which a dancer dances on a stage. The method can be performed by a three-dimensional reconstruction device of a dynamic scene, the device can be implemented in software and/or hardware, and can be integrated in a server. As shown in FIG. 1, the method may include, for example, steps S110 to S140.
在步骤S110中,获取动态场景的多个连续深度图像序列。In step S110, a plurality of consecutive depth image sequences of the dynamic scene are acquired.
其中,所述多个连续深度图像序列是由搭载深度相机的无人机阵列拍摄得到。The plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera.
其中,无人机阵列中可包括多台无人机,例如3台或5台,例如可根据动态场景的实际需要进行配置,本发明实施例对此并不做任何限定。无人机阵列中的不同无人机可以位于不同的视点同时对动态场景进行拍摄,从而从不同角度获取动态场景的深度图像序列,以便更好地进行三维重建。The UAS array may include multiple UAVs, for example, 3 or 5 U.S. units, for example, may be configured according to the actual needs of the dynamic scenario, which is not limited in this embodiment of the present invention. Different drones in the UAV array can simultaneously capture dynamic scenes at different viewpoints to obtain depth image sequences of dynamic scenes from different angles for better 3D reconstruction.
每台无人机都搭载有深度相机,用于拍摄动态场景的深度图像。深度图像是包含与视点的场景对象表面的距离有关的信息的图像或图像通道,深度图像中每个像素值是相机距离物体的实际距离,通过该距离可以构建三维模型。而通过搭载深度相机的无人机阵列来拍摄,不会如相关技术中的固定相机阵列那样受到拍摄空间的制约,而且还可以控制无人机阵列自动进行拍摄。Each drone is equipped with a depth camera for taking deep images of dynamic scenes. A depth image is an image or image channel that contains information about the distance from the surface of the scene object of the viewpoint. Each pixel value in the depth image is the actual distance of the camera from the object, by which a three-dimensional model can be constructed. The shooting by the drone array equipped with the depth camera is not restricted by the shooting space as in the fixed camera array of the related art, and the drone array can be controlled to automatically take the image.
初始时,可控制无人机阵列位于动态场景上方的初始位置并同时进行拍摄,因为要对动态场景进行三维重建,动态场景中人物或景物的位置或姿态等是实 时发生变化的,所以每个无人机会持续不断的进行拍摄,并实时地将拍摄到的连续深度图像序列发送至三维重建装置进行处理。其中,所述连续深度图像序列是指按照时间顺序连续拍摄到的深度图像序列,通常地,深度相机可以在每秒内连续拍摄30帧图像,每个图像按照时间顺序排列得到图像序列。Initially, the drone array can be controlled to be located at the initial position above the dynamic scene and simultaneously photographed, because the dynamic scene is three-dimensionally reconstructed, and the position or posture of the person or scene in the dynamic scene changes in real time, so each The unmanned person continues to shoot and sends the captured continuous depth image sequence to the 3D reconstruction device for processing in real time. The continuous depth image sequence refers to a sequence of depth images continuously captured in chronological order. Generally, the depth camera can continuously capture 30 frames of images per second, and each image is arranged in chronological order to obtain a sequence of images.
在一实施例中,所述获取动态场景的多个连续深度图像序列包括:In an embodiment, the acquiring a plurality of consecutive depth image sequences of the dynamic scene includes:
获取所述无人机阵列拍摄得到的所述动态场景的多个原始深度图像序列;Obtaining a plurality of original depth image sequences of the dynamic scene captured by the UAV array;
根据同步时间戳对齐所述多个原始深度图像序列,得到所述多个连续深度图像序列。Aligning the plurality of original depth image sequences according to a synchronization time stamp to obtain the plurality of consecutive depth image sequences.
其中,所述多个原始深度图像序列即为不同的无人机分别拍摄到的原始图像序列,而虽然无人机阵列是同时拍摄的,但是在时间上也会存在一定的误差,因此,需要根据同步时间戳对这些原始深度图像序列进行对齐,从而确保不同的无人机从不同视角拍摄到的深度图像序列具有时间上的一致性,由此可以提高重建模型的准确性。Wherein, the plurality of original depth image sequences are original image sequences captured by different drones, and although the UAV array is simultaneously photographed, there is a certain error in time, therefore, it is required These original depth image sequences are aligned according to the synchronization time stamp, thereby ensuring temporal consistency of different depth image sequences captured by different drones from different perspectives, thereby improving the accuracy of the reconstruction model.
在步骤S120中,对所述多个连续深度图像序列进行融合,建立所述动态场景的三维重建模型。In step S120, the plurality of consecutive depth image sequences are fused to establish a three-dimensional reconstruction model of the dynamic scene.
在一实施例中,可以是根据深度相机的内参矩阵将连续深度图像序列投影到三维空间中获得三维点云,然后对三维点云进行配准和融合,最终建立出三维重建模型。实现时,可以利用相关技术中的配准和融合的算法来进行,此处不再一一赘述。In an embodiment, the continuous depth image sequence may be projected into the three-dimensional space according to the internal reference matrix of the depth camera to obtain a three-dimensional point cloud, and then the three-dimensional point cloud is registered and fused, and finally a three-dimensional reconstruction model is established. When implemented, it can be performed by using the registration and fusion algorithms in the related art, and will not be repeated here.
在步骤S130中,根据所述三维重建模型和无人机阵列当前的位姿确定无人机阵列的目标观测点。In step S130, the target observation point of the UAV array is determined according to the three-dimensional reconstruction model and the current pose of the UAV array.
由于动态场景中的人物或景物的位置或姿态等会实时发生变化,因此,如何才能让无人机阵列顺应动态场景的变化在最合适的位置进行拍摄是决定重建模型效果的重要问题。因此,在本发明实施例中,通过实时地计算出下一时刻无人机阵列的目标观测点,该目标观测点即为最佳观测点,从而实时地指示无人机阵列移动至该目标观测点进行拍摄,从而更新重建模型,实现动态场景的精确重现。Since the position or posture of a person or a scene in a dynamic scene changes in real time, how to make the drone array conform to the change of the dynamic scene at the most suitable position is an important issue that determines the effect of reconstructing the model. Therefore, in the embodiment of the present invention, the target observation point of the UAV array at the next moment is calculated in real time, and the target observation point is the optimal observation point, thereby indicating that the UAV array moves to the target observation in real time. The point is taken to update the reconstructed model to achieve accurate reproduction of the dynamic scene.
其中,所述位姿可以由无人机的旋转角度和平移距离来表示,相应的,对无人机的控制也可包括旋转和平移两个参数。当确定出目标观测点后,则可以控制无人机移动到该最佳的目标观测点,并控制无人机从当前观测点至目标观测点的旋转角度,即控制无人机的最佳拍摄视点。而目标观测点的确定可以是 按照预先设定的标准,对无人机可能存在的拍摄点进行评估,将评估结果符合该标准的拍摄点确定为最佳的目标观测点。其中,可能存在的拍摄点可以根据无人机阵列当前的位姿来确定,而评估的过程可以根据所述可能存在的观测点和建立得到的三维重建模型来进行,评估在不同的可能存在的观测点中,依据哪一个观测点拍摄到的深度图像序列建立的三维重建模型的效果符合预设的标准,例如可以通过计算可能存在的观测点的能量函数来进行评估。Wherein, the pose can be represented by a rotation angle and a translation distance of the drone, and correspondingly, the control of the drone can also include two parameters of rotation and translation. After the target observation point is determined, the drone can be controlled to move to the optimal target observation point, and the rotation angle of the drone from the current observation point to the target observation point is controlled, that is, the optimal shooting of the drone is controlled. Viewpoint. The target observation point can be determined according to a preset standard, and the shooting points that may exist in the drone are evaluated, and the shooting point whose evaluation result meets the standard is determined as the best target observation point. Wherein, the possible shooting points may be determined according to the current pose of the UAV array, and the evaluation process may be performed according to the possible existing observation points and the established three-dimensional reconstruction model, and the evaluation may exist in different possible In the observation point, the effect of the three-dimensional reconstruction model established by the depth image sequence captured by which observation point conforms to a preset criterion, for example, by calculating the energy function of the observation point that may exist.
在步骤S140中,指示无人机阵列移动至所述目标观测点进行拍摄,并根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新所述三维重建模型。In step S140, the drone array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
本实施例的技术方案利用无人机阵列对动态场景进行拍摄,根据拍摄到的多个连续深度图像序列进行图像融合,得到动态场景的三维重建模型,因此,不需要依赖额外的设备,确保了采集者的舒适度。并且,在重建过程中,实时地通过目标观测点的计算指示无人机阵列移动至该目标观测点进行拍摄,根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新模型,从而获得更加准确的三维重建模型,而且不受拍摄空间的制约,能够自动完成重建过程。The technical solution of the embodiment uses the UAV array to capture the dynamic scene, and performs image fusion according to the captured multiple consecutive depth image sequences to obtain a three-dimensional reconstruction model of the dynamic scene. Therefore, it is not necessary to rely on additional equipment to ensure the The comfort of the collector. Moreover, in the reconstruction process, the calculation of the target observation point in real time indicates that the UAV array moves to the target observation point for shooting, and the model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point, thereby A more accurate 3D reconstruction model is obtained, and the reconstruction process can be completed automatically without being restricted by the shooting space.
实施例二 Embodiment 2
图2是本发明实施例二提供的动态场景的三维重建方法的流程图。如图2所示,该方法例如可以包括:步骤S210至步骤S270。FIG. 2 is a flowchart of a three-dimensional reconstruction method of a dynamic scene according to Embodiment 2 of the present invention. As shown in FIG. 2, the method may include, for example, steps S210 to S270.
在步骤S210中,获取动态场景的多个连续深度图像序列。In step S210, a plurality of consecutive depth image sequences of the dynamic scene are acquired.
其中,所述多个连续深度图像序列是由搭载深度相机的无人机阵列拍摄得到。The plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera.
在步骤S220中,对所述多个连续深度图像序列进行融合,按照预设周期确定关键帧重建体。In step S220, the plurality of consecutive depth image sequences are fused, and the key frame reconstructed body is determined according to a preset period.
在步骤S230中,在每一预设周期内,确定当前关键帧重建体中的非刚性变形节点的形变参数,根据所述形变参数将当前关键帧重建体里的重建模型更新至当前数据帧重建体中。In step S230, in each preset period, determining a deformation parameter of the non-rigid deformation node in the current key frame reconstruction body, and updating the reconstruction model in the current key frame reconstruction body to the current data frame reconstruction according to the deformation parameter In the body.
其中,所述当前数据帧重建体是指每一时刻实时的重建体。The current data frame reconstructed body refers to a reconstructed body in real time at each moment.
在步骤S240中,从当前数据帧重建体中提取出所述动态场景的三维重建模型。In step S240, a three-dimensional reconstruction model of the dynamic scene is extracted from the current data frame reconstructed body.
在步骤S250中,用当前数据帧重建体替代当前关键帧重建体,作为下一预设周期内的关键帧重建体。In step S250, the current key frame reconstructed body is replaced with the current data frame reconstructed body as the key frame reconstructed body in the next preset period.
其中,在步骤S220-步骤S250实际上是利用关键帧策略融合深度图像,并获得动态场景实时的重建模型。Wherein, in step S220-step S250, the depth image is actually merged by using a key frame strategy, and a real-time reconstruction model of the dynamic scene is obtained.
在一实施例中,根据多个连续深度图像序列通过图像融合可以建立初始的三维重建模型,然后按照预设周期,例如100帧,确定三维重建模型的关键帧重建体,并在每一预设周期内,执行步骤S230-步骤S250的操作。其中,非刚性变形节点可以表示动态场景中人物或景物发生变化的节点,通过确定非刚性变形节点的形变参数将当前关键帧重建体里的重建模型更新至当前数据帧重建体中,进而从当前数据帧重建体中提取出三维重建模型,从而可以捕捉到动态场景的变化细节,提高重建模型的精确度,避免出现错误和混乱,同时也可以避免卡顿。最后,用当前数据帧重建体替代当前关键帧重建体,作为下一预设周期内的关键帧重建体,由此通过当前数据帧重建体和关键帧重建体的迭代,实现对动态场景中每一个动态变化场景的重现。In an embodiment, an initial three-dimensional reconstruction model may be established by image fusion according to a plurality of consecutive depth image sequences, and then a key frame reconstruction body of the three-dimensional reconstruction model is determined according to a preset period, for example, 100 frames, and each preset is During the period, the operations of steps S230 to S250 are performed. The non-rigid deformation node may represent a node in which a character or a scene changes in a dynamic scene, and the reconstruction model in the current key frame reconstruction body is updated to the current data frame reconstruction body by determining the deformation parameter of the non-rigid deformation node, and then from the current The 3D reconstruction model is extracted from the data frame reconstruction body, so that the details of the dynamic scene can be captured, the accuracy of the reconstruction model can be improved, errors and confusion can be avoided, and the jam can be avoided. Finally, the current data frame reconstruction body is used to replace the current key frame reconstruction body as the key frame reconstruction body in the next preset period, thereby implementing the iteration of the current data frame reconstruction body and the key frame reconstruction body, thereby realizing each of the dynamic scenes. A recurrence of a dynamically changing scene.
其中,重建体可以理解为在三维重建过程中的一种假设,假设重建体可以包围住整个动态场景(或重建对象),重建体由很多个均匀的体素构成,通过配准、融合等算法在这个重建体之上建立出动态场景的三维重建模型。重建体中的节点以及节点的形变参数表征了动态场景的特征,因此,从重建体中可以提取出动态场景的三维重建模型。在本实施例中,通过上述关键帧策略融合深度图像,可以避免当点云配准不准确时数据融合产生的误差。Among them, the reconstructed body can be understood as a hypothesis in the process of 3D reconstruction. It is assumed that the reconstructed body can surround the entire dynamic scene (or reconstructed object). The reconstructed body is composed of many uniform voxels, through registration, fusion and other algorithms. A three-dimensional reconstruction model of the dynamic scene is built on top of this reconstructed body. The nodes in the reconstructed body and the deformation parameters of the nodes represent the characteristics of the dynamic scene. Therefore, the three-dimensional reconstruction model of the dynamic scene can be extracted from the reconstructed body. In this embodiment, by integrating the depth image by the key frame strategy described above, errors caused by data fusion when the point cloud registration is inaccurate can be avoided.
其中,所述形变参数包括每个形变节点的旋转及平移参数,其计算过程例如可以通过求解非刚性运动的能量方程来获得,该能量方程由非刚性运动约束项和局部刚性运动约束项构成,分别由如下公式表示:Wherein, the deformation parameter includes rotation and translation parameters of each deformation node, and the calculation process thereof can be obtained, for example, by solving an energy equation of a non-rigid motion constraint, the energy equation being composed of a non-rigid motion constraint term and a local rigid motion constraint term. They are represented by the following formulas:
Figure PCTCN2019083816-appb-000001
Figure PCTCN2019083816-appb-000001
Figure PCTCN2019083816-appb-000002
Figure PCTCN2019083816-appb-000002
其中,在非刚性运动约束项E n中,
Figure PCTCN2019083816-appb-000003
Figure PCTCN2019083816-appb-000004
分别表示经过非刚性运动驱动后的模型顶点坐标及其法向,u i表示同一匹配点对中三维点云的位置坐标,c i表示匹配点对集合中的第i个元素,
Figure PCTCN2019083816-appb-000005
Figure PCTCN2019083816-appb-000006
分别表示经过非刚性运动驱动后的模型顶点坐标及其法向。非刚性运动约束项E n保证经过非刚性运动驱动后的模型与从深度图获得的三维点云尽可能的对齐。
Wherein, in the non-rigid motion constraint terms in E n,
Figure PCTCN2019083816-appb-000003
with
Figure PCTCN2019083816-appb-000004
Representing the vertex coordinates of the model and its normal direction driven by the non-rigid motion, respectively, u i represents the position coordinates of the 3D point cloud in the same matching point pair, and c i represents the i-th element in the set of matching point pairs.
Figure PCTCN2019083816-appb-000005
with
Figure PCTCN2019083816-appb-000006
Represents the vertex coordinates of the model and its normal direction driven by non-rigid motion. Non-rigid motion constraints guarantee term E n through the non-rigid motion model driven and aligned as the three-dimensional point cloud obtained from the depth map.
在局部非刚性运动约束项E g中,i表示模型上第i个顶点,
Figure PCTCN2019083816-appb-000007
表示模型上 第i个顶点周围的邻近顶点的集合,
Figure PCTCN2019083816-appb-000008
Figure PCTCN2019083816-appb-000009
分别代表已知非刚性运动对模型表面顶点v i和v j的驱动作用,
Figure PCTCN2019083816-appb-000010
Figure PCTCN2019083816-appb-000011
代表作用在v i和v j上的非刚性运动同时作用在v j上的位置变换效果,即要保证模型上邻近顶点的非刚性驱动效果要尽可能的一致。
In the local non-rigid motion constraint term E g , i represents the i-th vertex on the model,
Figure PCTCN2019083816-appb-000007
Represents a collection of adjacent vertices around the ith vertex on the model,
Figure PCTCN2019083816-appb-000008
with
Figure PCTCN2019083816-appb-000009
Representing the driving effects of known non-rigid motion on the surface vertices v i and v j of the model, respectively.
Figure PCTCN2019083816-appb-000010
with
Figure PCTCN2019083816-appb-000011
It represents the positional transformation effect of the non-rigid motion acting on v i and v j simultaneously on v j , that is, to ensure that the non-rigid driving effects of adjacent vertices on the model are as uniform as possible.
非刚性运动约束项E n保证经过非刚性运动驱动后的模型与从深度图像获得的三维点云尽可能的对齐,局部刚性运动约束项E g可以在使模型整体受局部刚性约束运动的同时保证较大幅度的合理的非刚性运动也能被很好的解算出来。为了利用指数映射方法对变形后的顶点做如下近似: The non-rigid motion constraint term E n ensures that the model driven by the non-rigid motion is aligned with the three-dimensional point cloud obtained from the depth image as much as possible, and the local rigid motion constraint E g can be guaranteed while the model as a whole is subjected to the local rigid constraint motion. A large amount of reasonable non-rigid motion can also be well solved. In order to make use of the exponential mapping method, the deformed vertices are approximated as follows:
Figure PCTCN2019083816-appb-000012
Figure PCTCN2019083816-appb-000012
其中,
Figure PCTCN2019083816-appb-000013
为截至上一帧的模型顶点v i的累积变换矩阵,为已知量;I为四维单位阵;
Figure PCTCN2019083816-appb-000014
Figure PCTCN2019083816-appb-000015
即上一帧变换后的模型顶点,则经过变换有:
among them,
Figure PCTCN2019083816-appb-000013
The cumulative transformation matrix of the model vertex v i up to the previous frame is a known quantity; I is a four-dimensional unit matrix;
Figure PCTCN2019083816-appb-000014
make
Figure PCTCN2019083816-appb-000015
That is, the model vertices after the previous frame transformation are transformed:
Figure PCTCN2019083816-appb-000016
Figure PCTCN2019083816-appb-000016
对于每个顶点,要求解的未知参数即为六维变换参数x=(v 1,v 2,v 3,w x,w y,w z) TFor each vertex, the unknown parameter that requires the solution is the six-dimensional transformation parameter x = (v 1 , v 2 , v 3 , w x , w y , w z ) T .
在步骤S260中,根据所述三维重建模型和无人机阵列当前的位姿确定无人机阵列的目标观测点。In step S260, the target observation point of the UAV array is determined according to the three-dimensional reconstruction model and the current pose of the UAV array.
在步骤S270中,指示无人机阵列移动至所述目标观测点进行拍摄,并根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新所述三维重建模型。In step S270, the drone array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the drone array at the target observation point.
本实施例的技术方案利用无人机阵列对动态场景进行拍摄,根据拍摄到的 多个连续深度图像序列进行图像融合,并采用关键帧策略,最终得到动态场景的三维重建模型,在不受拍摄空间的制约、自动完成重建过程的基础上,提高了重建模型的准确性,避免出现因点云配准不准确而造成的误差。同时,不需要依赖额外的设备,确保了采集者的舒适度。The technical solution of the embodiment utilizes an unmanned aerial vehicle array to capture a dynamic scene, performs image fusion according to a plurality of consecutive continuous depth image sequences, and adopts a key frame strategy to finally obtain a three-dimensional reconstruction model of the dynamic scene, which is not subject to shooting. On the basis of space constraints and automatic completion of the reconstruction process, the accuracy of the reconstruction model is improved, and errors caused by inaccurate registration of point clouds are avoided. At the same time, there is no need to rely on additional equipment to ensure the comfort of the collector.
实施例三Embodiment 3
图3是本发明实施例三提供的动态场景的三维重建方法的流程图,本实施例是在上述实施例的基础上,对目标观测点的计算操作进一步进行优化。如图3所示,该方法例如可以包括:步骤S310至步骤S360。FIG. 3 is a flowchart of a three-dimensional reconstruction method of a dynamic scene according to Embodiment 3 of the present invention. In this embodiment, the calculation operation of the target observation point is further optimized based on the foregoing embodiment. As shown in FIG. 3, the method may include, for example, steps S310 to S360.
在步骤S310中,获取动态场景的多个连续深度图像序列。In step S310, a plurality of consecutive depth image sequences of the dynamic scene are acquired.
其中,所述多个连续深度图像序列是由搭载深度相机的无人机阵列拍摄得到。The plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera.
在步骤S320中,对所述多个连续深度图像序列进行融合,建立所述动态场景的三维重建模型。In step S320, the plurality of consecutive depth image sequences are fused to establish a three-dimensional reconstruction model of the dynamic scene.
在步骤S330中,根据无人机阵列当前的位姿,对其空间邻域进行栅格化,建立候选观测点集合。In step S330, according to the current pose of the UAV array, the spatial neighborhood is rasterized to establish a set of candidate observation points.
无人机阵列当前的位姿表征了无人机阵列中每个无人机当前的观测视点,包括当前的坐标和拍摄角度等。根据当前的位姿以及预设的距离划定空间邻域范围,并通过对空间邻域栅格化的方式确定候选观测点,即栅格化后每一个节点代表一个候选观测点。The current pose of the UAV array characterizes the current observation point of each UAV in the UAV array, including the current coordinates and shooting angle. The spatial neighborhood range is delineated according to the current pose and the preset distance, and the candidate observation points are determined by rasterizing the spatial neighborhood, that is, each node represents a candidate observation point after rasterization.
在步骤S340中,利用有效性能量函数确定候选观测点集合中每一个候选观测点的总能量值。In step S340, the total energy value of each of the candidate observation points in the set of candidate observation points is determined using the validity energy function.
在步骤S350中,将总能量值符合预设标准的候选观测点作为所述目标观测点。In step S350, candidate observation points whose total energy values meet the preset criteria are taken as the target observation points.
在一实施例中,所述有效性能量函数包括深度能量项、中心能量项和运动能量项;In an embodiment, the validity energy function includes a depth energy term, a central energy term, and a motion energy term;
其中,深度能量项用于确定候选观测点的平均深度值接近目标深度值的程度;Wherein, the depth energy term is used to determine the extent to which the average depth value of the candidate observation point is close to the target depth value;
中心能量项用于确定候选观测点观测到的重建模型与采集的图像画幅的中心部位的接近程度;The central energy term is used to determine how close the reconstructed model observed by the candidate observation point is to the central portion of the captured image frame;
运动能量项用于确定候选观测点观测到的动态场景中发生运动的部分的多 少。The kinetic energy term is used to determine the amount of motion occurring in the dynamic scene observed by the candidate observation point.
在一实施例中,所述有效性能量函数由如下公式表示:In an embodiment, the validity energy function is represented by the following formula:
E t=λ dE dcE cmE m E td E dc E cm E m
其中,E t为总能量项,Ed为深度能量项,Ec为中心能量项,Em为运动能量项,λ d、λ c和λ m分别为与深度能量项、中心能量项和运动能量项对应的权重系数; Where E t is the total energy term, Ed is the depth energy term, Ec is the central energy term, Em is the motion energy term, and λ d , λ c and λ m correspond to the depth energy term, the central energy term and the motion energy term, respectively. Weight coefficient;
所述深度能量项、中心能量项和运动能量项分别由如下公式表示:The depth energy term, the center energy term, and the motion energy term are respectively represented by the following formula:
E d=ψ(d avg-d o) E d =ψ(d avg -d o )
Figure PCTCN2019083816-appb-000017
Figure PCTCN2019083816-appb-000017
Figure PCTCN2019083816-appb-000018
Figure PCTCN2019083816-appb-000018
Figure PCTCN2019083816-appb-000019
Figure PCTCN2019083816-appb-000019
Figure PCTCN2019083816-appb-000020
Figure PCTCN2019083816-appb-000020
其中,T c和T V分别为无人机阵列和候选观测点在重建模型中的位姿;t v为候选观测点位姿的平移分量;x n为射线所击中的重建模型的体素;N x为该体素的法向;x i表示重建模型发生非刚性变形的节点;x′ i表示非刚性变形后的节点;π()表示从三维空间到二维像平面的投影透视变换;d avg和d o分别表示候选观测点的平均深度值和目标深度值;ψ()函数表示对距离的惩罚项;r为候选观测点投射并穿过重建模型的光线;du和dv分别表示重建模型在候选观测点的平均投影像素横坐标和纵坐标;λ为阻尼因子;φ1和φ2分别用于统计候选观测点的所有射线的运动信息和所有观察到的变形节点的运动信息。 Where T c and T V are the poses of the UAV array and the candidate observation points in the reconstruction model respectively; t v is the translation component of the pose of the candidate observation point; x n is the voxel of the reconstruction model hit by the ray N x is the normal of the voxel; x i is the node where the reconstructed model is non-rigid deformed; x' i is the node after the non-rigid deformation; π() is the projection perspective transform from the three-dimensional space to the two-dimensional image plane d avg and d o represent the average depth value and the target depth value of the candidate observation point respectively; the ψ() function represents the penalty term for the distance; r is the light of the candidate observation point projected through the reconstructed model; du and dv represent respectively The average projected pixel abscissa and ordinate of the reconstructed model at the candidate observation points; λ is the damping factor; φ1 and φ2 are used to calculate the motion information of all the rays of the candidate observation points and the motion information of all the observed deformation nodes, respectively.
通过对深度能量项、中心能量项和运动能量项的加权求和,可以对候选观测点进行综合评估,确定无人机在哪个观测点拍摄到的深度图像序列建立的三维重建模型的效果符合预设的标准,即综合考虑了在候选观测点采集的深度图像的平均深度、平均中心程度和累计运动信息,使得在目标观测点获取的深度图像更有利于当前的动态场景重建。在一实施例中,可以选择总能量值最大的候选观测点作为最优的目标观测点。Through the weighted summation of the depth energy term, the central energy term and the motion energy term, the candidate observation points can be comprehensively evaluated to determine the effect of the 3D reconstruction model established by the depth image sequence captured by the UAV at which observation point. The standard is to comprehensively consider the average depth, average center degree and cumulative motion information of the depth image collected at the candidate observation points, so that the depth image acquired at the target observation point is more favorable for the current dynamic scene reconstruction. In an embodiment, candidate observation points with the largest total energy value may be selected as the optimal target observation points.
在步骤S360中,指示无人机阵列移动至所述目标观测点进行拍摄,并根据 无人机阵列在目标观测点拍摄的多个连续深度图像序列更新所述三维重建模型。In step S360, the drone array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the drone array at the target observation point.
本实施例的技术方案利用无人机阵列对动态场景进行拍摄,根据拍摄到的多个连续深度图像序列进行图像融合,得到动态场景的三维重建模型,并通过有效性能量函数对候选观测点进行计算和评估,确定最优的目标观测点,并指示无人机阵列移动至该目标观测点进行拍摄,由此,不仅实现了自动拍摄和重建,还可以提高三维模型的重建效果,而且简单易行,具有广阔的应用前景。The technical solution of the embodiment uses a UAV array to capture a dynamic scene, performs image fusion according to the captured multiple consecutive depth image sequences, obtains a three-dimensional reconstruction model of the dynamic scene, and performs candidate observation points through a validity energy function. Calculate and evaluate, determine the optimal target observation point, and instruct the drone array to move to the target observation point for shooting, thereby not only achieving automatic shooting and reconstruction, but also improving the reconstruction effect of the three-dimensional model, and is simple and easy Line, has broad application prospects.
实施例四Embodiment 4
图4是本发明实施例四提供的动态场景的三维重建装置的结构示意图。本实施例可适用于对动态场景进行三维重建的情况,所述动态场景例如舞者在舞台跳舞的场景。本发明实施例所提供的动态场景的三维重建装置可执行本申请任意实施例所提供的动态场景的三维重建方法,具备执行方法相应的功能模块和有益效果。如图4所示,该装置包括:图像序列获取模块410、图像融合模块420、目标观测点计算模块430和重建模型更新模块440。4 is a schematic structural diagram of a three-dimensional reconstruction apparatus for a dynamic scene according to Embodiment 4 of the present invention. This embodiment can be applied to a case where a dynamic scene is three-dimensionally reconstructed, such as a scene in which a dancer dances on a stage. The three-dimensional reconstruction device of the dynamic scene provided by the embodiment of the present invention can perform the three-dimensional reconstruction method of the dynamic scene provided by any embodiment of the present application, and has the corresponding functional modules and beneficial effects of the execution method. As shown in FIG. 4, the apparatus includes an image sequence acquisition module 410, an image fusion module 420, a target observation point calculation module 430, and a reconstruction model update module 440.
图像序列获取模块410,设置为获取动态场景的多个连续深度图像序列,其中,所述多个连续深度图像序列是由搭载深度相机的无人机阵列拍摄得到。The image sequence acquisition module 410 is configured to acquire a plurality of consecutive depth image sequences of the dynamic scene, wherein the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera.
图像融合模块420,设置为对所述多个连续深度图像序列进行融合,建立所述动态场景的三维重建模型。The image fusion module 420 is configured to fuse the plurality of consecutive depth image sequences to establish a three-dimensional reconstruction model of the dynamic scene.
目标观测点计算模块430,设置为根据所述三维重建模型和无人机阵列当前的位姿确定无人机阵列的目标观测点。The target observation point calculation module 430 is configured to determine a target observation point of the UAV array according to the three-dimensional reconstruction model and the current pose of the UAV array.
重建模型更新模块440,设置为指示无人机阵列移动至所述目标观测点进行拍摄,并根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新所述三维重建模型。The reconstruction model update module 440 is configured to instruct the UAV array to move to the target observation point for shooting, and update the three-dimensional reconstruction model according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
在一实施例中,图像序列获取模块410包括:In an embodiment, the image sequence acquisition module 410 includes:
原始图像序列获取单元,设置为获取所述无人机阵列拍摄得到的所述动态场景的多个原始深度图像序列;An original image sequence acquiring unit, configured to acquire a plurality of original depth image sequences of the dynamic scene captured by the UAV array;
图像序列对齐单元,设置为根据同步时间戳对齐所述多个原始深度图像序列,得到所述多个连续深度图像序列。And an image sequence alignment unit configured to align the plurality of original depth image sequences according to a synchronization time stamp to obtain the plurality of consecutive depth image sequences.
在一实施例中,图像融合模块420还设置为:In an embodiment, the image fusion module 420 is further configured to:
对所述多个连续深度图像序列进行融合,按照预设周期确定关键帧重建体,并在每一预设周期内,执行如下操作:The plurality of consecutive depth image sequences are fused, and the key frame reconstructed body is determined according to a preset period, and in each preset period, the following operations are performed:
确定当前关键帧重建体中的非刚性变形节点的形变参数,根据所述形变参数将当前关键帧重建体里的重建模型更新至当前数据帧重建体中,其中,所述当前数据帧重建体是指每一时刻实时的重建体;Determining a deformation parameter of the non-rigid deformation node in the current key frame reconstruction body, and updating the reconstruction model in the current key frame reconstruction body to the current data frame reconstruction body according to the deformation parameter, wherein the current data frame reconstruction body is Refers to the real-time reconstructed body at each moment;
从当前数据帧重建体中提取出所述动态场景的三维重建模型;Extracting a three-dimensional reconstruction model of the dynamic scene from a current data frame reconstructed body;
用当前数据帧重建体替代当前关键帧重建体,作为下一预设周期内的关键帧重建体。The current data frame reconstructed body is replaced with the current data frame reconstructed body as a key frame reconstructed body in the next preset period.
在一实施例中,目标观测点计算模块430包括:In an embodiment, the target observation point calculation module 430 includes:
候选观测点建立单元,设置为根据无人机阵列当前的位姿,对其空间邻域进行栅格化,建立候选观测点集合;The candidate observation point establishing unit is configured to rasterize the spatial neighborhood according to the current pose of the UAV array to establish a set of candidate observation points;
能量值计算单元,设置为利用有效性能量函数确定候选观测点集合中每一个候选观测点的总能量值;An energy value calculation unit configured to determine a total energy value of each of the candidate observation points in the set of candidate observation points by using a validity energy function;
目标观测点确定单元,设置为将总能量值符合预设标准的候选观测点作为所述目标观测点。The target observation point determining unit is configured to select a candidate observation point whose total energy value conforms to a preset criterion as the target observation point.
在一实施例中,所述有效性能量函数包括深度能量项、中心能量项和运动能量项;In an embodiment, the validity energy function includes a depth energy term, a central energy term, and a motion energy term;
其中,深度能量项用于确定候选观测点的平均深度值接近目标深度值的程度;Wherein, the depth energy term is used to determine the extent to which the average depth value of the candidate observation point is close to the target depth value;
中心能量项用于确定候选观测点观测到的重建模型与采集的图像画幅的中心部位的接近程度;The central energy term is used to determine how close the reconstructed model observed by the candidate observation point is to the central portion of the captured image frame;
运动能量项用于确定候选观测点观测到的动态场景中发生运动的部分的多少。The kinetic energy term is used to determine the amount of motion occurring in the dynamic scene observed by the candidate observation point.
在一实施例中,所述有效性能量函数由如下公式表示:In an embodiment, the validity energy function is represented by the following formula:
E t=λ dE dcE cmE m E td E dc E cm E m
其中,E t为总能量项,Ed为深度能量项,Ec为中心能量项,Em为运动能量项,λ d、λ c和λ m分别为与深度能量项、中心能量项和运动能量项对应的权重系数; Where E t is the total energy term, Ed is the depth energy term, Ec is the central energy term, Em is the motion energy term, and λ d , λ c and λ m correspond to the depth energy term, the central energy term and the motion energy term, respectively. Weight coefficient;
所述深度能量项、中心能量项和运动能量项分别由如下公式表示:The depth energy term, the center energy term, and the motion energy term are respectively represented by the following formula:
E d=ψ(d avg-d o) E d =ψ(d avg -d o )
Figure PCTCN2019083816-appb-000021
Figure PCTCN2019083816-appb-000021
Figure PCTCN2019083816-appb-000022
Figure PCTCN2019083816-appb-000022
Figure PCTCN2019083816-appb-000023
Figure PCTCN2019083816-appb-000023
Figure PCTCN2019083816-appb-000024
Figure PCTCN2019083816-appb-000024
其中,T c和T V分别为无人机阵列和候选观测点在重建模型中的位姿;t v为候选观测点位姿的平移分量;x n为射线所击中的重建模型的体素;N x为该体素的法向;x i表示重建模型发生非刚性变形的节点;x′ i表示非刚性变形后的节点;π()表示从三维空间到二维像平面的投影透视变换;d avg和d o分别表示候选观测点的平均深度值和目标深度值;ψ()函数表示对距离的惩罚项;r为候选观测点投射并穿过重建模型的光线;du和dv分别表示重建模型在候选观测点的平均投影像素横坐标和纵坐标;λ为阻尼因子;φ1和φ2分别用于统计候选观测点的所有射线的运动信息和所有观察到的变形节点的运动信息。 Where T c and T V are the poses of the UAV array and the candidate observation points in the reconstruction model respectively; t v is the translation component of the pose of the candidate observation point; x n is the voxel of the reconstruction model hit by the ray N x is the normal of the voxel; x i is the node where the reconstructed model is non-rigid deformed; x' i is the node after the non-rigid deformation; π() is the projection perspective transform from the three-dimensional space to the two-dimensional image plane d avg and d o represent the average depth value and the target depth value of the candidate observation point respectively; the ψ() function represents the penalty term for the distance; r is the light of the candidate observation point projected through the reconstructed model; du and dv represent respectively The average projected pixel abscissa and ordinate of the reconstructed model at the candidate observation points; λ is the damping factor; φ1 and φ2 are used to calculate the motion information of all the rays of the candidate observation points and the motion information of all the observed deformation nodes, respectively.
本实施例的技术方案利用无人机阵列对动态场景进行拍摄,根据拍摄到的多个连续深度图像序列进行图像融合,得到动态场景的三维重建模型,因此,不需要依赖额外的设备,确保了采集者的舒适度。并且,在重建过程中,实时地通过目标观测点的计算指示无人机阵列移动至该目标观测点进行拍摄,根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新模型,从而获得更加准确的三维重建模型,而且不受拍摄空间的制约,能够自动完成重建过程。The technical solution of the embodiment uses the UAV array to capture the dynamic scene, and performs image fusion according to the captured multiple consecutive depth image sequences to obtain a three-dimensional reconstruction model of the dynamic scene. Therefore, it is not necessary to rely on additional equipment to ensure the The comfort of the collector. Moreover, in the reconstruction process, the calculation of the target observation point in real time indicates that the UAV array moves to the target observation point for shooting, and the model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point, thereby A more accurate 3D reconstruction model is obtained, and the reconstruction process can be completed automatically without being restricted by the shooting space.
实施例五Embodiment 5
图5是本发明实施例五提供的动态场景的三维重建系统的结构示意图,如图5所示,包括无人机阵列1和三维重建平台2。FIG. 5 is a schematic structural diagram of a three-dimensional reconstruction system for a dynamic scene according to Embodiment 5 of the present invention. As shown in FIG. 5, the UAV array 1 and the three-dimensional reconstruction platform 2 are included.
其中,无人机阵列1中的每个无人机搭载有深度相机,设置为拍摄动态场景的深度图像序列。示例性的,图5中示出了无人机阵列1包括三个无人机,分别是无人机11、无人机12和无人机13,但本发明实施例对无人机阵列中无人机的数量并不做任何限定,可以根据要重建的动态场景的实际情况进行配置。Among them, each drone in the UAV array 1 is equipped with a depth camera, which is set to capture a depth image sequence of a dynamic scene. Exemplarily, FIG. 5 shows that the UAV array 1 includes three UAVs, namely the UAV 11, the UAV 12 and the UAV 13, but the embodiment of the present invention is in the UAV array. The number of drones is not limited, and can be configured according to the actual situation of the dynamic scene to be reconstructed.
三维重建平台2包括上述任一实施例中所述的动态场景的三维重建装置21,设置为根据无人机阵列拍摄的多个连续深度图像序列生成动态场景的三维重建模型。The three-dimensional reconstruction platform 2 includes the three-dimensional reconstruction device 21 of the dynamic scene described in any of the above embodiments, and is configured to generate a three-dimensional reconstruction model of the dynamic scene according to the plurality of consecutive depth image sequences captured by the UAV array.
在一实施例中,三维重建平台2还包括无线通信模块22,与无人机阵列1无线连接,设置为接收无人机阵列拍摄的多个连续深度图像序列,还设置为将三维重建装置22计算得到的目标观测点的位置信息发送至无人机阵列1。In an embodiment, the three-dimensional reconstruction platform 2 further includes a wireless communication module 22, and is wirelessly connected to the UAV array 1 and configured to receive a plurality of consecutive depth image sequences captured by the UAV array, and is further configured to set the three-dimensional reconstruction device 22 The calculated position information of the target observation point is sent to the drone array 1.
相应的,无人机阵列1中的每个无人机还包括导航模块,设置为根据所述位置信息控制无人机移动到目标观测点对动态场景进行拍摄。Correspondingly, each of the UAV arrays 1 further includes a navigation module configured to control the drone to move to the target observation point to capture the dynamic scene according to the position information.
本实施例的技术方案利用无人机阵列对动态场景进行拍摄,根据拍摄到的多个连续深度图像序列进行图像融合,得到动态场景的三维重建模型,因此,不需要依赖额外的设备,确保了采集者的舒适度。并且,在重建过程中,实时地通过目标观测点的计算指示无人机阵列移动至该目标观测点进行拍摄,根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新模型,从而获得更加准确的三维重建模型,而且不受拍摄空间的制约,能够自动完成重建过程。The technical solution of the embodiment uses the UAV array to capture the dynamic scene, and performs image fusion according to the captured multiple consecutive depth image sequences to obtain a three-dimensional reconstruction model of the dynamic scene. Therefore, it is not necessary to rely on additional equipment to ensure the The comfort of the collector. Moreover, in the reconstruction process, the calculation of the target observation point in real time indicates that the UAV array moves to the target observation point for shooting, and the model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point, thereby A more accurate 3D reconstruction model is obtained, and the reconstruction process can be completed automatically without being restricted by the shooting space.
实施例六Embodiment 6
图6是本发明实施例六提供的一种服务器的结构示意图。图6示出了适于用来实现本发明实施方式的示例性服务器612的框图。图6显示的服务器612仅仅是一个示例,不应对本发明实施例的功能和使用范围带来任何限制。FIG. 6 is a schematic structural diagram of a server according to Embodiment 6 of the present invention. FIG. 6 shows a block diagram of an exemplary server 612 suitable for use in implementing embodiments of the present invention. The server 612 shown in FIG. 6 is merely an example and should not impose any limitation on the function and scope of use of the embodiments of the present invention.
如图6所示,服务器612以通用服务器的形式表现。服务器612的组件可以包括但不限于:一个或者多个处理器616,存储装置628,连接不同系统组件(包括存储装置628和处理器616)的总线618。As shown in Figure 6, server 612 is represented in the form of a general purpose server. Components of server 612 may include, but are not limited to, one or more processors 616, storage 628, and bus 618 that connect different system components, including storage 628 and processor 616.
总线618表示几类总线结构中的一种或多种,包括存储装置总线或者存储装置控制器,外围总线,图形加速端口,处理器或者使用多种总线结构中的任意总线结构的局域总线。举例来说,这些体系结构包括但不限于工业标准体系结构(Industry Subversive Alliance,ISA)总线,微通道体系结构(Micro Channel Architecture,MAC)总线,增强型ISA总线、视频电子标准协会(Video Electronics Standards Association,VESA)局域总线以及外围组件互连(Peripheral Component Interconnect,PCI)总线。 Bus 618 represents one or more of several types of bus structures, including a memory device bus or memory device controller, a peripheral bus, a graphics acceleration port, a processor, or a local bus using any of a variety of bus structures. For example, these architectures include, but are not limited to, the Industry Subversive Alliance (ISA) bus, the Micro Channel Architecture (MAC) bus, the Enhanced ISA Bus, and the Video Electronics Standards Association. Association, VESA) Local Bus and Peripheral Component Interconnect (PCI) bus.
服务器612典型地包括多种计算机系统可读介质。这些介质可以是任何能够被服务器612访问的可用介质,包括易失性和非易失性介质,可移动的和不可移动的介质。 Server 612 typically includes a variety of computer system readable media. These media can be any available media that can be accessed by server 612, including volatile and non-volatile media, removable and non-removable media.
存储装置628可以包括易失性存储器形式的计算机系统可读介质,例如随机存取存储器(Random Access Memory,RAM)630和/或高速缓存存储器632。 服务器612可以进一步包括其它可移动/不可移动的、易失性/非易失性计算机系统存储介质。仅作为举例,存储系统634可以用于读写不可移动的、非易失性磁介质(图6未显示,通常称为“硬盘驱动器”)。尽管图6中未示出,可以提供用于对可移动非易失性磁盘(例如“软盘”)读写的磁盘驱动器,以及对可移动非易失性光盘,例如只读光盘(Compact Disc Read-Only Memory,CD-ROM),数字视盘(Digital Video Disc-Read Only Memory,DVD-ROM)或者其它光介质)读写的光盘驱动器。在这些情况下,每个驱动器可以通过一个或者多个数据介质接口与总线618相连。存储装置628可以包括至少一个程序产品,该程序产品具有一组(例如至少一个)程序模块,这些程序模块被配置以执行本申请各实施例的功能。 Storage device 628 can include computer system readable media in the form of volatile memory, such as random access memory (RAM) 630 and/or cache memory 632. Server 612 may further include other removable/non-removable, volatile/non-volatile computer system storage media. By way of example only, storage system 634 can be used to read and write non-removable, non-volatile magnetic media (not shown in Figure 6, commonly referred to as "hard disk drives"). Although not shown in FIG. 6, a disk drive for reading and writing to a removable non-volatile disk (for example, a "floppy disk"), and a removable non-volatile disk such as a read-only disk (Compact Disc Read) may be provided. -Only Memory, CD-ROM), Digital Video Disc-Read Only Memory (DVD-ROM) or other optical media. In these cases, each drive can be coupled to bus 618 via one or more data medium interfaces. Storage device 628 can include at least one program product having a set (eg, at least one) of program modules configured to perform the functions of the various embodiments of the present application.
具有一组(至少一个)程序模块642的程序/实用工具640,可以存储在例如存储装置628中,这样的程序模块642包括但不限于操作系统、一个或者多个应用程序、其它程序模块以及程序数据,这些示例中的每一个或某种组合中可能包括网络环境的实现。程序模块642通常执行本申请所描述的实施例中的功能和/或方法。A program/utility 640 having a set (at least one) of program modules 642, which may be stored, for example, in storage device 628, such program program 642 includes but is not limited to an operating system, one or more applications, other program modules, and programs Data, each of these examples or some combination may include an implementation of a network environment. Program module 642 typically performs the functions and/or methods of the embodiments described herein.
服务器612也可以与一个或多个外部设备614(例如键盘、指向设备、显示器624等)通信,还可与一个或者多个使得用户能与该服务器612交互的设备通信,和/或与使得该服务器612能与一个或多个其它计算设备进行通信的任何设备(例如网卡,调制解调器等等)通信。这种通信可以通过输入/输出(I/O)接口622进行。并且,服务器612还可以通过网络适配器620与一个或者多个网络(例如局域网(Local Area Network,LAN),广域网(Wide Area Network,WAN)和/或公共网络,例如因特网)通信。如图6所示,网络适配器620通过总线618与服务器612的其它模块通信。应当明白,尽管图中未示出,可以结合服务器612使用其它硬件和/或软件模块,包括但不限于:微代码、设备驱动器、冗余处理器、外部磁盘驱动阵列、磁盘阵列(Redundant Arrays of Independent Disks,RAID)系统、磁带驱动器以及数据备份存储系统等。 Server 612 may also be in communication with one or more external devices 614 (eg, a keyboard, pointing device, display 624, etc.), and may also be in communication with one or more devices that enable a user to interact with the server 612, and/or Server 612 can communicate with any device (e.g., network card, modem, etc.) that is in communication with one or more other computing devices. This communication can take place via an input/output (I/O) interface 622. Moreover, the server 612 can also communicate with one or more networks (eg, a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet) through the network adapter 620. As shown in FIG. 6, network adapter 620 communicates with other modules of server 612 via bus 618. It should be understood that although not shown in the figures, other hardware and/or software modules may be utilized in connection with server 612, including but not limited to: microcode, device drivers, redundant processors, external disk drive arrays, disk arrays (Redundant Arrays of Independent Disks, RAID) systems, tape drives, and data backup storage systems.
处理器616通过运行存储在存储装置628中的程序,从而执行各种功能应用以及数据处理,例如实现本发明实施例所提供的动态场景的三维重建方法,包括:The processor 616 performs a three-dimensional reconstruction method of the dynamic scene provided by the embodiment of the present invention by executing a program stored in the storage device 628, and the method includes:
获取动态场景的多个连续深度图像序列,其中,所述多个连续深度图像序列是由搭载深度相机的无人机阵列拍摄得到;Acquiring a plurality of consecutive depth image sequences of the dynamic scene, wherein the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera;
对所述多个连续深度图像序列进行融合,建立所述动态场景的三维重建模型;Performing fusion on the plurality of consecutive depth image sequences to establish a three-dimensional reconstruction model of the dynamic scene;
根据所述三维重建模型和无人机阵列当前的位姿确定无人机阵列的目标观测点;Determining a target observation point of the UAV array according to the three-dimensional reconstruction model and the current pose of the UAV array;
指示无人机阵列移动至所述目标观测点进行拍摄,并根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新所述三维重建模型。The UAV array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
实施例七Example 7
本发明实施例七还提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现如本发明实施例所提供的动态场景的三维重建方法,包括:The seventh embodiment of the present invention further provides a computer readable storage medium, where the computer program is stored, and the program is executed by the processor to implement a three-dimensional reconstruction method of a dynamic scene according to the embodiment of the present invention, including:
获取动态场景的多个连续深度图像序列,其中,所述多个连续深度图像序列是由搭载深度相机的无人机阵列拍摄得到;Acquiring a plurality of consecutive depth image sequences of the dynamic scene, wherein the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera;
对所述多个连续深度图像序列进行融合,建立所述动态场景的三维重建模型;Performing fusion on the plurality of consecutive depth image sequences to establish a three-dimensional reconstruction model of the dynamic scene;
根据所述三维重建模型和无人机阵列当前的位姿确定无人机阵列的目标观测点;Determining a target observation point of the UAV array according to the three-dimensional reconstruction model and the current pose of the UAV array;
指示无人机阵列移动至所述目标观测点进行拍摄,并根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新所述三维重建模型。The UAV array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
本发明实施例的计算机存储介质,可以采用一个或多个计算机可读的介质的任意组合。计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子(非穷举的列表)包括:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(Erasable Programmable Read Only Memory,EPROM,或闪存)、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本文件中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。The computer storage medium of the embodiments of the present invention may employ any combination of one or more computer readable mediums. The computer readable medium can be a computer readable signal medium or a computer readable storage medium. The computer readable storage medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the above. More specific examples (non-exhaustive lists) of computer readable storage media include: electrical connections having one or more wires, a portable computer disk, a hard disk, a random access memory (RAM), a read only memory (ROM), Erasable Programmable Read Only Memory (EPROM), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the foregoing . In this document, a computer readable storage medium can be any tangible medium that can contain or store a program, which can be used by or in connection with an instruction execution system, apparatus or device.
计算机可读的信号介质可以包括在基带中或者作为载波一部分传播的数据 信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读的信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。The computer readable signal medium may comprise a data signal that is propagated in the baseband or as part of a carrier, carrying computer readable program code. Such propagated data signals can take a variety of forms including, but not limited to, electromagnetic signals, optical signals, or any suitable combination of the foregoing. The computer readable signal medium can also be any computer readable medium other than a computer readable storage medium, which can transmit, propagate, or transport a program for use by or in connection with the instruction execution system, apparatus, or device. .
计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括——但不限于无线、电线、光缆、射频(Radio Frequency,RF)等等,或者上述的任意合适的组合。Program code embodied on a computer readable medium can be transmitted by any suitable medium, including but not limited to wireless, wire, optical cable, radio frequency (RF), and the like, or any suitable combination of the foregoing.
可以以一种或多种程序设计语言或其组合来编写用于执行本申请操作的计算机程序代码,所述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。Computer program code for performing the operations of the present application may be written in one or more programming languages, or a combination thereof, including an object oriented programming language such as Java, Smalltalk, C++, and conventional Procedural programming language—such as the "C" language or a similar programming language. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer, partly on the remote computer, or entirely on the remote computer or server. In the case of a remote computer, the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or can be connected to an external computer (eg, using an Internet service provider) Internet connection).

Claims (16)

  1. 一种动态场景的三维重建方法,包括:A three-dimensional reconstruction method for a dynamic scene, comprising:
    获取动态场景的多个连续深度图像序列,其中,所述多个连续深度图像序列是由搭载深度相机的无人机阵列拍摄得到;Acquiring a plurality of consecutive depth image sequences of the dynamic scene, wherein the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera;
    对所述多个连续深度图像序列进行融合,建立所述动态场景的三维重建模型;Performing fusion on the plurality of consecutive depth image sequences to establish a three-dimensional reconstruction model of the dynamic scene;
    根据所述三维重建模型和无人机阵列当前的位姿确定无人机阵列的目标观测点;Determining a target observation point of the UAV array according to the three-dimensional reconstruction model and the current pose of the UAV array;
    指示无人机阵列移动至所述目标观测点进行拍摄,并根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新所述三维重建模型。The UAV array is instructed to move to the target observation point for shooting, and the three-dimensional reconstruction model is updated according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
  2. 根据权利要求1所述的方法,其中,所述获取动态场景的多个连续深度图像序列包括:The method of claim 1, wherein the acquiring a plurality of consecutive depth image sequences of the dynamic scene comprises:
    获取所述无人机阵列拍摄得到的所述动态场景的多个原始深度图像序列;Obtaining a plurality of original depth image sequences of the dynamic scene captured by the UAV array;
    根据同步时间戳对齐所述多个原始深度图像序列,得到所述多个连续深度图像序列。Aligning the plurality of original depth image sequences according to a synchronization time stamp to obtain the plurality of consecutive depth image sequences.
  3. 根据权利要求1或2所述的方法,其中,对所述多个连续深度图像序列进行融合,建立所述动态场景的三维重建模型,包括:The method according to claim 1 or 2, wherein the merging of the plurality of consecutive depth image sequences to establish a three-dimensional reconstruction model of the dynamic scene comprises:
    对所述多个连续深度图像序列进行融合,按照预设周期确定关键帧重建体,并在每一预设周期内,执行如下操作:The plurality of consecutive depth image sequences are fused, and the key frame reconstructed body is determined according to a preset period, and in each preset period, the following operations are performed:
    确定当前关键帧重建体中的非刚性变形节点的形变参数,根据所述形变参数将当前关键帧重建体里的重建模型更新至当前数据帧重建体中,其中,所述当前数据帧重建体是指每一时刻实时的重建体;Determining a deformation parameter of the non-rigid deformation node in the current key frame reconstruction body, and updating the reconstruction model in the current key frame reconstruction body to the current data frame reconstruction body according to the deformation parameter, wherein the current data frame reconstruction body is Refers to the real-time reconstructed body at each moment;
    从当前数据帧重建体中提取出所述动态场景的三维重建模型;Extracting a three-dimensional reconstruction model of the dynamic scene from a current data frame reconstructed body;
    用当前数据帧重建体替代当前关键帧重建体,作为下一预设周期内的关键帧重建体。The current data frame reconstructed body is replaced with the current data frame reconstructed body as a key frame reconstructed body in the next preset period.
  4. 根据权利要求1所述的方法,其中,根据所述三维重建模型和无人机阵列当前的位姿确定无人机阵列的目标观测点,包括:The method according to claim 1, wherein the target observation points of the UAV array are determined according to the three-dimensional reconstruction model and the current pose of the UAV array, including:
    根据无人机阵列当前的位姿,对无人机阵列的空间邻域进行栅格化,建立候选观测点集合;According to the current pose of the UAV array, the spatial neighborhood of the UAV array is rasterized to establish a set of candidate observation points;
    利用有效性能量函数确定候选观测点集合中每一个候选观测点的总能量值;Determining a total energy value of each candidate observation point in the set of candidate observation points by using a validity energy function;
    将总能量值符合预设标准的候选观测点作为所述目标观测点。A candidate observation point whose total energy value conforms to a preset criterion is taken as the target observation point.
  5. 根据权利要求4所述的方法,其中,所述有效性能量函数包括深度能量 项、中心能量项和运动能量项;The method of claim 4 wherein said effectiveness energy function comprises a depth energy term, a central energy term, and a kinematic energy term;
    其中,深度能量项用于确定候选观测点的平均深度值接近目标深度值的程度;Wherein, the depth energy term is used to determine the extent to which the average depth value of the candidate observation point is close to the target depth value;
    中心能量项用于确定候选观测点观测到的重建模型与采集的图像画幅的中心部位的接近程度;The central energy term is used to determine how close the reconstructed model observed by the candidate observation point is to the central portion of the captured image frame;
    运动能量项用于确定候选观测点观测到的动态场景中发生运动的部分的多少。The kinetic energy term is used to determine the amount of motion occurring in the dynamic scene observed by the candidate observation point.
  6. 根据权利要求5所述的方法,其中,所述有效性能量函数由如下公式表示:The method of claim 5 wherein said validity energy function is represented by the following formula:
    E t=λ dE dcE cmE m E td E dc E cm E m
    其中,E t为总能量项,Ed为深度能量项,Ec为中心能量项,Em为运动能量项,λ d、λ c和λ m分别为与深度能量项、中心能量项和运动能量项对应的权重系数; Where E t is the total energy term, Ed is the depth energy term, Ec is the central energy term, Em is the motion energy term, and λ d , λ c and λ m correspond to the depth energy term, the central energy term and the motion energy term, respectively. Weight coefficient;
    所述深度能量项、中心能量项和运动能量项分别由如下公式表示:The depth energy term, the center energy term, and the motion energy term are respectively represented by the following formula:
    E d=ψ(d avg-d o) E d =ψ(d avg -d o )
    Figure PCTCN2019083816-appb-100001
    Figure PCTCN2019083816-appb-100001
    Figure PCTCN2019083816-appb-100002
    Figure PCTCN2019083816-appb-100002
    Figure PCTCN2019083816-appb-100003
    Figure PCTCN2019083816-appb-100003
    Figure PCTCN2019083816-appb-100004
    Figure PCTCN2019083816-appb-100004
    其中,T c和T V分别为无人机阵列和候选观测点在重建模型中的位姿;t v为候选观测点位姿的平移分量;x n为射线所击中的重建模型的体素;N x为该体素的法向;x i表示重建模型发生非刚性变形的节点;x′ i表示非刚性变形后的节点;π()表示从三维空间到二维像平面的投影透视变换;d avg和d o分别表示候选观测点的平均深度值和目标深度值;ψ()函数表示对距离的惩罚项;r为候选观测点投射并穿过重建模型的光线;du和dv分别表示重建模型在候选观测点的平均投影像素横坐标和纵坐标;λ为阻尼因子;φ1和φ2分别用于统计候选观测点的所有射线的运动信息和所有观察到的变形节点的运动信息。 Where T c and T V are the poses of the UAV array and the candidate observation points in the reconstruction model respectively; t v is the translation component of the pose of the candidate observation point; x n is the voxel of the reconstruction model hit by the ray N x is the normal of the voxel; x i is the node where the reconstructed model is non-rigid deformed; x' i is the node after the non-rigid deformation; π() is the projection perspective transform from the three-dimensional space to the two-dimensional image plane d avg and d o represent the average depth value and the target depth value of the candidate observation point respectively; the ψ() function represents the penalty term for the distance; r is the light of the candidate observation point projected through the reconstructed model; du and dv represent respectively The average projected pixel abscissa and ordinate of the reconstructed model at the candidate observation points; λ is the damping factor; φ1 and φ2 are used to calculate the motion information of all the rays of the candidate observation points and the motion information of all the observed deformation nodes, respectively.
  7. 一种动态场景的三维重建装置,包括:A three-dimensional reconstruction device for a dynamic scene, comprising:
    图像序列获取模块,设置为获取动态场景的多个连续深度图像序列,其中,所述多个连续深度图像序列是由搭载深度相机的无人机阵列拍摄得到;An image sequence acquisition module, configured to acquire a plurality of consecutive depth image sequences of the dynamic scene, wherein the plurality of consecutive depth image sequences are captured by a drone array equipped with a depth camera;
    图像融合模块,设置为对所述多个连续深度图像序列进行融合,建立所述动态场景的三维重建模型;An image fusion module is configured to fuse the plurality of consecutive depth image sequences to establish a three-dimensional reconstruction model of the dynamic scene;
    目标观测点计算模块,设置为根据所述三维重建模型和无人机阵列当前的位姿确定无人机阵列的目标观测点;a target observation point calculation module, configured to determine a target observation point of the UAV array according to the three-dimensional reconstruction model and the current pose of the UAV array;
    重建模型更新模块,设置为指示无人机阵列移动至所述目标观测点进行拍摄,并根据无人机阵列在目标观测点拍摄的多个连续深度图像序列更新所述三维重建模型。And a reconstruction model updating module is configured to instruct the UAV array to move to the target observation point for shooting, and update the three-dimensional reconstruction model according to a plurality of consecutive depth image sequences captured by the UAV array at the target observation point.
  8. 根据权利要求7所述的装置,其中,所述图像序列获取模块包括:The apparatus of claim 7, wherein the image sequence acquisition module comprises:
    原始图像序列获取单元,设置为获取所述无人机阵列拍摄得到的所述动态场景的多个原始深度图像序列;An original image sequence acquiring unit, configured to acquire a plurality of original depth image sequences of the dynamic scene captured by the UAV array;
    图像序列对齐单元,设置为根据同步时间戳对齐所述多个原始深度图像序列,得到所述多个连续深度图像序列。And an image sequence alignment unit configured to align the plurality of original depth image sequences according to a synchronization time stamp to obtain the plurality of consecutive depth image sequences.
  9. 根据权利要求7或8所述的装置,其中,所述图像融合模块还设置为:The apparatus according to claim 7 or 8, wherein the image fusion module is further configured to:
    对所述多个连续深度图像序列进行融合,按照预设周期确定关键帧重建体,并在每一预设周期内,执行如下操作:The plurality of consecutive depth image sequences are fused, and the key frame reconstructed body is determined according to a preset period, and in each preset period, the following operations are performed:
    确定当前关键帧重建体中的非刚性变形节点的形变参数,根据所述形变参数将当前关键帧重建体里的重建模型更新至当前数据帧重建体中,其中,所述当前数据帧重建体是指每一时刻实时的重建体;Determining a deformation parameter of the non-rigid deformation node in the current key frame reconstruction body, and updating the reconstruction model in the current key frame reconstruction body to the current data frame reconstruction body according to the deformation parameter, wherein the current data frame reconstruction body is Refers to the real-time reconstructed body at each moment;
    从当前数据帧重建体中提取出所述动态场景的三维重建模型;Extracting a three-dimensional reconstruction model of the dynamic scene from a current data frame reconstructed body;
    用当前数据帧重建体替代当前关键帧重建体,作为下一预设周期内的关键帧重建体。The current data frame reconstructed body is replaced with the current data frame reconstructed body as a key frame reconstructed body in the next preset period.
  10. 根据权利要求7所述的装置,其中,所述目标观测点计算模块包括:The apparatus of claim 7, wherein the target observation point calculation module comprises:
    候选观测点建立单元,设置为根据无人机阵列当前的位姿,对其空间邻域进行栅格化,建立候选观测点集合;The candidate observation point establishing unit is configured to rasterize the spatial neighborhood according to the current pose of the UAV array to establish a set of candidate observation points;
    能量值计算单元,设置为利用有效性能量函数确定候选观测点集合中每一个候选观测点的总能量值;An energy value calculation unit configured to determine a total energy value of each of the candidate observation points in the set of candidate observation points by using a validity energy function;
    目标观测点确定单元,设置为将总能量值符合预设标准的候选观测点作为所述目标观测点。The target observation point determining unit is configured to select a candidate observation point whose total energy value conforms to a preset criterion as the target observation point.
  11. 根据权利要求10所述的装置,其中,所述有效性能量函数包括深度能量项、中心能量项和运动能量项;The apparatus of claim 10, wherein the validity energy function comprises a depth energy term, a center energy term, and a motion energy term;
    其中,深度能量项用于确定候选观测点的平均深度值接近目标深度值的程度;Wherein, the depth energy term is used to determine the extent to which the average depth value of the candidate observation point is close to the target depth value;
    中心能量项用于确定候选观测点观测到的重建模型与采集的图像画幅的中心部位的接近程度;The central energy term is used to determine how close the reconstructed model observed by the candidate observation point is to the central portion of the captured image frame;
    运动能量项用于确定候选观测点观测到的动态场景中发生运动的部分的多少。The kinetic energy term is used to determine the amount of motion occurring in the dynamic scene observed by the candidate observation point.
  12. 根据权利要求11所述的装置,其中,所述有效性能量函数由如下公式表示:The apparatus of claim 11 wherein said validity energy function is represented by the following formula:
    E t=λ dE dcE cmE m E td E dc E cm E m
    其中,E t为总能量项,Ed为深度能量项,Ec为中心能量项,Em为运动能量项,λ d、λ c和λ m分别为与深度能量项、中心能量项和运动能量项对应的权重系数; Where E t is the total energy term, Ed is the depth energy term, Ec is the central energy term, Em is the motion energy term, and λ d , λ c and λ m correspond to the depth energy term, the central energy term and the motion energy term, respectively. Weight coefficient;
    所述深度能量项、中心能量项和运动能量项分别由如下公式表示:The depth energy term, the center energy term, and the motion energy term are respectively represented by the following formula:
    E d=ψ(d avg-d o) E d =ψ(d avg -d o )
    Figure PCTCN2019083816-appb-100005
    Figure PCTCN2019083816-appb-100005
    Figure PCTCN2019083816-appb-100006
    Figure PCTCN2019083816-appb-100006
    Figure PCTCN2019083816-appb-100007
    Figure PCTCN2019083816-appb-100007
    Figure PCTCN2019083816-appb-100008
    Figure PCTCN2019083816-appb-100008
    其中,T c和T V分别为无人机阵列和候选观测点在重建模型中的位姿;t v为候选观测点位姿的平移分量;x n为射线所击中的重建模型的体素;N x为该体素的法向;x i表示重建模型发生非刚性变形的节点;x′ i表示非刚性变形后的节点;π()表示从三维空间到二维像平面的投影透视变换;d avg和d o分别表示候选观测点的平均深度值和目标深度值;ψ()函数表示对距离的惩罚项;r为候选观测点投射并穿过重建模型的光线;du和dv分别表示重建模型在候选观测点的平均投影像素横坐标和纵坐标;λ为阻尼因子;φ1和φ2分别用于统计候选观测点的所有 射线的运动信息和所有观察到的变形节点的运动信息。 Where T c and T V are the poses of the UAV array and the candidate observation points in the reconstruction model respectively; t v is the translation component of the pose of the candidate observation point; x n is the voxel of the reconstruction model hit by the ray N x is the normal of the voxel; x i is the node where the reconstructed model is non-rigid deformed; x' i is the node after the non-rigid deformation; π() is the projection perspective transform from the three-dimensional space to the two-dimensional image plane d avg and d o represent the average depth value and the target depth value of the candidate observation point respectively; the ψ() function represents the penalty term for the distance; r is the light of the candidate observation point projected through the reconstructed model; du and dv represent respectively The average projected pixel abscissa and ordinate of the reconstructed model at the candidate observation points; λ is the damping factor; φ1 and φ2 are used to calculate the motion information of all the rays of the candidate observation points and the motion information of all the observed deformation nodes, respectively.
  13. 一种动态场景的三维重建系统,包括无人机阵列和三维重建平台;A three-dimensional reconstruction system for dynamic scenes, including a drone array and a three-dimensional reconstruction platform;
    其中,无人机阵列中的每个无人机搭载有深度相机,所述深度相机设置为拍摄动态场景的深度图像序列;Wherein each of the drone arrays is equipped with a depth camera, and the depth camera is set to capture a depth image sequence of the dynamic scene;
    三维重建平台包括如权利要求7-12任一项所述的动态场景的三维重建装置,设置为根据无人机阵列拍摄的多个连续深度图像序列生成动态场景的三维重建模型。The three-dimensional reconstruction platform includes the three-dimensional reconstruction device of the dynamic scene according to any one of claims 7 to 12, configured to generate a three-dimensional reconstruction model of the dynamic scene according to the plurality of consecutive depth image sequences captured by the drone array.
  14. 根据权利要求13所述的系统,其中,The system of claim 13 wherein
    所述三维重建平台还包括无线通信模块,与无人机阵列无线连接,所述无线通信模块设置为接收无人机阵列拍摄的多个连续深度图像序列,还设置为将所述三维重建装置确定的目标观测点的位置信息发送至无人机阵列;The three-dimensional reconstruction platform further includes a wireless communication module wirelessly connected to the UAV array, the wireless communication module being configured to receive a plurality of consecutive depth image sequences captured by the UAV array, and further configured to determine the three-dimensional reconstruction device The location information of the target observation point is sent to the drone array;
    所述无人机阵列中的每个无人机还包括导航模块,所述导航模块设置为根据所述位置信息控制无人机移动到目标观测点对动态场景进行拍摄。Each of the UAV arrays further includes a navigation module, and the navigation module is configured to control the drone to move to the target observation point to capture the dynamic scene according to the location information.
  15. 一种服务器,包括:A server that includes:
    至少一个处理器;At least one processor;
    存储装置,设置为存储至少一个程序,a storage device configured to store at least one program,
    当所述至少一个程序被所述至少一个处理器执行,使得所述至少一个处理器实现如权利要求1-6中任一项所述的动态场景的三维重建方法。When the at least one program is executed by the at least one processor, the at least one processor implements the three-dimensional reconstruction method of the dynamic scene according to any one of claims 1-6.
  16. 一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现如权利要求1-6中任一项所述的动态场景的三维重建方法。A computer readable storage medium having stored thereon a computer program, the program being executed by a processor to implement a three-dimensional reconstruction method of a dynamic scene according to any one of claims 1-6.
PCT/CN2019/083816 2018-02-23 2019-04-23 Dynamic scene three-dimensional reconstruction method, apparatus and system, server, and medium WO2019161813A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/975,242 US11954870B2 (en) 2018-02-23 2019-04-23 Dynamic scene three-dimensional reconstruction method, apparatus and system, server, and medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201810155616.4 2018-02-23
CN201810155616.4A CN108335353B (en) 2018-02-23 2018-02-23 Three-dimensional reconstruction method, device and system of dynamic scene, server and medium

Publications (1)

Publication Number Publication Date
WO2019161813A1 true WO2019161813A1 (en) 2019-08-29

Family

ID=62929819

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/083816 WO2019161813A1 (en) 2018-02-23 2019-04-23 Dynamic scene three-dimensional reconstruction method, apparatus and system, server, and medium

Country Status (3)

Country Link
US (1) US11954870B2 (en)
CN (1) CN108335353B (en)
WO (1) WO2019161813A1 (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111462298A (en) * 2020-02-24 2020-07-28 西安电子科技大学 Method for reconstructing underwater three-dimensional scene
WO2021051358A1 (en) * 2019-09-19 2021-03-25 Beijing Voyager Technology Co., Ltd. Systems and methods for generating pose graph
CN112766215A (en) * 2021-01-29 2021-05-07 北京字跳网络技术有限公司 Face fusion method and device, electronic equipment and storage medium
CN113409444A (en) * 2021-05-21 2021-09-17 北京达佳互联信息技术有限公司 Three-dimensional reconstruction method and device, electronic equipment and storage medium
CN113808253A (en) * 2021-08-31 2021-12-17 武汉理工大学 Dynamic object processing method, system, device and medium for scene three-dimensional reconstruction
CN114022537A (en) * 2021-10-29 2022-02-08 浙江东鼎电子股份有限公司 Vehicle loading rate and unbalance loading rate analysis method for dynamic weighing area
CN114373041A (en) * 2021-12-15 2022-04-19 聚好看科技股份有限公司 Three-dimensional reconstruction method and equipment
CN114812540A (en) * 2022-06-23 2022-07-29 深圳市普渡科技有限公司 Picture construction method and device and computer equipment
CN116824070A (en) * 2023-08-31 2023-09-29 江西求是高等研究院 Real-time three-dimensional reconstruction method and system based on depth image

Families Citing this family (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108335353B (en) * 2018-02-23 2020-12-22 清华-伯克利深圳学院筹备办公室 Three-dimensional reconstruction method, device and system of dynamic scene, server and medium
CN109064555B (en) * 2018-08-06 2023-06-06 百度在线网络技术(北京)有限公司 Method, apparatus and storage medium for 3D modeling
CN109155846B (en) * 2018-08-14 2020-12-18 深圳前海达闼云端智能科技有限公司 Three-dimensional reconstruction method and device of scene, electronic equipment and storage medium
CN109040730B (en) * 2018-08-20 2020-03-17 武汉理工大学 Dynamic flower sea scene system and working method thereof
CN109410316B (en) * 2018-09-21 2023-07-07 达闼机器人股份有限公司 Method for three-dimensional reconstruction of object, tracking method, related device and storage medium
WO2020113423A1 (en) * 2018-12-04 2020-06-11 深圳市大疆创新科技有限公司 Target scene three-dimensional reconstruction method and system, and unmanned aerial vehicle
CN109708655A (en) * 2018-12-29 2019-05-03 百度在线网络技术(北京)有限公司 Air navigation aid, device, vehicle and computer readable storage medium
CN109816778B (en) * 2019-01-25 2023-04-07 北京百度网讯科技有限公司 Three-dimensional reconstruction method and device for material pile, electronic equipment and computer readable medium
CN109903326B (en) * 2019-02-28 2022-02-22 北京百度网讯科技有限公司 Method and device for determining a rotation angle of a construction machine
CN110874851A (en) * 2019-10-25 2020-03-10 深圳奥比中光科技有限公司 Method, device, system and readable storage medium for reconstructing three-dimensional model of human body
WO2021134376A1 (en) * 2019-12-30 2021-07-08 深圳市大疆创新科技有限公司 Control method, movable platform, and storage medium
CN111489392B (en) * 2020-03-30 2022-09-09 清华大学 Single target human motion posture capturing method and system in multi-person environment
CN112233228B (en) * 2020-10-28 2024-02-20 五邑大学 Unmanned aerial vehicle-based urban three-dimensional reconstruction method, device and storage medium
CN112530014B (en) * 2020-12-18 2023-07-25 北京理工大学重庆创新中心 Three-dimensional reconstruction method and device for indoor scene of multiple unmanned aerial vehicles
CN112865897B (en) * 2021-01-13 2022-08-02 山东师范大学 Non-stationary channel simulation method and system for ground scene by unmanned aerial vehicle
CN113096144B (en) * 2021-03-23 2022-07-29 清华大学 Method and device for generating dynamic human body free viewpoint video based on neural network
CN113094457B (en) * 2021-04-15 2023-11-03 成都纵横自动化技术股份有限公司 Incremental generation method of digital orthophoto map and related components
CN113436318A (en) * 2021-06-30 2021-09-24 北京市商汤科技开发有限公司 Scene reconstruction method and device, electronic equipment and computer storage medium
CN113643421B (en) * 2021-07-06 2023-08-25 北京航空航天大学 Three-dimensional reconstruction method and three-dimensional reconstruction device for image
CN114299252B (en) * 2021-12-30 2024-04-02 中国电信股份有限公司 Method and device for reconstructing universal three-dimensional model, storage medium and electronic equipment
CN114429495B (en) * 2022-03-14 2022-08-30 荣耀终端有限公司 Three-dimensional scene reconstruction method and electronic equipment
CN114812503A (en) * 2022-04-14 2022-07-29 湖北省水利水电规划勘测设计院 Cliff point cloud extraction method based on airborne laser scanning
CN114494984B (en) * 2022-04-18 2022-07-22 四川腾盾科技有限公司 Random static target three-dimensional reconstruction and positioning method based on unmanned aerial vehicle aerial photography data
CN115222875A (en) * 2022-06-01 2022-10-21 支付宝(杭州)信息技术有限公司 Model determination method, local scene reconstruction method, medium, device and product
CN115252992B (en) * 2022-07-28 2023-04-07 北京大学第三医院(北京大学第三临床医学院) Trachea cannula navigation system based on structured light stereoscopic vision
CN115048820B (en) * 2022-08-15 2022-11-04 中国长江三峡集团有限公司 Three-dimensional temperature field dynamic reconstruction method and device for electrochemical energy storage container
CN115223067B (en) * 2022-09-19 2022-12-09 季华实验室 Point cloud fusion method, device and equipment applied to unmanned aerial vehicle and storage medium
CN116437063A (en) * 2023-06-15 2023-07-14 广州科伊斯数字技术有限公司 Three-dimensional image display system and method
CN116721143B (en) * 2023-08-04 2023-10-20 南京诺源医疗器械有限公司 Depth information processing device and method for 3D medical image
CN116778066B (en) * 2023-08-24 2024-01-26 先临三维科技股份有限公司 Data processing method, device, equipment and medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105184863A (en) * 2015-07-23 2015-12-23 同济大学 Unmanned aerial vehicle aerial photography sequence image-based slope three-dimension reconstruction method
CN105225241A (en) * 2015-09-25 2016-01-06 广州极飞电子科技有限公司 The acquisition methods of unmanned plane depth image and unmanned plane
CN106780729A (en) * 2016-11-10 2017-05-31 中国人民解放军理工大学 A kind of unmanned plane sequential images batch processing three-dimensional rebuilding method
CN108335353A (en) * 2018-02-23 2018-07-27 清华-伯克利深圳学院筹备办公室 Three-dimensional rebuilding method, device and system, server, the medium of dynamic scene

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10192354B2 (en) * 2015-04-14 2019-01-29 ETAK Systems, LLC Systems and methods for obtaining accurate 3D modeling data using UAVS for cell sites
CN105698762B (en) * 2016-01-15 2018-02-23 中国人民解放军国防科学技术大学 Target method for rapidly positioning based on observation station at different moments on a kind of unit flight path
CN105786016B (en) * 2016-03-31 2019-11-05 深圳奥比中光科技有限公司 The processing method of unmanned plane and RGBD image
US10520943B2 (en) * 2016-08-12 2019-12-31 Skydio, Inc. Unmanned aerial image capture platform
WO2018063911A1 (en) * 2016-09-28 2018-04-05 Federal Express Corporation Systems and methods for monitoring the internal storage contents of a shipment storage using one or more internal monitor drones
CN106698762B (en) 2017-01-16 2023-07-04 陕西省石油化工研究设计院 Emulsion paint wastewater pretreatment device
US11740630B2 (en) * 2018-06-12 2023-08-29 Skydio, Inc. Fitness and sports applications for an autonomous unmanned aerial vehicle

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105184863A (en) * 2015-07-23 2015-12-23 同济大学 Unmanned aerial vehicle aerial photography sequence image-based slope three-dimension reconstruction method
CN105225241A (en) * 2015-09-25 2016-01-06 广州极飞电子科技有限公司 The acquisition methods of unmanned plane depth image and unmanned plane
CN106780729A (en) * 2016-11-10 2017-05-31 中国人民解放军理工大学 A kind of unmanned plane sequential images batch processing three-dimensional rebuilding method
CN108335353A (en) * 2018-02-23 2018-07-27 清华-伯克利深圳学院筹备办公室 Three-dimensional rebuilding method, device and system, server, the medium of dynamic scene

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021051358A1 (en) * 2019-09-19 2021-03-25 Beijing Voyager Technology Co., Ltd. Systems and methods for generating pose graph
CN113557548A (en) * 2019-09-19 2021-10-26 北京航迹科技有限公司 System and method for generating pose graph
CN111462298B (en) * 2020-02-24 2023-03-28 西安电子科技大学 Method for reconstructing underwater three-dimensional scene
CN111462298A (en) * 2020-02-24 2020-07-28 西安电子科技大学 Method for reconstructing underwater three-dimensional scene
CN112766215A (en) * 2021-01-29 2021-05-07 北京字跳网络技术有限公司 Face fusion method and device, electronic equipment and storage medium
CN113409444A (en) * 2021-05-21 2021-09-17 北京达佳互联信息技术有限公司 Three-dimensional reconstruction method and device, electronic equipment and storage medium
CN113409444B (en) * 2021-05-21 2023-07-11 北京达佳互联信息技术有限公司 Three-dimensional reconstruction method, three-dimensional reconstruction device, electronic equipment and storage medium
CN113808253B (en) * 2021-08-31 2023-08-15 武汉理工大学 Method, system, equipment and medium for processing dynamic object of three-dimensional reconstruction of scene
CN113808253A (en) * 2021-08-31 2021-12-17 武汉理工大学 Dynamic object processing method, system, device and medium for scene three-dimensional reconstruction
CN114022537B (en) * 2021-10-29 2023-05-05 浙江东鼎电子股份有限公司 Method for analyzing loading rate and unbalanced loading rate of vehicle in dynamic weighing area
CN114022537A (en) * 2021-10-29 2022-02-08 浙江东鼎电子股份有限公司 Vehicle loading rate and unbalance loading rate analysis method for dynamic weighing area
CN114373041A (en) * 2021-12-15 2022-04-19 聚好看科技股份有限公司 Three-dimensional reconstruction method and equipment
CN114373041B (en) * 2021-12-15 2024-04-02 聚好看科技股份有限公司 Three-dimensional reconstruction method and device
CN114812540A (en) * 2022-06-23 2022-07-29 深圳市普渡科技有限公司 Picture construction method and device and computer equipment
CN116824070A (en) * 2023-08-31 2023-09-29 江西求是高等研究院 Real-time three-dimensional reconstruction method and system based on depth image
CN116824070B (en) * 2023-08-31 2023-11-24 江西求是高等研究院 Real-time three-dimensional reconstruction method and system based on depth image

Also Published As

Publication number Publication date
CN108335353B (en) 2020-12-22
CN108335353A (en) 2018-07-27
US20210074012A1 (en) 2021-03-11
US11954870B2 (en) 2024-04-09

Similar Documents

Publication Publication Date Title
WO2019161813A1 (en) Dynamic scene three-dimensional reconstruction method, apparatus and system, server, and medium
EP3786890B1 (en) Method and apparatus for determining pose of image capture device, and storage medium therefor
CN110427917B (en) Method and device for detecting key points
US11064178B2 (en) Deep virtual stereo odometry
US10789765B2 (en) Three-dimensional reconstruction method
WO2019238114A1 (en) Three-dimensional dynamic model reconstruction method, apparatus and device, and storage medium
WO2021226876A1 (en) Target detection method and apparatus
JP2020507850A (en) Method, apparatus, equipment, and storage medium for determining the shape of an object in an image
US20170330375A1 (en) Data Processing Method and Apparatus
CN111415409B (en) Modeling method, system, equipment and storage medium based on oblique photography
CN112444242A (en) Pose optimization method and device
WO2021035731A1 (en) Control method and apparatus for unmanned aerial vehicle, and computer readable storage medium
CN111127524A (en) Method, system and device for tracking trajectory and reconstructing three-dimensional image
CN112927362A (en) Map reconstruction method and device, computer readable medium and electronic device
CN112927363A (en) Voxel map construction method and device, computer readable medium and electronic equipment
US20220114785A1 (en) Three-dimensional model generation method and three-dimensional model generation device
CN115035235A (en) Three-dimensional reconstruction method and device
CN111915723A (en) Indoor three-dimensional panorama construction method and system
CN112270702A (en) Volume measurement method and device, computer readable medium and electronic equipment
WO2020019175A1 (en) Image processing method and apparatus, and photographing device and unmanned aerial vehicle
WO2021185036A1 (en) Point cloud data generation and real-time display method and apparatus, device, and medium
WO2022077790A1 (en) Representation data generation of three-dimensional mapping data
CN114627438A (en) Target detection model generation method, target detection method, device and medium
WO2021051220A1 (en) Point cloud fusion method, device, and system, and storage medium
US20230047211A1 (en) Method and system for automatic characterization of a three-dimensional (3d) point cloud

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19758152

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19758152

Country of ref document: EP

Kind code of ref document: A1