WO2018107679A1 - Method and device for acquiring dynamic three-dimensional image - Google Patents

Method and device for acquiring dynamic three-dimensional image Download PDF

Info

Publication number
WO2018107679A1
WO2018107679A1 PCT/CN2017/088162 CN2017088162W WO2018107679A1 WO 2018107679 A1 WO2018107679 A1 WO 2018107679A1 CN 2017088162 W CN2017088162 W CN 2017088162W WO 2018107679 A1 WO2018107679 A1 WO 2018107679A1
Authority
WO
WIPO (PCT)
Prior art keywords
terminal device
image
depth
matching
dynamic
Prior art date
Application number
PCT/CN2017/088162
Other languages
French (fr)
Chinese (zh)
Inventor
邵明明
钟小飞
王金波
王林
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to CN201780076051.4A priority Critical patent/CN110169056B/en
Publication of WO2018107679A1 publication Critical patent/WO2018107679A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/221Image signal generators using stereoscopic image cameras using a single 2D image sensor using the relative movement between cameras and objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/254Image signal generators using stereoscopic image cameras in combination with electromagnetic radiation sources for illuminating objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/271Image signal generators wherein the generated image signals comprise depth maps or disparity maps
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/296Synchronisation thereof; Control thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/698Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/32Indexing scheme for image data processing or generation, in general involving image mosaicing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20024Filtering details
    • G06T2207/20028Bilateral filtering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0092Image segmentation from stereoscopic image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2213/00Details of stereoscopic systems
    • H04N2213/003Aspects relating to the "2D+depth" image format

Definitions

  • the present application relates to the field of image recognition, and more particularly to a method and apparatus for dynamic three-dimensional image acquisition.
  • images are captured by input devices such as cameras to describe the real world.
  • camera devices have been able to provide increasingly finer image quality and larger and larger image resolution.
  • image algorithms are generated to assist the camera to produce more diverse images, such as panoramic photos, panoramic selfies, skin photos, audio photos, face recognition and smile recognition. It makes the photo more interesting and enriches the form of the real world.
  • the existing two-dimensional camera device can acquire a scene inside and outside a fixed-size area at a certain moment, and generate a two-dimensional, static description of the real world, and the acquired data is a two-dimensional matrix in units of pixels, which is compressed. After the algorithm is processed and saved, the compressed image is extracted and decompressed and pushed to the display device cache. Since the real world is three-dimensional and dynamic, designing a camera system capable of acquiring dynamic three-dimensional images and a series of storage display methods will open a new camera revolution.
  • the panoramic image acquisition mode is horizontally moved or horizontally rotated by the user's handheld terminal device, and the newly acquired image is stitched into the existing image by the internal stitching algorithm in real time, and can be viewed by sliding, zooming, etc. after completion.
  • the method is simple in operation, and can obtain a static image in a wider horizontal direction, which broadens the imaging range of the conventional two-dimensional image.
  • the surrounding image acquisition mode is moved or rotated by the user handheld terminal device in one of four directions of up, down, left, and right, and the internal terminal device records the current terminal device posture and the acquired scene image and performs inter-frame feature region matching. With compression, you can view it by sliding or rotating the device.
  • the method is simple in operation, and can obtain a dynamic image of a wider area in a certain direction, and realizes acquisition of a local dynamic three-dimensional image in one direction.
  • the panoramic image acquisition method limits the user to move or rotate a fixed distance in one direction.
  • the shooting process is prone to jitter and affect the shooting effect.
  • the final stitched image is curved and deformed, and it is difficult to restore the real scene.
  • the surround image acquisition method can only shoot around in a single direction after shooting starts. When shooting close-up shots, it is not possible to handle changes in the distance between the device and the subject. When a captured image is displayed, it cannot be freely enlarged and reduced. Storage and display do not form an industry standard, and the acquired images can only be viewed within the shooting software.
  • the present application provides a method and a terminal device for dynamic three-dimensional image acquisition, which can improve user experience.
  • a method for dynamic three-dimensional image acquisition which can be applied to a terminal device, and the method includes:
  • the overlapping region is fused by the fusion algorithm to generate a dynamic three-dimensional image.
  • the three-dimensional gesture unlocking method of the embodiment of the present application the three-dimensional gesture image presented by the user in the three-dimensional space in front of the camera is acquired in real time, the gesture of the user in the gesture image is extracted, and the unlocking gesture is matched with the previously set unlocking gesture of the user to achieve the unlocking.
  • the purpose of the terminal device the user is provided with a new, interesting, accurate and fast unlocking method.
  • the motion posture of the terminal device is acquired by the accelerometer, the gyroscope and the electronic compass of the terminal device.
  • performing fast segmentation matching with the depth image according to a motion posture of the terminal device including:
  • the feature region of the depth map of the end time point is calculated based on the completed segmented feature region of the depth map of the start time point and the pose change of the terminal device.
  • the result of the fast segmentation matching is accurately matched according to the color image, including:
  • the overlapping region is fused by the fusion algorithm to generate a dynamic three-dimensional image, including:
  • the fusion processing is performed to update the overlapping region.
  • a terminal device for performing the method of the first aspect or any possible implementation of the first aspect.
  • the terminal device may comprise means for performing the method of the first aspect or any of the possible implementations of the first aspect.
  • a third aspect provides a terminal device including a memory, a processor, and a display, the memory being used to store a computer program, the processor is configured to call and run a computer program from the memory, and when the program is executed, the processor executes the above The method of any of the first aspect or any of the possible implementations of the first aspect.
  • a computer readable medium for storing a computer program comprising instructions for performing the method of the first aspect or any of the possible implementations of the first aspect.
  • FIG. 1 is a schematic diagram of a minimum hardware system for implementing a terminal device of an embodiment of the present application.
  • FIG. 2 is a schematic flowchart of a dynamic three-dimensional image acquisition method according to an embodiment of the present application.
  • FIG. 3 is a block diagram showing the design of motion pose recognition and trajectory acquisition in accordance with one embodiment of the present application.
  • FIG. 4 is a schematic diagram of data fusion of a gyroscope and an accelerometer according to an embodiment of the present application.
  • FIG. 5 is a schematic diagram of a method for data fusion of a gyroscope and an electronic compass according to an embodiment of the present application.
  • FIG. 6 is a schematic flowchart of performing fast segmentation matching according to a motion posture and a depth image of a terminal device according to an embodiment of the present application.
  • FIG. 7 is an accurate match of a result of fast segmentation matching according to the color image according to an embodiment of the present application.
  • FIG. 8 is a schematic flowchart of a method for image overlap region fusion according to an embodiment of the present application.
  • FIG. 9 is a schematic flowchart of a user capturing a dynamic three-dimensional image according to an embodiment of the present application.
  • FIG. 10 is a schematic flowchart of a user viewing a dynamic three-dimensional image according to an embodiment of the present application.
  • FIG. 11 is a schematic block diagram of an example of a terminal device according to an embodiment of the present application.
  • the terminal device in this embodiment may be an access terminal, a user equipment (UE), a subscriber unit, a subscriber station, a mobile station, a mobile station, a remote station, a remote terminal, a mobile device, a user terminal, and a wireless communication device. , user agent or user device.
  • the terminal device may be a cellular phone, a cordless phone, a session initiation protocol (SIP) phone, a wireless local loop (WLL) station, a personal digital assistant (PDA), and a wireless communication function.
  • FIG. 1 is a schematic diagram of a minimum hardware system 100 of a terminal device implementing the three-dimensional gesture unlocking method of the present application.
  • the system 100 shown in FIG. 1 includes a light source transmitter 110, a depth camera 120, a spectrum analysis module 130, a color camera 140, a processor 150, a display unit 160, a nonvolatile memory 170, a memory 180, and a sensing unit 190.
  • the color camera 140, the light source emitter 110 and the depth camera 120 constitute a spectral input module, and the spectral analysis module 130 constitutes an image generation module.
  • the light source emitter 110, color camera 140, and depth camera 120 can be mounted side by side over the device (eg, directly above the device).
  • the light source emitter 110 can be an infrared emitter
  • the depth camera 120 can be an infrared camera
  • the spectrum analysis module 130 can be an infrared spectrum analysis module. In this case, the light source emitter 110 cooperates with the depth camera 120 to project the scene through the infrared light encoded image.
  • the light source emitter 110 outputs a common laser light source, which is filtered by a frosted glass and an infrared filter to form near-infrared light. Wherein, the light source emitter 110 can continuously output infrared light having a wavelength of 840 nanometers (nm).
  • the depth camera 120 is a Complementary Metal Oxide Semiconductor (CMOS) image sensor for receiving an excitation light source reflected from the outside, such as infrared light, and digitally encoding the excitation light source to form a digital image for transmission to the spectrum analysis module. 130.
  • CMOS Complementary Metal Oxide Semiconductor
  • the spectral analysis module 130 analyzes the speckles, calculates the distance between the corresponding pixel points of the image and the depth camera 120, and forms a depth data matrix for the driver to read.
  • the sensing unit 190 is connected to the processor 150, detects location information of the terminal device or a change in the surrounding environment, and transmits the sensed information to the processor 150.
  • the sensing unit 190 includes at least one of: a gyro sensor for detecting rotation, rotational movement, angular displacement, tilt, or any other non-linear motion, for A triaxial acceleration sensor that senses acceleration in one or more directions, an electronic compass that senses the earth's magnetic field to determine the north-south direction.
  • the sensing unit 190 operates under the control of the processor 150.
  • the terminal device may receive motion sensor data generated by a motion sensor (eg, a gyro sensor or an acceleration sensor) in the sensing unit 190, and process the generated motion sensor data using the motion sensing application.
  • a motion sensor eg, a gyro sensor or an acceleration sensor
  • a processor running a motion sensing application can analyze motion sensor data to identify specific types of motion events.
  • Display unit 160 is configured to display graphics, images or data to a user.
  • the display unit 160 is configured to provide various screens associated with the operation of the terminal device.
  • the display unit 160 provides a home screen, a message composing screen, a phone screen, a game screen, a music playing screen, and a video playing screen.
  • the display unit 160 can be implemented using a flat display panel such as a liquid crystal display (LCD), an organic light emitting diode (OLED), and an active matrix OLED (AMOLED).
  • LCD liquid crystal display
  • OLED organic light emitting diode
  • AMOLED active matrix OLED
  • the display unit 160 can operate as an input device.
  • the display unit 160 includes a touch panel for detecting a touch gesture.
  • the touch panel is configured to convert a pressure applied to a specific position of the display unit 160 or a capacitance change at a specific area of the display unit 160 into an electric input signal.
  • the touch panel can be implemented in one of add-on or on-cell (or in-cell).
  • the touch panel can be implemented in one of the following panels: a resistive touch panel, a capacitive touch panel, an electromagnetic induction touch panel, and a pressure touch panel.
  • the touch panel is configured to detect the pressure of the touch as well as the location and area being touched. If a touch gesture is made on the touch panel, a corresponding input signal is generated to the processor 150. The processor 150 then checks the user's touch input information to perform the corresponding function.
  • the processor 150 can be responsible for executing various software programs (e.g., applications and operating systems) to provide computing and processing operations for the terminal devices.
  • the non-volatile memory 170 is used to store program files, system files, and data.
  • Memory 180 is used for system and program running caches.
  • FIG. 2 is a schematic flow chart of a method for dynamic three-dimensional image acquisition according to an embodiment of the present application. The method shown in FIG. 2 can be performed by the terminal device shown in FIG. 1.
  • the following describes the acquisition method of the motion posture of the device.
  • a “pose” or “motion pose” as referred to herein is a set of motions of a device, which may be a set of motions included, such as a swing or a circular motion, or may be a simple movement of the device, eg, the device is specific The tilt of the axis or angle.
  • Figure 3 shows a block diagram of a design for motion pose recognition and acquisition of trajectories.
  • the sampling unit 310 can receive motion data from the gyroscope, the accelerometer, and the electronic compass and sample.
  • the attitude solving unit 320 reads the data of the gyroscope, the accelerometer and the electronic compass, calculates the triaxial angular velocity of the device, calculates the angular increment matrix, solves the attitude differential equation, and finally updates the attitude quaternion.
  • the data fusion unit 330 filters the noise in the correlation output based on the Kalman filter algorithm and finally outputs the device pose and trajectory.
  • the error model used in the gyroscope or accelerometer error calibration process can be expressed by equation (1).
  • [D x D y D z ] T is the true value of the physical quantity measured by the gyroscope or accelerometer
  • [M x M y M z ] T is the actual measured value of the gyroscope or accelerometer
  • [B x B y B z ] T is the sensor bias.
  • D x , D y , and D z are all 0, and for the accelerometers D x and D y in the horizontal stationary state, both are 0, and D z is a gravitational acceleration value.
  • x 1 and y 1 are the outputs of the calibrated electronic compass, and x and y are the outputs when the electronic compass is deviated.
  • This application can obtain x 0 , y 0 , ⁇ , a, b by least square fitting. Eliminate errors.
  • the quaternion is used to describe the attitude of the device.
  • the gyro data is read, the three-axis angular velocity of the device is calculated, the angular increment matrix is calculated, the attitude differential equation is solved, and the attitude quaternion is finally updated.
  • the rotation quaternion from the inertial coordinate system a to the device coordinate system b is:
  • the Pika method can be used to solve the quaternion differential equation.
  • the process is to first calculate the corresponding quaternion Q(t) when the carrier moves, and then according to the quaternion and the attitude matrix.
  • the rollover angle calculated by the accelerometer and the rollover angular velocity of the gyroscope test the pitch angle data calculated by the accelerometer and the gyroscope test pitch rate data are respectively filtered, and the accelerometer and the gyroscope can be made.
  • the data compensates each other, reduces the measurement noise, and the pitch angle and roll angle test values are more accurate, which makes the magnetic sensor tilt angle compensation effect better, can perform static calibration, and can also perform dynamic calibration.
  • the noise variance matrix of these two sensors is set as a variable, and the external disturbance is monitored in real time, and the noise variance matrix of the accelerometer and the electronic compass is dynamically changed, and then the gain in the Kalman filter is corrected.
  • the values of the accelerometer and the electronic compass are read to obtain the observation, and the a priori quaternion is used as the initial value of the state quantity, and the formula of the Kalman filter is brought. Get the final pose quaternion.
  • the gyroscope is integrated with the accelerometer, and the pitch angle ⁇ and the roll angle ⁇ are estimated, and the gyroscope is integrated with the electronic compass to estimate the heading angle.
  • the data fusion process of gyroscope and accelerometer is shown in Figure 4.
  • the data fusion process of gyroscope and electronic compass is shown in Figure 5.
  • S220 Collect a depth image and a color image respectively by using a depth camera and a color camera.
  • the depth image is also referred to as a range image, and refers to an image from a image collector (for example, the depth camera 120 in the present application) to a point (depth) of each point in the scene as a pixel value. It directly reflects the geometry of the visible surface of the scene.
  • the depth camera 120 digitally encodes the excitation light source by receiving an excitation light source reflected by the outside, such as infrared light, to form a digital image and transmit it to the spectrum analysis module 130.
  • the spectral analysis module 130 analyzes the speckles and calculates a distance z between the corresponding pixel point (x, y) in the current image and the depth camera 120, so that the current depth image can be acquired.
  • S230 Perform fast segmentation matching with the depth image according to the motion posture of the terminal device.
  • FIG. 6 shows a fast matching method of a central region three-dimensional object in which a motion figure and a depth image of a device are fused, the method tracking a device state change in real time, and extracting a start and end time point corresponding to each fixed time period when the device smoothly moves.
  • Depth map frame Based on the start time point, the depth map has completed the segmentation of the feature region and the device pose change in the close-up mode, and the approximate range of each feature region of the depth map at the end of the time is inferred, thereby performing fast segmentation matching.
  • the depth image Since the value of each pixel of the depth image is the linear distance between the object and the camera, there is a similarity between the distance from the same object to the camera. Therefore, the coarse segmentation based on the depth map adopts the region growing method. However, since the depth image has noise and is easy to lose information, the depth image is first filtered to achieve image smoothing and loss depth filling.
  • the specific implementation manner is as follows:
  • the image is filtered using a bilateral filter defined as:
  • I is the original image
  • I' is the filtered image
  • is the neighborhood of (x, y)
  • w(i, j) is the weight of the filter at the corresponding coordinates.
  • Pixels with similar depths in the image are combined to form a similar feature region.
  • the specific implementation is:
  • the selection of the starting point is very important for the efficiency of depth image segmentation. If properly selected, the segmentation can be speeded up.
  • This application is based on the relative posture and trajectory of the device to roughly estimate the position of the feature region in the depth map at the end time point to accelerate. . Since the distance of the captured object from the camera is usually close, the minimum value region in the depth map is selected, and a multi-fork tree of the image minimum value region is established to realize the selection of the starting point.
  • the similarity criterion is used to distinguish the object from the background. Select the average depth and difference mean of the pixel points in the previous comparison, and select the average depth and difference mean of the pixel points. The difference between the mean value of the pixel depth and the difference between the two is It is determined to be the same area within 5%.
  • the present application accurately matches the result of the fast matching for the close-up mode and the image segmentation method of the color image, and directly draws the frame for the device to smoothly move or rotate for the perspective mode, and color
  • the images are accurately matched, and the precise matching is mainly optimized at the edge of the feature region obtained by the fast matching.
  • the matching process is shown in FIG. 7 .
  • the depth image and the color image are separately filtered, and then the feature regions are quickly matched according to the posture information and the depth information, and a series of feature regions are obtained, and the representative pixel points of each feature region are provided to the color image for segmentation.
  • the image segmentation adopts the watershed algorithm, and after filtering, the grayscale image after the coloration is generated, and the water injection operation is directly performed according to the provided characteristic pixel points, and finally the boundary of each feature region is obtained. Based on the segmentation of the boundary region of the color image, the boundary points of the feature regions obtained by the fast matching are compared with each other. If there is no deviation, the matching results are both normal.
  • the color image segmentation result is the final result. If there is a deviation, and the neighborhood depth data is missing or the drop is not clear, the color image segmentation result is the final result. If there is a deviation, and the neighborhood depth data is perfect and the drop is not clear, the depth image segmentation result is the final result. If there is a deviation, and the neighborhood depth data is perfect and the difference is clear, the color image segmentation result is the final result.
  • Accurate matching results can provide feedback for device attitude and trajectory information, making gesture recognition more accurate.
  • the process of obtaining attitude information based on the exact matching result is as follows:
  • the present application allows the device to acquire an image in all directions, and the overlapping regions need to be merged when the acquired image overlaps with the captured image.
  • the overlap region fusion is based on the historical feature matrix and the current device pose information.
  • the present application associates each feature matrix with the device pose, so that the previous device pose information of the historical feature matrix can be obtained.
  • the specific integration process is shown in Figure 8.
  • the posture information of the device during the motion shooting will be continuously recorded and saved, and the attitude information may be extracted at regular intervals as the iconic data for comparison with the device posture in the future. Since different device poses may also have overlapping regions, the method records the field of view that can be captured by each gesture experienced by the device, and stores the feature region and posture information in combination. The current relative pose and trajectory are calculated in real time during the movement of the device, and the comparable iconic data is extracted from the historical feature matrix for matching. If the matching result indicates that the current image frame overlaps with the historical image frame, the fusion processing is performed. , updated to the overlap area, while recording the current device pose.
  • the method provided by the present application can dynamically identify whether the current scene is a distant view or a close-up view.
  • the determining method is to scan the depth image matrix, calculate the number of pixel points whose depth value is less than the threshold value, and determine the foreground when the number is less than the threshold value.
  • the depth camera is automatically turned off and periodically activated to detect if it is in close-up mode, which reduces power consumption.
  • the method tracks the change of the depth state of the feature area.
  • the distance is stored in the feature matrix, so that the display can recognize whether there is a close to the moving action when shooting, thereby prompting the user to Zoom out and zoom in.
  • the method does not activate the depth camera, so there is no need to perform fast matching of the central region of the depth map, and only the color image and the attitude information are accurately matched to obtain the feature matrix.
  • the method prompts the user to move or rotate the mobile phone in any direction to shoot the target object.
  • the simultaneous starting attitude sensor, depth camera and color camera work are triggered.
  • the device attitude and trajectory identification module reads the data of the gyroscope, accelerometer and electronic compass in real time, performs attitude calculation on it, and then fuses the multi-sensor data to obtain the posture and trajectory of the device.
  • the depth camera collects the depth data in real time, and after filtering, recognizes the near and far modes.
  • the depth image is roughly segmented, and the device region and the trajectory data are quickly matched to the central region to speed up the matching.
  • the color camera collects the color data in real time, and after filtering, performs the water injection operation according to the feature area provided by the fast matching result, and finally obtains the boundary of the scene, compares and determines the boundary of the feature area obtained by the fast matching, and adjusts the feature area.
  • the boundary eventually produces a subtle matching feature area.
  • the method is based on the device posture and the feature area in the device.
  • the position in the coordinate system is calculated, and the coincident area is fused and updated, and the posture of the device taken twice is recorded, so that it can be cyclically viewed when supplied to the display.
  • the coincidence region is merged, the final feature region set of the image is generated.
  • the posture sensor When the user clicks on the picture to view, the posture sensor is activated to acquire the device posture, and the image feature region set is read to obtain the final image frame.
  • the method supports the user to rotate the mobile phone to view the picture, and the gesture of the mobile phone is corresponding to the posture of the mobile phone when the picture is taken.
  • the device attitude and trajectory identification module reads the data of the gyroscope, accelerometer and electronic compass in real time, performs attitude calculation on it, and then fuses the multi-sensor data to obtain the posture of the device.
  • the image frames After reading the image feature region set, the image frames are synthesized according to the coordinates in the image frame in which each feature region is located, and finally the image frames are buffered one by one to be read.
  • the current pose of the device After the current pose of the device is generated, it needs to correspond to the initial device pose of the captured photo. After that, the change of the gesture of the mobile phone will trigger the display of the corresponding state picture frame. The change of the device pose will trigger the selection of the image frame of the corresponding pose and submit the display. After the image is implemented, it is judged whether the current frame is scalable. If possible, the screen prompts that the current zoom is possible, and then the posture sensor data is acquired to start a new cycle. If you can't zoom, you also get the attitude sensor data to start a new loop.
  • the application is equipped with a gyroscope, an accelerometer and an electronic compass sensor on the terminal device for providing device attitude information; an infrared transmitter and an infrared camera for providing depth image data; and a color camera for use
  • a gyroscope, an accelerometer and an electronic compass sensor on the terminal device for providing device attitude information
  • an infrared transmitter and an infrared camera for providing depth image data
  • a color camera for use
  • the combination of the three provides raw data support for the acquisition of dynamic three-dimensional images.
  • the method for gesture recognition of a three-dimensional space terminal device can obtain an initial posture of a device by sampling, posture settlement, and data fusion of three attitude sensors.
  • the attitude generation algorithm is compensated according to the change of the depth image and the color image, and the closed-loop tracking of the attitude detection is completed.
  • the fast matching method of the three-dimensional object in the central region of the device posture and depth image fusion in the present application can provide a strategy of speeding up matching in the form of frame drawing when the device posture changes at a constant speed, and realize the same three-dimensional object between multiple frames of images. Quick match.
  • the fine matching method based on the color image and the fast matching result of the same three-dimensional object between the multi-frame images according to the present application can compensate and optimize the fast matching result according to the data information of the corresponding position of the color image for each feature region. , get the most detailed image feature description.
  • the present application can realize 360-degree omnidirectional imaging, and supports shooting of an already photographed object, and can dynamically recognize the overlapped area that has been photographed according to the current posture information of the device and the historical feature matrix information. Data fusion is performed on the overlapping regions, and the overlapping data information is added, so that the display can be smoothly switched according to the posture.
  • This application can dynamically recognize the distant and near mode of the subject, and achieve a full range of camera for the near view.
  • Scenery the integration of the panorama.
  • the Vision automatically turns off the depth camera to reduce power consumption.
  • FIG. 11 is another schematic block diagram of a terminal device according to an embodiment of the present application.
  • the terminal device 1100 shown in FIG. 11 includes: a radio frequency (RF) circuit 1110, a memory 1120, other input devices 1130, a display screen 1140, a sensor 1150, an audio circuit 1160, an I/O subsystem 1170, and a processor 1180. And power supply 1190 and other components.
  • RF radio frequency
  • the terminal device structure shown in FIG. 11 does not constitute a limitation of the terminal device, and may include more or less components than those illustrated, or combine some components or split some components. , or different parts layout.
  • the display screen 1140 belongs to a User Interface (UI), and the terminal device 1100 may include a user interface that is smaller than illustrated or less.
  • UI User Interface
  • terminal device 1100 The specific components of the terminal device 1100 are specifically described below with reference to FIG. 11:
  • the RF circuit 1110 can be used for receiving and transmitting signals during and after receiving or transmitting information, in particular, after receiving the downlink information of the base station, and processing it to the processor 1180; in addition, transmitting the designed uplink data to the base station.
  • RF circuits include, but are not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a Low Noise Amplifier (LNA), a duplexer, and the like.
  • LNA Low Noise Amplifier
  • RF circuitry 1110 can also communicate with the network and other devices via wireless communication.
  • the wireless communication may use any communication standard or protocol, including but not limited to Global System of Mobile communication (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (Code). Division Multiple Access (CDMA), Wideband Code Division Multiple Access (WCDMA), Long Term Evolution (LTE), E-mail, Short Messaging Service (SMS), etc.
  • GSM Global System of Mobile communication
  • GPRS General Packet Radio Service
  • the memory 1120 can be used to store software programs and modules, and the processor 1180 executes various functional applications and data processing of the terminal device 1100 by running software programs and modules stored in the memory 1120.
  • the memory 1120 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application required for at least one function (such as a sound playing function, an image playing function, etc.), and the like; the storage data area may be stored according to Data (such as audio data, phone book, etc.) created by the use of the terminal device 1100.
  • memory 1120 can include high speed random access memory, and can also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
  • Other input devices 1130 can be used to receive input numeric or character information, as well as to generate key signal inputs related to user settings and function control of terminal device 1100.
  • other input devices 1130 may include, but are not limited to, physical keyboards, function keys (such as volume control buttons, switch buttons, etc.), trackballs, mice, joysticks, and light mice (the light mouse is not sensitive to display visual output).
  • function keys such as volume control buttons, switch buttons, etc.
  • trackballs mice, joysticks, and light mice (the light mouse is not sensitive to display visual output).
  • Other input devices 1130 are coupled to other input device controllers 1171 of I/O subsystem 1170 for signal interaction with processor 1180 under the control of other device input controllers 1171.
  • the display 1140 can be used to display information entered by the user or information provided to the user as well as various menus of the terminal device 1100, and can also accept user input.
  • the specific display screen 1140 can include a display panel 1141 and a touch panel 1142.
  • the display panel 1141 can be configured by using a liquid crystal display (LCD), an organic light-emitting diode (OLED), or the like.
  • the touch panel 1142 also referred to as a touch screen, a touch sensitive screen, etc., can collect contact or non-contact operations on or near the user (eg, the user uses any suitable object or accessory such as a finger, a stylus, etc. on the touch panel 1142.
  • the operation in the vicinity of the touch panel 1142 may also include a somatosensory operation; the operation includes a single-point control operation, a multi-point control operation, and the like, and the corresponding connection device is driven according to a preset program.
  • the touch panel 1142 may include two parts: a touch detection device and a touch controller. Wherein, the touch detection device detects a user's touch orientation and posture, and Detecting a signal brought by the touch operation, transmitting the signal to the touch controller; the touch controller receives the touch information from the touch detection device, and converts it into information that the processor can process, and sends the information to the processor 1180, and can receive the processing The command sent by the device 1180 is executed.
  • the touch panel 1142 can be implemented by using various types such as resistive, capacitive, infrared, and surface acoustic waves, and the touch panel 1142 can be implemented by any technology developed in the future.
  • the touch panel 1142 can cover the display panel 1141, and the user can display the content according to the display panel 1141 (the display content includes, but is not limited to, a soft keyboard, a virtual mouse, a virtual button, an icon, etc.) on the display panel 1141. Operation is performed on or near the covered touch panel 1142. After detecting the operation on or near the touch panel 1142, the touch panel 1142 transmits to the processor 1180 through the I/O subsystem 1170 to determine user input, and then the processor 1180 is based on the user.
  • the input provides a corresponding visual output on display panel 1141 via I/O subsystem 1170.
  • the touch panel 1142 and the display panel 1141 are two independent components to implement the input and input functions of the terminal device 1100 , in some embodiments, the touch panel 1142 and the display panel 1141 may be The input and output functions of the terminal device 1100 are implemented integrated.
  • the terminal device 1100 may also include at least one type of sensor 1150, such as a light sensor, a motion sensor, and other sensors.
  • the light sensor may include an ambient light sensor and a proximity sensor, wherein the ambient light sensor may adjust the brightness of the display panel 1141 according to the brightness of the ambient light, and the proximity sensor may close the display panel 1141 when the terminal device 1100 moves to the ear. And / or backlight.
  • the accelerometer sensor can detect the magnitude of acceleration in all directions (usually three axes). When it is stationary, it can detect the magnitude and direction of gravity.
  • the terminal device 1100 can also be configured with gyroscopes, barometers, hygrometers, thermometers, infrared sensors and other sensors, here No longer.
  • An audio circuit 1160, a speaker 1161, and a microphone 1162 can provide an audio interface between the user and the terminal device 1100.
  • the audio circuit 1160 can transmit the converted audio data to the speaker 1161, and convert it into a sound signal output by the speaker 1161; on the other hand, the microphone 1162 converts the collected sound signal into a signal, which is received by the audio circuit 1160.
  • the audio data is converted, and the audio data is output to the RF circuit 1110 for transmission to, for example, another mobile phone, or the audio data is output to the memory 1120 for further processing.
  • the I/O subsystem 1170 is used to control external devices for input and output, and may include other device input controllers 1171, sensor controllers 1172, and display controllers 1173.
  • one or more other input control device controllers 1171 receive signals from other input devices 1130 and/or send signals to other input devices 1130, and other input devices 1130 may include physical buttons (press buttons, rocker buttons, etc.) , dial, slide switch, joystick, click wheel, light mouse (light mouse is a touch-sensitive surface that does not display visual output, or an extension of a touch-sensitive surface formed by a touch screen). It is worth noting that other input control device controllers 1171 can be connected to any one or more of the above devices.
  • Display controller 1173 in I/O subsystem 1170 receives signals from display 1140 and/or transmits signals to display 1140. After the display 1140 detects the user input, the display controller 1173 converts the detected user input into an interaction with the user interface object displayed on the display 1140, ie, implements human-computer interaction. Sensor controller 1172 can receive signals from one or more sensors 1150 and/or send signals to one or more sensors 1150.
  • the processor 1180 is a control center of the terminal device 1100 that connects various portions of the entire terminal device using various interfaces and lines, by running or executing software programs and/or modules stored in the memory 1120, and recalling stored in the memory 1120.
  • the data performs various functions and processing data of the terminal device 1100, thereby performing overall monitoring of the terminal device.
  • the processor 1180 can include one or more processing units; optionally, the processor 1180
  • the application processor and the modem processor can be integrated, wherein the application processor mainly processes an operating system, a user interface, an application, etc., and the modem processor mainly processes wireless communication. It will be appreciated that the above described modem processor may also not be integrated into the processor 1180.
  • the processor 1180 is configured to: acquire a motion posture of the terminal device; separately acquire a depth image and a color image by using the depth camera and the color camera; perform fast segmentation matching with the depth image according to the motion posture of the terminal device; The result of the fast segmentation matching is accurately matched; if the acquired current image overlaps with the captured image, the overlapping region is fused by the fusion algorithm to generate a dynamic three-dimensional image. .
  • the terminal device 1100 further includes a power source 1190 (such as a battery) for supplying power to the various components.
  • a power source 1190 such as a battery
  • the power source can be logically connected to the processor 1180 through the power management system to manage functions such as charging, discharging, and power consumption through the power management system. .
  • the terminal device 1100 may further include a camera (a depth camera and a color camera), a Bluetooth module, and the like, and details are not described herein again.
  • a camera a depth camera and a color camera
  • a Bluetooth module a Bluetooth module
  • the terminal device 1100 may correspond to a terminal device in a dynamic three-dimensional image acquisition method according to an embodiment of the present application, and the terminal device 1100 may include a physical unit for performing a method performed by a terminal device or an electronic device in the above method. .
  • the physical units in the terminal device 1100 and the other operations and/or functions described above are respectively used for the corresponding processes of the foregoing methods, and are not described herein for brevity.
  • the terminal device 1100 can include a physical unit in a method for performing the above-described dynamic three-dimensional image acquisition.
  • the physical units in the terminal device 1100 and the other operations and/or functions described above are respectively used for the corresponding processes of the foregoing methods, and are not described herein for brevity.
  • the processor in the embodiment of the present application may be an integrated circuit chip with signal processing capability.
  • each step of the foregoing method embodiment may be completed by an integrated logic circuit of hardware in a processor or an instruction in a form of software.
  • the processor may be a central processing unit (CPU), the processor may be another general-purpose processor, a digital signal processor (DSP), or an application specific integrated circuit (ASIC). ), Field Programmable Gate Array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
  • the methods, steps, and logical block diagrams disclosed in the embodiments of the present application can be implemented or executed.
  • the general purpose processor may be a microprocessor or the processor or any conventional processor or the like.
  • the steps of the method disclosed in the embodiments of the present application may be directly implemented by the hardware decoding processor, or may be performed by a combination of hardware and software in the decoding processor.
  • the software can be located in a random storage medium, such as a flash memory, a read only memory, a programmable read only memory or an electrically erasable programmable memory, a register, and the like.
  • the storage medium is located in the memory, and the processor reads the information in the memory and combines the hardware to complete the steps of the above method.
  • the memory in the embodiments of the present application may be a volatile memory or a non-volatile memory, or may include both volatile and non-volatile memory.
  • the non-volatile memory may be a read-only memory (ROM), a programmable read only memory (PROM), an erasable programmable read only memory (Erasable PROM, EPROM), or an electric Erase programmable read only memory (EEPROM) or flash memory.
  • the volatile memory can be a Random Access Memory (RAM) that acts as an external cache.
  • RAM Random Access Memory
  • many forms of RAM are available, such as static random access memory (SRAM), dynamic random access memory (DRAM), synchronous dynamic random access memory (Synchronous DRAM).
  • SDRAM Double Data Rate Synchronous Dynamic Random Access Memory
  • ESDRAM Enhanced Synchronous Dynamic Random Access Memory
  • SDRAM Synchronous Connection Dynamic Random Access Memory
  • DR RAM Direct Memory Bus Random Access Memory
  • bus system may include a power bus, a control bus, a status signal bus, and the like in addition to the data bus.
  • bus systems may include a power bus, a control bus, a status signal bus, and the like in addition to the data bus.
  • various buses are labeled as bus systems in the figure.
  • B corresponding to A means that B is associated with A, and B can be determined according to A.
  • determining B from A does not mean that B is only determined based on A, and that B can also be determined based on A and/or other information.
  • the term "and/or” herein is merely an association relationship describing an associated object, indicating that there may be three relationships, for example, A and/or B, which may indicate that A exists separately while 10 is stored in A. And B, there are three cases of B alone.
  • the character "/" in this article generally indicates that the contextual object is an "or" relationship.
  • each step of the above method may be completed by an integrated logic circuit of hardware in a processor or an instruction in a form of software.
  • the steps of the method for transmitting an uplink signal disclosed in the embodiments of the present application may be directly implemented as hardware processor execution completion, or performed by a combination of hardware and software in a processor.
  • the software can be located in a random storage medium, such as a flash memory, a read only memory, a programmable read only memory or an electrically erasable programmable memory, a register, and the like.
  • the storage medium is located in the memory, and the processor reads the information in the memory and combines the hardware to complete the steps of the above method. To avoid repetition, it will not be described in detail here.
  • the embodiment of the present application further provides a computer readable storage medium storing one or more programs, the one or more programs including instructions, when the portable electronic device is included in a plurality of applications When executed, the portable electronic device can be caused to perform the method of the embodiment shown in Figures 2 and/or 3.
  • the disclosed systems, devices, and methods may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the embodiments of the present application may be integrated into one processing unit, It may be that each unit physically exists alone, or two or more units may be integrated into one unit.
  • the functions may be stored in a computer readable storage medium if implemented in the form of a software functional unit and sold or used as a standalone product. Based on such understanding, the technical solution of the embodiments of the present application, or the part contributing to the prior art or the part of the technical solution, may be embodied in the form of a software product stored in a storage medium.
  • the instructions include a plurality of instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to perform all or part of the steps of the methods described in the various embodiments of the embodiments of the present application.
  • the foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like, which can store program codes. .

Abstract

Provided is a method for acquiring a dynamic three-dimensional image. The method for acquiring a dynamic three-dimensional image comprises: acquiring a movement posture of a terminal device; respectively collecting a depth image and a colour image by means of a depth camera and a colour camera; according to the movement posture of the terminal device and the depth image, performing quick segmentation and matching; according to the colour image, performing exact matching on a result of the quick segmentation and matching; and if there is an overlap between the acquired current image and a photographed image, performing fusion on the overlapping area through a fusion algorithm so as to generate a dynamic three-dimensional image. The present application can realise, with respect to the currently occurring defects existing in panoramic photography and surround photography, by means of adding a depth camera to a device and combining data of a mobile phone posture sensor and a colour image sensor, a method capable of recording a scenery appearance from various directions at once so as to acquire a dynamic three-dimensional image and support storage display.

Description

一种动态三维图像获取的方法和设备Method and device for acquiring dynamic three-dimensional image
本申请要求于2016年12月12日提交中国专利局、申请号为201611142062.1、发明名称为“一种动态三维图像获取的方法和设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。The present application claims priority to Chinese Patent Application No. 201611142062.1, entitled "A Method and Apparatus for Dynamic Three-Dimensional Image Acquisition", filed on Dec. 12, 2016, the entire contents of In this application.
技术领域Technical field
本申请涉及图像识别领域,并且更具体地涉及一种动态三维图像获取的方法和设备。The present application relates to the field of image recognition, and more particularly to a method and apparatus for dynamic three-dimensional image acquisition.
背景技术Background technique
通常,图像由摄像机等输入设备捕捉产生,用于描述真实世界。随着科技的发展,摄像装置已经能够提供越来越精细的图像质量,越来越大的图像分辨率。在此基础上,产生了大量的图像算法以辅助摄像装置产生更多样化的图片,如全景拍照、全景自拍、美肤拍照、有声照片、人脸识别和笑脸识别等,这些图像算法的运用使得拍照具备了更多的趣味性,也更加丰富了对真实世界展现的形式。Typically, images are captured by input devices such as cameras to describe the real world. With the development of technology, camera devices have been able to provide increasingly finer image quality and larger and larger image resolution. On this basis, a large number of image algorithms are generated to assist the camera to produce more diverse images, such as panoramic photos, panoramic selfies, skin photos, audio photos, face recognition and smile recognition. It makes the photo more interesting and enriches the form of the real world.
现有的二维的摄像装置能够获取某一时刻固定大小的区域内外部的景象,产生对真实世界二维、静态的描述,其获取的数据是以像素点为单位的二维矩阵,经过压缩算法处理后保存,需要显示时取出压缩后的图像进行解压并推送到显示设备缓存。由于真实世界是三维的、动态的,设计一种能够采集动态三维图像的摄像系统与一系列的存储显示方法将会开启一场全新的摄像革命。The existing two-dimensional camera device can acquire a scene inside and outside a fixed-size area at a certain moment, and generate a two-dimensional, static description of the real world, and the acquired data is a two-dimensional matrix in units of pixels, which is compressed. After the algorithm is processed and saved, the compressed image is extracted and decompressed and pushed to the display device cache. Since the real world is three-dimensional and dynamic, designing a camera system capable of acquiring dynamic three-dimensional images and a series of storage display methods will open a new camera revolution.
目前针对三维、动态图像获取的方法包括:全景图像获取和环绕图像获取。Current methods for three-dimensional, dynamic image acquisition include: panoramic image acquisition and surround image acquisition.
全景图像获取方式由用户手持终端设备水平移动或水平旋转,由内部拼接算法实时得将刚采集到的图像拼接到已有的图像中,完成后可通过滑动、缩放等方式查看。该方式操作简单,能够获得水平方向更广区域的静态图像,拓宽了传统二维图像的摄像范围。The panoramic image acquisition mode is horizontally moved or horizontally rotated by the user's handheld terminal device, and the newly acquired image is stitched into the existing image by the internal stitching algorithm in real time, and can be viewed by sliding, zooming, etc. after completion. The method is simple in operation, and can obtain a static image in a wider horizontal direction, which broadens the imaging range of the conventional two-dimensional image.
环绕图像获取方式由用户手持终端设备朝上、下、左、右四个方向中某个方向上移动或旋转,由内部算法记录当前终端设备姿态与所获取的景象图片并进行帧间特征区域匹配与压缩,完成后可通过滑动或旋转设备查看。该方式操作简单,能够获得某一方向上更广区域的动态图像,实现了单方向上局部动态三维图像的获取。The surrounding image acquisition mode is moved or rotated by the user handheld terminal device in one of four directions of up, down, left, and right, and the internal terminal device records the current terminal device posture and the acquired scene image and performs inter-frame feature region matching. With compression, you can view it by sliding or rotating the device. The method is simple in operation, and can obtain a dynamic image of a wider area in a certain direction, and realizes acquisition of a local dynamic three-dimensional image in one direction.
全景图像获取方式限定了用户只能朝一个方向移动或旋转固定的距离,拍摄过程容易出现抖动而影响拍摄效果,最终拼接的图像弯曲变形,难以还原真实的景象。The panoramic image acquisition method limits the user to move or rotate a fixed distance in one direction. The shooting process is prone to jitter and affect the shooting effect. The final stitched image is curved and deformed, and it is difficult to restore the real scene.
环绕图像获取方式在拍摄开始后只能朝单一方向环绕拍摄。在对近景拍摄时,无法处理设备与景物距离的变化。在显示已拍摄图像时,无法对其自由放大和缩小。存储与显示没有形成行业标准,所获取的图像只能在该拍摄软件内查看。The surround image acquisition method can only shoot around in a single direction after shooting starts. When shooting close-up shots, it is not possible to handle changes in the distance between the device and the subject. When a captured image is displayed, it cannot be freely enlarged and reduced. Storage and display do not form an industry standard, and the acquired images can only be viewed within the shooting software.
在交互式终端设备越来越智能的今天,人们对终端设备本身操作的趣味性、准确性、快速性要求越来越高。由此,针对目前终端设备全景摄像或环绕摄像存在的缺陷,需要一种从各个方位记录景物外观获取动态、三维图像并支持存储、显示的方法。丰富图片的表现形式,改变大众对图像的认知和对拍摄的体验。Today, as interactive terminal devices become more and more intelligent, people are increasingly demanding the fun, accuracy, and rapidity of the operation of the terminal devices themselves. Therefore, in view of the defects of the current terminal device panoramic imaging or surround imaging, a method for acquiring dynamic and three-dimensional images from various orientations of the scene and supporting storage and display is needed. Enrich the representation of the picture, changing the public's perception of the image and the experience of shooting.
发明内容Summary of the invention
本申请提供了一种动态三维图像获取的方法和终端设备,能够提高用户体验。 The present application provides a method and a terminal device for dynamic three-dimensional image acquisition, which can improve user experience.
第一方面,提供了一种动态三维图像获取的方法,该方法可以应用于终端设备,该方法包括:In a first aspect, a method for dynamic three-dimensional image acquisition is provided, which can be applied to a terminal device, and the method includes:
获取终端设备的运动姿态;Obtaining a motion posture of the terminal device;
通过深度摄像头和彩色摄像头分别采集深度图像和彩色图像;Collecting depth images and color images by a depth camera and a color camera, respectively;
根据终端设备的运动姿态与所述深度图像进行快速分割匹配;Performing fast segmentation matching with the depth image according to the motion posture of the terminal device;
根据所述彩色图像对所述快速分割匹配的结果进行精准匹配;Performing exact matching on the result of the fast segmentation matching according to the color image;
如果获取的当前图像与已拍摄图像存在重叠,通过融合算法对重叠区域进行融合以生成动态三维图像。If the acquired current image overlaps with the captured image, the overlapping region is fused by the fusion algorithm to generate a dynamic three-dimensional image.
根据本申请实施例的三维手势解锁方法,通过实时获取用户在摄像头前方三维空间内所呈现的立体手势图像,提取手势图像中用户的手势,并通过与用户之前设定的解锁手势匹配,达到解锁终端设备的目的。从而,为用户提供了一种全新的趣味性强、准确性高和快速性好的解锁方式。According to the three-dimensional gesture unlocking method of the embodiment of the present application, the three-dimensional gesture image presented by the user in the three-dimensional space in front of the camera is acquired in real time, the gesture of the user in the gesture image is extracted, and the unlocking gesture is matched with the previously set unlocking gesture of the user to achieve the unlocking. The purpose of the terminal device. Thus, the user is provided with a new, interesting, accurate and fast unlocking method.
在一种可能的实现方式中,通过终端设备的加速度计、陀螺仪和电子罗盘获取终端设备的运动姿态。In a possible implementation, the motion posture of the terminal device is acquired by the accelerometer, the gyroscope and the electronic compass of the terminal device.
在一种可能的实现方式中,根据终端设备的运动姿态与所述深度图像进行快速分割匹配,包括:In a possible implementation manner, performing fast segmentation matching with the depth image according to a motion posture of the terminal device, including:
当根据终端设备的运动姿态确定所述终端设备平滑移动时,获取第一时间段内起始时间点与结束时间点对应的深度图;When determining that the terminal device moves smoothly according to the motion posture of the terminal device, acquiring a depth map corresponding to the start time point and the end time point in the first time period;
基于起始时间点的深度图的已完成分割的特征区域和所述终端设备的姿态变化计算结束时间点的深度图的特征区域范围。The feature region of the depth map of the end time point is calculated based on the completed segmented feature region of the depth map of the start time point and the pose change of the terminal device.
在一种可能的实现方式中,根据所述彩色图像对所述快速分割匹配的结果进行精准匹配,包括:In a possible implementation, the result of the fast segmentation matching is accurately matched according to the color image, including:
根据所述彩色图像对所述特征区域范围执行补偿优化,获取细粒度的图片特征描述。Performing compensation optimization on the feature region range according to the color image to obtain a fine-grained picture feature description.
在一种可能的实现方式中,通过融合算法对重叠区域进行融合以生成动态三维图像,包括:In a possible implementation, the overlapping region is fused by the fusion algorithm to generate a dynamic three-dimensional image, including:
实时计算终端设备的当前姿态,从历史特征矩阵中提取可比较的标志性数据进行匹配;Calculating the current posture of the terminal device in real time, and extracting comparable iconic data from the historical feature matrix for matching;
如果匹配结果表明当前图像帧与历史某图像帧有重叠的特征区域则进行融合处理,更新重叠区域。If the matching result indicates that the current image frame overlaps with the historical image frame, the fusion processing is performed to update the overlapping region.
第二方面,提供了一种终端设备,用于执行第一方面或第一方面任意可能的实现方式中的方法。具体地,该终端设备可以包括用于执行第一方面或第一方面任意可能的实现方式中的方法的单元。In a second aspect, a terminal device is provided for performing the method of the first aspect or any possible implementation of the first aspect. In particular, the terminal device may comprise means for performing the method of the first aspect or any of the possible implementations of the first aspect.
第三方面,提供了一种终端设备,包括存储器、处理器和显示器,该存储器用于存储计算机程序,处理器用于从存储器中调用并运行计算机程序,当程序被运行时,该处理器执行上述第一方面或第一方面任意可能的实现方式中的方法。A third aspect provides a terminal device including a memory, a processor, and a display, the memory being used to store a computer program, the processor is configured to call and run a computer program from the memory, and when the program is executed, the processor executes the above The method of any of the first aspect or any of the possible implementations of the first aspect.
第四方面,提供一种计算机可读介质,用于存储计算机程序,该计算机程序包括用于执行第一方面或第一方面的任意可能的实现方式中的方法的指令。In a fourth aspect, a computer readable medium is provided for storing a computer program comprising instructions for performing the method of the first aspect or any of the possible implementations of the first aspect.
附图说明DRAWINGS
图1是实现本申请一个实施例的终端设备的一个最少硬件系统的示意图。 1 is a schematic diagram of a minimum hardware system for implementing a terminal device of an embodiment of the present application.
图2是根据本申请实施例的动态三维图像获取方法的示意性流程图。FIG. 2 is a schematic flowchart of a dynamic three-dimensional image acquisition method according to an embodiment of the present application.
图3是根据本申请一个实施例的运动姿态识别和轨迹的获取的设计方框图。3 is a block diagram showing the design of motion pose recognition and trajectory acquisition in accordance with one embodiment of the present application.
图4是根据本申请一个实施例的陀螺仪与加速度计数据融合示意图。4 is a schematic diagram of data fusion of a gyroscope and an accelerometer according to an embodiment of the present application.
图5是根据本申请一个实施例的陀螺仪与电子罗盘数据融合流程的方法示意图。FIG. 5 is a schematic diagram of a method for data fusion of a gyroscope and an electronic compass according to an embodiment of the present application.
图6是根据本申请实施例的根据终端设备的运动姿态与深度图像进行快速分割匹配的示意性流程图。FIG. 6 is a schematic flowchart of performing fast segmentation matching according to a motion posture and a depth image of a terminal device according to an embodiment of the present application.
图7是根据本申请实施例的根据所述彩色图像对快速分割匹配的结果进行精准匹配FIG. 7 is an accurate match of a result of fast segmentation matching according to the color image according to an embodiment of the present application.
的示意性流程图。Schematic flow chart.
图8根据本申请实施例的图像重叠区域融合的方法的示意性流程图。FIG. 8 is a schematic flowchart of a method for image overlap region fusion according to an embodiment of the present application.
图9是根据本申请实施例的用户拍摄动态三维图像的示意性流程图。FIG. 9 is a schematic flowchart of a user capturing a dynamic three-dimensional image according to an embodiment of the present application.
图10是根据本申请实施例的用户查看动态三维图像的示意性流程图。FIG. 10 is a schematic flowchart of a user viewing a dynamic three-dimensional image according to an embodiment of the present application.
图11是根据本申请实施例的终端设备的一例的示意性框图。FIG. 11 is a schematic block diagram of an example of a terminal device according to an embodiment of the present application.
具体实施方式detailed description
下面结合附图,对本申请的实施例进行描述。Embodiments of the present application will be described below with reference to the accompanying drawings.
本申请是实施例的终端设备可以是接入终端、用户设备(user equipment,UE)、用户单元、用户站、移动站、移动台、远方站、远程终端、移动设备、用户终端、无线通信设备、用户代理或用户装置。终端设备可以是蜂窝电话、无绳电话、会话启动协议(session initiation protocol,SIP)电话、无线本地环路(wireless local loop,WLL)站、个人数字处理(personal digital assistant,PDA)、具有无线通信功能的手持设备、计算设备或连接到无线调制解调器的其它处理设备、车载设备、可穿戴设备等。The terminal device in this embodiment may be an access terminal, a user equipment (UE), a subscriber unit, a subscriber station, a mobile station, a mobile station, a remote station, a remote terminal, a mobile device, a user terminal, and a wireless communication device. , user agent or user device. The terminal device may be a cellular phone, a cordless phone, a session initiation protocol (SIP) phone, a wireless local loop (WLL) station, a personal digital assistant (PDA), and a wireless communication function. Handheld device, computing device or other processing device connected to a wireless modem, in-vehicle device, wearable device, and the like.
图1是实现本申请的三维手势解锁方法的终端设备的一个最少硬件系统100的示意图。图1所示的系统100包括:光源发射器110、深度摄像头120、光谱分析模块130、彩色摄像头140、处理器150、显示单元160、非易失性存储器170、内存180和感测单元190。1 is a schematic diagram of a minimum hardware system 100 of a terminal device implementing the three-dimensional gesture unlocking method of the present application. The system 100 shown in FIG. 1 includes a light source transmitter 110, a depth camera 120, a spectrum analysis module 130, a color camera 140, a processor 150, a display unit 160, a nonvolatile memory 170, a memory 180, and a sensing unit 190.
彩色摄像头140,光源发射器110和深度摄像头120组成光谱输入模块,光谱分析模块130构成图像生成模块。光源发射器110、彩色摄像头140和深度摄像头120可以并排安装于设备上方(例如,设备正上方中央位置)。光源发射器110可以是红外发射器,深度摄像头120可以是红外摄像头,光谱分析模块130可以是红外光谱分析模块。在该情况下,光源发射器110与深度摄像头120配合工作,通过红外光编码影像放映场景。光源发射器110输出普通激光光源,经磨砂玻璃与红外滤光片过滤后形成近红外光。其中,光源发射器110可以持续全面输出波长为840纳米(nm)的红外光。The color camera 140, the light source emitter 110 and the depth camera 120 constitute a spectral input module, and the spectral analysis module 130 constitutes an image generation module. The light source emitter 110, color camera 140, and depth camera 120 can be mounted side by side over the device (eg, directly above the device). The light source emitter 110 can be an infrared emitter, the depth camera 120 can be an infrared camera, and the spectrum analysis module 130 can be an infrared spectrum analysis module. In this case, the light source emitter 110 cooperates with the depth camera 120 to project the scene through the infrared light encoded image. The light source emitter 110 outputs a common laser light source, which is filtered by a frosted glass and an infrared filter to form near-infrared light. Wherein, the light source emitter 110 can continuously output infrared light having a wavelength of 840 nanometers (nm).
深度摄像头120是一个互补金属氧化物半导体(Complementary Metal Oxide Semiconductor,CMOS)图像传感器,用于接收外界反射来的激励光源,比如红外光,将激励光源进行数字编码后形成数字影像传输给光谱分析模块130。光谱分析模块130分析散斑,计算出图像对应像素点与深度摄像头120的距离,并构成深度数据矩阵供驱动程序读取。The depth camera 120 is a Complementary Metal Oxide Semiconductor (CMOS) image sensor for receiving an excitation light source reflected from the outside, such as infrared light, and digitally encoding the excitation light source to form a digital image for transmission to the spectrum analysis module. 130. The spectral analysis module 130 analyzes the speckles, calculates the distance between the corresponding pixel points of the image and the depth camera 120, and forms a depth data matrix for the driver to read.
感测单元190连接到处理器150,检测终端设备的位置信息或周围环境的改变,并将感测的信息发送到处理器150。具体地,感测单元190包括以下项中的至少一个:用于通过检测旋转、旋转移动、角位移、倾斜或者任何其它非线性运动的陀螺仪传感器、用于 感测一个或多个方向的加速度的三轴加速度传感器、用于感测地球磁场确定南北方向的的电子罗盘。感测单元190在处理器150的控制下进行操作。The sensing unit 190 is connected to the processor 150, detects location information of the terminal device or a change in the surrounding environment, and transmits the sensed information to the processor 150. Specifically, the sensing unit 190 includes at least one of: a gyro sensor for detecting rotation, rotational movement, angular displacement, tilt, or any other non-linear motion, for A triaxial acceleration sensor that senses acceleration in one or more directions, an electronic compass that senses the earth's magnetic field to determine the north-south direction. The sensing unit 190 operates under the control of the processor 150.
终端设备可接收感测单元190中的运动传感器(例如,陀螺仪传感器或加速度传感器)生成的运动传感器数据,利用运动感测应用处理生成的运动传感器数据。例如,运行运动感测应用的处理器可以分析运动传感器数据,从而辨别具体类型的运动事件。The terminal device may receive motion sensor data generated by a motion sensor (eg, a gyro sensor or an acceleration sensor) in the sensing unit 190, and process the generated motion sensor data using the motion sensing application. For example, a processor running a motion sensing application can analyze motion sensor data to identify specific types of motion events.
显示单元160被配置为将图形、图像或数据显示给用户。显示单元160被配置为提供与终端设备的操作相关联的各种屏幕。显示单元160提供主屏幕、消息编写屏幕、电话屏幕、游戏屏幕、音乐播放屏幕和视频播放屏幕。显示单元160可利用平面显示面板(诸如,液晶显示器(LCD)、有机发光二极管(OLED)和有源矩阵OLED(AMOLED))来实现。 Display unit 160 is configured to display graphics, images or data to a user. The display unit 160 is configured to provide various screens associated with the operation of the terminal device. The display unit 160 provides a home screen, a message composing screen, a phone screen, a game screen, a music playing screen, and a video playing screen. The display unit 160 can be implemented using a flat display panel such as a liquid crystal display (LCD), an organic light emitting diode (OLED), and an active matrix OLED (AMOLED).
在以触摸屏的形式实现显示单元160的情况下,显示单元160可作为输入装置进行工作。在以触摸屏的形式实现显示单元160的情况下,显示单元160包括用于检测触摸手势的触摸面板。触摸面板被配置为将施加到显示单元160的特定位置的压力或在显示单元160的特定区域的电容变化转换为电输入信号。触摸面板可按照add-on(附加)式或者on-cell式(或in-cell式)之一来实现。In the case where the display unit 160 is implemented in the form of a touch screen, the display unit 160 can operate as an input device. In the case where the display unit 160 is implemented in the form of a touch screen, the display unit 160 includes a touch panel for detecting a touch gesture. The touch panel is configured to convert a pressure applied to a specific position of the display unit 160 or a capacitance change at a specific area of the display unit 160 into an electric input signal. The touch panel can be implemented in one of add-on or on-cell (or in-cell).
触摸面板可按照以下面板之一来实现:电阻式触摸面板、电容式触摸面板、电磁感应式触摸面板和压力式触摸面板。触摸面板被配置为检测触摸的压力以及被触摸的位置和区域。如果在触摸面板上做出触摸手势,则向处理器150产生相应的输入信号。然后,处理器150检查用户的触摸输入信息以执行相应的功能。The touch panel can be implemented in one of the following panels: a resistive touch panel, a capacitive touch panel, an electromagnetic induction touch panel, and a pressure touch panel. The touch panel is configured to detect the pressure of the touch as well as the location and area being touched. If a touch gesture is made on the touch panel, a corresponding input signal is generated to the processor 150. The processor 150 then checks the user's touch input information to perform the corresponding function.
处理器150可以负责执行各种软件程序(例如,应用程序和操作系统),以便提供用于终端设备的计算和处理操作。非易失性存储器170用于保存程序文件、系统文件和数据。内存180用于系统与程序运行缓存。The processor 150 can be responsible for executing various software programs (e.g., applications and operating systems) to provide computing and processing operations for the terminal devices. The non-volatile memory 170 is used to store program files, system files, and data. Memory 180 is used for system and program running caches.
下面,将详细描述根据本申请实施例的终端设备的动态三维图像获取的方法。Hereinafter, a method of dynamic three-dimensional image acquisition of a terminal device according to an embodiment of the present application will be described in detail.
图2是根据本申请一个实施例的动态三维图像获取的方法的示意性流程图。图2所示的方法可以由图1所示的终端设备执行。2 is a schematic flow chart of a method for dynamic three-dimensional image acquisition according to an embodiment of the present application. The method shown in FIG. 2 can be performed by the terminal device shown in FIG. 1.
S210,获取终端设备的运动姿态。S210. Acquire a motion posture of the terminal device.
下面,具体介绍设备运动姿态的获取方法。The following describes the acquisition method of the motion posture of the device.
本文提到的“姿态”或“运动姿态”是设备的一组运动,该运动可以是所包含的一组运动,例如,摆动或圆周运动,或者可以是设备的简单移动,例如,设备在特定的轴或角度上的倾斜。A "pose" or "motion pose" as referred to herein is a set of motions of a device, which may be a set of motions included, such as a swing or a circular motion, or may be a simple movement of the device, eg, the device is specific The tilt of the axis or angle.
图3示出了用于运动姿态识别和轨迹的获取的设计方框图。采样单元310可以从陀螺仪、加速度计和电子罗盘接收运动数据并进行采样。姿态解算单元320读取陀螺仪、加速度计和电子罗盘的数据,计算设备三轴角速度,计算角增量矩阵,求解姿态微分方程,最终更新姿态四元数。数据融合单元330基于卡尔曼滤波算法对相关输出中的噪声进行滤波并最终输出设备姿态和轨迹。Figure 3 shows a block diagram of a design for motion pose recognition and acquisition of trajectories. The sampling unit 310 can receive motion data from the gyroscope, the accelerometer, and the electronic compass and sample. The attitude solving unit 320 reads the data of the gyroscope, the accelerometer and the electronic compass, calculates the triaxial angular velocity of the device, calculates the angular increment matrix, solves the attitude differential equation, and finally updates the attitude quaternion. The data fusion unit 330 filters the noise in the correlation output based on the Kalman filter algorithm and finally outputs the device pose and trajectory.
由于陀螺仪与加速度计会受到外界因素(如摩擦力、不稳定力矩等)的影响,因此在采集陀螺仪与加速度计传感器数据之前需要在第一次上电前对之进行校准,以消除静态误差。Since gyroscopes and accelerometers are affected by external factors (such as friction, unstable torque, etc.), it is necessary to calibrate the gyroscope and accelerometer sensor data before the first power-on to eliminate static. error.
陀螺仪或加速度计误差校准过程中采用的误差模型可以用式(1)表示。 The error model used in the gyroscope or accelerometer error calibration process can be expressed by equation (1).
Figure PCTCN2017088162-appb-000001
Figure PCTCN2017088162-appb-000001
其中,[Dx Dy Dz]T是陀螺仪或加速度计所测物理量的真实值,[Mx My Mz]T是陀螺仪或加速度计的实际测量值,[Bx By Bz]T是传感器零偏。对于处于静止状态的陀螺仪,Dx、Dy、Dz均为0,而对于处于水平静止状态的加速度计Dx、Dy、均为0,Dz为重力加速度值。Where [D x D y D z ] T is the true value of the physical quantity measured by the gyroscope or accelerometer, [M x M y M z ] T is the actual measured value of the gyroscope or accelerometer, [B x B y B z ] T is the sensor bias. For the gyroscope in a stationary state, D x , D y , and D z are all 0, and for the accelerometers D x and D y in the horizontal stationary state, both are 0, and D z is a gravitational acceleration value.
利用电子罗盘进行地磁测量时,导致电子罗盘方向误差的因素很多,比如环境磁干扰因素,如电流、铁质材料、永久性磁铁等,使得磁传感器测量值偏离了地磁真值,从而使得方向计算时会产生导航偏差;同时,罗盘倾斜角度,其依赖于地理位置和方位,会导致较大的方向误差。因此,磁传感器测量数据的误差校准是不可缺少的重要环节。When using the electronic compass for geomagnetic measurement, there are many factors that cause the error of the electronic compass direction, such as environmental magnetic interference factors, such as current, ferrous materials, permanent magnets, etc., so that the magnetic sensor measurement value deviates from the geomagnetic true value, thus making the direction calculation Navigation deviation occurs at the same time; at the same time, the compass tilt angle, which depends on the geographical position and orientation, leads to a large directional error. Therefore, error calibration of magnetic sensor measurement data is an indispensable important link.
针对电子罗盘,重点在于消除XY平面的误差,当没有误差的时候,它的测量值在XY平面上表现为一个圆形。当单位圆在X、Y轴经过a、b两个比例值变换,经过θ角的旋转和(x0,y0)的平移后,形成式(2)所示的椭圆方程:For the electronic compass, the focus is on eliminating the error in the XY plane. When there is no error, its measured value appears as a circle on the XY plane. When the unit circle is transformed by the two ratios a and b on the X and Y axes, after the rotation of the θ angle and the translation of (x 0 , y 0 ), the elliptic equation shown by the formula (2) is formed:
Figure PCTCN2017088162-appb-000002
Figure PCTCN2017088162-appb-000002
其中x1、y1为校准后的电子罗盘的输出,x、y为电子罗盘有偏差时的输出,本申请通过最小二乘拟合求出x0、y0、θ、a、b即可消除误差。Where x 1 and y 1 are the outputs of the calibrated electronic compass, and x and y are the outputs when the electronic compass is deviated. This application can obtain x 0 , y 0 , θ, a, b by least square fitting. Eliminate errors.
式(2)方程中的参数x0、y0、θ、a、b的计算如式(3)所示:The calculation of the parameters x 0 , y 0 , θ, a, b in the equation (2) is as shown in equation (3):
Figure PCTCN2017088162-appb-000003
Figure PCTCN2017088162-appb-000003
其中式(3)中相关参数的计算公式如下: The formula for calculating the relevant parameters in equation (3) is as follows:
Figure PCTCN2017088162-appb-000004
Figure PCTCN2017088162-appb-000004
U=α22γ2,V=2(β2γ22γ),W=α2γ22,U = α 2 + β 2 γ 2 , V = 2 (β 2 γ 2 - α 2 γ), W = α 2 γ 2 + β 2 ,
Figure PCTCN2017088162-appb-000005
Figure PCTCN2017088162-appb-000005
误差校准后,通过四元数描述设备的姿态,首先读取陀螺仪的数据,计算设备三轴角速度,计算角增量矩阵,求解姿态微分方程,最终更新姿态四元数。After the error is calibrated, the quaternion is used to describe the attitude of the device. First, the gyro data is read, the three-axis angular velocity of the device is calculated, the angular increment matrix is calculated, the attitude differential equation is solved, and the attitude quaternion is finally updated.
从惯性坐标系a转换到设备坐标系b的旋转四元数为:The rotation quaternion from the inertial coordinate system a to the device coordinate system b is:
Figure PCTCN2017088162-appb-000006
Figure PCTCN2017088162-appb-000006
其中θ是旋转的角度,μR是旋转轴在惯性坐标系中的表示,从式(4)可得到:Where θ is the angle of rotation and μ R is the representation of the axis of rotation in the inertial coordinate system. From equation (4):
Figure PCTCN2017088162-appb-000007
Figure PCTCN2017088162-appb-000007
通过将
Figure PCTCN2017088162-appb-000008
转换为陀螺仪可测的
Figure PCTCN2017088162-appb-000009
可得四元数:
Through
Figure PCTCN2017088162-appb-000008
Convert to gyroscope measurable
Figure PCTCN2017088162-appb-000009
Can get quaternion:
Figure PCTCN2017088162-appb-000010
Figure PCTCN2017088162-appb-000010
可采用毕卡法求解四元数微分方程,其过程是先计算出载体运动时对应的四元数Q(t),再根据四元数和姿态矩阵
Figure PCTCN2017088162-appb-000011
的对应关系,分别求出姿态矩阵和姿态角设Δθx、Δθy、Δθz为陀螺仪x、y、z轴在[tk,tk+1]采样间隔时间内的角变化量,Δθ2=Δθx 2+Δθy 2+Δθz 2,得四元数的各阶近似算法为:
The Pika method can be used to solve the quaternion differential equation. The process is to first calculate the corresponding quaternion Q(t) when the carrier moves, and then according to the quaternion and the attitude matrix.
Figure PCTCN2017088162-appb-000011
Correspondence relationship, respectively determine the attitude matrix and attitude angles Δθ x , Δθ y , Δθ z are the angular changes of the gyroscope x, y, and z axes at the sampling interval of [t k , t k+1 ], Δθ 2 = Δθ x 2 + Δθ y 2 + Δθ z 2 , the order approximation algorithm for the quaternion is:
Figure PCTCN2017088162-appb-000012
Figure PCTCN2017088162-appb-000012
Figure PCTCN2017088162-appb-000013
Figure PCTCN2017088162-appb-000013
采用卡尔曼滤波算法,对加速度计解算的翻滚角与陀螺仪测试的翻滚角速度、加速度计解算的俯仰角数据与陀螺仪测试俯仰角速度数据,分别进行滤波处理,可以让加速度计与陀螺仪数据互相补偿,减小测量噪音,俯仰角、翻滚角测试值更准确,使得磁传感器倾角补偿效果好,能够进行静态校准,同时也可进行动态校准。 Using the Kalman filter algorithm, the rollover angle calculated by the accelerometer and the rollover angular velocity of the gyroscope test, the pitch angle data calculated by the accelerometer and the gyroscope test pitch rate data are respectively filtered, and the accelerometer and the gyroscope can be made. The data compensates each other, reduces the measurement noise, and the pitch angle and roll angle test values are more accurate, which makes the magnetic sensor tilt angle compensation effect better, can perform static calibration, and can also perform dynamic calibration.
将这两个传感器的噪声方差矩阵设为变量,并实时监测外界干扰,动态改变加速度计和电子罗盘的噪声方差矩阵,进而修正其在卡尔曼滤波器中的增益。The noise variance matrix of these two sensors is set as a variable, and the external disturbance is monitored in real time, and the noise variance matrix of the accelerometer and the electronic compass is dynamically changed, and then the gain in the Kalman filter is corrected.
在姿态解算步骤得到先验姿态四元数后,读取加速度计和电子罗盘的数值,得到观测量,将先验姿态四元数作为状态量的初值,带入卡尔曼滤波器的公式中得到最终姿态四元数。本申请将陀螺仪与加速度计融合,估计出俯仰角θ和横滚角γ,将陀螺仪与电子罗盘融合,估计航向角
Figure PCTCN2017088162-appb-000014
After obtaining the a priori quaternion in the attitude solving step, the values of the accelerometer and the electronic compass are read to obtain the observation, and the a priori quaternion is used as the initial value of the state quantity, and the formula of the Kalman filter is brought. Get the final pose quaternion. In the present application, the gyroscope is integrated with the accelerometer, and the pitch angle θ and the roll angle γ are estimated, and the gyroscope is integrated with the electronic compass to estimate the heading angle.
Figure PCTCN2017088162-appb-000014
陀螺仪与加速度计数据融合流程如图4所示,陀螺仪与电子罗盘数据融合流程如图5所示。The data fusion process of gyroscope and accelerometer is shown in Figure 4. The data fusion process of gyroscope and electronic compass is shown in Figure 5.
通过姿态解算与数据融合以后的数据是由四元数表示的,可通过式(9)转换为方向余弦阵:The data after the fusion of the attitude solution and the data is represented by a quaternion, which can be converted into a direction cosine matrix by the formula (9):
Figure PCTCN2017088162-appb-000015
Figure PCTCN2017088162-appb-000015
通过式(10)和式(11)转换为欧拉角:Converted to Euler angles by equations (10) and (11):
Figure PCTCN2017088162-appb-000016
Figure PCTCN2017088162-appb-000016
Figure PCTCN2017088162-appb-000017
Figure PCTCN2017088162-appb-000017
S220,通过深度摄像头和彩色摄像头分别采集深度图像和彩色图像。S220: Collect a depth image and a color image respectively by using a depth camera and a color camera.
其中,深度图像也被称为距离映像(range image),是指将从图像采集器(例如,本申请中的深度摄像头120)到场景中各点的距离(深度)作为像素值的图像,它直接反映了景物可见表面的几何形状。The depth image is also referred to as a range image, and refers to an image from a image collector (for example, the depth camera 120 in the present application) to a point (depth) of each point in the scene as a pixel value. It directly reflects the geometry of the visible surface of the scene.
举例来说,在该方法由图1所示的终端设备执行时,深度摄像头120通过接收外界反射来的激励光源,比如红外光,将激励光源进行数字编码后形成数字影像传输给光谱分析模块130。光谱分析模块130分析散斑,计算出当前图像中对应像素点(x,y)与深度摄像头120的距离z,从而可以获取当前深度图像。For example, when the method is performed by the terminal device shown in FIG. 1 , the depth camera 120 digitally encodes the excitation light source by receiving an excitation light source reflected by the outside, such as infrared light, to form a digital image and transmit it to the spectrum analysis module 130. . The spectral analysis module 130 analyzes the speckles and calculates a distance z between the corresponding pixel point (x, y) in the current image and the depth camera 120, so that the current depth image can be acquired.
S230,根据终端设备的运动姿态与所述深度图像进行快速分割匹配。S230: Perform fast segmentation matching with the depth image according to the motion posture of the terminal device.
图6示出了设备运动姿态与深度图像融合的中心区域三维物体快速匹配方法,该方法实时跟踪设备状态变化,当设备平滑移动时,抽取每个固定时间段内起始与结束时间点对应的深度图帧。基于起始时间点的深度图已完成分割的特征区域和近景模式下设备姿态变化推断出在时间结束点的深度图各特征区域大致范围,进而进行快速分割匹配。6 shows a fast matching method of a central region three-dimensional object in which a motion figure and a depth image of a device are fused, the method tracking a device state change in real time, and extracting a start and end time point corresponding to each fixed time period when the device smoothly moves. Depth map frame. Based on the start time point, the depth map has completed the segmentation of the feature region and the device pose change in the close-up mode, and the approximate range of each feature region of the depth map at the end of the time is inferred, thereby performing fast segmentation matching.
由于深度图像每个像素点的值都是物体到摄像头之间的直线距离,所以同一物体到摄像头之间距离存在相似性。因此基于深度图的粗分割采用区域增长法,但由于深度图像存在噪声且容易丢失信息,所以首先对深度图像进行滤波,来实现图像的平滑和丢失深度的填充,具体实现方式如下:Since the value of each pixel of the depth image is the linear distance between the object and the camera, there is a similarity between the distance from the same object to the camera. Therefore, the coarse segmentation based on the depth map adopts the region growing method. However, since the depth image has noise and is easy to lose information, the depth image is first filtered to achieve image smoothing and loss depth filling. The specific implementation manner is as follows:
采用双边滤波器对图像进行滤波,滤波器定义为: The image is filtered using a bilateral filter defined as:
Figure PCTCN2017088162-appb-000018
Figure PCTCN2017088162-appb-000018
其中I为原始图像,I'为滤波后的图像,Ω为(x,y)的邻域,w(i,j)为滤波器在相应坐标处的权值,
Figure PCTCN2017088162-appb-000019
Where I is the original image, I' is the filtered image, Ω is the neighborhood of (x, y), and w(i, j) is the weight of the filter at the corresponding coordinates.
Figure PCTCN2017088162-appb-000019
对图像中深度相似的像素点合并构成相似特征区域。具体实现方式是:Pixels with similar depths in the image are combined to form a similar feature region. The specific implementation is:
1)选择起始像素点;1) Select the starting pixel point;
2)根据相似准则,将起始像素点与周围像素点进行深度值比较;2) comparing the starting pixel point with the surrounding pixel point according to the similarity criterion;
3)如果两者满足相似性的条件,就将该像素点合并到起始像素点,形成新的起始点区域;3) If the two satisfy the condition of similarity, the pixel is merged into the starting pixel to form a new starting point region;
4)当周围像素点不满足相似性条件时,则停止该方向上的增长。4) When the surrounding pixel points do not satisfy the similarity condition, the growth in the direction is stopped.
其中,起始点的选取对深度图像分割的效率至关重要,如果选择得当,可以加快分割,本申请就是根据设备相对姿态与轨迹来大致估算特征区域在结束时间点深度图内的位置来加速的。由于所拍摄物体离摄像头距离通常较近,选取深度图中的极小值区域,建立图像极小值区域的多叉树,实现起始点的选取。设深度图像D,像素点深度集Λ={d1,d2,...,dN},将像素点根据深度值进行从小到大的排序,从前往后找到图像中的所有极小值区域:
Figure PCTCN2017088162-appb-000020
Among them, the selection of the starting point is very important for the efficiency of depth image segmentation. If properly selected, the segmentation can be speeded up. This application is based on the relative posture and trajectory of the device to roughly estimate the position of the feature region in the depth map at the end time point to accelerate. . Since the distance of the captured object from the camera is usually close, the minimum value region in the depth map is selected, and a multi-fork tree of the image minimum value region is established to realize the selection of the starting point. Let the depth image D, the pixel depth set Λ={d 1 , d 2 ,..., d N }, sort the pixels according to the depth value from small to large, and find all the minimum values in the image from the arrival. region:
Figure PCTCN2017088162-appb-000020
相似准则用于区分物体与背景,选取前面几次比较的像素点深度均值与差异均值,选取后几次像素点深度均值与差异均值,当本次像素点深度均值与差异均值与两者差异在5%以内判定为同一区域。The similarity criterion is used to distinguish the object from the background. Select the average depth and difference mean of the pixel points in the previous comparison, and select the average depth and difference mean of the pixel points. The difference between the mean value of the pixel depth and the difference between the two is It is determined to be the same area within 5%.
运用设备姿态信息对时间结束点的深度图特征区域位置推断的过程如下:The process of inferring the location of the feature area of the depth map using the device attitude information at the end of time is as follows:
1)记录时间起始点设备姿态信息,根据特征区域像素点在图像中的坐标与自身的深度值确定其在设备自身坐标系中的坐标;1) recording the device attitude information of the starting point of the time, and determining the coordinates in the coordinate system of the device according to the coordinates of the pixel of the feature area in the image and the depth value of the image;
2)获取时间结束点相较于开始点的设备姿态变化,并将姿态变化折算到设备原来自身坐标系;2) Obtaining the change of the attitude of the device at the end of the time compared to the starting point, and converting the change of the posture to the original coordinate system of the device;
3)将起始点特征区域像素点坐标折算到时间结束点设备自身的坐标系。3) Convert the pixel coordinates of the starting point feature area to the coordinate system of the time end point device itself.
S240,根据所述彩色图像对所述快速分割匹配的结果进行精准匹配。S240. Perform accurate matching on the fast segmentation matching result according to the color image.
深度图像经过有效滤波后不仅能平滑噪声,更能填充缺失深度的像素点,但其精度并不高,因此基于深度图像的特征区域匹配并不能有效匹配实际物体特征区域。由于彩色图像分割方法能够有效的提取边界信息,本申请针对近景模式结合彩色图像的图像分割方法对快速匹配的结果进行精准匹配,针对远景模式直接对设备平滑移动或旋转时抽帧,并对彩色图像进行精准匹配,精准匹配主要是在快速匹配得到的特征区域边缘处进行优化,匹配流程如图7所示。After the depth image is effectively filtered, not only can the noise be smoothed, but also the pixels with missing depth can be filled, but the accuracy is not high. Therefore, the feature region matching based on the depth image cannot effectively match the actual object feature region. Since the color image segmentation method can effectively extract the boundary information, the present application accurately matches the result of the fast matching for the close-up mode and the image segmentation method of the color image, and directly draws the frame for the device to smoothly move or rotate for the perspective mode, and color The images are accurately matched, and the precise matching is mainly optimized at the edge of the feature region obtained by the fast matching. The matching process is shown in FIG. 7 .
首先对深度图像和彩色图像分别滤波,然后根据姿态信息和深度信息进行特征区域的快速匹配,得到一系列的特征区域,将各特征区域的代表性像素点提供给彩色图像进行分割。图像分割采用分水岭算法,经过滤波后生成发色后的灰度图,根据提供的特征像素点直接进行注水操作,最终得到各个特征区域的边界。以彩色图像分割边界区域为基础,快速匹配得到的特征区域边界点与之比较,如果不存在偏差,则两者匹配结果均 正常。如果存在偏差,且邻域深度数据缺失或落差不分明,则以彩色图像分割结果为最终结果。如果存在偏差,且邻域深度数据完善且落差不分明,则以深度图像分割结果为最终结果。如果存在偏差,且邻域深度数据完善且落差分明,则以彩色图像分割结果为最终结果。Firstly, the depth image and the color image are separately filtered, and then the feature regions are quickly matched according to the posture information and the depth information, and a series of feature regions are obtained, and the representative pixel points of each feature region are provided to the color image for segmentation. The image segmentation adopts the watershed algorithm, and after filtering, the grayscale image after the coloration is generated, and the water injection operation is directly performed according to the provided characteristic pixel points, and finally the boundary of each feature region is obtained. Based on the segmentation of the boundary region of the color image, the boundary points of the feature regions obtained by the fast matching are compared with each other. If there is no deviation, the matching results are both normal. If there is a deviation, and the neighborhood depth data is missing or the drop is not clear, the color image segmentation result is the final result. If there is a deviation, and the neighborhood depth data is perfect and the drop is not clear, the depth image segmentation result is the final result. If there is a deviation, and the neighborhood depth data is perfect and the difference is clear, the color image segmentation result is the final result.
精准匹配结果可以为设备姿态与轨迹信息提供反馈,使得姿态的识别更加精准。根据精准匹配结果得到姿态信息的过程如下:Accurate matching results can provide feedback for device attitude and trajectory information, making gesture recognition more accurate. The process of obtaining attitude information based on the exact matching result is as follows:
1)记录当前时刻设备姿态信息与当前各特征区域代表性的像素点在深度图像中的坐标,记录特征区域像素点的深度值,确定其在设备自身坐标系中的坐标;1) recording the coordinates of the device at the current time and the coordinates of the representative pixel points of the current feature regions in the depth image, recording the depth values of the pixel points of the feature region, and determining the coordinates thereof in the coordinate system of the device itself;
2)获取当前时刻相较于时间开始点的设备姿态变化,并将姿态变化折算到设备原来自身坐标系;2) Obtain the change of the device posture of the current time compared with the start time of the time, and convert the posture change to the original coordinate system of the device;
3)将当前时刻特征区域像素点坐标折算到时间结束点设备自身的坐标系,若存在偏差,则修正当前时刻设备姿态信息并重新折算,直至两者没有偏差为止。3) Convert the current point feature area pixel point coordinate to the coordinate system of the time end point device itself. If there is a deviation, correct the current time device posture information and re-convert until the two have no deviation.
S250,如果获取的当前图像与已拍摄图像存在重叠,通过融合算法对重叠区域进行融合。S250: If the acquired current image overlaps with the captured image, the overlapping region is merged by a fusion algorithm.
本申请允许设备全方位获取图像,在获取的图像与已拍摄图像存在重叠的时候需要对重叠区域进行融合。重叠区域融合基于历史特征矩阵和当前设备姿态信息,本申请将各个特征矩阵与设备姿态关联存储,因此可以获取历史特征矩阵前一次的设备姿态信息。具体的融合流程如图8所示。The present application allows the device to acquire an image in all directions, and the overlapping regions need to be merged when the acquired image overlaps with the captured image. The overlap region fusion is based on the historical feature matrix and the current device pose information. The present application associates each feature matrix with the device pose, so that the previous device pose information of the historical feature matrix can be obtained. The specific integration process is shown in Figure 8.
为了能够判定当前图像与历史图像存在重叠,设备在运动拍摄过程中的姿态信息将不断被记录、保存,同时也会定时得提取姿态信息作为与以后设备姿态进行比较的标志性数据。由于不同的设备姿态也可能存在重叠区域,本方法将设备所经历的每个姿态所能拍摄到的视野范围进行记录,与特征区域和姿态信息联合存储。在设备运动过程中实时得计算当前相对姿态与轨迹,从历史特征矩阵中提取可比较的标志性数据进行匹配,如果匹配结果表明当前图像帧与历史某图像帧有重叠的特征区域则进行融合处理,更新到重叠区域,同时记录当前设备姿态。In order to be able to determine that there is overlap between the current image and the historical image, the posture information of the device during the motion shooting will be continuously recorded and saved, and the attitude information may be extracted at regular intervals as the iconic data for comparison with the device posture in the future. Since different device poses may also have overlapping regions, the method records the field of view that can be captured by each gesture experienced by the device, and stores the feature region and posture information in combination. The current relative pose and trajectory are calculated in real time during the movement of the device, and the comparable iconic data is extracted from the historical feature matrix for matching. If the matching result indicates that the current image frame overlaps with the historical image frame, the fusion processing is performed. , updated to the overlap area, while recording the current device pose.
为了实现近景三维图像的准确获取和远景条件下功耗的优化,实现远、近景模式图像的关联。本申请提供的方法能够动态识别当前所摄景物是远景还是近景,判定方法是扫描深度图像矩阵,计算深度值小于阈值的像素点个数,当个数小于阈值时判定为远景。当处于远景模式时,自动关闭深度摄像头,并定时启动它以检测是否处于近景模式,这样可以降低功耗。In order to achieve accurate acquisition of close-range 3D images and optimization of power consumption under long-range conditions, the association of far and near-field mode images is realized. The method provided by the present application can dynamically identify whether the current scene is a distant view or a close-up view. The determining method is to scan the depth image matrix, calculate the number of pixel points whose depth value is less than the threshold value, and determine the foreground when the number is less than the threshold value. When in Vision mode, the depth camera is automatically turned off and periodically activated to detect if it is in close-up mode, which reduces power consumption.
在近景模式下,本方法跟踪特征区域的深度状态变化,在距离有变化时,通过将该距离存储到特征矩阵中,使得显示时可以识别到拍摄时是否有接近于远离动作,从而提示用户可以缩小和放大。In the close-up mode, the method tracks the change of the depth state of the feature area. When the distance changes, the distance is stored in the feature matrix, so that the display can recognize whether there is a close to the moving action when shooting, thereby prompting the user to Zoom out and zoom in.
在远景模式下,本方法不启动深度摄像头,因此无需进行深度图的中心区域快速匹配,仅根据彩色图像与姿态信息进行精准匹配进而得到特征矩阵。In the remote mode, the method does not activate the depth camera, so there is no need to perform fast matching of the central region of the depth map, and only the color image and the attitude information are accurately matched to obtain the feature matrix.
下面介绍用户拍摄流程,为保证拍摄效果,用户按下按键启动拍摄后,本方法提示用户可以向任意方向移动或旋转手机来对目标物体拍摄。启动拍摄后会触发同时出发姿态传感器、深度摄像头和彩色摄像头工作。设备姿态与轨迹识别模块会实时读取陀螺仪、加速度计和电子罗盘的数据,对之进行姿态解算,然后进行多传感器数据的融合,得到设备的姿态和轨迹。深度摄像头实时采集深度数据,经过滤波后进行远近模式的识别, 如果是近景模式,则对深度图像粗分割,结合设备姿态与轨迹数据进行中心区域快速匹配以加快匹配速度。彩色摄像头实时采集彩色数据,经过滤波后根据快速匹配结果提供的特征区域代表像素点实施注水操作,最终得到景物的各边界,与快速匹配所得到的特征区域边界进行比较并决策,调整特征区域的边界最终生成细微匹配的特征区域。The following describes the user shooting process. In order to ensure the shooting effect, after the user presses the button to start shooting, the method prompts the user to move or rotate the mobile phone in any direction to shoot the target object. When the shooting is started, the simultaneous starting attitude sensor, depth camera and color camera work are triggered. The device attitude and trajectory identification module reads the data of the gyroscope, accelerometer and electronic compass in real time, performs attitude calculation on it, and then fuses the multi-sensor data to obtain the posture and trajectory of the device. The depth camera collects the depth data in real time, and after filtering, recognizes the near and far modes. If it is a close-up mode, the depth image is roughly segmented, and the device region and the trajectory data are quickly matched to the central region to speed up the matching. The color camera collects the color data in real time, and after filtering, performs the water injection operation according to the feature area provided by the fast matching result, and finally obtains the boundary of the scene, compares and determines the boundary of the feature area obtained by the fast matching, and adjusts the feature area. The boundary eventually produces a subtle matching feature area.
为解决在拍摄过程中所拍摄图像与已经拍摄过的区域再次重合的情况下,能够在显示时识别到并认为是同一个区域,从而实现循环的查看,本方法根据设备姿态与特征区域在设备坐标系中的位置进行计算,对重合区域进行融合更新并记录两次拍摄的设备姿态,使得提供给显示时可以使之循环查看。重合区域融合后就生成了图像最终的特征区域集合。In order to solve the problem that the image captured during the shooting coincides with the already photographed area, it can be recognized and considered to be the same area at the time of display, thereby realizing the cycle view, and the method is based on the device posture and the feature area in the device. The position in the coordinate system is calculated, and the coincident area is fused and updated, and the posture of the device taken twice is recorded, so that it can be cyclically viewed when supplied to the display. After the coincidence region is merged, the final feature region set of the image is generated.
在用户松开按键时,认为用户希望停止拍摄,此时停止传感器与摄像头的工作,并等待最后一帧图像处理完成,然后清理中间缓存并释放资源,最终将所得特征区域集合写入非易失性存储器完成拍摄。用户拍摄图像过程如图9所示:When the user releases the button, the user is expected to stop shooting. At this time, the sensor and the camera are stopped, and the last frame of image processing is completed, then the intermediate buffer is cleared and the resources are released, and finally the set of the obtained feature regions is written to the nonvolatile. The memory finishes shooting. The process of shooting images by the user is shown in Figure 9:
在用户点击图片进行查看时启动姿态传感器进行设备姿态获取,同时读取图像特征区域集合取得最终图像帧。本方法支持用户旋转手机来查看图片,打开图片时手机的姿态被对应成开始拍摄图片时手机的姿态。设备姿态与轨迹识别模块会实时读取陀螺仪、加速度计和电子罗盘的数据,对之进行姿态解算,然后进行多传感器数据的融合,得到设备的姿态。读取图像特征区域集合后,根据每个特征区域所在图像帧中的坐标进行图像帧的合成,最后将图像帧一一缓存等待被读取。在设备当前姿态生成后需跟拍摄照片是的初始设备姿态对应,此后手机姿态的变化将会触发对应状态图片帧的显示。设备姿态的变化将触发选取对应姿态的图像帧并提交显示,在实现图像后判断当前帧是否可缩放,如果可以,就在屏幕提示当前可缩放,然后获取姿态传感器数据开始新的循环。如果不可以缩放,同样获取姿态传感器数据开始新的循环。When the user clicks on the picture to view, the posture sensor is activated to acquire the device posture, and the image feature region set is read to obtain the final image frame. The method supports the user to rotate the mobile phone to view the picture, and the gesture of the mobile phone is corresponding to the posture of the mobile phone when the picture is taken. The device attitude and trajectory identification module reads the data of the gyroscope, accelerometer and electronic compass in real time, performs attitude calculation on it, and then fuses the multi-sensor data to obtain the posture of the device. After reading the image feature region set, the image frames are synthesized according to the coordinates in the image frame in which each feature region is located, and finally the image frames are buffered one by one to be read. After the current pose of the device is generated, it needs to correspond to the initial device pose of the captured photo. After that, the change of the gesture of the mobile phone will trigger the display of the corresponding state picture frame. The change of the device pose will trigger the selection of the image frame of the corresponding pose and submit the display. After the image is implemented, it is judged whether the current frame is scalable. If possible, the screen prompts that the current zoom is possible, and then the posture sensor data is acquired to start a new cycle. If you can't zoom, you also get the attitude sensor data to start a new loop.
在用户点击返回时,认为用户不再查看图片,此时停止传感器与摄像头的工作,清理中间缓存并释放资源。用户查看图像过程如图10所示。When the user clicks back, the user is considered to no longer view the picture. At this time, the sensor and the camera are stopped, the intermediate cache is cleared, and resources are released. The process of viewing images by the user is shown in FIG.
(1)本申请在终端设备上配备陀螺仪、加速度计和电子罗盘传感器,用于提供设备姿态信息;配备一个红外发射器和一个红外摄像头,用于提供深度图像数据;配备一个彩色摄像头,用于提供彩色图像数据,三者结合为动态三维图像的获取提供原始数据支撑。(1) The application is equipped with a gyroscope, an accelerometer and an electronic compass sensor on the terminal device for providing device attitude information; an infrared transmitter and an infrared camera for providing depth image data; and a color camera for use In order to provide color image data, the combination of the three provides raw data support for the acquisition of dynamic three-dimensional images.
(2)本申请所涉及的三维空间终端设备姿态识别的方法,能够通过对三个姿态传感器采样、姿态结算和数据融合,得到设备初始姿态。根据深度图像和彩色图像的变化对姿态生成算法进行补偿,完成对姿态检测的闭环跟踪。(2) The method for gesture recognition of a three-dimensional space terminal device according to the present application can obtain an initial posture of a device by sampling, posture settlement, and data fusion of three attitude sensors. The attitude generation algorithm is compensated according to the change of the depth image and the color image, and the closed-loop tracking of the attitude detection is completed.
(3)本申请所涉及的设备姿态与深度图像融合的中心区域三维物体快速匹配方法,能够在设备姿态匀速变化时提供一种抽帧形式加速匹配的策略,实现多帧图像间同一三维物体的快速匹配。(3) The fast matching method of the three-dimensional object in the central region of the device posture and depth image fusion in the present application can provide a strategy of speeding up matching in the form of frame drawing when the device posture changes at a constant speed, and realize the same three-dimensional object between multiple frames of images. Quick match.
(4)本申请所涉及的基于彩色图像的,对多帧图像间同一三维物体的快速匹配结果的细微匹配方法,能够针对各特征区域根据彩色图像对应位置的数据信息对快速匹配结果进行补偿优化,得到最细粒度的图片特征描述。(4) The fine matching method based on the color image and the fast matching result of the same three-dimensional object between the multi-frame images according to the present application can compensate and optimize the fast matching result according to the data information of the corresponding position of the color image for each feature region. , get the most detailed image feature description.
(5)本申请能实现360度全方位摄像,支持对已经拍摄过的物体在此拍摄,能够根据设备当前姿态信息与历史特征矩阵信息匹配,实现动态识别已经拍摄过的重叠区域。对重叠区域进行数据融合,增加重叠数据信息,使得显示时可以根据姿态平滑切换。(5) The present application can realize 360-degree omnidirectional imaging, and supports shooting of an already photographed object, and can dynamically recognize the overlapped area that has been photographed according to the current posture information of the device and the historical feature matrix information. Data fusion is performed on the overlapping regions, and the overlapping data information is added, so that the display can be smoothly switched according to the posture.
(6)本申请能够动态识别所摄景物远近模式,针对近景,实现全方位的摄像,针对远 景,实现全景的融合。远景自动关闭深度摄像头以降低功耗。(6) This application can dynamically recognize the distant and near mode of the subject, and achieve a full range of camera for the near view. Scenery, the integration of the panorama. The Vision automatically turns off the depth camera to reduce power consumption.
图11是根据本申请实施例的终端设备的另一示意性框图。图11所示的终端设备1100包括:射频(Radio Frequency,RF)电路1110、存储器1120、其他输入设备1130、显示屏1140、传感器1150、音频电路1160、I/O子系统1170、处理器1180、以及电源1190等部件。本领域技术人员可以理解,图11中示出的终端设备结构并不构成对终端设备的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者拆分某些部件,或者不同的部件布置。本领领域技术人员可以理解显示屏1140属于用户界面(User Interface,UI),且终端设备1100可以包括比图示或者更少的用户界面。FIG. 11 is another schematic block diagram of a terminal device according to an embodiment of the present application. The terminal device 1100 shown in FIG. 11 includes: a radio frequency (RF) circuit 1110, a memory 1120, other input devices 1130, a display screen 1140, a sensor 1150, an audio circuit 1160, an I/O subsystem 1170, and a processor 1180. And power supply 1190 and other components. It will be understood by those skilled in the art that the terminal device structure shown in FIG. 11 does not constitute a limitation of the terminal device, and may include more or less components than those illustrated, or combine some components or split some components. , or different parts layout. Those skilled in the art will appreciate that the display screen 1140 belongs to a User Interface (UI), and the terminal device 1100 may include a user interface that is smaller than illustrated or less.
下面结合图11对终端设备1100的各个构成部件进行具体的介绍:The specific components of the terminal device 1100 are specifically described below with reference to FIG. 11:
RF电路1110可用于收发信息或通话过程中,信号的接收和发送,特别地,将基站的下行信息接收后,给处理器1180处理;另外,将设计上行的数据发送给基站。通常,RF电路包括但不限于天线、至少一个放大器、收发信机、耦合器、低噪声放大器(Low Noise Amplifier,LNA)、双工器等。此外,RF电路1110还可以通过无线通信与网络和其他设备通信。所述无线通信可以使用任一通信标准或协议,包括但不限于全球移动通讯系统(Global System of Mobile communication,GSM)、通用分组无线服务(General Packet Radio Service,GPRS)、码分多址(Code Division Multiple Access,CDMA)、宽带码分多址(Wideband Code Division Multiple Access,WCDMA)、长期演进(Long Term Evolution,LTE)、电子邮件、短消息服务(Short Messaging Service,SMS)等。The RF circuit 1110 can be used for receiving and transmitting signals during and after receiving or transmitting information, in particular, after receiving the downlink information of the base station, and processing it to the processor 1180; in addition, transmitting the designed uplink data to the base station. Generally, RF circuits include, but are not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a Low Noise Amplifier (LNA), a duplexer, and the like. In addition, RF circuitry 1110 can also communicate with the network and other devices via wireless communication. The wireless communication may use any communication standard or protocol, including but not limited to Global System of Mobile communication (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (Code). Division Multiple Access (CDMA), Wideband Code Division Multiple Access (WCDMA), Long Term Evolution (LTE), E-mail, Short Messaging Service (SMS), etc.
存储器1120可用于存储软件程序以及模块,处理器1180通过运行存储在存储器1120的软件程序以及模块,从而执行终端设备1100的各种功能应用以及数据处理。存储器1120可主要包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需的应用程序(比如声音播放功能、图像播放功能等)等;存储数据区可存储根据终端设备1100的使用所创建的数据(比如音频数据、电话本等)等。此外,存储器1120可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件、闪存器件、或其他易失性固态存储器件。The memory 1120 can be used to store software programs and modules, and the processor 1180 executes various functional applications and data processing of the terminal device 1100 by running software programs and modules stored in the memory 1120. The memory 1120 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application required for at least one function (such as a sound playing function, an image playing function, etc.), and the like; the storage data area may be stored according to Data (such as audio data, phone book, etc.) created by the use of the terminal device 1100. Moreover, memory 1120 can include high speed random access memory, and can also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
其他输入设备1130可用于接收输入的数字或字符信息,以及产生与终端设备1100的用户设置以及功能控制有关的键信号输入。具体地,其他输入设备1130可包括但不限于物理键盘、功能键(比如音量控制按键、开关按键等)、轨迹球、鼠标、操作杆、光鼠(光鼠是不显示可视输出的触摸敏感表面,或者是由触摸屏形成的触摸敏感表面的延伸)等中的一种或多种。其他输入设备1130与I/O子系统1170的其他输入设备控制器1171相连接,在其他设备输入控制器1171的控制下与处理器1180进行信号交互。 Other input devices 1130 can be used to receive input numeric or character information, as well as to generate key signal inputs related to user settings and function control of terminal device 1100. Specifically, other input devices 1130 may include, but are not limited to, physical keyboards, function keys (such as volume control buttons, switch buttons, etc.), trackballs, mice, joysticks, and light mice (the light mouse is not sensitive to display visual output). One or more of a surface, or an extension of a touch sensitive surface formed by a touch screen. Other input devices 1130 are coupled to other input device controllers 1171 of I/O subsystem 1170 for signal interaction with processor 1180 under the control of other device input controllers 1171.
显示屏1140可用于显示由用户输入的信息或提供给用户的信息以及终端设备1100的各种菜单,还可以接受用户输入。具体的显示屏1140可包括显示面板1141,以及触控面板1142。其中显示面板1141可以采用液晶显示器(Liquid Crystal Display,LCD)、有机发光二极管(Organic Light-Emitting Diode,OLED)等形式来配置显示面板1141。触控面板1142,也称为触摸屏、触敏屏等,可收集用户在其上或附近的接触或者非接触操作(比如用户使用手指、触笔等任何适合的物体或附件在触控面板1142上或在触控面板1142附近的操作,也可以包括体感操作;该操作包括单点控制操作、多点控制操作等操作类型),并根据预先设定的程式驱动相应的连接装置。可选的,触控面板1142可包括触摸检测装置和触摸控制器两个部分。其中,触摸检测装置检测用户的触摸方位、姿势,并 检测触摸操作带来的信号,将信号传送给触摸控制器;触摸控制器从触摸检测装置上接收触摸信息,并将它转换成处理器能够处理的信息,再送给处理器1180,并能接收处理器1180发来的命令并加以执行。此外,可以采用电阻式、电容式、红外线以及表面声波等多种类型实现触控面板1142,也可以采用未来发展的任何技术实现触控面板1142。进一步的,触控面板1142可覆盖显示面板1141,用户可以根据显示面板1141显示的内容(该显示内容包括但不限于,软键盘、虚拟鼠标、虚拟按键、图标等等),在显示面板1141上覆盖的触控面板1142上或者附近进行操作,触控面板1142检测到在其上或附近的操作后,通过I/O子系统1170传送给处理器1180以确定用户输入,随后处理器1180根据用户输入通过I/O子系统1170在显示面板1141上提供相应的视觉输出。虽然在图11中,触控面板1142与显示面板1141是作为两个独立的部件来实现终端设备1100的输入和输入功能,但是在某些实施例中,可以将触控面板1142与显示面板1141集成而实现终端设备1100的输入和输出功能。The display 1140 can be used to display information entered by the user or information provided to the user as well as various menus of the terminal device 1100, and can also accept user input. The specific display screen 1140 can include a display panel 1141 and a touch panel 1142. The display panel 1141 can be configured by using a liquid crystal display (LCD), an organic light-emitting diode (OLED), or the like. The touch panel 1142, also referred to as a touch screen, a touch sensitive screen, etc., can collect contact or non-contact operations on or near the user (eg, the user uses any suitable object or accessory such as a finger, a stylus, etc. on the touch panel 1142. Or the operation in the vicinity of the touch panel 1142 may also include a somatosensory operation; the operation includes a single-point control operation, a multi-point control operation, and the like, and the corresponding connection device is driven according to a preset program. Optionally, the touch panel 1142 may include two parts: a touch detection device and a touch controller. Wherein, the touch detection device detects a user's touch orientation and posture, and Detecting a signal brought by the touch operation, transmitting the signal to the touch controller; the touch controller receives the touch information from the touch detection device, and converts it into information that the processor can process, and sends the information to the processor 1180, and can receive the processing The command sent by the device 1180 is executed. In addition, the touch panel 1142 can be implemented by using various types such as resistive, capacitive, infrared, and surface acoustic waves, and the touch panel 1142 can be implemented by any technology developed in the future. Further, the touch panel 1142 can cover the display panel 1141, and the user can display the content according to the display panel 1141 (the display content includes, but is not limited to, a soft keyboard, a virtual mouse, a virtual button, an icon, etc.) on the display panel 1141. Operation is performed on or near the covered touch panel 1142. After detecting the operation on or near the touch panel 1142, the touch panel 1142 transmits to the processor 1180 through the I/O subsystem 1170 to determine user input, and then the processor 1180 is based on the user. The input provides a corresponding visual output on display panel 1141 via I/O subsystem 1170. Although in FIG. 11 , the touch panel 1142 and the display panel 1141 are two independent components to implement the input and input functions of the terminal device 1100 , in some embodiments, the touch panel 1142 and the display panel 1141 may be The input and output functions of the terminal device 1100 are implemented integrated.
终端设备1100还可包括至少一种传感器1150,比如光传感器、运动传感器以及其他传感器。具体地,光传感器可包括环境光传感器及接近传感器,其中,环境光传感器可根据环境光线的明暗来调节显示面板1141的亮度,接近传感器可在终端设备1100移动到耳边时,关闭显示面板1141和/或背光。作为运动传感器的一种,加速计传感器可检测各个方向上(一般为三轴)加速度的大小,静止时可检测出重力的大小及方向,可用于识别手机姿态的应用(比如横竖屏切换、相关游戏、磁力计姿态校准)、振动识别相关功能(比如计步器、敲击)等;至于终端设备1100还可配置的陀螺仪、气压计、湿度计、温度计、红外线传感器等其他传感器,在此不再赘述。The terminal device 1100 may also include at least one type of sensor 1150, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor may include an ambient light sensor and a proximity sensor, wherein the ambient light sensor may adjust the brightness of the display panel 1141 according to the brightness of the ambient light, and the proximity sensor may close the display panel 1141 when the terminal device 1100 moves to the ear. And / or backlight. As a kind of motion sensor, the accelerometer sensor can detect the magnitude of acceleration in all directions (usually three axes). When it is stationary, it can detect the magnitude and direction of gravity. It can be used to identify the gesture of the mobile phone (such as horizontal and vertical screen switching, related Game, magnetometer attitude calibration), vibration recognition related functions (such as pedometer, tapping), etc.; as for the terminal device 1100 can also be configured with gyroscopes, barometers, hygrometers, thermometers, infrared sensors and other sensors, here No longer.
音频电路1160、扬声器1161,麦克风1162可提供用户与终端设备1100之间的音频接口。音频电路1160可将接收到的音频数据转换后的信号,传输到扬声器1161,由扬声器1161转换为声音信号输出;另一方面,麦克风1162将收集的声音信号转换为信号,由音频电路1160接收后转换为音频数据,再将音频数据输出至RF电路1110以发送给比如另一手机,或者将音频数据输出至存储器1120以便进一步处理。An audio circuit 1160, a speaker 1161, and a microphone 1162 can provide an audio interface between the user and the terminal device 1100. The audio circuit 1160 can transmit the converted audio data to the speaker 1161, and convert it into a sound signal output by the speaker 1161; on the other hand, the microphone 1162 converts the collected sound signal into a signal, which is received by the audio circuit 1160. The audio data is converted, and the audio data is output to the RF circuit 1110 for transmission to, for example, another mobile phone, or the audio data is output to the memory 1120 for further processing.
I/O子系统1170用来控制输入输出的外部设备,可以包括其他设备输入控制器1171、传感器控制器1172、显示控制器1173。可选的,一个或多个其他输入控制设备控制器1171从其他输入设备1130接收信号和/或者向其他输入设备1130发送信号,其他输入设备1130可以包括物理按钮(按压按钮、摇臂按钮等)、拨号盘、滑动开关、操纵杆、点击滚轮、光鼠(光鼠是不显示可视输出的触摸敏感表面,或者是由触摸屏形成的触摸敏感表面的延伸)。值得说明的是,其他输入控制设备控制器1171可以与任一个或者多个上述设备连接。所述I/O子系统1170中的显示控制器1173从显示屏1140接收信号和/或者向显示屏1140发送信号。显示屏1140检测到用户输入后,显示控制器1173将检测到的用户输入转换为与显示在显示屏1140上的用户界面对象的交互,即实现人机交互。传感器控制器1172可以从一个或者多个传感器1150接收信号和/或者向一个或者多个传感器1150发送信号。The I/O subsystem 1170 is used to control external devices for input and output, and may include other device input controllers 1171, sensor controllers 1172, and display controllers 1173. Optionally, one or more other input control device controllers 1171 receive signals from other input devices 1130 and/or send signals to other input devices 1130, and other input devices 1130 may include physical buttons (press buttons, rocker buttons, etc.) , dial, slide switch, joystick, click wheel, light mouse (light mouse is a touch-sensitive surface that does not display visual output, or an extension of a touch-sensitive surface formed by a touch screen). It is worth noting that other input control device controllers 1171 can be connected to any one or more of the above devices. Display controller 1173 in I/O subsystem 1170 receives signals from display 1140 and/or transmits signals to display 1140. After the display 1140 detects the user input, the display controller 1173 converts the detected user input into an interaction with the user interface object displayed on the display 1140, ie, implements human-computer interaction. Sensor controller 1172 can receive signals from one or more sensors 1150 and/or send signals to one or more sensors 1150.
处理器1180是终端设备1100的控制中心,利用各种接口和线路连接整个终端设备的各个部分,通过运行或执行存储在存储器1120内的软件程序和/或模块,以及调用存储在存储器1120内的数据,执行终端设备1100的各种功能和处理数据,从而对终端设备进行整体监控。可选的,处理器1180可包括一个或多个处理单元;可选地,处理器1180 可集成应用处理器和调制解调处理器,其中,应用处理器主要处理操作系统、用户界面和应用程序等,调制解调处理器主要处理无线通信。可以理解的是,上述调制解调处理器也可以不集成到处理器1180中。The processor 1180 is a control center of the terminal device 1100 that connects various portions of the entire terminal device using various interfaces and lines, by running or executing software programs and/or modules stored in the memory 1120, and recalling stored in the memory 1120. The data performs various functions and processing data of the terminal device 1100, thereby performing overall monitoring of the terminal device. Optionally, the processor 1180 can include one or more processing units; optionally, the processor 1180 The application processor and the modem processor can be integrated, wherein the application processor mainly processes an operating system, a user interface, an application, etc., and the modem processor mainly processes wireless communication. It will be appreciated that the above described modem processor may also not be integrated into the processor 1180.
处理器1180用于:获取终端设备的运动姿态;通过深度摄像头和彩色摄像头分别采集深度图像和彩色图像;根据终端设备的运动姿态与所述深度图像进行快速分割匹配;根据所述彩色图像对所述快速分割匹配的结果进行精准匹配;如果获取的当前图像与已拍摄图像存在重叠,通过融合算法对重叠区域进行融合以生成动态三维图像。。The processor 1180 is configured to: acquire a motion posture of the terminal device; separately acquire a depth image and a color image by using the depth camera and the color camera; perform fast segmentation matching with the depth image according to the motion posture of the terminal device; The result of the fast segmentation matching is accurately matched; if the acquired current image overlaps with the captured image, the overlapping region is fused by the fusion algorithm to generate a dynamic three-dimensional image. .
终端设备1100还包括给各个部件供电的电源1190(比如电池),可选的,电源可以通过电源管理系统与处理器1180逻辑相连,从而通过电源管理系统实现管理充电、放电、以及功耗等功能。The terminal device 1100 further includes a power source 1190 (such as a battery) for supplying power to the various components. Optionally, the power source can be logically connected to the processor 1180 through the power management system to manage functions such as charging, discharging, and power consumption through the power management system. .
尽管未示出,终端设备1100还可以包括摄像头(深度摄像头和彩色摄像头)、蓝牙模块等,在此不再赘述。Although not shown, the terminal device 1100 may further include a camera (a depth camera and a color camera), a Bluetooth module, and the like, and details are not described herein again.
应理解,该终端设备1100可对应于根据本申请实施例的动态三维图像获取方法中的终端设备,该终端设备1100可以包括用于执行上述方法中的终端设备或电子设备执行的方法的实体单元。并且,该终端设备1100中的各实体单元和上述其他操作和/或功能分别为了上述方法的相应流程,为了简洁,在此不再赘述。It should be understood that the terminal device 1100 may correspond to a terminal device in a dynamic three-dimensional image acquisition method according to an embodiment of the present application, and the terminal device 1100 may include a physical unit for performing a method performed by a terminal device or an electronic device in the above method. . In addition, the physical units in the terminal device 1100 and the other operations and/or functions described above are respectively used for the corresponding processes of the foregoing methods, and are not described herein for brevity.
还应理解,该终端设备1100可以包括用于执行上述动态三维图像获取的方法中的实体单元。并且,该终端设备1100中的各实体单元和上述其他操作和/或功能分别为了上述方法的相应流程,为了简洁,在此不再赘述。It should also be understood that the terminal device 1100 can include a physical unit in a method for performing the above-described dynamic three-dimensional image acquisition. In addition, the physical units in the terminal device 1100 and the other operations and/or functions described above are respectively used for the corresponding processes of the foregoing methods, and are not described herein for brevity.
还应理解,本申请实施例中的处理器可以是一种集成电路芯片,具有信号的处理能力。在实现过程中,上述方法实施例的各步骤可以通过处理器中的硬件的集成逻辑电路或者软件形式的指令完成。上述的处理器可以是中央处理单元(Central Processing Unit,CPU)、该处理器还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件。可以实现或者执行本申请实施例中的公开的各方法、步骤及逻辑框图。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。结合本申请实施例所公开的方法的步骤可以直接体现为硬件译码处理器执行完成,或者用译码处理器中的硬件及软件器组合执行完成。软件器可以位于随机存储器,闪存、只读存储器,可编程只读存储器或者电可擦写可编程存储器、寄存器等本领域成熟的存储介质中。该存储介质位于存储器,处理器读取存储器中的信息,结合其硬件完成上述方法的步骤。It should also be understood that the processor in the embodiment of the present application may be an integrated circuit chip with signal processing capability. In the implementation process, each step of the foregoing method embodiment may be completed by an integrated logic circuit of hardware in a processor or an instruction in a form of software. The processor may be a central processing unit (CPU), the processor may be another general-purpose processor, a digital signal processor (DSP), or an application specific integrated circuit (ASIC). ), Field Programmable Gate Array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components. The methods, steps, and logical block diagrams disclosed in the embodiments of the present application can be implemented or executed. The general purpose processor may be a microprocessor or the processor or any conventional processor or the like. The steps of the method disclosed in the embodiments of the present application may be directly implemented by the hardware decoding processor, or may be performed by a combination of hardware and software in the decoding processor. The software can be located in a random storage medium, such as a flash memory, a read only memory, a programmable read only memory or an electrically erasable programmable memory, a register, and the like. The storage medium is located in the memory, and the processor reads the information in the memory and combines the hardware to complete the steps of the above method.
还应理解,本申请实施例中的存储器可以是易失性存储器或非易失性存储器,或可包括易失性和非易失性存储器两者。其中,非易失性存储器可以是只读存储器(Read-Only Memory,ROM)、可编程只读存储器(Programmable ROM,PROM)、可擦除可编程只读存储器(Erasable PROM,EPROM)、电可擦除可编程只读存储器(Electrically EPROM,EEPROM)或闪存。易失性存储器可以是随机存取存储器(Random Access Memory,RAM),其用作外部高速缓存。通过示例性但不是限制性说明,许多形式的RAM可用,例如静态随机存取存储器(Static RAM,SRAM)、动态随机存取存储器(Dynamic RAM,DRAM)、同步动态随机存取存储器(Synchronous DRAM,SDRAM)、 双倍数据速率同步动态随机存取存储器(Double Data Rate SDRAM,DDR SDRAM)、增强型同步动态随机存取存储器(Enhanced SDRAM,ESDRAM)、同步连接动态随机存取存储器(Synchlink DRAM,SLDRAM)和直接内存总线随机存取存储器(Direct Rambus RAM,DR RAM)。应注意,本文描述的系统和方法的存储器旨在包括但不限于这些和任意其它适合类型的存储器。It should also be understood that the memory in the embodiments of the present application may be a volatile memory or a non-volatile memory, or may include both volatile and non-volatile memory. The non-volatile memory may be a read-only memory (ROM), a programmable read only memory (PROM), an erasable programmable read only memory (Erasable PROM, EPROM), or an electric Erase programmable read only memory (EEPROM) or flash memory. The volatile memory can be a Random Access Memory (RAM) that acts as an external cache. By way of example and not limitation, many forms of RAM are available, such as static random access memory (SRAM), dynamic random access memory (DRAM), synchronous dynamic random access memory (Synchronous DRAM). SDRAM), Double Data Rate Synchronous Dynamic Random Access Memory (DDR), Enhanced Synchronous Dynamic Random Access Memory (ESDRAM), Synchronous Connection Dynamic Random Access Memory (SDRAM) and Direct Memory Bus Random Access Memory (DR RAM). It should be noted that the memories of the systems and methods described herein are intended to comprise, without being limited to, these and any other suitable types of memory.
还应理解,该总线系统除包括数据总线之外,还可以包括电源总线、控制总线和状态信号总线等。但是为了清楚说明起见,在图中将各种总线都标为总线系统。It should also be understood that the bus system may include a power bus, a control bus, a status signal bus, and the like in addition to the data bus. However, for the sake of clarity, the various buses are labeled as bus systems in the figure.
还应理解,在本申请实施例中,“与A相应的B”表示B与A相关联,根据A可以确定B。但还应理解,根据A确定B并不意味着仅仅根据A确定B,还可以根据A和/或其它信息确定B。应理解,本文中术语“和/或”,仅仅是一种描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存10在A和B,单独存在B这三种情况。另外,本文中字符“/”,一般表示前后关联对象是一种“或”的关系。It should also be understood that in the embodiment of the present application, "B corresponding to A" means that B is associated with A, and B can be determined according to A. However, it should also be understood that determining B from A does not mean that B is only determined based on A, and that B can also be determined based on A and/or other information. It should be understood that the term "and/or" herein is merely an association relationship describing an associated object, indicating that there may be three relationships, for example, A and/or B, which may indicate that A exists separately while 10 is stored in A. And B, there are three cases of B alone. In addition, the character "/" in this article generally indicates that the contextual object is an "or" relationship.
在实现过程中,上述方法的各步骤可以通过处理器中的硬件的集成逻辑电路或者软件形式的指令完成。结合本申请实施例所公开的用于传输上行信号的方法的步骤可以直接体现为硬件处理器执行完成,或者用处理器中的硬件及软件器组合执行完成。软件器可以位于随机存储器,闪存、只读存储器,可编程只读存储器或者电可擦写可编程存储器、寄存器等本领域成熟的存储介质中。该存储介质位于存储器,处理器读取存储器中的信息,结合其硬件完成上述方法的步骤。为避免重复,这里不再详细描述。In the implementation process, each step of the above method may be completed by an integrated logic circuit of hardware in a processor or an instruction in a form of software. The steps of the method for transmitting an uplink signal disclosed in the embodiments of the present application may be directly implemented as hardware processor execution completion, or performed by a combination of hardware and software in a processor. The software can be located in a random storage medium, such as a flash memory, a read only memory, a programmable read only memory or an electrically erasable programmable memory, a register, and the like. The storage medium is located in the memory, and the processor reads the information in the memory and combines the hardware to complete the steps of the above method. To avoid repetition, it will not be described in detail here.
本申请实施例还提出了一种计算机可读存储介质,该计算机可读存储介质存储一个或多个程序,该一个或多个程序包括指令,该指令当被包括多个应用程序的便携式电子设备执行时,能够使该便携式电子设备执行图2和/或图3所示实施例的方法。The embodiment of the present application further provides a computer readable storage medium storing one or more programs, the one or more programs including instructions, when the portable electronic device is included in a plurality of applications When executed, the portable electronic device can be caused to perform the method of the embodiment shown in Figures 2 and/or 3.
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请实施例的范围。Those of ordinary skill in the art will appreciate that the elements and algorithm steps of the various examples described in connection with the embodiments disclosed herein can be implemented in electronic hardware or a combination of computer software and electronic hardware. Whether these functions are performed in hardware or software depends on the specific application and design constraints of the solution. A person skilled in the art can use different methods to implement the described functions for each particular application, but such implementation should not be considered to be beyond the scope of the embodiments of the present application.
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统、装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。A person skilled in the art can clearly understand that for the convenience and brevity of the description, the specific working process of the system, the device and the unit described above can refer to the corresponding process in the foregoing method embodiment, and details are not described herein again.
在本申请所提供的几个实施例中,应该理解到,所揭露的系统、装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。In the several embodiments provided by the present application, it should be understood that the disclosed systems, devices, and methods may be implemented in other manners. For example, the device embodiments described above are merely illustrative. For example, the division of the unit is only a logical function division. In actual implementation, there may be another division manner, for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed. In addition, the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
另外,在本申请实施例各个实施例中的各功能单元可以集成在一个处理单元中,也 可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。In addition, each functional unit in each embodiment of the embodiments of the present application may be integrated into one processing unit, It may be that each unit physically exists alone, or two or more units may be integrated into one unit.
所述功能如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请实施例的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本申请实施例各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。The functions may be stored in a computer readable storage medium if implemented in the form of a software functional unit and sold or used as a standalone product. Based on such understanding, the technical solution of the embodiments of the present application, or the part contributing to the prior art or the part of the technical solution, may be embodied in the form of a software product stored in a storage medium. The instructions include a plurality of instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to perform all or part of the steps of the methods described in the various embodiments of the embodiments of the present application. The foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like, which can store program codes. .
以上所述,仅为本申请实施例的具体实施方式,但本申请实施例的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请实施例揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请实施例的保护范围之内。因此,本申请实施例的保护范围应以所述权利要求的保护范围为准。 The foregoing is only a specific embodiment of the embodiments of the present application, but the scope of protection of the embodiments of the present application is not limited thereto, and any person skilled in the art can easily adopt the technical scope disclosed in the embodiments of the present application. All changes or substitutions are contemplated to be within the scope of the embodiments of the present application. Therefore, the scope of protection of the embodiments of the present application is subject to the scope of protection of the claims.

Claims (11)

  1. 一种动态三维图像获取的方法,其特征在于,所述方法应用于终端设备,所述方法包括:A method for acquiring a dynamic three-dimensional image, wherein the method is applied to a terminal device, and the method includes:
    获取终端设备的运动姿态;Obtaining a motion posture of the terminal device;
    通过深度摄像头和彩色摄像头分别采集深度图像和彩色图像;Collecting depth images and color images by a depth camera and a color camera, respectively;
    根据终端设备的运动姿态与所述深度图像进行快速分割匹配;Performing fast segmentation matching with the depth image according to the motion posture of the terminal device;
    根据所述彩色图像对所述快速分割匹配的结果进行精准匹配;Performing exact matching on the result of the fast segmentation matching according to the color image;
    如果获取的当前图像与已拍摄图像存在重叠,通过融合算法对重叠区域进行融合以生成动态三维图像。If the acquired current image overlaps with the captured image, the overlapping region is fused by the fusion algorithm to generate a dynamic three-dimensional image.
  2. 如权利要求1所述的方法,其特征在于,所述获取终端设备的运动姿态,包括:The method of claim 1, wherein the obtaining the motion gesture of the terminal device comprises:
    通过终端设备的加速度计、陀螺仪和电子罗盘获取终端设备的运动姿态。The motion posture of the terminal device is obtained by the accelerometer, the gyroscope and the electronic compass of the terminal device.
  3. 如权利要求1或2所述的方法,其特征在于,所述根据终端设备的运动姿态与所述深度图像进行快速分割匹配,包括:The method according to claim 1 or 2, wherein the fast segmentation matching with the depth image according to the motion posture of the terminal device comprises:
    当根据终端设备的运动姿态确定所述终端设备平滑移动时,获取第一时间段内起始时间点与结束时间点对应的深度图;When determining that the terminal device moves smoothly according to the motion posture of the terminal device, acquiring a depth map corresponding to the start time point and the end time point in the first time period;
    基于起始时间点的深度图的已完成分割的特征区域和所述终端设备的姿态变化计算结束时间点的深度图的特征区域范围。The feature region of the depth map of the end time point is calculated based on the completed segmented feature region of the depth map of the start time point and the pose change of the terminal device.
  4. 如权利要求3所述的方法,其特征在于,所述根据所述彩色图像对所述快速分割匹配的结果进行精准匹配,包括:The method according to claim 3, wherein the accurately matching the results of the fast segmentation matching according to the color image comprises:
    根据所述彩色图像对所述特征区域范围执行补偿优化,获取细粒度的图片特征描述。Performing compensation optimization on the feature region range according to the color image to obtain a fine-grained picture feature description.
  5. 如权利要求1至4中任一项所述的方法,其特征在于,所述通过融合算法对重叠区域进行融合以生成动态三维图像,包括:The method according to any one of claims 1 to 4, wherein the merging the overlapping regions by the fusion algorithm to generate a dynamic three-dimensional image comprises:
    实时计算终端设备的当前姿态,从历史特征矩阵中提取可比较的标志性数据进行匹配;Calculating the current posture of the terminal device in real time, and extracting comparable iconic data from the historical feature matrix for matching;
    如果匹配结果表明当前图像帧与历史某图像帧有重叠的特征区域则进行融合处理,更新重叠区域。If the matching result indicates that the current image frame overlaps with the historical image frame, the fusion processing is performed to update the overlapping region.
  6. 一种终端设备,其特征在于,包括:A terminal device, comprising:
    获取单元,用于获取终端设备的运动姿态;An acquiring unit, configured to acquire a motion posture of the terminal device;
    采集单元,用于通过深度摄像头和彩色摄像头分别采集深度图像和彩色图像;a collecting unit, configured to separately collect a depth image and a color image by using a depth camera and a color camera;
    处理单元,用于根据终端设备的运动姿态与所述深度图像进行快速分割匹配;根据所述彩色图像对所述快速分割匹配的结果进行精准匹配;如果获取的当前图像与已拍摄图像存在重叠,通过融合算法对重叠区域进行融合以生成动态三维图像。a processing unit, configured to perform fast segmentation matching with the depth image according to the motion posture of the terminal device; and accurately match the result of the fast segmentation matching according to the color image; if the acquired current image overlaps with the captured image, The overlapping regions are fused by a fusion algorithm to generate a dynamic three-dimensional image.
  7. 如权利要求6所述的终端设备,其特征在于,所述获取单元具体用于:The terminal device according to claim 6, wherein the obtaining unit is specifically configured to:
    通过终端设备的加速度计、陀螺仪和电子罗盘获取终端设备的运动姿态。The motion posture of the terminal device is obtained by the accelerometer, the gyroscope and the electronic compass of the terminal device.
  8. 如权利要求6或7所述的终端设备,其特征在于,所述处理单元具体用于:The terminal device according to claim 6 or 7, wherein the processing unit is specifically configured to:
    当根据终端设备的运动姿态确定所述终端设备平滑移动时,获取第一时间段内起始时间点与结束时间点对应的深度图;When determining that the terminal device moves smoothly according to the motion posture of the terminal device, acquiring a depth map corresponding to the start time point and the end time point in the first time period;
    基于起始时间点的深度图的已完成分割的特征区域和所述终端设备的姿态变化计算结束时间点的深度图的特征区域范围。The feature region of the depth map of the end time point is calculated based on the completed segmented feature region of the depth map of the start time point and the pose change of the terminal device.
  9. 如权利要求8所述的终端设备,其特征在于,所述处理单元具体用于: The terminal device according to claim 8, wherein the processing unit is specifically configured to:
    根据所述彩色图像对所述特征区域范围执行补偿优化,获取细粒度的图片特征描述。Performing compensation optimization on the feature region range according to the color image to obtain a fine-grained picture feature description.
  10. 如权利要求6至9任一所述的终端设备,其特征在于,所述处理单元具体用于:The terminal device according to any one of claims 6 to 9, wherein the processing unit is specifically configured to:
    实时计算终端设备的当前姿态,从历史特征矩阵中提取可比较的标志性数据进行匹配;Calculating the current posture of the terminal device in real time, and extracting comparable iconic data from the historical feature matrix for matching;
    如果匹配结果表明当前图像帧与历史某图像帧有重叠的特征区域则进行融合处理,更新重叠区域。If the matching result indicates that the current image frame overlaps with the historical image frame, the fusion processing is performed to update the overlapping region.
  11. 一种终端设备,其特征在于,包括:存储器、处理器以及显示器;A terminal device, comprising: a memory, a processor, and a display;
    存储器,用于存储程序;Memory for storing programs;
    所述处理器,用于执行所述存储器存储的所述程序,当所述程序被执行时,所述处理器用于执行如权利要求1-5中任意一项所述的方法。 The processor is configured to execute the program stored by the memory, and the processor is configured to perform the method of any one of claims 1-5 when the program is executed.
PCT/CN2017/088162 2016-12-12 2017-06-13 Method and device for acquiring dynamic three-dimensional image WO2018107679A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201780076051.4A CN110169056B (en) 2016-12-12 2017-06-13 Method and equipment for acquiring dynamic three-dimensional image

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201611142062 2016-12-12
CN201611142062.1 2016-12-12

Publications (1)

Publication Number Publication Date
WO2018107679A1 true WO2018107679A1 (en) 2018-06-21

Family

ID=62557853

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/088162 WO2018107679A1 (en) 2016-12-12 2017-06-13 Method and device for acquiring dynamic three-dimensional image

Country Status (2)

Country Link
CN (2) CN112132881A (en)
WO (1) WO2018107679A1 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109118581A (en) * 2018-08-22 2019-01-01 Oppo广东移动通信有限公司 Image processing method and device, electronic equipment, computer readable storage medium
CN109272576A (en) * 2018-09-30 2019-01-25 Oppo广东移动通信有限公司 A kind of data processing method, MEC server, terminal device and device
CN109583411A (en) * 2018-12-09 2019-04-05 大连海事大学 The online checking method of tourist's classification based on TOF camera
CN109685042A (en) * 2019-02-03 2019-04-26 同方威视技术股份有限公司 A kind of 3-D image identification device and its recognition methods
CN111145100A (en) * 2018-11-02 2020-05-12 深圳富泰宏精密工业有限公司 Dynamic image generation method and system, computer device and readable storage medium
CN111695459A (en) * 2020-05-28 2020-09-22 腾讯科技(深圳)有限公司 State information prompting method and related equipment
CN111739146A (en) * 2019-03-25 2020-10-02 华为技术有限公司 Object three-dimensional model reconstruction method and device
CN112382374A (en) * 2020-11-25 2021-02-19 华南理工大学 Tumor segmentation device and segmentation method
CN112710250A (en) * 2020-11-23 2021-04-27 武汉光谷卓越科技股份有限公司 Three-dimensional measurement method based on line structured light and sensor
CN112766066A (en) * 2020-12-31 2021-05-07 北京小白世纪网络科技有限公司 Method and system for processing and displaying dynamic video stream and static image
CN113490054A (en) * 2021-07-01 2021-10-08 网易(杭州)网络有限公司 Virtual role control method, device, equipment and storage medium
CN113743237A (en) * 2021-08-11 2021-12-03 北京奇艺世纪科技有限公司 Follow-up action accuracy determination method and device, electronic device and storage medium
CN114302214A (en) * 2021-01-18 2022-04-08 海信视像科技股份有限公司 Virtual reality equipment and anti-jitter screen recording method
CN114763994A (en) * 2021-05-06 2022-07-19 苏州精源创智能科技有限公司 Inertial attitude navigation system applied to floor sweeping robot

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112132881A (en) * 2016-12-12 2020-12-25 华为技术有限公司 Method and equipment for acquiring dynamic three-dimensional image
CN110933275B (en) * 2019-12-09 2021-07-23 Oppo广东移动通信有限公司 Photographing method and related equipment
CN111045518B (en) * 2019-12-09 2023-06-30 上海瑾盛通信科技有限公司 Method and related device for acquiring attitude data
CN111175248B (en) * 2020-01-22 2021-03-30 中国农业科学院农产品加工研究所 Intelligent meat quality online detection method and detection system
CN111583317B (en) * 2020-04-29 2024-02-09 深圳市优必选科技股份有限公司 Image alignment method and device and terminal equipment
CN112261303B (en) * 2020-11-19 2021-08-20 贝壳技术有限公司 Three-dimensional color panoramic model generation device and method, storage medium and processor
CN113658229B (en) * 2021-08-13 2024-02-02 杭州华橙软件技术有限公司 Method and device for determining abnormal object, storage medium and electronic device
CN114283195B (en) * 2022-03-03 2022-07-26 荣耀终端有限公司 Method for generating dynamic image, electronic device and readable storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101341512A (en) * 2005-11-22 2009-01-07 索尼爱立信移动通讯股份有限公司 Method for obtaining enhanced photography and device therefor
CN101577795A (en) * 2009-06-17 2009-11-11 深圳华为通信技术有限公司 Method and device for realizing real-time viewing of panoramic picture
CN104519340A (en) * 2014-12-30 2015-04-15 余俊池 Panoramic video stitching method based on multi-depth image transformation matrix
US20150281678A1 (en) * 2014-03-25 2015-10-01 Samsung Electronics Co., Ltd. Image generating device, 3d image display system having the same and control methods thereof

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102420985B (en) * 2011-11-29 2014-01-22 宁波大学 Multi-view video object extraction method
CN102761765B (en) * 2012-07-16 2014-08-20 清华大学 Deep and repaid frame inserting method for three-dimensional video
US20140104394A1 (en) * 2012-10-15 2014-04-17 Intel Corporation System and method for combining data from multiple depth cameras
CN103400409B (en) * 2013-08-27 2016-08-10 华中师范大学 A kind of coverage 3D method for visualizing based on photographic head attitude Fast estimation
CN103559737A (en) * 2013-11-12 2014-02-05 中国科学院自动化研究所 Object panorama modeling method
CN103796001B (en) * 2014-01-10 2015-07-29 深圳奥比中光科技有限公司 A kind of method of synchronous acquisition degree of depth and color information and device
CN105282375B (en) * 2014-07-24 2019-12-31 钰立微电子股份有限公司 Attached stereo scanning module
CN104517289B (en) * 2014-12-12 2017-08-08 浙江大学 A kind of indoor scene localization method based on hybrid camera
CN104794722A (en) * 2015-04-30 2015-07-22 浙江大学 Dressed human body three-dimensional bare body model calculation method through single Kinect
JP6570327B2 (en) * 2015-06-05 2019-09-04 キヤノン株式会社 Control device, imaging device, control method, program, and storage medium
CN105225269B (en) * 2015-09-22 2018-08-17 浙江大学 Object modelling system based on motion
CN106203390B (en) * 2016-07-22 2019-09-24 杭州视氪科技有限公司 A kind of intelligent blind auxiliary system
CN112132881A (en) * 2016-12-12 2020-12-25 华为技术有限公司 Method and equipment for acquiring dynamic three-dimensional image

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101341512A (en) * 2005-11-22 2009-01-07 索尼爱立信移动通讯股份有限公司 Method for obtaining enhanced photography and device therefor
CN101577795A (en) * 2009-06-17 2009-11-11 深圳华为通信技术有限公司 Method and device for realizing real-time viewing of panoramic picture
US20150281678A1 (en) * 2014-03-25 2015-10-01 Samsung Electronics Co., Ltd. Image generating device, 3d image display system having the same and control methods thereof
CN104519340A (en) * 2014-12-30 2015-04-15 余俊池 Panoramic video stitching method based on multi-depth image transformation matrix

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109118581A (en) * 2018-08-22 2019-01-01 Oppo广东移动通信有限公司 Image processing method and device, electronic equipment, computer readable storage medium
CN109118581B (en) * 2018-08-22 2023-04-11 Oppo广东移动通信有限公司 Image processing method and device, electronic equipment and computer readable storage medium
CN109272576A (en) * 2018-09-30 2019-01-25 Oppo广东移动通信有限公司 A kind of data processing method, MEC server, terminal device and device
CN109272576B (en) * 2018-09-30 2023-03-24 Oppo广东移动通信有限公司 Data processing method, MEC server, terminal equipment and device
CN111145100A (en) * 2018-11-02 2020-05-12 深圳富泰宏精密工业有限公司 Dynamic image generation method and system, computer device and readable storage medium
CN111145100B (en) * 2018-11-02 2023-01-20 深圳富泰宏精密工业有限公司 Dynamic image generation method and system, computer device and readable storage medium
CN109583411B (en) * 2018-12-09 2022-10-21 大连海事大学 TOF camera-based tourist category online auditing method
CN109583411A (en) * 2018-12-09 2019-04-05 大连海事大学 The online checking method of tourist's classification based on TOF camera
CN109685042A (en) * 2019-02-03 2019-04-26 同方威视技术股份有限公司 A kind of 3-D image identification device and its recognition methods
CN111739146A (en) * 2019-03-25 2020-10-02 华为技术有限公司 Object three-dimensional model reconstruction method and device
CN111695459A (en) * 2020-05-28 2020-09-22 腾讯科技(深圳)有限公司 State information prompting method and related equipment
CN111695459B (en) * 2020-05-28 2023-04-18 腾讯科技(深圳)有限公司 State information prompting method and related equipment
CN112710250A (en) * 2020-11-23 2021-04-27 武汉光谷卓越科技股份有限公司 Three-dimensional measurement method based on line structured light and sensor
CN112382374A (en) * 2020-11-25 2021-02-19 华南理工大学 Tumor segmentation device and segmentation method
CN112382374B (en) * 2020-11-25 2024-04-12 华南理工大学 Tumor segmentation device and segmentation method
CN112766066A (en) * 2020-12-31 2021-05-07 北京小白世纪网络科技有限公司 Method and system for processing and displaying dynamic video stream and static image
CN114302214A (en) * 2021-01-18 2022-04-08 海信视像科技股份有限公司 Virtual reality equipment and anti-jitter screen recording method
CN114763994A (en) * 2021-05-06 2022-07-19 苏州精源创智能科技有限公司 Inertial attitude navigation system applied to floor sweeping robot
CN114763994B (en) * 2021-05-06 2024-01-30 苏州精源创智能科技有限公司 Inertial attitude navigation system applied to sweeping robot
CN113490054A (en) * 2021-07-01 2021-10-08 网易(杭州)网络有限公司 Virtual role control method, device, equipment and storage medium
CN113743237A (en) * 2021-08-11 2021-12-03 北京奇艺世纪科技有限公司 Follow-up action accuracy determination method and device, electronic device and storage medium
CN113743237B (en) * 2021-08-11 2023-06-02 北京奇艺世纪科技有限公司 Method and device for judging accuracy of follow-up action, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN110169056A (en) 2019-08-23
CN110169056B (en) 2020-09-04
CN112132881A (en) 2020-12-25

Similar Documents

Publication Publication Date Title
WO2018107679A1 (en) Method and device for acquiring dynamic three-dimensional image
US11158083B2 (en) Position and attitude determining method and apparatus, smart device, and storage medium
US11798190B2 (en) Position and pose determining method, apparatus, smart device, and storage medium
CN108682038B (en) Pose determination method, pose determination device and storage medium
CN110555883B (en) Repositioning method and device for camera attitude tracking process and storage medium
WO2019223468A1 (en) Camera orientation tracking method and apparatus, device, and system
US9417689B1 (en) Robust device motion detection
KR102114377B1 (en) Method for previewing images captured by electronic device and the electronic device therefor
CN109101120B (en) Method and device for displaying image
CN110148178B (en) Camera positioning method, device, terminal and storage medium
US20140253592A1 (en) Method for providing augmented reality, machine-readable storage medium, and portable terminal
CN110986930B (en) Equipment positioning method and device, electronic equipment and storage medium
WO2020192535A1 (en) Distance measurement method and electronic device
JP2022511427A (en) How to determine motion information of image feature points, task execution method and device
CN108776822B (en) Target area detection method, device, terminal and storage medium
CN111768454A (en) Pose determination method, device, equipment and storage medium
CN111724412A (en) Method and device for determining motion trail and computer storage medium
CN111862148B (en) Method, device, electronic equipment and medium for realizing visual tracking
CN113253908A (en) Key function execution method, device, equipment and storage medium
CN111928861B (en) Map construction method and device
CN108196701B (en) Method and device for determining posture and VR equipment
CN110633336B (en) Method and device for determining laser data search range and storage medium
WO2021218926A1 (en) Image display method and apparatus, and computer device
KR102084161B1 (en) Electro device for correcting image and method for controlling thereof
CN112861565A (en) Method and device for determining track similarity, computer equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17881176

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17881176

Country of ref document: EP

Kind code of ref document: A1