WO2019127192A1 - Image processing method and apparatus - Google Patents

Image processing method and apparatus Download PDF

Info

Publication number
WO2019127192A1
WO2019127192A1 PCT/CN2017/119291 CN2017119291W WO2019127192A1 WO 2019127192 A1 WO2019127192 A1 WO 2019127192A1 CN 2017119291 W CN2017119291 W CN 2017119291W WO 2019127192 A1 WO2019127192 A1 WO 2019127192A1
Authority
WO
WIPO (PCT)
Prior art keywords
living body
target
determining
joint
image
Prior art date
Application number
PCT/CN2017/119291
Other languages
French (fr)
Chinese (zh)
Inventor
周游
朱振宇
刘洁
Original Assignee
深圳市大疆创新科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市大疆创新科技有限公司 filed Critical 深圳市大疆创新科技有限公司
Priority to PCT/CN2017/119291 priority Critical patent/WO2019127192A1/en
Priority to CN201780022779.9A priority patent/CN109074661A/en
Publication of WO2019127192A1 publication Critical patent/WO2019127192A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Definitions

  • Embodiments of the present application relate to the field of image processing, and more particularly, to an image processing method and apparatus.
  • Computer vision relies on the imaging system instead of the visual organ as an input sensitive means.
  • the most common is the camera, which can form a basic vision system by a dual camera.
  • the binocular camera system can capture two photos at the same moment and at different angles through two cameras, and then through the difference between the two photos and the position and angle relationship between the two cameras, the triangle relationship can be used to calculate the scene and The distance map of the camera, that is, the depth map can be obtained. In the final analysis, the binocular camera system obtains the depth information of the scene by the difference of two photos at different angles at the same time.
  • the embodiment of the present application provides an image processing method and device, which can fully consider the features of the limbs or joints of the living body in a high dynamic scene with a living body, so that the depth map calculation of the living body is more accurate.
  • an image processing method including: determining a pointing vector of a target point on a target living body in an image to at least one joint, and determining a positional relationship between the target point and at least one pixel point; a pointing vector, and the positional relationship, adjusting a penalty coefficient of a global energy function of a Semi-Global Matching (SGM) algorithm; based on a disparity of the at least one pixel, using the penalty coefficient adjusted
  • SGM Semi-Global Matching
  • an image processing apparatus including a determining unit and a calculating unit; wherein the determining unit is configured to: determine a pointing vector of a target point on a target living body in the image to at least one joint, and determine a positional relationship between the target point and at least one pixel point; the calculating unit is configured to: adjust a penalty coefficient of a global energy function of the SGM algorithm according to the pointing vector, and the positional relationship; based on the at least one pixel point The parallax of the target point is calculated by using the global energy function adjusted by the penalty coefficient.
  • an image processing apparatus comprising a memory and a processor, the memory storing code, the processor being capable of calling code in the memory to perform a method of determining a target point on a target living body in the image to at least a pointing vector at a joint, and determining a positional relationship of the target point with at least one pixel; adjusting a penalty coefficient of a global energy function of the SGM algorithm according to the pointing vector, and the positional relationship; based on the at least one The parallax of the pixel is calculated by using the global energy function adjusted by the penalty coefficient to calculate the parallax of the target point.
  • a computer storage medium storing code for determining a pointing vector of a target point on a target living body in the image to at least one joint, and determining the target point and a positional relationship of at least one pixel; adjusting a penalty coefficient of a global energy function of the SGM algorithm according to the pointing vector, and the positional relationship; and adjusting a penalty coefficient based on the parallax of the at least one pixel
  • the global energy function calculates the parallax of the target point.
  • a computer program product comprising code for determining a pointing vector of a target point on a target living body in an image to at least one joint, and determining the target point a positional relationship with at least one pixel; adjusting a penalty coefficient of a global energy function of the SGM algorithm according to the pointing vector, and the positional relationship; and adjusting the penalty coefficient based on the parallax of the at least one pixel
  • the global energy function calculates the parallax of the target point.
  • a pointing vector of a target point on a target living body in an image to at least one joint is determined, and a positional relationship between the target point and at least one pixel point is determined; according to the pointing vector, And the positional relationship, adjusting a penalty coefficient of the global energy function of the SGM algorithm; calculating a parallax of the target point by using a global energy function adjusted by the penalty coefficient based on the parallax of the at least one pixel point,
  • the penalty coefficient in the semi-global matching algorithm is adjusted by fully considering the features of the limb or joint of the living body, and the fixed penalty coefficient is avoided, so that the depth map calculation of the living body is more accurate.
  • FIG. 1 is a schematic diagram of the disconnection of depth information in a high dynamic scenario.
  • FIG. 2 is a schematic diagram of an image processing method according to an embodiment of the present application.
  • FIG. 3 is a schematic diagram of segmenting a living body using a PAF algorithm according to an embodiment of the present application.
  • Figure 4 is a schematic view of the ground.
  • Figure 5 is a schematic illustration of a limb vector field.
  • Figure 6 is a schematic illustration of a limb vector field.
  • Figure 7 is an image of a thermography.
  • FIG. 8 is a schematic diagram of an image processing apparatus according to an embodiment of the present application.
  • FIG. 9 is a schematic diagram of an image processing apparatus according to an embodiment of the present application.
  • FIG. 10 is a schematic view of a drone according to an embodiment of the present application.
  • a component when a component is “fixedly connected” or “connected” to another component in the embodiment of the present application, or when one component is “fixed” to another component, it may be directly on another component, or There can be a centered component.
  • the route planning obstacle avoidance needs to use the 3D depth map, that is, the quality of the depth map directly affects the success or failure of the obstacle avoidance.
  • the embodiments of the present application provide the following solutions, and better depth information can be obtained.
  • the depth information of the embodiment of the present application may be used for the obstacle avoidance of the UAV, and may also be used for other scenarios, which is not specifically limited in this embodiment of the present application.
  • the embodiment of the present application may use the image captured by the binocular camera to calculate the depth information, and may also use the image captured by the monocular camera to calculate the depth information, which is not specifically limited in the embodiment of the present application.
  • the embodiments of the present application can be used for aerial vehicles or other vehicles with multiple cameras, such as unmanned cars, auto-flying drones, VR/AR glasses, dual-camera mobile phones, smart cars with visual systems, etc. device.
  • FIG. 2 is a schematic flowchart of an image processing method 100 according to an embodiment of the present application.
  • the method 100 includes at least a portion of the following.
  • a pointing vector of a target point on the target living body in the image to at least one joint is determined, and a positional relationship of the target point with the at least one pixel point is determined.
  • each limb mentioned in the embodiment of the present application may be divided by a joint.
  • the limb mentioned in the embodiment of the present application may include a head, a hand, an upper arm, a lower arm, a thigh, and a lower leg. Department and so on.
  • the living body mentioned in the embodiment of the present application may be a person, and of course, other living bodies, such as cats, dogs, elephants, and birds.
  • the target living body may be segmented from the image in advance. Further, a pointing vector of the target point on the target living body to the at least one joint is determined, and a positional relationship between the target point and the at least one pixel point is determined.
  • the target point refers to a target pixel point
  • the at least one pixel point may be an adjacent pixel point of the target pixel point.
  • the limb joint of the living body can be determined; according to the vector field of the limb joint of the living body, the connection relationship of the limb joint of the living body is determined; according to the connection relationship of the limb joint of the living body, from the image Split the target living body.
  • the living body can be segmented by using Part Affinity Fields (PAFs).
  • PAFs Part Affinity Fields
  • a in FIG. 3 may be configured as a Part Confidence Maps as shown by b in FIG. 3, and further configured as a limb-associated field as shown by c in FIG. 3 ( Part Affinity Fields (PAF), so that the human body can be segmented according to the limb-associated field shown by c in FIG. 3, as shown by d in FIG.
  • PAF Part Affinity Fields
  • the joint portion of the human body for example, the wrist joint, the elbow joint, the shoulder joint, etc.
  • CNN Convolutional Neural Network
  • the body joint confidence map can be obtained for the human body k Value at position p It can be constructed as follows:
  • x j,k is the groundtruth position of the limb j of the human body k in the image
  • defines an extension of the peak, wherein the peak corresponds to each visible limb of each human body.
  • the limb-associated field refers to the pointing relationship between the limb joints, for example, the shoulder joint points to the elbow joint, and the elbow joint points to the wrist joint.
  • x j1,k and x j2,k are the true positions of the joints j 1 and j 2 of the limb c connecting the human body k
  • the limb-associated vector field of the point p can be defined according to the following formula 2:
  • the set of points on the limb c can be defined as points within the line segment within the distance threshold, which satisfy the following Equation 3 and Equation 4:
  • limb width ⁇ l is the pixel distance
  • I the length of the limb
  • v ⁇ is a vector perpendicular to v.
  • connection relationship of the limb joint of the living body can be determined; and the connection relationship of the limb joint of the living body is segmented from the image according to the connection relationship of the limb joint of the living body The target living body.
  • the pointing vector of the target point to the at least one joint may be determined according to the pointing relationship of the limb joint of the target living body.
  • a pointing vector at any point in the lower arm to the joint can be obtained.
  • the target living body before the target living body is segmented from the image according to the connection relationship of the limb joint of the living body, the target is initially segmented from the image by using thermal imaging.
  • FIG. 7 shows a human body image acquired by means of thermal imaging.
  • the target living body may be initially segmented without using thermal imaging.
  • the living body may be segmented based on a minimum spanning tree graph segmentation method. Specifically, the following operations may be included.
  • graph G there are n vertex v and m edges e.
  • Step 3 Starting from the block area S 0 , in the set S 0 , each vertex is a self-forming subset (each pixel is a separate area).
  • Step 5 Calculate S q by S q-1 as described below.
  • Step 6 After traversing the m connection edge lines, return the final result S m , which is S.
  • the penalty coefficient of the global energy function of the Semi-Global Matching (SGM) algorithm is adjusted.
  • the global energy function adjusted by the penalty coefficient is used to calculate the parallax of the target point.
  • the SGM in addition to using the SGM algorithm to calculate the parallax of the target point, other algorithms may be used to calculate the parallax of the target point.
  • the SGM is used as an example to illustrate how to use the SGM to calculate the target. The parallax of the point.
  • a disparity map can be formed by selecting the disparity of each pixel, and a global energy function related to the disparity map can be set to minimize the energy function to achieve optimal solution for each pixel.
  • the energy function can be as shown in the following formula 6:
  • D refers to the disparity map
  • E(D) is the energy function corresponding to the disparity map
  • p, q represents a certain pixel in the image
  • N p refers to the adjacent pixel of the pixel p
  • C(p, D p ) Refers to the cost of the pixel when the current pixel disparity is D p
  • P 1 is a penalty coefficient, which is applicable to those pixels in which the disparity value of the pixel adjacent to the pixel p differs by 1 from the disparity value of p
  • P 2 It is a penalty coefficient, which is applicable to those pixels in which the disparity values of pixels adjacent to the pixel p differ by more than one from the disparity values of p.
  • Equation 7 For each one-dimensional solution, Equation 7 below can be used:
  • r refers to a direction pointing to the current pixel p, which can be understood here as an adjacent pixel point of the pixel p in the direction.
  • L r (p,d) represents the minimum cost value when the parallax of the current pixel point p is d along the current direction.
  • the minimum value can be selected from the minimum of 4 possible candidate values:
  • the first possible value is the smallest value when the current pixel is equal to the previous pixel's disparity value.
  • the second and third types may be the smallest cost + penalty coefficient P 1 when the current pixel is 1 (more than 1 or less) from the previous pixel disparity value.
  • the fourth possibility is that the difference between the current pixel point and the previous pixel disparity value is greater than 1, and its minimum cost value + penalty coefficient P 2 .
  • the generation value of the current pixel point can also be subtracted from the minimum value of the previous pixel point when taking different disparity values. This is because L r (p,d) will grow with the right shift of the current pixel. To prevent the value from overflowing, you can keep it at a small value.
  • C(p,d) can be calculated by the following formulas 8 and 9:
  • the multiple values for example, the cumulative value of the value of the eight directions, and the disparity value with the smallest accumulated value may be selected as the final disparity value of the pixel, for example, Accumulate by Equation 10 below:
  • the penalty coefficient P 1 and the penalty coefficient P 2 can be set larger, so that the target pixel point and the adjacent pixel can be increased.
  • the pixel points adopt the same probability of parallax. If the difference between the target pixel point and the adjacent pixel point is expected to be large, the penalty coefficient P 2 can be set smaller, and the penalty coefficient P 1 can be set larger.
  • the probability that the large target pixel points and the adjacent pixel points adopt a larger difference of the parallax if the difference between the target pixel point and the adjacent pixel point is expected to be small, the penalty coefficient P 1 can be set smaller, and the penalty will be The coefficient P 2 is set larger, which increases the probability that the difference between the target pixel and the adjacent pixel takes a smaller parallax.
  • the ground shown in FIG. 5 can be used as an example for description.
  • the ground in the 2D image, the ground is in the up and down direction, the depth changes sequentially, and in the left and right direction, the depth is basically the same.
  • the penalty coefficient is adjusted according to an angle between the pointing vector and a vector corresponding to the positional relationship.
  • the modulus of the difference between the angle and the 90 degree is positively correlated with the penalty coefficient.
  • the image shown in Fig. 6 is a bevel in the direction of extension of the arm, that is, in the direction from the elbow joint to the wrist joint, the distance from the lens is gradually changed, and the depth is changed, analogous to Fig. 5
  • the algorithm tends to choose different parallaxes in the direction of the elbow joint to the wrist joint.
  • the direction perpendicular to the arm, perpendicular to the elbow joint to the wrist joint is basically at a distance, that is, the depth is substantially the same, analogous to the left and right direction of the ground in Figure 5, dynamically increasing the penalty parameters P 1 and P 2
  • the algorithm tends to choose similar or even the same parallax.
  • determining, according to the limb edge of the target living body, the at least one pixel point is on the target living body.
  • the method 100 of the embodiment of the present application can be used to calculate the parallax of the target point.
  • the target point is a pixel point at the edge of the limb, and the pixel used to calculate the depth information is a pixel point other than the target living body, a smaller penalty P 2 can be set, and a larger penalty coefficient is further set. P 1 , thus allowing the calculated parallax jump.
  • only the target living body segmented by the PAF method in the embodiment of the present application may be referred to, and the pixel point of the edge of the target living body is determined, and the positional relationship between the pixel based on the edge and the adjacent pixel point is determined. , adjusting the penalty coefficient, regardless of the pointing vector of the target point to at least one joint.
  • the depth information of the target living body may be calculated according to the disparity of the at least one target point.
  • the depth information of the living body can be calculated.
  • the parallax can be inversely proportional to the depth.
  • the depth can be calculated by the following formula 11:
  • d is the depth
  • b is the distance between the left and right cameras
  • f is the focal length of the camera
  • d p is the disparity
  • the first speed may be determined according to the depth of the target living body, where the direction of the first speed is a direction from the target living body to the unmanned driving device; according to the first speed And a second speed determining a control speed for controlling the unmanned device, the second speed being a speed input by the controller; and controlling the flight of the unmanned device according to the control speed.
  • the magnitude of the first velocity is inversely proportional to the depth.
  • the unmanned device may be a drone or a driverless car. The following description will be made by taking an unmanned aerial vehicle as an example.
  • the Repulsive Force Field is used to "bounce" the drone to achieve the obstacle bypass.
  • the repulsion field can be constructed by referring to the following universal gravitation formula 12, and the specific expression can refer to Equation 13.
  • D x is the depth information of the living body, and the depth of each pixel of the living body can be averaged.
  • the speed corresponding to the repulsive field points away from the direction of the living body, and the user-controlled drone has a speed.
  • the two speeds are superimposed by the vector to generate a new speed.
  • the final planned speed as the speed loop command of the control system, Eventually, the obstacles will be bypassed.
  • a pointing vector of a target point on a target living body in an image to at least one joint is determined, and a positional relationship between the target point and at least one pixel point is determined; according to the pointing vector, And the positional relationship, adjusting a penalty coefficient of the global energy function of the SGM algorithm; calculating a parallax of the target point by using a global energy function adjusted by the penalty coefficient based on the parallax of the at least one pixel point,
  • the penalty coefficient in the semi-global matching algorithm is adjusted by fully considering the features of the limb or joint of the living body, and the fixed penalty coefficient is avoided, so that the depth map calculation of the living body is more accurate.
  • a complete strategy is proposed for the low-altitude flying drone, which is used for detecting the human body and other animals, and is used for optimizing the calculation of the depth map, thereby obtaining more accurate. Obstacle observation can guarantee the safety of people, and can better realize the planning of bypass for unmanned obstacle avoidance routes.
  • FIG. 8 is a schematic block diagram of an image processing apparatus 200 according to an embodiment of the present application. As shown in FIG. 8, the device 200 includes a determining unit 210 and a calculating unit 220;
  • the determining unit 210 is configured to: determine a pointing vector of the target point on the target living body in the image to the at least one joint, and determine a positional relationship between the target point and the at least one pixel point;
  • the calculating unit 220 is configured to: adjust a penalty coefficient of a global energy function of the semi-global matching SGM algorithm according to the pointing vector and the position relationship;
  • a parallax of the target point is calculated based on a parallax of the at least one pixel point using a global energy function adjusted by the penalty coefficient.
  • the calculating unit 220 is further configured to:
  • the penalty coefficient is adjusted according to an angle between the pointing vector and a vector corresponding to the positional relationship.
  • a modulus of the difference between the angle and 90 degrees is positively correlated with the penalty coefficient.
  • the device 200 further includes a first dividing unit 230, configured to:
  • the target living body is segmented from the image according to a connection relationship of the limb joints of the living body.
  • the determining unit 210 is further configured to:
  • the determining unit 210 is further configured to:
  • the device 200 further includes a second dividing unit 240, configured to:
  • the target living body is initially segmented from the image by means of thermal imaging.
  • the device 200 further includes a control unit 250, configured to:
  • the unmanned device is controlled according to the control speed.
  • the magnitude of the first velocity is inversely proportional to the depth.
  • image processing device may implement corresponding operations in the method 100, and for brevity, no further details are provided herein.
  • FIG. 9 is a schematic block diagram of an image processing apparatus 400 according to an embodiment of the present application.
  • the image processing device 400 may include a plurality of different components that may be integrated circuits (ICs), or portions of integrated circuits, discrete electronic devices, or other suitable for use in a circuit board (such as a motherboard). Modules, or additional boards, may also be incorporated as part of a computer system.
  • ICs integrated circuits
  • circuit board such as a motherboard
  • the image processing device can include a processor 410 and a storage medium 420 coupled to the processor 410.
  • Processor 410 may include one or more general purpose processors, such as a central processing unit (CPU), or a processing device or the like.
  • the processor 410 may be a complex instruction set computing (CISC) microprocessor, a very long instruction word (VLIW) microprocessor, and implements micro-processing of multiple instruction set combinations.
  • the processor may also be one or more dedicated processors, such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), and a digital signal processor. , DSP).
  • ASIC application specific integrated circuit
  • FPGA field programmable gate array
  • DSP digital signal processor
  • Processor 410 can be in communication with storage medium 420.
  • the storage medium 420 may be a magnetic disk, an optical disk, a read only memory (ROM), a flash memory, or a phase change memory.
  • the storage medium 420 can store instructions stored by the processor and/or can cache some of the information stored from the external storage device.
  • the image processing apparatus may include a display controller and/or display device unit 430, a transceiver 440, a video input output unit 450, an audio input output unit 460, and other input and output units 470.
  • These components included in image processing device 400 may be interconnected by a bus or internal connection.
  • the transceiver 440 can be a wired transceiver or a wireless transceiver, such as a WIFI transceiver, a satellite transceiver, a Bluetooth transceiver, a wireless cellular telephone transceiver, or combinations thereof.
  • a wireless transceiver such as a WIFI transceiver, a satellite transceiver, a Bluetooth transceiver, a wireless cellular telephone transceiver, or combinations thereof.
  • the video input and output unit 450 may include an image processing subsystem such as a camera, including a photo sensor, a charge coupled device (CCD) or a complementary metal-oxide semiconductor (CMOS) light. Sensor for use in shooting functions.
  • an image processing subsystem such as a camera, including a photo sensor, a charge coupled device (CCD) or a complementary metal-oxide semiconductor (CMOS) light. Sensor for use in shooting functions.
  • CCD charge coupled device
  • CMOS complementary metal-oxide semiconductor
  • the audio input and output unit 460 may include a speaker, a microphone, an earpiece, and the like.
  • other input and output devices 470 may include a storage device, a universal serial bus (USB) port, a serial port, a parallel port, a printer, a network interface, and the like.
  • USB universal serial bus
  • the image processing device 400 can perform the operations shown in the method 100.
  • the image processing device 400 can perform the operations shown in the method 100.
  • details are not described herein again.
  • the image processing device 400 or 400 may be located in a mobile device.
  • the mobile device can be moved in any suitable environment, for example, in the air (eg, a fixed-wing aircraft, a rotorcraft, or an aircraft with neither a fixed wing nor a rotor), in water (eg, a ship or submarine), on land. (for example, a car or train), space (for example, a space plane, satellite or detector), and any combination of the above.
  • the mobile device can be an unmanned car, an auto-flying drone, a VR/AR glasses, a dual-camera mobile phone, a smart car with a vision system, and the like.
  • FIG. 10 is a schematic block diagram of a removable device 500 in accordance with an embodiment of the present application.
  • the mobile device 500 includes a carrier 510 and a load 520.
  • the description of the mobile device in Figure 14 as a drone is for illustrative purposes only.
  • the load 520 may not be connected to the mobile device via the carrier 510.
  • the removable device 500 can also include a power system 530, a sensing system 540 and a communication system 550 and an image processing device 562 and a photographing system 564.
  • Power system 530 can include an electronic governor (referred to as an ESC), one or more propellers, and one or more electric machines corresponding to one or more propellers.
  • the motor and the propeller are disposed on the corresponding arm; the electronic governor is configured to receive a driving signal generated by the flight controller, and provide a driving current to the motor according to the driving signal to control the rotation speed and/or steering of the motor.
  • the motor is used to drive the propeller to rotate to power the UAV's flight, which enables the UAV to achieve one or more degrees of freedom of motion.
  • the UAV can be rotated about one or more axes of rotation.
  • the above-described rotating shaft may include a roll axis, a pan axis, and a pitch axis.
  • the motor can be a DC motor or an AC motor.
  • the motor can be a brushless motor or a brush motor.
  • the sensing system 540 is used to measure the attitude information of the UAV, that is, the position information and state information of the UAV in space, for example, three-dimensional position, three-dimensional angle, three-dimensional speed, three-dimensional acceleration, and three-dimensional angular velocity.
  • the sensing system may include, for example, a gyroscope, an electronic compass, an Inertial Measurement Unit ("IMU"), a vision sensor, a Global Positioning System (GPS), and a barometer. At least one of them.
  • the flight controller is used to control the flight of the UAV, for example, the UAV flight can be controlled based on the attitude information measured by the sensing system. It should be understood that the flight controller may control the UAV in accordance with pre-programmed program instructions, or may control the UAV in response to one or more control commands from the operating device.
  • Communication system 550 is capable of communicating with wireless terminal 590 via a terminal device 580 having communication system 570.
  • Communication system 550 and communication system 570 can include a plurality of transmitters, receivers, and/or transceivers for wireless communication.
  • the wireless communication herein may be one-way communication, for example, only the mobile device 500 may transmit data to the terminal device 580.
  • the wireless communication may be two-way communication, and the data may be transmitted from the mobile device 500 to the terminal device 580 or may be transmitted by the terminal device 580 to the mobile device 500.
  • terminal device 580 can provide control data for one or more of mobile device 500, carrier 510, and load 520, and can receive information transmitted by mobile device 500, carrier 510, and load 520.
  • the control data provided by terminal device 580 can be used to control the status of one or more of mobile device 500, carrier 510, and load 520.
  • a carrier 510 and load 520 include a communication module for communicating with the terminal device 580.
  • the image processing device 562 included in the mobile device shown in FIG. 10 can perform the method 100, which is not described herein for brevity.

Abstract

Embodiments of the present application provide an image processing method and apparatus. In a highly dynamic scenario with a human subject, the present application fully considers the features of limbs or joints of the human subject, thereby improving computation accuracy of a depth map of the human subject. The method comprises: determining a pointing vector pointing from a target point on a target human subject in an image to at least one joint, and determining a positional relationship between the target point and at least one pixel point; and calculating a parallax of the target point according to the pointing vector, the positional relationship, and a parallax of the at least one pixel point, wherein a penalty coefficient of a global energy function of an SGM algorithm is adjusted according to the pointing vector and the positional relationship, and the parallax of the target point is calculated on the basis of the parallax of the at least one pixel point by using the global energy function with the adjusted penalty coefficient.

Description

图像处理方法和设备Image processing method and device
版权申明Copyright statement
本专利文件披露的内容包含受版权保护的材料。该版权为版权所有人所有。版权所有人不反对任何人复制专利与商标局的官方记录和档案中所存在的该专利文件或者该专利披露。The disclosure of this patent document contains material that is subject to copyright protection. This copyright is the property of the copyright holder. The copyright owner has no objection to the reproduction of the patent document or the patent disclosure in the official records and files of the Patent and Trademark Office.
技术领域Technical field
本申请实施例涉及图像处理领域,并且更具体地,涉及一种图像处理方法和设备。Embodiments of the present application relate to the field of image processing, and more particularly, to an image processing method and apparatus.
背景技术Background technique
人类正在进入信息时代,计算机越来越广泛地进入几乎所有领域。作为智能计算的重要领域,计算机视觉得到了极大的开发应用。计算机视觉是依靠成像系统代替视觉器官作为输入敏感手段,最常用的是摄像头,由双摄像头即可组成一个基础的视觉系统。Human beings are entering the information age, and computers are increasingly entering almost all areas. As an important area of intelligent computing, computer vision has been greatly developed and applied. Computer vision relies on the imaging system instead of the visual organ as an input sensitive means. The most common is the camera, which can form a basic vision system by a dual camera.
双目摄像头系统可以通过两个摄像头,拍摄同一时刻,不同角度的两张照片,再通过两张照片的差异,以及双摄像头之间的位置、角度关系,利用三角关系,即可计算出场景与摄像头的距离关系,即可以得到深度图。归根结底,双目摄像头系统是通过同一时刻不同角度的两张照片的差异,来获取场景的深度信息。The binocular camera system can capture two photos at the same moment and at different angles through two cameras, and then through the difference between the two photos and the position and angle relationship between the two cameras, the triangle relationship can be used to calculate the scene and The distance map of the camera, that is, the depth map can be obtained. In the final analysis, the binocular camera system obtains the depth information of the scene by the difference of two photos at different angles at the same time.
但对于高动态场景,会有一些深度图无效的情况,即使有针对于前景的动态调节曝光策略,仍有一些情况下并不能很好的工作。However, for high dynamic scenes, there will be some cases where the depth map is invalid. Even if there is a dynamic adjustment exposure strategy for the foreground, there are still some cases that do not work well.
发明内容Summary of the invention
本申请实施例提供了一种图像处理方法和设备,可以实现在具有生命体的高动态场景下,充分考虑生命体的肢体或关节等特征,使得生命体的深度图计算的更加精准。The embodiment of the present application provides an image processing method and device, which can fully consider the features of the limbs or joints of the living body in a high dynamic scene with a living body, so that the depth map calculation of the living body is more accurate.
一方面,提供了一种图像处理方法,包括:确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定所述目标点与至少一个像素点的位置关系;根据所述指向向量,以及所述位置关系,调整半全局匹配(Semi-Global Matching,SGM)算法的全局能量函数的惩罚系数;基于所 述至少一个像素点的视差,利用调整了所述惩罚系数后的所全局能量函数,计算所述目标点的视差。In one aspect, an image processing method is provided, including: determining a pointing vector of a target point on a target living body in an image to at least one joint, and determining a positional relationship between the target point and at least one pixel point; a pointing vector, and the positional relationship, adjusting a penalty coefficient of a global energy function of a Semi-Global Matching (SGM) algorithm; based on a disparity of the at least one pixel, using the penalty coefficient adjusted The global energy function calculates the parallax of the target point.
另一方面,提供了一种图像处理设备,包括确定单元和计算单元;其中,所述确定单元用于:确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定所述目标点与至少一个像素点的位置关系;所述计算单元用于:根据所述指向向量,以及所述位置关系,调整SGM算法的全局能量函数的惩罚系数;基于所述至少一个像素点的视差,利用调整了所述惩罚系数后的所全局能量函数,计算所述目标点的视差。In another aspect, an image processing apparatus is provided, including a determining unit and a calculating unit; wherein the determining unit is configured to: determine a pointing vector of a target point on a target living body in the image to at least one joint, and determine a positional relationship between the target point and at least one pixel point; the calculating unit is configured to: adjust a penalty coefficient of a global energy function of the SGM algorithm according to the pointing vector, and the positional relationship; based on the at least one pixel point The parallax of the target point is calculated by using the global energy function adjusted by the penalty coefficient.
另一方面,提供了一种图像处理设备,包括存储器和处理器,该存储器存储有代码,该处理器可以调用存储器中的代码执行以下方法:确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定所述目标点与至少一个像素点的位置关系;根据所述指向向量,以及所述位置关系,调整SGM算法的全局能量函数的惩罚系数;基于所述至少一个像素点的视差,利用调整了所述惩罚系数后的所全局能量函数,计算所述目标点的视差。In another aspect, an image processing apparatus is provided, comprising a memory and a processor, the memory storing code, the processor being capable of calling code in the memory to perform a method of determining a target point on a target living body in the image to at least a pointing vector at a joint, and determining a positional relationship of the target point with at least one pixel; adjusting a penalty coefficient of a global energy function of the SGM algorithm according to the pointing vector, and the positional relationship; based on the at least one The parallax of the pixel is calculated by using the global energy function adjusted by the penalty coefficient to calculate the parallax of the target point.
另一方面,提供了一种计算机存储介质,该介质存储有代码,该代码可以用于确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定所述目标点与至少一个像素点的位置关系;根据所述指向向量,以及所述位置关系,调整SGM算法的全局能量函数的惩罚系数;基于所述至少一个像素点的视差,利用调整了所述惩罚系数后的所全局能量函数,计算所述目标点的视差。In another aspect, a computer storage medium is provided, the medium storing code for determining a pointing vector of a target point on a target living body in the image to at least one joint, and determining the target point and a positional relationship of at least one pixel; adjusting a penalty coefficient of a global energy function of the SGM algorithm according to the pointing vector, and the positional relationship; and adjusting a penalty coefficient based on the parallax of the at least one pixel The global energy function calculates the parallax of the target point.
另一方面,提供了一种计算机程序产品,该计算机程序产品包括代码,该代码可以用于确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定所述目标点与至少一个像素点的位置关系;根据所述指向向量,以及所述位置关系,调整SGM算法的全局能量函数的惩罚系数;基于所述至少一个像素点的视差,利用调整了所述惩罚系数后的所全局能量函数,计算所述目标点的视差。In another aspect, a computer program product is provided, the computer program product comprising code for determining a pointing vector of a target point on a target living body in an image to at least one joint, and determining the target point a positional relationship with at least one pixel; adjusting a penalty coefficient of a global energy function of the SGM algorithm according to the pointing vector, and the positional relationship; and adjusting the penalty coefficient based on the parallax of the at least one pixel The global energy function calculates the parallax of the target point.
因此,在本申请实施例中,确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定所述目标点与至少一个像素点的位置关系;根据所述指向向量,以及所述位置关系,调整SGM算法的全局能量函数的惩罚系数;基于所述至少一个像素点的视差,利用调整了所述惩罚系数后的所全局能量函数,计算所述目标点的视差,可以实现在具有生命体的高 动态场景下,充分考虑生命体的肢体或关节等特征来调整半全局匹配算法中的惩罚系数,避免采用固定的惩罚系数,使得生命体的深度图计算的更加精准。Therefore, in the embodiment of the present application, a pointing vector of a target point on a target living body in an image to at least one joint is determined, and a positional relationship between the target point and at least one pixel point is determined; according to the pointing vector, And the positional relationship, adjusting a penalty coefficient of the global energy function of the SGM algorithm; calculating a parallax of the target point by using a global energy function adjusted by the penalty coefficient based on the parallax of the at least one pixel point, In the high dynamic scene with living body, the penalty coefficient in the semi-global matching algorithm is adjusted by fully considering the features of the limb or joint of the living body, and the fixed penalty coefficient is avoided, so that the depth map calculation of the living body is more accurate.
附图说明DRAWINGS
为了更清楚地说明本申请实施例的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings to be used in the embodiments or the prior art description will be briefly described below. Obviously, the drawings in the following description are only some of the present application. For the embodiments, those skilled in the art can obtain other drawings according to the drawings without any creative work.
图1是在高动态场景下,深度信息出现断连的示意性图。FIG. 1 is a schematic diagram of the disconnection of depth information in a high dynamic scenario.
图2是根据本申请实施例的图像处理方法的示意性图。2 is a schematic diagram of an image processing method according to an embodiment of the present application.
图3是根据本申请实施例的利用PAF算法分割出生命体的示意性图。FIG. 3 is a schematic diagram of segmenting a living body using a PAF algorithm according to an embodiment of the present application.
图4是地面的示意性图。Figure 4 is a schematic view of the ground.
图5是肢体向量场的示意性图。Figure 5 is a schematic illustration of a limb vector field.
图6是肢体向量场的示意性图。Figure 6 is a schematic illustration of a limb vector field.
图7是热成像的图像。Figure 7 is an image of a thermography.
图8是根据本申请实施例的图像处理设备的示意性图。FIG. 8 is a schematic diagram of an image processing apparatus according to an embodiment of the present application.
图9是根据本申请实施例的图像处理设备的示意性图。FIG. 9 is a schematic diagram of an image processing apparatus according to an embodiment of the present application.
图10是根据本申请实施例的无人机的示意性图。FIG. 10 is a schematic view of a drone according to an embodiment of the present application.
具体实施方式Detailed ways
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The technical solutions in the embodiments of the present application are described in conjunction with the accompanying drawings in the embodiments of the present application. It is obvious that the described embodiments are a part of the embodiments of the present application, and not all of the embodiments. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of the present application without departing from the inventive scope are the scope of the present application.
需要说明的是,本申请实施例中当一组件与另一组件“固定连接”或“连接”,或者,一组件“固定于”另一组件时,它可以直接在另一组件上,或者也可以存在居中的组件。It should be noted that, when a component is "fixedly connected" or "connected" to another component in the embodiment of the present application, or when one component is "fixed" to another component, it may be directly on another component, or There can be a centered component.
除非另有说明,本申请实施例所使用的所有技术和科学术语与本申请的技术领域的技术人员通常理解的含义相同。本申请中所使用的术语只是为了 描述具体的实施例的目的,不是旨在限制本申请的范围。本申请所使用的术语“和/或”包括一个或多个相关的所列项的任意的和所有的组合。Unless otherwise indicated, all technical and scientific terms used in the embodiments of the present application have the same meaning The terminology used in the present application is for the purpose of describing particular embodiments and is not intended to limit the scope of the application. The term "and/or" used in this application includes any and all combinations of one or more of the associated listed.
对于生命体(例如,人)而言,由于处在一个高动态环境中,在获取深度信息时,则会出现一些深度图无效的情况,即使采用针对前景的动态调节曝光策略,仍存在一些情况出现深度图无效,例如,在场景切换的时候,动态曝光策略仍然需要收敛时间,则会造成深度图无效,例如,如图1针对人的手臂部分,有较多的无效深度信息,导致3D深度图上出现肢体断连的情况,其中,图1左侧是拍摄出的照片,右侧为深度信息。For living organisms (for example, people), due to the fact that in a highly dynamic environment, when the depth information is acquired, some depth maps are invalid. Even with the dynamic adjustment exposure strategy for the foreground, there are still some cases. The depth map is invalid. For example, when the scene is switched, the dynamic exposure strategy still needs convergence time, which will cause the depth map to be invalid. For example, as shown in Figure 1 for the human arm part, there is more invalid depth information, resulting in 3D depth. There are cases of limb disconnection on the map. Among them, the left side of Figure 1 is the photograph taken, and the right side is the depth information.
对于无人机而语言,航线规划避障需要使用3D深度图,即深度图的质量直接影响了避障的成败与效果。For the UAV and the language, the route planning obstacle avoidance needs to use the 3D depth map, that is, the quality of the depth map directly affects the success or failure of the obstacle avoidance.
因此,本申请实施例提供了了以下的方案,可以得到更优的深度信息。Therefore, the embodiments of the present application provide the following solutions, and better depth information can be obtained.
应理解,本申请实施例得到深度信息可以用于无人机进行避障,也可以用于其他场景,本申请实施例对此不作具体限定。It should be understood that the depth information of the embodiment of the present application may be used for the obstacle avoidance of the UAV, and may also be used for other scenarios, which is not specifically limited in this embodiment of the present application.
还应理解,本申请实施例可以采用双目摄像头拍摄的图像进行计算深度信息,也可以利用单目摄像头拍摄的图像计算深度信息,本申请实施例对此不作具体限定。It should be understood that the embodiment of the present application may use the image captured by the binocular camera to calculate the depth information, and may also use the image captured by the monocular camera to calculate the depth information, which is not specifically limited in the embodiment of the present application.
本申请实施例可以用于航拍飞行器或其他带有多摄像头的载具,例如无人驾驶的汽车、自动飞行的无人机、VR/AR眼镜、双摄像头的手机、有视觉系统的智能小车等设备。The embodiments of the present application can be used for aerial vehicles or other vehicles with multiple cameras, such as unmanned cars, auto-flying drones, VR/AR glasses, dual-camera mobile phones, smart cars with visual systems, etc. device.
图2是根据本申请实施例的图像处理方法100的示意性流程图。该方法100包括以下内容中的至少部分内容。FIG. 2 is a schematic flowchart of an image processing method 100 according to an embodiment of the present application. The method 100 includes at least a portion of the following.
在110中,确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定该目标点与至少一个像素点的位置关系。In 110, a pointing vector of a target point on the target living body in the image to at least one joint is determined, and a positional relationship of the target point with the at least one pixel point is determined.
可选地,本申请实施例提到的各个肢体之间可以通过关节来分割,例如,本申请实施例提到的肢体可以包括头部、手部、上臂、下臂、大腿部和小腿部等。Optionally, each limb mentioned in the embodiment of the present application may be divided by a joint. For example, the limb mentioned in the embodiment of the present application may include a head, a hand, an upper arm, a lower arm, a thigh, and a lower leg. Department and so on.
可选地,本申请实施例提到的生命体可以是人,当然,也可以是其他生命体,例如,猫,狗,大象和鸟类等。Optionally, the living body mentioned in the embodiment of the present application may be a person, and of course, other living bodies, such as cats, dogs, elephants, and birds.
可选地,在本申请实施例中,可以事先从图像中分割出目标生命体。进而确定目标生命体上的目标点向至少一个关节处的指向向量,以及确定该目标点与至少一个像素点的位置关系。其中,该目标点是指目标像素点,该至 少一个像素点可以是该目标像素点的相邻像素点。Optionally, in the embodiment of the present application, the target living body may be segmented from the image in advance. Further, a pointing vector of the target point on the target living body to the at least one joint is determined, and a positional relationship between the target point and the at least one pixel point is determined. Wherein, the target point refers to a target pixel point, and the at least one pixel point may be an adjacent pixel point of the target pixel point.
具体地,在该图像上,可以确定生命体的肢体关节;根据生命体的肢体关节的向量场,确定生命体的肢体关节的连接关系;根据生命体的肢体关节的连接关系,从该图像中分割出该目标生命体。Specifically, on the image, the limb joint of the living body can be determined; according to the vector field of the limb joint of the living body, the connection relationship of the limb joint of the living body is determined; according to the connection relationship of the limb joint of the living body, from the image Split the target living body.
具体地,可以采用肢体关联向量场(Part Affinity Fields,PAFs),分割出生命体。Specifically, the living body can be segmented by using Part Affinity Fields (PAFs).
具体地,可以将图3中的a,构造成如图3中的b所示的肢体关节置信图(Part Confidence Maps),以及进一步地构造成如图3中的c所示的肢体关联场(Part Affinity Fields,PAF),从而可以根据图3中的c所示的肢体关联场,对人体进行分割,具体可以如图3中的d所示。Specifically, a in FIG. 3 may be configured as a Part Confidence Maps as shown by b in FIG. 3, and further configured as a limb-associated field as shown by c in FIG. 3 ( Part Affinity Fields (PAF), so that the human body can be segmented according to the limb-associated field shown by c in FIG. 3, as shown by d in FIG.
在构造成如图3中的b所示的肢体关节置信图的过程中,可以通过卷积神经网络(Convolutional Neural Network,CNN)找到人体的关节部分,例如,腕关节,肘关节和肩关节等。其中,可以针对人体k得到肢体关节置信图
Figure PCTCN2017119291-appb-000001
在位置p处的值
Figure PCTCN2017119291-appb-000002
可以按照以下公式1构造:
In the process of constructing a joint articulation map of the limb as shown by b in FIG. 3, the joint portion of the human body, for example, the wrist joint, the elbow joint, the shoulder joint, etc., can be found by the Convolutional Neural Network (CNN). . Among them, the body joint confidence map can be obtained for the human body k
Figure PCTCN2017119291-appb-000001
Value at position p
Figure PCTCN2017119291-appb-000002
It can be constructed as follows:
Figure PCTCN2017119291-appb-000003
Figure PCTCN2017119291-appb-000003
其中,x j,k是图像中人体k的肢体j的真实位置(groundtruth position),σ限定峰值的扩展,其中,峰值对应到每个人体的每个可视的肢体。 Where x j,k is the groundtruth position of the limb j of the human body k in the image, and σ defines an extension of the peak, wherein the peak corresponds to each visible limb of each human body.
肢体关联场指的是肢体关节之间的指向关系,例如,肩关节指向肘关节,肘关节指向腕关节等。其中,如图4所示,x j1,k和x j2,k是连接人体k的肢体c的关节j 1和j 2的真实位置,点p的肢体关联向量场可以按照以下式2进行定义: The limb-associated field refers to the pointing relationship between the limb joints, for example, the shoulder joint points to the elbow joint, and the elbow joint points to the wrist joint. Wherein, as shown in FIG. 4, x j1,k and x j2,k are the true positions of the joints j 1 and j 2 of the limb c connecting the human body k, and the limb-associated vector field of the point p can be defined according to the following formula 2:
Figure PCTCN2017119291-appb-000004
Figure PCTCN2017119291-appb-000004
从以上式2可以看出,如果点p在肢体c上,则
Figure PCTCN2017119291-appb-000005
的值是从j1到j2的点的单位向量,如果不在肢体c上,则
Figure PCTCN2017119291-appb-000006
的值为0。其中,
Figure PCTCN2017119291-appb-000007
It can be seen from the above formula 2 that if the point p is on the limb c, then
Figure PCTCN2017119291-appb-000005
The value is the unit vector of the point from j1 to j2, if not on limb c, then
Figure PCTCN2017119291-appb-000006
The value is 0. among them,
Figure PCTCN2017119291-appb-000007
肢体c上点的集合可以定义为线段内在距离阈值内的点,这些点p满足以下公式3和公式4:The set of points on the limb c can be defined as points within the line segment within the distance threshold, which satisfy the following Equation 3 and Equation 4:
Figure PCTCN2017119291-appb-000008
Figure PCTCN2017119291-appb-000008
Figure PCTCN2017119291-appb-000009
Figure PCTCN2017119291-appb-000009
其中,肢体宽度σ l是像素距离,
Figure PCTCN2017119291-appb-000010
是肢体长度,v 是垂直于v的向量。
Where the limb width σ l is the pixel distance,
Figure PCTCN2017119291-appb-000010
Is the length of the limb, and v is a vector perpendicular to v.
因此,在按照上述介绍的方案或相似方案确定出生命体的肢体关节的向量场后,可以确定生命体的肢体关节的连接关系;根据生命体的肢体关节的连接关系,从该图像中分割出该目标生命体。Therefore, after determining the vector field of the limb joint of the living body according to the above-described scheme or the similar scheme, the connection relationship of the limb joint of the living body can be determined; and the connection relationship of the limb joint of the living body is segmented from the image according to the connection relationship of the limb joint of the living body The target living body.
可选地,在本申请实施例中,可以根据该目标生命体的肢体关节的指向关系,确定该目标点向至少一个关节处的指向向量。Optionally, in the embodiment of the present application, the pointing vector of the target point to the at least one joint may be determined according to the pointing relationship of the limb joint of the target living body.
具体地,如图4所示,在确定出了从肘关节向腕指向关系之后,可以得到该下臂中的任意一点向关节处的指向向量。Specifically, as shown in FIG. 4, after the relationship from the elbow joint to the wrist is determined, a pointing vector at any point in the lower arm to the joint can be obtained.
可选地,在本申请实施例中,在根据该生命体的肢体关节的连接关系,从该图像中分割出该目标生命体之前,采用热成像的方式,从该图像中初始分割出该目标生命体。其中,图7所示为热成像的方式获取的人体图像。当然,在本申请实施例,也可以不采用热成像的方式,初始分割出目标生命体。Optionally, in the embodiment of the present application, before the target living body is segmented from the image according to the connection relationship of the limb joint of the living body, the target is initially segmented from the image by using thermal imaging. Life form. Among them, FIG. 7 shows a human body image acquired by means of thermal imaging. Of course, in the embodiment of the present application, the target living body may be initially segmented without using thermal imaging.
可选地,可以基于最小生成树(minimum spanning tree)图分割法,分割出生命体。具体可以包括以下操作。Alternatively, the living body may be segmented based on a minimum spanning tree graph segmentation method. Specifically, the following operations may be included.
步骤1:先把图像(image)转换为曲线图(graph),得到graph G=(V,E),针对graph G,具有n个顶点v和m条边缘e。Step 1: First convert the image into a graph, and get a graph G=(V, E). For graph G, there are n vertex v and m edges e.
通过以下步骤得到的分割区域的集合S=(C 1,...,C r),其中C 1,...,C r为顶点组成的子集。 The set of divided regions obtained by the following steps S = (C 1 , ..., C r ), where C 1 , ..., Cr are a subset of the vertices.
步骤2:将边缘E的权重w按照非递减排列,得到集合π=(o 1,...,o m)。 Step 2: Arrange the weights w of the edges E in a non-decreasing manner to obtain a set π=(o 1 , . . . , o m ).
步骤3:从分块区域S 0开始,在集合S 0中,每个顶点都是自成子集(每个像素都是单独的区域)。 Step 3: Starting from the block area S 0 , in the set S 0 , each vertex is a self-forming subset (each pixel is a separate area).
步骤4:重复步骤4,对于S q,q=1,...,m。 Step 4: Repeat step 4 for S q , q=1,...,m.
步骤5:如下所述,通过S q-1计算得到S q。这里用v i和v j表示第q条边缘连接的两个顶点,记作o q=(v i,v j)。如果v i和v j在S q-1中是独立、尚未连接的两个顶点,且连线的边缘权重w(o q)小于一定的阈值
Figure PCTCN2017119291-appb-000011
则连接合并v i和v j所在的子集,否则保持不变。具体来说,S q-1中包含vi的区域子集记为
Figure PCTCN2017119291-appb-000012
而包含v j的记为
Figure PCTCN2017119291-appb-000013
如果
Figure PCTCN2017119291-appb-000014
Figure PCTCN2017119291-appb-000015
则合并子集
Figure PCTCN2017119291-appb-000016
Figure PCTCN2017119291-appb-000017
得到新的集合S q,否则S q=S q-1保持不变。
Step 5: Calculate S q by S q-1 as described below. Here, v i and v j are used to represent the two vertices of the q- th edge connection, denoted as o q = (v i , v j ). If v i and v j are independent and unconnected vertices in S q-1 , and the edge weight w(o q ) of the connection is less than a certain threshold
Figure PCTCN2017119291-appb-000011
Then join the subset where v i and v j are merged, otherwise it remains unchanged. Specifically, the subset of regions containing vi in S q-1 is recorded as
Figure PCTCN2017119291-appb-000012
And the record containing v j is
Figure PCTCN2017119291-appb-000013
in case
Figure PCTCN2017119291-appb-000014
And
Figure PCTCN2017119291-appb-000015
Merge subset
Figure PCTCN2017119291-appb-000016
versus
Figure PCTCN2017119291-appb-000017
A new set S q is obtained , otherwise S q =S q-1 remains unchanged.
其中,
Figure PCTCN2017119291-appb-000018
among them,
Figure PCTCN2017119291-appb-000018
其中,int定义为内差,即为子集C 1构成的最小小生成树MST中最大的权重,而权重定义为w(e)=w(v i,v j)=|I(p i)-I(p j)|。 Where int is defined as the inner difference, that is, the largest weight in the minimum small spanning tree MST composed of the subset C 1 , and the weight is defined as w(e)=w(v i , v j )=|I(p i ) -I(p j )|.
步骤6:遍历完m条连接边缘线后,返回最终结果S m,即为S。 Step 6: After traversing the m connection edge lines, return the final result S m , which is S.
在120中,根据该指向向量,以及该位置关系,调整半全局匹配(Semi-Global Matching,SGM)算法的全局能量函数的惩罚系数。In 120, according to the pointing vector and the positional relationship, the penalty coefficient of the global energy function of the Semi-Global Matching (SGM) algorithm is adjusted.
在130中,基于该至少一个像素点的视差,利用调整了该惩罚系数后的所全局能量函数,计算该目标点的视差。In 130, based on the parallax of the at least one pixel, the global energy function adjusted by the penalty coefficient is used to calculate the parallax of the target point.
应理解,在本申请实施例中,除了采用SGM算法计算目标点的视差外,也可以采用其他的算法,计算目标点的视差,本申请实施例以下将以SGM为例说明如何利用SGM计算目标点的视差。It should be understood that, in the embodiment of the present application, in addition to using the SGM algorithm to calculate the parallax of the target point, other algorithms may be used to calculate the parallax of the target point. In the following example, the SGM is used as an example to illustrate how to use the SGM to calculate the target. The parallax of the point.
为了便于更加清楚地理解本申请,以下将对SGM算法进行介绍。In order to facilitate a clearer understanding of the present application, the SGM algorithm will be described below.
可以通过选取每个像素点的视差,组成一个视差地图(disparity map),设置一个和视差地图(disparity map)相关的全局能量函数,使这个能量函数最小化,以达到求解每个像素点最优视差的目的。其中,能量函数可以如下式6所示:A disparity map can be formed by selecting the disparity of each pixel, and a global energy function related to the disparity map can be set to minimize the energy function to achieve optimal solution for each pixel. The purpose of parallax. Wherein, the energy function can be as shown in the following formula 6:
Figure PCTCN2017119291-appb-000019
Figure PCTCN2017119291-appb-000019
其中,D指视差地图,E(D)是该视差地图对应的能量函数;p,q代表图像中的某个像素点;N p指像素p的相邻像素点;C(p,D p)指当前像素点视差为D p时,该像素点的代价;P 1是一个惩罚系数,它适用于像素点p相邻像素中视差值与p的视差值相差1的那些像素点;P 2是一个惩罚系数,它适用于像素点p相邻像素中视差值与p的视差值相差大于1的那些像素点。 Where D refers to the disparity map, E(D) is the energy function corresponding to the disparity map; p, q represents a certain pixel in the image; N p refers to the adjacent pixel of the pixel p; C(p, D p ) Refers to the cost of the pixel when the current pixel disparity is D p ; P 1 is a penalty coefficient, which is applicable to those pixels in which the disparity value of the pixel adjacent to the pixel p differs by 1 from the disparity value of p; P 2 It is a penalty coefficient, which is applicable to those pixels in which the disparity values of pixels adjacent to the pixel p differ by more than one from the disparity values of p.
利用上述函数式6在二维图像中寻找最优解耗时较大,因此该问题被近似分解为多个一维问题,即线性问题。而且每个一维问题都可以用动态规划来解决。因为1个像素点通常有8个相邻像素点(当然也可以有其他数量的相邻像素点),因此一般分解为8个一维问题。Using the above function formula 6 to find the optimal solution time in the two-dimensional image is large, so the problem is approximately decomposed into a plurality of one-dimensional problems, that is, a linear problem. And every one-dimensional problem can be solved with dynamic programming. Since 1 pixel usually has 8 adjacent pixels (of course, there may be other numbers of adjacent pixels), it is generally decomposed into 8 one-dimensional problems.
对于每一个一维的求解,可以采用以下式7:For each one-dimensional solution, Equation 7 below can be used:
Figure PCTCN2017119291-appb-000020
Figure PCTCN2017119291-appb-000020
其中,r指某个指向当前像素p的方向,在此可以理解为像素p在该方向的相邻像素点。Where r refers to a direction pointing to the current pixel p, which can be understood here as an adjacent pixel point of the pixel p in the direction.
L r(p,d)表示沿着当前方向,当目前像素点p的视差取值为d时,其最 小代价值。 L r (p,d) represents the minimum cost value when the parallax of the current pixel point p is d along the current direction.
其中,这个最小值可以从4种可能的候选值中选取的最小值:Wherein, the minimum value can be selected from the minimum of 4 possible candidate values:
第1种可能时当前像素点与前一个像素视差值相等时,其最小的代价值。The first possible value is the smallest value when the current pixel is equal to the previous pixel's disparity value.
第2和3种可能是当前像素点与前一个像素视差值差1(多1或少1)时,其最小的代价值+惩罚系数P 1The second and third types may be the smallest cost + penalty coefficient P 1 when the current pixel is 1 (more than 1 or less) from the previous pixel disparity value.
第4种可能是当前像素点与前一个像素视差值的差大于1时,其最小的代价值+惩罚系数P 2The fourth possibility is that the difference between the current pixel point and the previous pixel disparity value is greater than 1, and its minimum cost value + penalty coefficient P 2 .
另外,当前像素点的代价值还可以减去前一个像素点取不同视差值时最小的代价值。这是因为L r(p,d)是会随着当前像素的右移不停增长的,为了防止数值溢出,可以让它维持在一个较小的数值。 In addition, the generation value of the current pixel point can also be subtracted from the minimum value of the previous pixel point when taking different disparity values. This is because L r (p,d) will grow with the right shift of the current pixel. To prevent the value from overflowing, you can keep it at a small value.
其中,可以采用以下公式8和9计算C(p,d):Among them, C(p,d) can be calculated by the following formulas 8 and 9:
C(p,d)=min(d(p,p-d,I L,I R),d(p-d,p,I R,I L))   式8 C(p,d)=min(d(p,pd,I L ,I R ),d(pd,p,I R ,I L ))
Figure PCTCN2017119291-appb-000021
Figure PCTCN2017119291-appb-000021
在分别计算了每个方向的代价值时,可以将该多个方向,例如8个方向的代价值累计值,选取累加代价值最小的视差值作为该像素的最终视差值,例如,可以通过以下公式10进行累加:When the cost value of each direction is calculated separately, the multiple values, for example, the cumulative value of the value of the eight directions, and the disparity value with the smallest accumulated value may be selected as the final disparity value of the pixel, for example, Accumulate by Equation 10 below:
S(p,d)=∑L r(p,d)   式10 S(p,d)=∑L r (p,d) Equation 10
针对以上式7可以看出,如果期望目标像素点与相邻像素点采用相同的视差,则可以将惩罚系数P 1和惩罚系数P 2设置的较大,这样可以增大目标像素点与相邻像素点采用相同的视差的概率,如果期望目标像素点与相邻像素点的视差相差较大,则可以将惩罚系数P 2设置的较小,将惩罚系数P 1设置的较大,这样可以增大目标像素点与相邻像素点采用较大的视差的差值的概率,如果期望目标像素点与相邻像素点的视差相差较小,则可以将惩罚系数P 1设置的较小,将惩罚系数P 2设置的较大,这样可以增大目标像素点与相邻像素点采用较小的视差的差值的概率。 It can be seen from the above Equation 7 that if the target pixel point and the adjacent pixel point are expected to adopt the same disparity, the penalty coefficient P 1 and the penalty coefficient P 2 can be set larger, so that the target pixel point and the adjacent pixel can be increased. The pixel points adopt the same probability of parallax. If the difference between the target pixel point and the adjacent pixel point is expected to be large, the penalty coefficient P 2 can be set smaller, and the penalty coefficient P 1 can be set larger. The probability that the large target pixel points and the adjacent pixel points adopt a larger difference of the parallax, if the difference between the target pixel point and the adjacent pixel point is expected to be small, the penalty coefficient P 1 can be set smaller, and the penalty will be The coefficient P 2 is set larger, which increases the probability that the difference between the target pixel and the adjacent pixel takes a smaller parallax.
这里可以采用如图5所示的地面为例进行说明。对于地面而言,在2D图像中,地面是沿着上下的方向,深度依次变化的,而左右方向上,深度基本是一致的。Here, the ground shown in FIG. 5 can be used as an example for description. For the ground, in the 2D image, the ground is in the up and down direction, the depth changes sequentially, and in the left and right direction, the depth is basically the same.
所以在图5中,对应着上下左右四个方向(path)的时候,由于左右方向上深度是一致的,所以在左右方向上惩罚参数P 1和P 2较大,算法会倾向 于在左右方向上选择相同的视差;而上下方向上,就会给予较小的惩罚参数P 1和P 2,算法就会倾向于在上下方向上选择不同的视差。 Therefore, in FIG. 5, the time corresponding to vertical and horizontal four directions (path), and since the depth of the left-right direction is the same, so the penalty parameter P 1 in the horizontal direction and P 2 is large, the algorithm will tend in the horizontal direction The same parallax is selected on the top; in the up and down direction, the smaller penalty parameters P 1 and P 2 are given , and the algorithm tends to select different parallaxes in the up and down direction.
可选地,根据该指向向量与该位置关系对应的向量的夹角,调整该惩罚系数。Optionally, the penalty coefficient is adjusted according to an angle between the pointing vector and a vector corresponding to the positional relationship.
在一种实现方式中,在该视差差值的绝对值大于或等于预定视差时,该夹角与90度的差值的模,与该惩罚系数正相关。In one implementation, when the absolute value of the disparity value is greater than or equal to the predetermined disparity, the modulus of the difference between the angle and the 90 degree is positively correlated with the penalty coefficient.
例如,如图6所示的图像,在手臂的延展方向上,即在从肘关节向腕关节的方向上,是一个斜面,离镜头的距离是逐渐变化的,深度是变化的,类比图5中地面部分的上下方向,给予较大的惩罚参数P 1和P 2,算法就会倾向于在肘关节向腕关节的方向上选择不同的视差。而垂直于手臂的方向,垂直于肘关节向腕关节的方向上,基本是在一个距离上,即深度是基本相同,类比图5中地面的左右方向,动态调大惩罚参数P 1和P 2,算法就会倾向于选择相近甚至相同的视差。 For example, the image shown in Fig. 6 is a bevel in the direction of extension of the arm, that is, in the direction from the elbow joint to the wrist joint, the distance from the lens is gradually changed, and the depth is changed, analogous to Fig. 5 In the up and down direction of the ground portion, given the larger penalty parameters P 1 and P 2 , the algorithm tends to choose different parallaxes in the direction of the elbow joint to the wrist joint. The direction perpendicular to the arm, perpendicular to the elbow joint to the wrist joint, is basically at a distance, that is, the depth is substantially the same, analogous to the left and right direction of the ground in Figure 5, dynamically increasing the penalty parameters P 1 and P 2 The algorithm tends to choose similar or even the same parallax.
可选地,根据该目标生命体的肢体边缘,确定该至少一个像素点处于该目标生命体上。Optionally, determining, according to the limb edge of the target living body, the at least one pixel point is on the target living body.
具体地,在相邻像素点位于生命体上时,利用相邻像素点与目标点之间的位置关系,以及目标点向至少一个关节的指向向量确定惩罚系数,更有意义,因此可以根据目标生命体的肢体边缘,确定出相邻点处于目标生命体时,可以采用本申请实施例的方法100计算目标点的视差。Specifically, when the adjacent pixel points are located on the living body, it is more meaningful to determine the penalty coefficient by using the positional relationship between the adjacent pixel points and the target point, and the target point to the pointing vector of the at least one joint, so that the target can be determined according to the target When the edge of the limb of the living body is determined to be the target living body, the method 100 of the embodiment of the present application can be used to calculate the parallax of the target point.
当然,如果目标点是肢体边缘的像素点,且用于计算深度信息的像素点是目标生命体之外的像素点时,则可以设置较小的惩罚P 2,进一步地设置较大的惩罚系数P 1,从而可以允许计算出的视差跳变。 Of course, if the target point is a pixel point at the edge of the limb, and the pixel used to calculate the depth information is a pixel point other than the target living body, a smaller penalty P 2 can be set, and a larger penalty coefficient is further set. P 1 , thus allowing the calculated parallax jump.
可选地,在本申请实施例,可以仅仅参考本申请实施例的PAF方式分割出的目标生命体,确定目标生命体的边缘的像素点,基于边缘的像素点与其相邻像素点的位置关系,调整惩罚系数,而不考虑目标点向至少一个关节处的指向向量。Optionally, in the embodiment of the present application, only the target living body segmented by the PAF method in the embodiment of the present application may be referred to, and the pixel point of the edge of the target living body is determined, and the positional relationship between the pixel based on the edge and the adjacent pixel point is determined. , adjusting the penalty coefficient, regardless of the pointing vector of the target point to at least one joint.
可选地,在本申请实施例中,可以根据至少一个目标点的视差,计算该目标生命体的深度信息。Optionally, in the embodiment of the present application, the depth information of the target living body may be calculated according to the disparity of the at least one target point.
具体地,在计算了目标生命体的各个像素点的视差之后,可以计算该生命体的深度信息。其中,视差可以是深度成反比关系。Specifically, after calculating the parallax of each pixel of the target living body, the depth information of the living body can be calculated. Among them, the parallax can be inversely proportional to the depth.
可选地,可以通过以下公式11计算深度:Alternatively, the depth can be calculated by the following formula 11:
Figure PCTCN2017119291-appb-000022
Figure PCTCN2017119291-appb-000022
其中,d是深度,b是左右相机之间的距离,f是相机的焦距,d p是视差(disparity)。 Where d is the depth, b is the distance between the left and right cameras, f is the focal length of the camera, and d p is the disparity.
从上式1可以看出,由于b和f是物理属性,一般保持不变,则d与d p成反比关系。对于近距离的物体而言,深度较小,那么视差较大,而对于远距离的物体而言,深度较大,对应的视差较小。 It can be seen from the above formula 1 that since b and f are physical properties and generally remain unchanged, d is inversely proportional to d p . For objects at close range, the depth is small, then the parallax is larger, while for distant objects, the depth is larger and the corresponding parallax is smaller.
可选地,在本申请实施例中,可以根据该目标生命体的深度,确定第一速度,该第一速度的方向为从该目标生命体向无人驾驶设备的方向;根据该第一速度和第二速度,确定对该无人驾驶设备进行控制的控制速度,该第二速度为控制器输入的速度;根据该控制速度,控制该无人驾驶设备的飞行。可选地,该第一速度的大小与深度成反比。其中,该无人驾驶设备可以是无人机或无人驾驶汽车等,以下将以无人机为例进行说明。Optionally, in the embodiment of the present application, the first speed may be determined according to the depth of the target living body, where the direction of the first speed is a direction from the target living body to the unmanned driving device; according to the first speed And a second speed determining a control speed for controlling the unmanned device, the second speed being a speed input by the controller; and controlling the flight of the unmanned device according to the control speed. Optionally, the magnitude of the first velocity is inversely proportional to the depth. The unmanned device may be a drone or a driverless car. The following description will be made by taking an unmanned aerial vehicle as an example.
具体地,无人机飞到距离生命体比较近时,利用斥力场(Repulsive Force Field)“弹开”无人机,实现对障碍物的绕行。Specifically, when the drone flies closer to the living body, the Repulsive Force Field is used to "bounce" the drone to achieve the obstacle bypass.
这里可以参照以下万有引力公式12,构造斥力场,具体表现形式可以参考式13。Here, the repulsion field can be constructed by referring to the following universal gravitation formula 12, and the specific expression can refer to Equation 13.
Figure PCTCN2017119291-appb-000023
Figure PCTCN2017119291-appb-000023
Figure PCTCN2017119291-appb-000024
Figure PCTCN2017119291-appb-000024
这里生命体质量m obstacle可以取一个较大的恒定值,其中,m drone是无人机的质量,而G也是一个恒定值,因此可以定义常数k=G·m obstacle,由此,可以得到下式14: Here, the life mass m obstacle can take a larger constant value, where m drone is the mass of the drone , and G is also a constant value, so the constant k=G·m obstacle can be defined, thereby obtaining the next Equation 14:
Figure PCTCN2017119291-appb-000025
Figure PCTCN2017119291-appb-000025
其中,D x是生命体的深度信息,可以对生命体的各个像素点的深度取平均。 Where D x is the depth information of the living body, and the depth of each pixel of the living body can be averaged.
然后,可以通过以下式15中的恒加速度公式,得出斥力场内规划的速度
Figure PCTCN2017119291-appb-000026
Then, the speed of the repulsive field planning can be obtained by the constant acceleration formula in Equation 15 below.
Figure PCTCN2017119291-appb-000026
Figure PCTCN2017119291-appb-000027
Figure PCTCN2017119291-appb-000027
斥力场对应的速度指向远离生命体的方向,而用户控制的无人机本身有一个速度,两速度通过矢量叠加后生成一个新的速度,作为最终规划的速度,当做控制系统的速度环指令,最终会实现对障碍物的绕行。The speed corresponding to the repulsive field points away from the direction of the living body, and the user-controlled drone has a speed. The two speeds are superimposed by the vector to generate a new speed. As the final planned speed, as the speed loop command of the control system, Eventually, the obstacles will be bypassed.
因此,在本申请实施例中,确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定所述目标点与至少一个像素点的位置关系;根据所述指向向量,以及所述位置关系,调整SGM算法的全局能量函数的惩罚系数;基于所述至少一个像素点的视差,利用调整了所述惩罚系数后的所全局能量函数,计算所述目标点的视差,可以实现在具有生命体的高动态场景下,充分考虑生命体的肢体或关节等特征来调整半全局匹配算法中的惩罚系数,避免采用固定的惩罚系数,使得生命体的深度图计算的更加精准。Therefore, in the embodiment of the present application, a pointing vector of a target point on a target living body in an image to at least one joint is determined, and a positional relationship between the target point and at least one pixel point is determined; according to the pointing vector, And the positional relationship, adjusting a penalty coefficient of the global energy function of the SGM algorithm; calculating a parallax of the target point by using a global energy function adjusted by the penalty coefficient based on the parallax of the at least one pixel point, In the high dynamic scene with living body, the penalty coefficient in the semi-global matching algorithm is adjusted by fully considering the features of the limb or joint of the living body, and the fixed penalty coefficient is avoided, so that the depth map calculation of the living body is more accurate.
并且,进一步地,在本申请实施例中,针对于低空飞行的无人机,提出了一套完全策略,针对于人体以及其他动物做检测,并用于优化深度图的计算,从而得到更加精准的障碍物观测,即能保证人的安全,也能更好的实现无人机避障航线规划绕行。Moreover, further, in the embodiment of the present application, a complete strategy is proposed for the low-altitude flying drone, which is used for detecting the human body and other animals, and is used for optimizing the calculation of the depth map, thereby obtaining more accurate. Obstacle observation can guarantee the safety of people, and can better realize the planning of bypass for unmanned obstacle avoidance routes.
图8是根据本申请实施例的图像处理设备200的示意性框图。如图8所示,该设备200包括确定单元210和计算单元220;其中,FIG. 8 is a schematic block diagram of an image processing apparatus 200 according to an embodiment of the present application. As shown in FIG. 8, the device 200 includes a determining unit 210 and a calculating unit 220;
所述确定单元210用于:确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定所述目标点与至少一个像素点的位置关系;The determining unit 210 is configured to: determine a pointing vector of the target point on the target living body in the image to the at least one joint, and determine a positional relationship between the target point and the at least one pixel point;
所述计算单元220用于:根据所述指向向量,以及所述位置关系,调整半全局匹配SGM算法的全局能量函数的惩罚系数;The calculating unit 220 is configured to: adjust a penalty coefficient of a global energy function of the semi-global matching SGM algorithm according to the pointing vector and the position relationship;
基于所述至少一个像素点的视差,利用调整了所述惩罚系数后的所全局能量函数,计算所述目标点的视差。A parallax of the target point is calculated based on a parallax of the at least one pixel point using a global energy function adjusted by the penalty coefficient.
可选地,所述计算单元220进一步用于:Optionally, the calculating unit 220 is further configured to:
根据所述指向向量与所述位置关系对应的向量的夹角,调整所述惩罚系数。The penalty coefficient is adjusted according to an angle between the pointing vector and a vector corresponding to the positional relationship.
可选地,在所述视差差值的绝对值大于或等于预定视差时,所述夹角与 90度的差值的模,与所述惩罚系数正相关。Optionally, when the absolute value of the disparity value is greater than or equal to a predetermined disparity, a modulus of the difference between the angle and 90 degrees is positively correlated with the penalty coefficient.
可选地,如图8所示,该设备200还包括第一分割单元230,用于:Optionally, as shown in FIG. 8, the device 200 further includes a first dividing unit 230, configured to:
在所述图像上,确定生命体的肢体关节;Determining a limb joint of a living body on the image;
根据所述生命体的肢体关节的向量场,确定所述生命体的肢体关节的连接关系;Determining a connection relationship of the limb joints of the living body according to a vector field of the limb joint of the living body;
根据所述生命体的肢体关节的连接关系,从所述图像中分割出所述目标生命体。The target living body is segmented from the image according to a connection relationship of the limb joints of the living body.
可选地,所述确定单元210进一步用于:Optionally, the determining unit 210 is further configured to:
根据所述目标生命体的肢体关节的指向关系,确定所述目标点向至少一个关节处的指向向量。Determining a pointing vector of the target point to at least one joint according to a pointing relationship of a limb joint of the target living body.
可选地,所述确定单元210进一步用于:Optionally, the determining unit 210 is further configured to:
根据所述目标生命体的肢体边缘,确定所述至少一个像素点处于所述目标生命体上。Determining that the at least one pixel point is on the target living body according to a limb edge of the target living body.
可选地,如图8所示,该设备200还包括第二分割单元240,用于:Optionally, as shown in FIG. 8, the device 200 further includes a second dividing unit 240, configured to:
采用热成像的方式,从所述图像中初始分割出所述目标生命体。The target living body is initially segmented from the image by means of thermal imaging.
可选地,如图8所示,该设备200还包括控制单元250,用于:Optionally, as shown in FIG. 8, the device 200 further includes a control unit 250, configured to:
根据至少一个所述目标点的视差,计算所述目标生命体的深度;Calculating a depth of the target living body according to a disparity of the at least one target point;
根据所述目标生命体的深度,确定第一速度,所述第一速度的方向为从所述目标生命体指向无人驾驶设备的方向;Determining a first speed according to a depth of the target living body, the direction of the first speed being a direction from the target living body to the unmanned device;
根据所述第一速度和第二速度,确定对所述无人驾驶设备进行控制的控制速度,其中,所述第二速度为控制器输入的速度;Determining, according to the first speed and the second speed, a control speed for controlling the unmanned device, wherein the second speed is a speed input by the controller;
根据所述控制速度,控制所述无人驾驶设备。The unmanned device is controlled according to the control speed.
可选地,所述第一速度的大小与所述深度成反比。Optionally, the magnitude of the first velocity is inversely proportional to the depth.
应理解,该图像处理设备可以实现方法100中的相应操作,为了简洁,在此不再赘述。It should be understood that the image processing device may implement corresponding operations in the method 100, and for brevity, no further details are provided herein.
图9是根据本申请实施例的图像处理设备400的示意性框图FIG. 9 is a schematic block diagram of an image processing apparatus 400 according to an embodiment of the present application.
可选地,该图像处理设备400可以包括多个不同的部件,这些部件可以作为集成电路(integrated circuits,ICs),或集成电路的部分,离散的电子设备,或其它适用于电路板(诸如主板,或附加板)的模块,也可以作为并入计算机系统的部件。Alternatively, the image processing device 400 may include a plurality of different components that may be integrated circuits (ICs), or portions of integrated circuits, discrete electronic devices, or other suitable for use in a circuit board (such as a motherboard). Modules, or additional boards, may also be incorporated as part of a computer system.
可选地,该图像处理设备可以包括处理器410和与处理器410耦合的存 储介质420。Alternatively, the image processing device can include a processor 410 and a storage medium 420 coupled to the processor 410.
处理器410可以包括一个或多个通用处理器,诸如中央处理单元(central processing unit,CPU),或处理设备等。具体地,该处理器410可以是复杂指令集处理(complex instruction set computing,CISC)微处理器,超长指令字(very long instruction word,VLIW)微处理器,实现多个指令集组合的微处理器。该处理器也可以是一个或多个专用处理器,诸如应用专用集成电路(application specific integrated circuit,ASIC),现场可编程门阵列(field programmable gate array,FPGA),数字信号处理器(digital signal processor,DSP)。Processor 410 may include one or more general purpose processors, such as a central processing unit (CPU), or a processing device or the like. Specifically, the processor 410 may be a complex instruction set computing (CISC) microprocessor, a very long instruction word (VLIW) microprocessor, and implements micro-processing of multiple instruction set combinations. Device. The processor may also be one or more dedicated processors, such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), and a digital signal processor. , DSP).
处理器410可以与存储介质420通信。该存储介质420可以为磁盘、光盘、只读存储器(read only memory,ROM),闪存,相变存储器。该存储介质420可以存储有处理器存储的指令,和/或,可以缓存一些从外部存储设备存储的信息。Processor 410 can be in communication with storage medium 420. The storage medium 420 may be a magnetic disk, an optical disk, a read only memory (ROM), a flash memory, or a phase change memory. The storage medium 420 can store instructions stored by the processor and/or can cache some of the information stored from the external storage device.
可选地,除了处理器420和存储介质420,图像处理设备可以包括显示控制器和/或显示设备单元430,收发器440,视频输入输出单元450,音频输入输出单元460,其他输入输出单元470。图像处理设备400包括的这些部件可以通过总线或内部连接互联。Optionally, in addition to the processor 420 and the storage medium 420, the image processing apparatus may include a display controller and/or display device unit 430, a transceiver 440, a video input output unit 450, an audio input output unit 460, and other input and output units 470. . These components included in image processing device 400 may be interconnected by a bus or internal connection.
可选地,该收发器440可以是有线收发器或无线收发器,诸如,WIFI收发器,卫星收发器,蓝牙收发器,无线蜂窝电话收发器或其组合等。Alternatively, the transceiver 440 can be a wired transceiver or a wireless transceiver, such as a WIFI transceiver, a satellite transceiver, a Bluetooth transceiver, a wireless cellular telephone transceiver, or combinations thereof.
可选地,视频输入输出单元450可以包括诸如摄像机的图像处理子系统,其包括光传感器,电荷耦合器件(charged coupled device,CCD)或互补金属氧化物半导体(complementary metal-oxide semiconductor,CMOS)光传感器,以用于实现拍摄功能。Optionally, the video input and output unit 450 may include an image processing subsystem such as a camera, including a photo sensor, a charge coupled device (CCD) or a complementary metal-oxide semiconductor (CMOS) light. Sensor for use in shooting functions.
可选地,该音频输入输出单元460可以包括扬声器,话筒,听筒等。Alternatively, the audio input and output unit 460 may include a speaker, a microphone, an earpiece, and the like.
可选地,其他输入输出设备470可以包括存储设备,universal serial bus(USB)端口,串行端口,并行端口,打印机,网络接口等。Alternatively, other input and output devices 470 may include a storage device, a universal serial bus (USB) port, a serial port, a parallel port, a printer, a network interface, and the like.
可选地,该图像处理设备400可以执行方法100所示的操作,为了简洁,在此不再赘述。Optionally, the image processing device 400 can perform the operations shown in the method 100. For brevity, details are not described herein again.
可选地,图像处理设备400或400可以位于可移动设备中。可移动设备可以在任何合适的环境下移动,例如,空气中(例如,定翼飞机、旋翼飞机,或既没有定翼也没有旋翼的飞机)、水中(例如,轮船或潜水艇)、陆地上(例 如,汽车或火车)、太空(例如,太空飞机、卫星或探测器),以及以上各种环境的任何组合。可移动设备可以无人驾驶的汽车、自动飞行的无人机、VR/AR眼镜、双摄像头的手机、有视觉系统的智能小车等设备。Alternatively, the image processing device 400 or 400 may be located in a mobile device. The mobile device can be moved in any suitable environment, for example, in the air (eg, a fixed-wing aircraft, a rotorcraft, or an aircraft with neither a fixed wing nor a rotor), in water (eg, a ship or submarine), on land. (for example, a car or train), space (for example, a space plane, satellite or detector), and any combination of the above. The mobile device can be an unmanned car, an auto-flying drone, a VR/AR glasses, a dual-camera mobile phone, a smart car with a vision system, and the like.
图10是根据本申请实施例的可移动设备500的示意性框图。如图10所示,可移动设备500包括载体510和负载520。图14中将可移动设备描述为无人机仅仅是为了描述方面。负载520可以不通过载体510连接到可移动设备上。可移动设备500还可以包括动力系统530、传感系统540和通信系统550和图像处理设备562和拍摄系统564。FIG. 10 is a schematic block diagram of a removable device 500 in accordance with an embodiment of the present application. As shown in FIG. 10, the mobile device 500 includes a carrier 510 and a load 520. The description of the mobile device in Figure 14 as a drone is for illustrative purposes only. The load 520 may not be connected to the mobile device via the carrier 510. The removable device 500 can also include a power system 530, a sensing system 540 and a communication system 550 and an image processing device 562 and a photographing system 564.
动力系统530可以包括电子调速器(简称为电调)、一个或多个螺旋桨以及与一个或多个螺旋桨相对应的一个或多个电机。电机和螺旋桨设置在对应的机臂上;电子调速器用于接收飞行控制器产生的驱动信号,并根据驱动信号提供驱动电流给电机,以控制电机的转速和/或转向。电机用于驱动螺旋桨旋转,从而为UAV的飞行提供动力,该动力使得UAV能够实现一个或多个自由度的运动。在某些实施例中,UAV可以围绕一个或多个旋转轴旋转。例如,上述旋转轴可以包括横滚轴、平移轴和俯仰轴。应理解,电机可以是直流电机,也可以交流电机。另外,电机可以是无刷电机,也可以有刷电机。 Power system 530 can include an electronic governor (referred to as an ESC), one or more propellers, and one or more electric machines corresponding to one or more propellers. The motor and the propeller are disposed on the corresponding arm; the electronic governor is configured to receive a driving signal generated by the flight controller, and provide a driving current to the motor according to the driving signal to control the rotation speed and/or steering of the motor. The motor is used to drive the propeller to rotate to power the UAV's flight, which enables the UAV to achieve one or more degrees of freedom of motion. In certain embodiments, the UAV can be rotated about one or more axes of rotation. For example, the above-described rotating shaft may include a roll axis, a pan axis, and a pitch axis. It should be understood that the motor can be a DC motor or an AC motor. In addition, the motor can be a brushless motor or a brush motor.
传感系统540用于测量UAV的姿态信息,即UAV在空间的位置信息和状态信息,例如,三维位置、三维角度、三维速度、三维加速度和三维角速度等。传感系统例如可以包括陀螺仪、电子罗盘、惯性测量单元(Inertial Measurement Unit,简称为“IMU”)、视觉传感器、全球定位系统(Global Positioning System,简称为“GPS”)和气压计等传感器中的至少一种。飞行控制器用于控制UAV的飞行,例如,可以根据传感系统测量的姿态信息控制UAV的飞行。应理解,飞行控制器可以按照预先编好的程序指令对UAV进行控制,也可以通过响应来自操纵设备的一个或多个控制指令对UAV进行控制。The sensing system 540 is used to measure the attitude information of the UAV, that is, the position information and state information of the UAV in space, for example, three-dimensional position, three-dimensional angle, three-dimensional speed, three-dimensional acceleration, and three-dimensional angular velocity. The sensing system may include, for example, a gyroscope, an electronic compass, an Inertial Measurement Unit ("IMU"), a vision sensor, a Global Positioning System (GPS), and a barometer. At least one of them. The flight controller is used to control the flight of the UAV, for example, the UAV flight can be controlled based on the attitude information measured by the sensing system. It should be understood that the flight controller may control the UAV in accordance with pre-programmed program instructions, or may control the UAV in response to one or more control commands from the operating device.
通信系统550能够与一个具有通信系统570的终端设备580通过无线信号590进行通信。通信系统550和通信系统570可以包括多个用于无线通信的发射机、接收机和/或收发机。这里的无线通信可以是单向通信,例如,只能是可移动设备500向终端设备580发送数据。或者无线通信可以是双向通信,数据即可以从可移动设备500发送给终端设备580,也可以由终端设备580发送给可移动设备500。 Communication system 550 is capable of communicating with wireless terminal 590 via a terminal device 580 having communication system 570. Communication system 550 and communication system 570 can include a plurality of transmitters, receivers, and/or transceivers for wireless communication. The wireless communication herein may be one-way communication, for example, only the mobile device 500 may transmit data to the terminal device 580. Alternatively, the wireless communication may be two-way communication, and the data may be transmitted from the mobile device 500 to the terminal device 580 or may be transmitted by the terminal device 580 to the mobile device 500.
可选地,终端设备580能够提供针对于一个或多个可移动设备500、载体510和负载520的控制数据,并能接收可移动设备500、载体510和负载520发送的信息。终端设备580提供的控制数据能够用于控制一个或多个可移动设备500、载体510和负载520的状态。可选地,载体510和负载520中包括用于与终端设备580进行通信的通信模块。Alternatively, terminal device 580 can provide control data for one or more of mobile device 500, carrier 510, and load 520, and can receive information transmitted by mobile device 500, carrier 510, and load 520. The control data provided by terminal device 580 can be used to control the status of one or more of mobile device 500, carrier 510, and load 520. Optionally, a carrier 510 and load 520 include a communication module for communicating with the terminal device 580.
可以理解的是,图10所示出的可移动设备包括的图像处理设备562能够执行方法100,为了简洁,在此不再赘述。It can be understood that the image processing device 562 included in the mobile device shown in FIG. 10 can perform the method 100, which is not described herein for brevity.
以上所述,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应所述以权利要求的保护范围为准。The foregoing is only a specific embodiment of the present application, but the scope of protection of the present application is not limited thereto, and any person skilled in the art can easily think of changes or substitutions within the technical scope disclosed in the present application. It should be covered by the scope of protection of this application. Therefore, the scope of protection of the present application should be determined by the scope of the claims.
以上,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以权利要求的保护范围为准。The above is only a specific embodiment of the present application, but the scope of protection of the present application is not limited thereto, and any person skilled in the art can easily think of changes or substitutions within the technical scope disclosed in the present application, and should cover Within the scope of protection of this application. Therefore, the scope of protection of this application should be determined by the scope of protection of the claims.

Claims (18)

  1. 一种图像处理方法,其特征在于,包括:An image processing method, comprising:
    确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定所述目标点与至少一个像素点的位置关系;Determining a pointing vector of the target point on the target living body in the image to the at least one joint, and determining a positional relationship of the target point and the at least one pixel point;
    根据所述指向向量,以及所述位置关系,调整半全局匹配SGM算法的全局能量函数的惩罚系数;Adjusting a penalty coefficient of a global energy function of the semi-global matching SGM algorithm according to the pointing vector and the positional relationship;
    基于所述至少一个像素点的视差,利用调整了所述惩罚系数后的所全局能量函数,计算所述目标点的视差。A parallax of the target point is calculated based on a parallax of the at least one pixel point using a global energy function adjusted by the penalty coefficient.
  2. 根据权利要求1所述的方法,其特征在于,根据所述指向向量,以及所述位置关系,调整半全局匹配SGM算法的全局能量函数的惩罚系数,包括:The method according to claim 1, wherein the penalty coefficient of the global energy function of the semi-global matching SGM algorithm is adjusted according to the pointing vector and the positional relationship, including:
    根据所述指向向量与所述位置关系对应的向量的夹角,调整所述惩罚系数。The penalty coefficient is adjusted according to an angle between the pointing vector and a vector corresponding to the positional relationship.
  3. 根据权利要求2所述的方法,其特征在于,在所述视差差值的绝对值大于或等于预定视差时,所述夹角与90度的差值的模,与所述惩罚系数正相关。The method according to claim 2, wherein, when the absolute value of the disparity value is greater than or equal to a predetermined disparity, a modulus of the difference between the angle and 90 degrees is positively correlated with the penalty coefficient.
  4. 根据权利要求1至3中任一项所述的方法,其特征在于,在所述确定图像中的目标生命体上的目标点向至少一个关节处的指向向量之前,所述方法还包括:The method according to any one of claims 1 to 3, wherein before the determining a target point on the target living body in the image to the pointing vector at the at least one joint, the method further comprises:
    在所述图像上,确定生命体的肢体关节;Determining a limb joint of a living body on the image;
    根据所述生命体的肢体关节的向量场,确定所述生命体的肢体关节的连接关系;Determining a connection relationship of the limb joints of the living body according to a vector field of the limb joint of the living body;
    根据所述生命体的肢体关节的连接关系,从所述图像中分割出所述目标生命体。The target living body is segmented from the image according to a connection relationship of the limb joints of the living body.
  5. 根据权利要求4所述的方法,其特征在于,所述确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,包括:The method according to claim 4, wherein the determining a pointing vector of the target point on the target living body in the image to the at least one joint comprises:
    根据所述目标生命体的肢体关节的指向关系,确定所述目标点向至少一个关节处的指向向量。Determining a pointing vector of the target point to at least one joint according to a pointing relationship of a limb joint of the target living body.
  6. 根据权利要求4或5所述的方法,其特征在于,在所述确定所述目标点与至少一个像素点的位置关系之前,所述方法还包括:The method according to claim 4 or 5, wherein before the determining the positional relationship between the target point and the at least one pixel, the method further comprises:
    根据所述目标生命体的肢体边缘,确定所述至少一个像素点处于所述目 标生命体上。Determining that the at least one pixel point is on the target living body based on a limb edge of the target living body.
  7. 根据权利要求4至6中任一项所述的方法,其特征在于,在所述根据所述生命体的肢体关节的连接关系,从所述图像中分割出所述目标生命体之前,所述方法还包括:The method according to any one of claims 4 to 6, wherein said said living body is segmented from said image based on said connection relationship of limb joints of said living body The method also includes:
    采用热成像的方式,从所述图像中初始分割出所述目标生命体。The target living body is initially segmented from the image by means of thermal imaging.
  8. 根据权利要求1至7中任一项所述的方法,其特征在于,所述方法还包括:The method according to any one of claims 1 to 7, wherein the method further comprises:
    根据至少一个所述目标点的视差,计算所述目标生命体的深度;Calculating a depth of the target living body according to a disparity of the at least one target point;
    根据所述目标生命体的深度,确定第一速度,所述第一速度的方向为从所述目标生命体指向无人驾驶设备的方向;Determining a first speed according to a depth of the target living body, the direction of the first speed being a direction from the target living body to the unmanned device;
    根据所述第一速度和第二速度,确定对所述无人驾驶设备进行控制的控制速度,其中,所述第二速度为控制器输入的速度;Determining, according to the first speed and the second speed, a control speed for controlling the unmanned device, wherein the second speed is a speed input by the controller;
    根据所述控制速度,控制所述无人驾驶设备。The unmanned device is controlled according to the control speed.
  9. 根据权利要求8所述的方法,其特征在于,所述第一速度的大小与所述深度成反比。The method of claim 8 wherein the magnitude of the first velocity is inversely proportional to the depth.
  10. 一种图像处理设备,其特征在于,包括确定单元和计算单元;其中,An image processing device, comprising: a determining unit and a calculating unit; wherein
    所述确定单元用于:确定图像中的目标生命体上的目标点向至少一个关节处的指向向量,以及确定所述目标点与至少一个像素点的位置关系;The determining unit is configured to: determine a pointing vector of the target point on the target living body in the image to the at least one joint, and determine a positional relationship between the target point and the at least one pixel point;
    所述计算单元用于:根据所述指向向量,以及所述位置关系,调整半全局匹配SGM算法的全局能量函数的惩罚系数;The calculating unit is configured to: adjust a penalty coefficient of a global energy function of the semi-global matching SGM algorithm according to the pointing vector and the positional relationship;
    基于所述至少一个像素点的视差,利用调整了所述惩罚系数后的所全局能量函数,计算所述目标点的视差。A parallax of the target point is calculated based on a parallax of the at least one pixel point using a global energy function adjusted by the penalty coefficient.
  11. 根据权利要求10所述的设备,其特征在于,所述计算单元进一步用于:The device according to claim 10, wherein the calculating unit is further configured to:
    根据所述指向向量与所述位置关系对应的向量的夹角,调整所述惩罚系数。The penalty coefficient is adjusted according to an angle between the pointing vector and a vector corresponding to the positional relationship.
  12. 根据权利要求11所述的设备,其特征在于,在所述视差差值的绝对值大于或等于预定视差时,所述夹角与90度的差值的模,与所述惩罚系数正相关。The apparatus according to claim 11, wherein a modulus of a difference between the included angle and 90 degrees is positively correlated with the penalty coefficient when an absolute value of the parallax difference value is greater than or equal to a predetermined parallax.
  13. 根据权利要求10至12中任一项所述的设备,其特征在于,还包括第一分割单元,用于:The device according to any one of claims 10 to 12, further comprising a first dividing unit for:
    在所述图像上,确定生命体的肢体关节;Determining a limb joint of a living body on the image;
    根据所述生命体的肢体关节的向量场,确定所述生命体的肢体关节的连接关系;Determining a connection relationship of the limb joints of the living body according to a vector field of the limb joint of the living body;
    根据所述生命体的肢体关节的连接关系,从所述图像中分割出所述目标生命体。The target living body is segmented from the image according to a connection relationship of the limb joints of the living body.
  14. 根据权利要求13所述的设备,其特征在于,所述确定单元进一步用于:The device according to claim 13, wherein the determining unit is further configured to:
    根据所述目标生命体的肢体关节的指向关系,确定所述目标点向至少一个关节处的指向向量。Determining a pointing vector of the target point to at least one joint according to a pointing relationship of a limb joint of the target living body.
  15. 根据权利要求13或14所述的设备,其特征在于,所述确定单元进一步用于:The device according to claim 13 or 14, wherein the determining unit is further configured to:
    根据所述目标生命体的肢体边缘,确定所述至少一个像素点处于所述目标生命体上。Determining that the at least one pixel point is on the target living body according to a limb edge of the target living body.
  16. 根据权利要求13至15中任一项所述的设备,其特征在于,还包括第二分割单元,用于:The device according to any one of claims 13 to 15, further comprising a second dividing unit for:
    采用热成像的方式,从所述图像中初始分割出所述目标生命体。The target living body is initially segmented from the image by means of thermal imaging.
  17. 根据权利要求10至16中任一项所述的设备,其特征在于,还包括控制单元,用于:The device according to any one of claims 10 to 16, further comprising a control unit for:
    根据至少一个所述目标点的视差,计算所述目标生命体的深度;Calculating a depth of the target living body according to a disparity of the at least one target point;
    根据所述目标生命体的深度,确定第一速度,所述第一速度的方向为从所述目标生命体指向无人驾驶设备的方向;Determining a first speed according to a depth of the target living body, the direction of the first speed being a direction from the target living body to the unmanned device;
    根据所述第一速度和第二速度,确定对所述无人驾驶设备进行控制的控制速度,其中,所述第二速度为控制器输入的速度;Determining, according to the first speed and the second speed, a control speed for controlling the unmanned device, wherein the second speed is a speed input by the controller;
    根据所述控制速度,控制所述无人驾驶设备。The unmanned device is controlled according to the control speed.
  18. 根据权利要求17所述的设备,其特征在于,所述第一速度的大小与所述深度成反比。The apparatus of claim 17 wherein the magnitude of the first velocity is inversely proportional to the depth.
PCT/CN2017/119291 2017-12-28 2017-12-28 Image processing method and apparatus WO2019127192A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2017/119291 WO2019127192A1 (en) 2017-12-28 2017-12-28 Image processing method and apparatus
CN201780022779.9A CN109074661A (en) 2017-12-28 2017-12-28 Image processing method and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/119291 WO2019127192A1 (en) 2017-12-28 2017-12-28 Image processing method and apparatus

Publications (1)

Publication Number Publication Date
WO2019127192A1 true WO2019127192A1 (en) 2019-07-04

Family

ID=64812376

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/119291 WO2019127192A1 (en) 2017-12-28 2017-12-28 Image processing method and apparatus

Country Status (2)

Country Link
CN (1) CN109074661A (en)
WO (1) WO2019127192A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112040214A (en) * 2019-06-04 2020-12-04 万维科研有限公司 Double-camera three-dimensional imaging system and processing method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120314031A1 (en) * 2011-06-07 2012-12-13 Microsoft Corporation Invariant features for computer vision
CN105222760A (en) * 2015-10-22 2016-01-06 一飞智控(天津)科技有限公司 The autonomous obstacle detection system of a kind of unmanned plane based on binocular vision and method
CN106022304A (en) * 2016-06-03 2016-10-12 浙江大学 Binocular camera-based real time human sitting posture condition detection method

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6561512B2 (en) * 2014-03-27 2019-08-21 株式会社リコー Parallax value deriving device, moving body, robot, parallax value deriving method, parallax value producing method, and program
CN104835165B (en) * 2015-05-12 2017-05-24 努比亚技术有限公司 Image processing method and image processing device
CN106815594A (en) * 2015-11-30 2017-06-09 展讯通信(上海)有限公司 Solid matching method and device
WO2018090250A1 (en) * 2016-11-16 2018-05-24 深圳市大疆创新科技有限公司 Three-dimensional point cloud generation method, device, computer system, and mobile apparatus
CN106931961B (en) * 2017-03-20 2020-06-23 成都通甲优博科技有限责任公司 Automatic navigation method and device
CN106981075A (en) * 2017-05-31 2017-07-25 江西制造职业技术学院 The skeleton point parameter acquisition devices of apery motion mimicry and its recognition methods
CN107214703B (en) * 2017-07-11 2020-08-14 江南大学 Robot self-calibration method based on vision-assisted positioning
CN107392898B (en) * 2017-07-20 2020-03-20 海信集团有限公司 Method and device for calculating pixel point parallax value applied to binocular stereo vision

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120314031A1 (en) * 2011-06-07 2012-12-13 Microsoft Corporation Invariant features for computer vision
CN105222760A (en) * 2015-10-22 2016-01-06 一飞智控(天津)科技有限公司 The autonomous obstacle detection system of a kind of unmanned plane based on binocular vision and method
CN106022304A (en) * 2016-06-03 2016-10-12 浙江大学 Binocular camera-based real time human sitting posture condition detection method

Also Published As

Publication number Publication date
CN109074661A (en) 2018-12-21

Similar Documents

Publication Publication Date Title
US11649052B2 (en) System and method for providing autonomous photography and videography
US20220374010A1 (en) User Interaction Paradigms For A Flying Digital Assistant
US10447912B2 (en) Systems, methods, and devices for setting camera parameters
US7974460B2 (en) Method and system for three-dimensional obstacle mapping for navigation of autonomous vehicles
WO2018098704A1 (en) Control method, apparatus, and system, unmanned aerial vehicle, and mobile platform
CN108235815B (en) Imaging control device, imaging system, moving object, imaging control method, and medium
US11057604B2 (en) Image processing method and device
US20220086362A1 (en) Focusing method and apparatus, aerial camera and unmanned aerial vehicle
WO2019104571A1 (en) Image processing method and device
TWI649721B (en) Panoramic photographing method of unmanned aerial vehicle and unmanned aerial vehicle using same
WO2018120350A1 (en) Method and device for positioning unmanned aerial vehicle
CN111226154B (en) Autofocus camera and system
WO2019227333A1 (en) Group photograph photographing method and apparatus
WO2020024182A1 (en) Parameter processing method and apparatus, camera device and aircraft
CN111699454B (en) Flight planning method and related equipment
WO2020019175A1 (en) Image processing method and apparatus, and photographing device and unmanned aerial vehicle
WO2020198963A1 (en) Data processing method and apparatus related to photographing device, and image processing device
WO2021217371A1 (en) Control method and apparatus for movable platform
JP6949930B2 (en) Control device, moving body and control method
WO2019127192A1 (en) Image processing method and apparatus
CN111344650B (en) Information processing device, flight path generation method, program, and recording medium
JP6790318B2 (en) Unmanned aerial vehicles, control methods, and programs
WO2019189381A1 (en) Moving body, control device, and control program
JP6515423B2 (en) CONTROL DEVICE, MOBILE OBJECT, CONTROL METHOD, AND PROGRAM
WO2020225979A1 (en) Information processing device, information processing method, program, and information processing system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17936322

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17936322

Country of ref document: EP

Kind code of ref document: A1