WO2019019160A1 - Method for acquiring image information, image processing device, and computer storage medium - Google Patents

Method for acquiring image information, image processing device, and computer storage medium Download PDF

Info

Publication number
WO2019019160A1
WO2019019160A1 PCT/CN2017/094932 CN2017094932W WO2019019160A1 WO 2019019160 A1 WO2019019160 A1 WO 2019019160A1 CN 2017094932 W CN2017094932 W CN 2017094932W WO 2019019160 A1 WO2019019160 A1 WO 2019019160A1
Authority
WO
WIPO (PCT)
Prior art keywords
point
image
detected
feature value
depth
Prior art date
Application number
PCT/CN2017/094932
Other languages
French (fr)
Chinese (zh)
Inventor
阳光
Original Assignee
深圳配天智能技术研究院有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳配天智能技术研究院有限公司 filed Critical 深圳配天智能技术研究院有限公司
Priority to PCT/CN2017/094932 priority Critical patent/WO2019019160A1/en
Priority to CN201780092646.9A priority patent/CN110800020B/en
Publication of WO2019019160A1 publication Critical patent/WO2019019160A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof

Definitions

  • the invention belongs to the technical field of information analysis, and in particular relates to an image information acquisition method, an image processing device and a computer storage medium.
  • Binocular stereo vision is an important branch of computer vision. Binocular stereo vision is a method of simulating the principle of human vision. It uses a computer to passively perceive distance. It uses two identical cameras to image the same object from different positions to obtain the stereoscopic view of the object. The image pair, according to the pixel matching relationship between the images, calculates the offset between the pixels by the principle of triangulation to obtain the three-dimensional information of the object, and obtains the depth information of the object, and can calculate the actual distance between the object and the camera. The three-dimensional size of the object, the actual distance between the two points.
  • the camera is generally used to perform imaging shooting from multiple angles, and then the depth information of the occluded pixels is restored according to multiple images captured by multiple cameras at multiple angles, thereby obtaining the depth of the occluded portion of the object. Eliminate dead ends.
  • the camera is added to eliminate the dead angle.
  • Embodiments of the present invention provide an image information acquisition method, an image processing device, and a computer storage medium for reducing unnecessary hardware expenditures in multi-view stereo vision detection.
  • a first aspect of the embodiments of the present invention provides a method for acquiring image information, including:
  • the to-be-detected point is a pixel in the first image or the second image
  • the to-be-detected point is not included in the matching area
  • the matching area is the first image
  • an area included in the second image the first image is captured by a first camera
  • the second image is captured by a second camera
  • the first image and the second image are taken at different angles Images obtained from the same target;
  • the depth of the target pixel is taken as the depth of the point to be detected.
  • a second aspect of the embodiments of the present invention provides an image processing device, where the image processing device includes:
  • the memory is configured to store an operation instruction
  • the processor is configured to acquire an actual image feature value of the to-be-detected point, where the to-be-detected point is a pixel in the first image or the second image, and the to-be-detected point is not included in the matching area, and the matching An area is an area included in the first image and the second image, the first image is captured by a first camera, and the second image is captured by a second camera, the first image and the first The image obtained by capturing the same target at different angles is used to obtain a set of feature values corresponding to the first image and the second image, where the set of feature values includes actual pixels of the matching region.
  • An image feature value configured to find a target pixel point in the matching area according to the actual image feature value of the to-be-detected point and the feature value set, and the actual image feature value of the target pixel point and the to-be-detected
  • the difference between the actual image feature values of the points is smaller than the first preset difference rate value; and the depth of the target pixel points is used as the depth of the to-be-detected point;
  • the sensor is configured to acquire the first image and the second image.
  • a third aspect of an embodiment of the present invention provides a computer storage medium comprising instructions which, when run on a computer, cause the computer to perform the methods described in the above aspects.
  • the first camera and the second camera respectively capture the same target at different angles to obtain the first image and the second image, and the pixels existing only in the first image or the second image
  • a point is called a point to be detected, and an area included in the first image and the second image is referred to as a matching area
  • an actual image feature value of the point to be detected is obtained by detecting an image of the point to be detected, by detecting the first image and the The second image obtains the actual image feature values of each pixel in the matching region to obtain a feature value set, and finds the target pixel according to the actual image feature value and the feature value set of the point to be detected.
  • the difference between the actual image feature value of the target pixel and the actual image feature value of the point to be detected is smaller than the first preset value, and the depth of the target pixel is obtained as the depth of the point to be detected.
  • the point to be tested reduces unnecessary hardware expenses.
  • FIG. 1 is a schematic diagram of an application scenario according to an embodiment of the present invention.
  • FIG. 2A is a flowchart of an embodiment of an image information acquiring method according to an embodiment of the present invention
  • 2B is a schematic technical diagram of depth calculation according to an embodiment of the present invention.
  • 3A is a flowchart of another embodiment of an image information acquiring method according to an embodiment of the present invention.
  • FIG. 3B is a schematic diagram of a technical method for determining a polar line according to an embodiment of the present invention.
  • FIG. 4 is a device diagram of an embodiment of an image processing apparatus according to an embodiment of the present invention.
  • the embodiment of the invention is applicable to the application scenario shown in FIG. 1 .
  • Point a and point b on object A are projected onto sensor 1 through lens 1, and point a and point b are occluded for the line of sight of lens 2, so the effective depths of point a and point b cannot be calculated.
  • an additional camera is used to perform from multiple angles. Shooting to eliminate dead ends, but when there are multiple directions in which objects are blocked, correspondingly, multiple cameras are required, which increases hardware costs.
  • the resulting image reduces the unnecessary hardware expenditure by restoring the to-be-detected point in the occluded portion by finding the target pixel point in the unoccluded portion, that is, the matching region.
  • an embodiment of the image information acquiring method in the embodiment of the present invention includes:
  • the first camera captures the first image
  • the second camera captures the second image. Since the first image and the second image are shot at different angles of the same target, the first image and the second image are The same area and different areas are also included. For convenience of description, the same area in the first image and the second image is referred to as a matching area, and may also be referred to as a coincident area in practical applications.
  • the device may detect an image of the point to be detected by an image analysis method to obtain an actual image feature value of the point to be detected, where the point to be detected is a pixel in the first image or the second image and is not included in the matching area, and
  • the actual image feature value may be an actual ambiguity value or an actual sharpness value, etc., which is not limited herein.
  • f(x, y) represents the gray value of the image at the point (x, y)
  • Nx, Ny respectively represent the width and height of the image
  • s represents the actual point (x, y) Image feature value.
  • the embodiment of the present invention can be applied to a multi-view stereo vision technology, that is, includes at least two
  • the embodiment of the present invention uses two cameras as an example for description.
  • step 202 the first image and the second image are detected by image analysis to obtain a feature value set corresponding to the first image and the second image, wherein the feature value set The actual image feature value of each pixel in the matching area, that is, the area included in the first image and the second image is included.
  • the device obtains the actual image feature value of the point to be detected in step 201, and obtains the feature value set according to step 202.
  • the two processes do not have a sequence relationship.
  • Step 201 may be performed first, or step 202 may be performed first. , or at the same time, specifically not limited here.
  • the device After obtaining the feature value set and the actual image feature value of the point to be detected, the device finds the target pixel point in the matching area according to the two, wherein the actual image feature value of the target pixel point and the actual image feature of the point to be detected
  • the actual image feature values of the pixel of the matching area and the point to be detected for example, the actual pixel point of the matching area is set. If the image feature value is a, and the actual image feature value of the point to be detected is b, the difference rate can be obtained by the formula (ab)/a, and the obtained difference rate is compared with the first preset difference rate value. If the difference is smaller than the first preset difference value, the pixel of the matching area is determined as the target pixel.
  • the depth of the target pixel is used as the depth of the point to be detected.
  • the device may determine the depth of the target pixel according to a preset algorithm, and the target pixel is The depth is taken as the depth of the point to be detected.
  • depth refers to the distance from a point in the scene to the XY plane where the camera center is located.
  • a depth map can be used to represent the depth information of each point in the scene, that is, each pixel in the depth map records the distance from a certain point in the scene to the XY plane where the camera center is located.
  • a special hardware device can be used to actively acquire depth information of each pixel in the image, such as using an infrared pulse light source to transmit a signal to a scene, and then detecting by using an infrared sensor.
  • the image or the plurality of viewpoint images are stereo-matched to restore the depth information of the object, including: (1) performing stereo matching on the image pair to obtain a parallax image of the corresponding point; and (2) calculating the depth according to the relationship between the parallax and the depth of the corresponding point.
  • Z is used to indicate the depth of the pixel point
  • B is used to indicate the distance between the optical center of the first camera and the optical center of the second camera
  • f is used to indicate the focal length of the first camera or the second camera, corresponding to x and x' It is the distance between the pixel point and the projection point of the camera center on the image plane, and the difference between them (x-x') is used to represent the parallax of the pixel point.
  • the device finds the target pixel point in the matching area by using the feature value set corresponding to the first image and the second image and the actual image feature value of the point to be detected, and The depth of the pixel is used as the depth of the point to be detected, and the depth of the point to be detected is calculated, and no additional camera is needed, which reduces unnecessary hardware expenditure.
  • FIG. 3A is a flowchart of another embodiment of an image information acquiring method according to an embodiment of the present invention.
  • steps 301 to 302 in FIG. 3A are similar to steps 201 to 202 in FIG. 2A, and details are not described herein again.
  • step 303 Determine whether the target actual image feature value exists in the feature value set; if yes, execute step 304; if no, perform step 306.
  • step 304 is performed; if not, Then step 306 is performed.
  • the device selects one pixel from the pixel corresponding to the target actual image feature value as the target pixel.
  • the device determines that there is a target actual image feature value in the feature value set, it can be understood that The target actual image feature value may be one or more, and the corresponding pixel point is also one or more, so the device may randomly select one pixel point from the corresponding pixel point as the target pixel point.
  • target pixel points there are various ways to select target pixel points. For example, a pixel point with the smallest difference between the actual image feature values of the point to be detected may be selected as the target pixel point, so the selection of the target pixel point.
  • the method is not limited here.
  • the depth of the target pixel is the depth of the point to be detected.
  • the step 305 in FIG. 3A is similar to the step 204 in FIG. 2A, and details are not described herein again.
  • the device selects one pixel from the matching region, which may be referred to as a first reference point in the embodiment of the present invention, and obtains a reference value of the first reference point, where
  • the reference value includes at least the reference actual image feature value and the reference theoretical image feature value, and the reference actual image feature value of the first reference point may be obtained by image detection technology, and the manner of obtaining the reference actual image feature value of the first reference point is compared with FIG. 2A
  • Step 201 of the illustrated embodiment obtains similar actual image feature values of the point to be detected, and details are not described herein again.
  • the reference theoretical image feature value of the first reference point can be obtained by a preset calculation formula.
  • the device calculates the theoretical image feature value of the point to be detected according to the detected actual image feature value of the to-be-detected point, and the calculation process may include the following process: setting the first reference point Referring to the actual image feature value R1, the reference theoretical image feature value of the first reference point is M1, the actual image feature value of the point to be detected is R2, and the theoretical image feature value of the point to be detected is M2. In practical applications, it can be considered R1/M1 ⁇ R2/M2, so the theoretical image feature value of the point to be detected can be estimated by this formula.
  • the depth of the point to be detected is calculated according to a preset formula.
  • the depth of the point to be detected may be calculated as follows:
  • n the aperture value of the camera
  • c the theoretical ambiguity value of the point to be detected
  • U the depth of the point to be detected
  • F the focal length of the lens of the camera
  • d the focal length of the lens of the camera
  • d is fixed when the camera system is fixed
  • m is the depth of field, where the depth of field can be understood as the distance between the front and back of the subject measured by the camera lens or other imager front edge. Therefore, n, c, F, d, and m are all known, so the depth U of the point to be detected can be calculated.
  • the depth of the point to be detected is verified to ensure the depth of the point to be detected.
  • the area of the matching area in the first image is referred to as a first area
  • the area of the matching area in the second image is referred to as a second area
  • the device may be used in the contour extraction method in the prior art.
  • the purpose of the contour extraction method is to obtain a peripheral contour feature of the image
  • the step of the contour extraction method may include first finding the extracted edge Any point on the contour of the image is used as the starting point, and from this starting point, the starting point field is searched in one direction, and the next contour boundary point of the detected image is continuously found, and finally the complete contour area is obtained, and the The closed edge of the outline area.
  • the first closed edge and the second closed edge match, the first world point and the second world point are found according to the first closed edge or the second closed edge, and the polar plane.
  • the depth of the target intersection is taken as the target value.
  • the device finds the first closed edge in the first region and the second closed edge in the second region, since the number of the first closed edge and the second closed edge may be one or more, it is necessary to determine and A closed edge that matches the second closed edge.
  • the first closed edge and the second closed edge may be matched by a preset matching algorithm, and specifically, the point on the first closed edge may be correlated with the point on the second closed edge, for example, The correlation value of each point on the first closed edge and each point on each second closed edge is accumulated to obtain an accumulated value, and the largest accumulated is found among the second closed edges The second closed edge corresponding to the value is considered to match the first closed edge.
  • the projection point on the imaging plane of the first camera is P 1
  • the projection point on the imaging plane of the second camera is P 2
  • C 1 and C 2 are the optical centers of the first camera and the second camera, respectively, that is, the origin of the camera coordinate system.
  • the line connecting C 1 and C 2 is the baseline.
  • the intersection of the baseline and the imaging plane of the first camera is called the intersection point e 1
  • the intersection point e 1 is the pole of the first camera.
  • the intersection of the baseline and the imaging plane of the second camera is called the intersection point e 2
  • the intersection point e 2 is the first point.
  • the poles of the two cameras which are the projection coordinates of the optical centers C 1 and C 2 of the two cameras on the corresponding camera imaging plane.
  • the triangular plane composed of P, C 1 and C 2 is called the polar plane ⁇ .
  • the intersection line ⁇ and the intersection planes of the two camera imaging planes l 1 and l 2 are called polar lines, and it can be said that l 1 is the polar line corresponding to the point P 1 , and l 2 is the polar line corresponding to the point P 2 .
  • a point M is taken from the first closed edge, and the polar plane formed by the point M, the optical center of the first camera, and the optical center of the second camera is determined.
  • the intersection line of the imaging plane of the first camera is an epipolar line. In the two-dimensional plane, the polar line has at least two intersection points with the first closed edge.
  • the at least two intersection points are referred to as a first world point and a second world point, and in a quadrilateral region composed of four points of the first world point, the second world point, the optical center of the first camera, and the optical center of the second camera, find a point where the diagonal intersects in the quadrilateral
  • the depth of the target intersection point is determined as the target value to verify the depth of the point to be detected. It can be understood that, in practical applications, a point may be taken from the second closed edge to form a plane with the optical center of the first camera and the optical center of the second camera, which is not limited herein.
  • step 314 Verify that the depth of the point to be detected is greater than the target value. If yes, go to step 314; if no, go to step 315.
  • step 314 is performed; when the depth of the point to be detected is not greater than the target value, step 315 is performed.
  • the device confirms that the depth of the point to be detected is reliable, that is, confirms that the depth of the point to be detected passes the verification.
  • the device When the depth of the point to be detected is not greater than the target value, the device will be deeper than the pixel point adjacent to the point to be detected
  • the average value of the degree is taken as the depth of the point to be detected, and the adjacent pixel points include a plurality of pixel points, which may include pixel points in the matching area, and the depth of the pixel points in the matching area is known, and may also include not matching
  • the pixel points in the area it should be noted that the pixel points in the non-matching area are points that have been estimated to have depth and whose depth has been verified, that is, if the pixel points adjacent to the detected point are not in the matching area, and the estimation is performed. If the depth is not verified, the depth of the pixel adjacent to the detected point is calculated, and the depth of the point is not considered.
  • the first image and the second image are combined into a depth image, that is, the image obtained by the matching of the matching regions in the two images.
  • the first region and the first image are After the second area of the two images is matched, the first area of the first image and the second area of the second image are overlapped to obtain a superimposed image, that is, a depth image, where the matching area and the occlusion area are included, and the occlusion is included.
  • the area includes an area in the first image from which the first area is removed, and an area in the second image from which the second area is removed.
  • the device Since the number of gray levels that each pixel of the gray image may have is determined by the depth of the pixel, when the depth distribution of a certain region in the occlusion region does not correspond to the gray value distribution of the corresponding image The device performs smooth pre-processing on the area.
  • the area is referred to as a sub-occlusion area, that is, the depth distribution of the sub-occlusion area does not correspond to the gray value distribution of the sub-occlusion area, for example,
  • the depth distribution of the sub-occlusion area is a progressively increasing distribution, and the gray value distribution of the sub-occlusion area is firstly smaller and then larger, so the device needs to perform smooth pre-processing on the sub-occlusion area, and the sub-occlusion is performed in the sub-occlusion area.
  • a target occlusion point is selected in the region, wherein a depth difference between a depth of the target occlusion point and an adjacent point of the target occlusion point is greater than a preset difference.
  • the device determines the target occlusion point
  • the depth of the adjacent point of the target occlusion point is obtained, and the depth of the adjacent point is averaged to obtain the depth average of the adjacent point, and the device can target the depth average The depth of the occlusion point.
  • the device passes the first image and the second image through steps 315 to 317.
  • the process is an optional step, which is not limited herein.
  • the present invention further provides an apparatus.
  • FIG. 4 it is a device diagram of a device according to an embodiment of the present invention.
  • the device 40 includes a memory 410, a processor 420, and a sensor 430.
  • the memory 410 is configured to store an operation instruction
  • the processor 420 is configured to perform the following steps by calling an operation instruction stored in the memory 410:
  • the sensor 430 is configured to acquire the first image and the second image.
  • processor 420 may also be referred to as a central processing unit (English full name: Central Processing Unit, English abbreviation: CPU).
  • the memory 410 is configured to store operation instructions and data, so that the processor 420 invokes the above operation instructions to implement corresponding operations, and may include a read only memory and a random access memory. A portion of the memory 410 may also include a non-volatile random access memory (English name: Non-Volatile Random Access Memory, English abbreviation: NVRAM).
  • NVRAM Non-Volatile Random Access Memory
  • the apparatus 40 also includes a bus system 440 that couples various components of the device 40, the various components including the sensor 410, the memory 420, and the processor 430, wherein the bus system 440 includes, in addition to the data bus, It can also include a power bus, a control bus, a status signal bus, and the like. However, for clarity of description, various buses are labeled as bus system 440 in the figure.
  • Processor 420 may be an integrated circuit chip with signal processing capabilities. In the implementation process, each step of the above method may pass through the processor 420. The integrated logic of the hardware or the instruction in the form of software is completed.
  • the processor 420 may be a general-purpose processor, a digital signal processor (English name: Digital Signal Processing, English abbreviation: DSP), an application specific integrated circuit (English name: Application Specific Integrated Circuit, English abbreviation: ASIC), ready-made programmable Gate array (English name: Field-Programmable Gate Array, English abbreviation: FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
  • DSP Digital Signal Processing
  • ASIC Application Specific Integrated Circuit
  • FPGA ready-made programmable Gate array
  • the methods, steps, and logical block diagrams disclosed in the embodiments of the present invention may be implemented or carried out.
  • the general purpose processor may be a microprocessor or the processor or any conventional processor or the like.
  • the steps of the method disclosed in the embodiments of the present invention may be directly implemented by the hardware decoding processor, or may be performed by a combination of hardware and software modules in the decoding processor.
  • the software modules can be located in a conventional computer storage medium of the art, such as random access memory, flash memory, read only memory, programmable read only memory or electrically erasable programmable memory, registers, and the like.
  • the computer storage medium is located in memory 410, and processor 420 reads the information in memory 410 and, in conjunction with its hardware, performs the steps of the above method.
  • the specific implementation of the processor 420 to find the target pixel in the matching area according to the first actual image feature value and the feature value set may be:
  • the target actual image feature value exists in the feature value set, the difference between the target actual image feature value and the actual image feature value is less than the first preset difference rate value; if yes, from the One pixel of the pixel corresponding to the target actual image feature value is selected as the target pixel.
  • processor 420 may also invoke an operation instruction in the memory 410 to perform the following steps:
  • the reference theoretical image feature value of the first reference point is obtained according to a preset calculation formula, and the first reference point is a pixel point in the matching region; according to the first reference Calculating a theoretical image feature value of the point to be detected by the reference value of the point and the actual image feature value of the point to be detected, the reference value of the first reference point includes the reference actual image feature value of the first reference point and the reference theoretical image feature value; The theoretical image feature value of the detection point is calculated to obtain the depth of the point to be detected.
  • processor 420 may also invoke an operation instruction in the memory 410 to perform the following steps:
  • the average value of the depths of the pixel points adjacent to the point to be detected is taken as the depth of the point to be detected.
  • processor 420 may also invoke an operation instruction in the memory 410 to perform the following steps:
  • the first closed edge and the second closed edge are respectively correspondingly found in the first region and the second region by the contour extraction method, the first region is a region of the matching region in the first image, and the second region is a matching region in the first The area in the second image;
  • the first closed edge and the second closed edge match, the first world point and the second world point are found according to the first closed edge or the second closed edge, and the polar plane;
  • the depth of the target intersection is taken as the target value.
  • processor 420 may also invoke an operation instruction in the memory 410 to perform the following steps:
  • the depth image including a matching area and an occlusion area, the occlusion area including an area in the first image from which the first area is removed and an area in the second image from which the second area is removed;
  • the occlusion region is preprocessed, and the sub-occlusion region is included in the occlusion region.
  • the specific implementation of the pre-processing of the sub-occlusion area by the processor 420 in the above embodiment may be:
  • the depth average of the adjacent points of the target occlusion point is taken as the depth of the target occlusion point.
  • the first camera and the second camera respectively capture the same target at different angles to obtain the first image and the second image, and the pixels existing only in the first image or the second image are referred to as to be detected.
  • a point, an area included in the first image and the second image is referred to as a matching area, and an actual image feature value of the point to be detected is obtained by detecting an image of the point to be detected, by detecting the first image
  • the difference between the actual image feature value and the point to be detected is smaller than the first preset value, and the depth of the target pixel is obtained as the depth of the point to be detected.
  • the camera is not required to be added to eliminate the occluded portion. That is, only the portion of the first image or the second image exists, and the target pixel to be detected in the occluded portion is restored by finding the target pixel in the unoccluded portion, that is, the matching region, thereby reducing unnecessary hardware expenditure. .
  • the disclosed system, apparatus, and method may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
  • the integrated unit if implemented in the form of a software functional unit and sold or used as a standalone product, may be stored in a computer readable storage medium.
  • the medium includes instructions for causing a computer device (which may be a personal computer, server, or network device, etc.) to perform all or part of the steps of the methods described in various embodiments of the present invention.
  • the foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)
  • Studio Devices (AREA)

Abstract

Disclosed in an embodiment of the present invention are a method for acquiring image information, an image processing device, and a computer storage medium, which are used during multi-view stereoscopic vision detection so as to reduce unnecessary hardware expenditure. The method in the embodiment of the present invention comprises: acquiring an actual image feature value of a point to be detected, the point to be detected being a pixel point in a first image or a second image, and the point to be detected not being contained within a matching region, wherein the matching region is a region comprised in both the first image and the second image; acquiring a feature value set corresponding to the first image and the second image, the feature value set comprising actual image feature values of each pixel point in the matching region; according to the actual image feature value of the point to be detected and the feature value set, searching in the matching region for a target pixel point, wherein the rate of difference between an actual image feature value of the target pixel point and the actual image feature value of the point to be detected is less than a first pre-configured rate of difference value; using the depth of the target pixel point as the depth of the point to be detected.

Description

一种图像信息获取方法、图像处理设备及计算机存储介质Image information acquisition method, image processing device and computer storage medium 技术领域Technical field
本发明属于信息分析技术领域,尤其涉及一种图像信息获取方法、图像处理设备及计算机存储介质。The invention belongs to the technical field of information analysis, and in particular relates to an image information acquisition method, an image processing device and a computer storage medium.
背景技术Background technique
双目立体视觉是计算机视觉的一个重要分支,双目立体视觉是模拟人类视觉原理,使用计算机被动感知距离的方法,即运用两个相同的摄像机对同一物体从不同的位置成像,获得物体的立体图像对,根据图像之间的像素匹配关系,通过三角测量原理计算出像素之间的偏移来获取物体的三维信息,得到了物体的深度信息,就可以计算出物体与相机之间的实际距离,物体三维大小,两点之间实际距离。Binocular stereo vision is an important branch of computer vision. Binocular stereo vision is a method of simulating the principle of human vision. It uses a computer to passively perceive distance. It uses two identical cameras to image the same object from different positions to obtain the stereoscopic view of the object. The image pair, according to the pixel matching relationship between the images, calculates the offset between the pixels by the principle of triangulation to obtain the three-dimensional information of the object, and obtains the depth information of the object, and can calculate the actual distance between the object and the camera. The three-dimensional size of the object, the actual distance between the two points.
然而实际应用中,当物体出现遮挡时,双摄像头视线被阻挡,无法有效计算深度,导致视觉误差很大。现有技术中,一般采用增加摄像头来从多角度进行成像拍摄,再根据多个摄像头在多个角度拍摄得到的多张图片来恢复被遮挡像素的深度信息,进而得到物体被遮挡部分的深度以消除死角。However, in practical applications, when an object is occluded, the line of sight of the dual camera is blocked, and the depth cannot be effectively calculated, resulting in a large visual error. In the prior art, the camera is generally used to perform imaging shooting from multiple angles, and then the depth information of the occluded pixels is restored according to multiple images captured by multiple cameras at multiple angles, thereby obtaining the depth of the occluded portion of the object. Eliminate dead ends.
但是,现有技术中增加摄像头来消除死角,当物体被遮挡的方向出现多个时,相应的也需要在物体被遮挡的方向上增加多个摄像头,增添了硬件成本。However, in the prior art, the camera is added to eliminate the dead angle. When there are multiple objects in the direction in which the object is blocked, correspondingly, it is also necessary to add a plurality of cameras in the direction in which the object is blocked, which increases the hardware cost.
发明内容Summary of the invention
本发明实施例提供了一种图像信息获取方法、图像处理设备及计算机存储介质,用于在多目立体视觉检测时减少不必要的硬件开支。Embodiments of the present invention provide an image information acquisition method, an image processing device, and a computer storage medium for reducing unnecessary hardware expenditures in multi-view stereo vision detection.
本发明实施例的第一方面提供一种图像信息获取方法,包括:A first aspect of the embodiments of the present invention provides a method for acquiring image information, including:
获取待检测点的实际图像特征值,所述待检测点为第一图像或者第二图像中的像素点,所述待检测点不包含在匹配区域内,所述匹配区域为所述第一图像和所述第二图像中都包括的区域,所述第一图像由第一摄像机拍摄,所述第二图像由第二摄像机拍摄,所述第一图像和所述第二图像为在不同角度拍摄同一目标所得到的图像; Acquiring an actual image feature value of the to-be-detected point, where the to-be-detected point is a pixel in the first image or the second image, the to-be-detected point is not included in the matching area, and the matching area is the first image And an area included in the second image, the first image is captured by a first camera, the second image is captured by a second camera, and the first image and the second image are taken at different angles Images obtained from the same target;
获取所述第一图像和所述第二图像对应的特征值集合,所述特征值集合包括所述匹配区域中各像素点的实际图像特征值;Acquiring a set of feature values corresponding to the first image and the second image, where the set of feature values includes actual image feature values of each pixel in the matching region;
根据所述待检测点的实际图像特征值和所述特征值集合在所述匹配区域中找出目标像素点,所述目标像素点的实际图像特征值与所述待检测点的实际图像特征值的差率小于第一预置差率值;Finding a target pixel point in the matching area according to the actual image feature value of the to-be-detected point and the feature value set, an actual image feature value of the target pixel point and an actual image feature value of the to-be-detected point The difference is less than the first preset difference value;
将所述目标像素点的深度作为所述待检测点的深度。The depth of the target pixel is taken as the depth of the point to be detected.
本发明实施例第二方面提供了一种图像处理设备,所述图像处理设备包括:A second aspect of the embodiments of the present invention provides an image processing device, where the image processing device includes:
存储器、处理器和传感器;Memory, processor and sensor;
所述存储器,用于存储操作指令;The memory is configured to store an operation instruction;
所述处理器,用于获取待检测点的实际图像特征值,所述待检测点为第一图像或者第二图像中的像素点,所述待检测点不包含在匹配区域内,所述匹配区域为所述第一图像和所述第二图像中都包括的区域,所述第一图像由第一摄像机拍摄,所述第二图像由第二摄像机拍摄,所述第一图像和所述第二图像为在不同角度拍摄同一目标所得到的图像;用于获取所述第一图像和所述第二图像对应的特征值集合,所述特征值集合包括所述匹配区域中各像素点的实际图像特征值;用于根据所述待检测点的实际图像特征值和所述特征值集合在所述匹配区域中找出目标像素点,所述目标像素点的实际图像特征值与所述待检测点的实际图像特征值的差率小于第一预置差率值;用于将所述目标像素点的深度作为所述待检测点的深度;The processor is configured to acquire an actual image feature value of the to-be-detected point, where the to-be-detected point is a pixel in the first image or the second image, and the to-be-detected point is not included in the matching area, and the matching An area is an area included in the first image and the second image, the first image is captured by a first camera, and the second image is captured by a second camera, the first image and the first The image obtained by capturing the same target at different angles is used to obtain a set of feature values corresponding to the first image and the second image, where the set of feature values includes actual pixels of the matching region. An image feature value, configured to find a target pixel point in the matching area according to the actual image feature value of the to-be-detected point and the feature value set, and the actual image feature value of the target pixel point and the to-be-detected The difference between the actual image feature values of the points is smaller than the first preset difference rate value; and the depth of the target pixel points is used as the depth of the to-be-detected point;
所述传感器,用于获取所述第一图像和所述第二图像。The sensor is configured to acquire the first image and the second image.
本发明实施例第三方面提供了一种计算机存储介质,包括指令,当其在计算机上运行时,使得计算机执行上述各方面所述的方法。A third aspect of an embodiment of the present invention provides a computer storage medium comprising instructions which, when run on a computer, cause the computer to perform the methods described in the above aspects.
本发明实施例提供的技术方案中,第一摄像机和第二摄像机分别在不同的角度对同一目标进行拍摄得到第一图像和第二图像,将仅存在于第一图像或者第二图像中的像素点称为待检测点,第一图像和第二图像均包括的区域称为匹配区域,并通过检测待检测点所在的图像得到该待检测点的实际图像特征值,通过检测第一图像和第二图像得到匹配区域中各像素点的实际图像特征值以得到特征值集合,根据待检测点的实际图像特征值和特征值集合找出目标像素 点,该目标像素点的实际图像特征值与待检测点的实际图像特征值的差率小于第一预设值,获得该目标像素点的深度并作为待检测点的深度,本实施例中,不需要额外增加摄像头来消除被遮挡的部分即只存在于第一图像或第二图像的部分,通过在将未被遮挡的部分即匹配区域中找出目标像素点,来还原被遮挡部分中的待检测点,减少了不必要的硬件开支。In the technical solution provided by the embodiment of the present invention, the first camera and the second camera respectively capture the same target at different angles to obtain the first image and the second image, and the pixels existing only in the first image or the second image A point is called a point to be detected, and an area included in the first image and the second image is referred to as a matching area, and an actual image feature value of the point to be detected is obtained by detecting an image of the point to be detected, by detecting the first image and the The second image obtains the actual image feature values of each pixel in the matching region to obtain a feature value set, and finds the target pixel according to the actual image feature value and the feature value set of the point to be detected. a point, the difference between the actual image feature value of the target pixel and the actual image feature value of the point to be detected is smaller than the first preset value, and the depth of the target pixel is obtained as the depth of the point to be detected. In this embodiment, There is no need to additionally add a camera to eliminate the occluded portion, that is, the portion existing only in the first image or the second image, and restore the occluded portion by finding the target pixel point in the unoccluded portion, that is, the matching region. The point to be tested reduces unnecessary hardware expenses.
附图说明DRAWINGS
图1为本发明实施例的一个应用场景示意图;FIG. 1 is a schematic diagram of an application scenario according to an embodiment of the present invention;
图2A为本发明实施例中图像信息获取方法的一个实施例的流程图;2A is a flowchart of an embodiment of an image information acquiring method according to an embodiment of the present invention;
图2B为本发明实施例深度计算的一个技术示意图;2B is a schematic technical diagram of depth calculation according to an embodiment of the present invention;
图3A为本发明实施例中图像信息获取方法的另一实施例的流程图;3A is a flowchart of another embodiment of an image information acquiring method according to an embodiment of the present invention;
图3B为本发明实施例中确定极线的一个技术示意图;FIG. 3B is a schematic diagram of a technical method for determining a polar line according to an embodiment of the present invention; FIG.
图4为本发明实施例中图像处理设备的一个实施例的装置图。4 is a device diagram of an embodiment of an image processing apparatus according to an embodiment of the present invention.
具体实施方式Detailed ways
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention are clearly and completely described in the following with reference to the accompanying drawings in the embodiments of the present invention. It is obvious that the described embodiments are only a part of the embodiments of the present invention, but not all embodiments. All other embodiments obtained by a person skilled in the art based on the embodiments of the present invention without creative efforts are within the scope of the present invention.
本发明的说明书和权利要求书及上述附图中的术语“第一”、“第二”、“第三”、“第四”等(如果存在)是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便这里描述的实施例能够以除了在这里图示或描述的内容以外的顺序实施。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元的过程、方法、系统、产品或设备不必限于清楚地列出的那些步骤或单元,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它步骤或单元。The terms "first", "second", "third", "fourth", etc. (if present) in the specification and claims of the present invention and the above figures are used to distinguish similar objects without having to use To describe a specific order or order. It is to be understood that the data so used may be interchanged where appropriate so that the embodiments described herein can be implemented in a sequence other than what is illustrated or described herein. In addition, the terms "comprises" and "comprises" and "the" and "the" are intended to cover a non-exclusive inclusion, for example, a process, method, system, product, or device that comprises a series of steps or units is not necessarily limited to Those steps or units may include other steps or units not explicitly listed or inherent to such processes, methods, products or devices.
本发明实施例适用于图1所示的应用场景。物体A上的点a和点b通过镜头1投射到传感器1上,而针对镜头2的视线,点a和点b被遮挡,因此无法计算点a和点b的有效深度。现有技术中,采用增加摄像头来从多角度进行 拍摄以消除死角,然而当物体被遮挡的方向有多个时,相应的也需要增加多个摄像头,增加了硬件成本。The embodiment of the invention is applicable to the application scenario shown in FIG. 1 . Point a and point b on object A are projected onto sensor 1 through lens 1, and point a and point b are occluded for the line of sight of lens 2, so the effective depths of point a and point b cannot be calculated. In the prior art, an additional camera is used to perform from multiple angles. Shooting to eliminate dead ends, but when there are multiple directions in which objects are blocked, correspondingly, multiple cameras are required, which increases hardware costs.
有鉴于此,本发明实施例中,不需要额外增加摄像头来消除被遮挡的部分即只存在于第一图像或第二图像的部分,该第一图像和第二图像为在不同角度拍摄同一目标所得到的图像,通过在将未被遮挡的部分即匹配区域中找出目标像素点,来还原被遮挡部分中的待检测点,减少了不必要的硬件开支。In view of this, in the embodiment of the present invention, there is no need to additionally add a camera to eliminate the occluded portion, that is, the portion existing only in the first image or the second image, the first image and the second image are the same target at different angles. The resulting image reduces the unnecessary hardware expenditure by restoring the to-be-detected point in the occluded portion by finding the target pixel point in the unoccluded portion, that is, the matching region.
为便于理解,下面对本发明实施例中的具体流程进行描述,请参阅图2A,本发明实施例中图像信息获取方法的一个实施例包括:For ease of understanding, the specific process in the embodiment of the present invention is described below. Referring to FIG. 2A, an embodiment of the image information acquiring method in the embodiment of the present invention includes:
201、获取待检测点的实际图像特征值。201. Acquire an actual image feature value of the point to be detected.
当需要测量距离和恢复3D景物时,双目立体视觉最较为常用的方法之一,它是利用两台摄像机得到同一目标的两幅图像的视差来计算深度。本发明实施例中,第一摄像机拍摄第一图像,第二摄像机拍摄第二图像,由于第一图像和第二图像是对同一目标的不同角度的拍摄,故第一图像和第二图像中有相同的区域,也有不同的区域,为便于表述,将第一图像和第二图像中相同的区域称为匹配区域,实际应用中也可以称为重合区域等。设备可通过图像分析法检测待检测点所在的图像得到待检测点的实际图像特征值,其中,该待检测点为第一图像或者第二图像中的像素点且不包含在匹配区域中,另外,本发明实施例中,实际图像特征值可以为实际模糊度值或者实际锐度值等,具体此处不做限定。One of the most common methods of binocular stereo vision when measuring distance and recovering 3D scenes is to calculate the depth using the parallax of two images of the same target from two cameras. In the embodiment of the present invention, the first camera captures the first image, and the second camera captures the second image. Since the first image and the second image are shot at different angles of the same target, the first image and the second image are The same area and different areas are also included. For convenience of description, the same area in the first image and the second image is referred to as a matching area, and may also be referred to as a coincident area in practical applications. The device may detect an image of the point to be detected by an image analysis method to obtain an actual image feature value of the point to be detected, where the point to be detected is a pixel in the first image or the second image and is not included in the matching area, and In the embodiment of the present invention, the actual image feature value may be an actual ambiguity value or an actual sharpness value, etc., which is not limited herein.
需要说明的是,实际应用中,获取实际图像特征值的方式有多种,例如灰度方差算法,具体可以采用如下公式进行计算:It should be noted that, in practical applications, there are various ways to obtain actual image feature values, such as a gray-scale variance algorithm, which can be calculated by using the following formula:
Figure PCTCN2017094932-appb-000001
Figure PCTCN2017094932-appb-000001
Figure PCTCN2017094932-appb-000002
Figure PCTCN2017094932-appb-000002
其中,
Figure PCTCN2017094932-appb-000003
表示图像的灰度值均值,f(x,y)表示图像在点(x,y)处的灰度值,Nx,Ny分别表示图像的宽度和高度,s表示点(x,y)的实际图像特征值。
among them,
Figure PCTCN2017094932-appb-000003
Indicates the mean value of the gray value of the image, f(x, y) represents the gray value of the image at the point (x, y), Nx, Ny respectively represent the width and height of the image, and s represents the actual point (x, y) Image feature value.
需要说明的是,本发明实施例可应用于多目立体视觉技术,即包括至少两 个摄像机,为便于描述,本发明实施例以两个摄像机为例进行说明。It should be noted that the embodiment of the present invention can be applied to a multi-view stereo vision technology, that is, includes at least two For the convenience of description, the embodiment of the present invention uses two cameras as an example for description.
202、获取第一图像和第二图像对应的特征值集合。202. Acquire a set of feature values corresponding to the first image and the second image.
与步骤201中获取待检测点的实际图像特征值的方式类似,步骤202中通过图像分析检测第一图像和第二图像得到第一图像和第二图像对应的特征值集合,其中该特征值集合包括匹配区域即第一图像和第二图像中都包括的区域中每个像素点的实际图像特征值。Similar to the manner of obtaining the actual image feature value of the point to be detected in step 201, in step 202, the first image and the second image are detected by image analysis to obtain a feature value set corresponding to the first image and the second image, wherein the feature value set The actual image feature value of each pixel in the matching area, that is, the area included in the first image and the second image is included.
需要说明的是,设备通过步骤201获得待检测点的实际图像特征值,根据步骤202获得特征值集合,而这两个过程并不存在先后关系,可以先执行步骤201,也可以先执行步骤202,或者同时执行,具体此处不做限定。It is to be noted that the device obtains the actual image feature value of the point to be detected in step 201, and obtains the feature value set according to step 202. The two processes do not have a sequence relationship. Step 201 may be performed first, or step 202 may be performed first. , or at the same time, specifically not limited here.
203、根据待检测点的实际图像特征值和特征值集合在匹配区域中找出目标像素点。203. Find a target pixel point in the matching area according to the actual image feature value and the feature value set of the point to be detected.
设备获得特征值集合和待检测点的实际图像特征值后,再根据两者在匹配区域中找出目标像素点,其中,目标像素点的实际图像特征值与所述待检测点的实际图像特征值的差率小于第一预置差率值,本发明实施例中,计算匹配区域的像素点与待检测点的实际图像特征值的方式有多种,例如,设匹配区域的像素点的实际图像特征值为a,待检测点的实际图像特征值为b,则差率可以通过公式(a-b)/a得到,并将得到的差率与第一预置差率值进行比较,若得到的差率小于第一预置差率值,则将该匹配区域的像素点确定为目标像素点。After obtaining the feature value set and the actual image feature value of the point to be detected, the device finds the target pixel point in the matching area according to the two, wherein the actual image feature value of the target pixel point and the actual image feature of the point to be detected In the embodiment of the present invention, there are various ways to calculate the actual image feature values of the pixel of the matching area and the point to be detected, for example, the actual pixel point of the matching area is set. If the image feature value is a, and the actual image feature value of the point to be detected is b, the difference rate can be obtained by the formula (ab)/a, and the obtained difference rate is compared with the first preset difference rate value. If the difference is smaller than the first preset difference value, the pixel of the matching area is determined as the target pixel.
204、将目标像素点的深度作为待检测点的深度。204. The depth of the target pixel is used as the depth of the point to be detected.
设备在匹配区域中确定了目标像素点后,由于目标像素点为第一图像和第二图像中包括的像素点,故设备可以根据预置算法确定目标像素点的深度,并将该目标像素点的深度作为待检测点的深度。其中,深度指的是场景中某个点到相机中心所在的XY平面的距离。实际应用中,可以用一张深度图来表示场景中各个点的深度信息,即深度图中的每一个像素记录了场景中的某一个点到相机中心所在的XY平面的距离。另外,确定像素点的深度的方式有多种,例如,可以利用特殊的硬件设备来主动获取图像中每个像素点的深度信息,如利用红外脉冲光源向场景发射信号,然后用红外传感器来检测场景中物体反射回来的红外光,从而确定图像中每一个像素点到摄像机的距离;或者,基于传统的计算机立体视觉方法,通过利用在两个不同的视点来获得的同一景物的两幅 图像或多个视点图像进行立体匹配来恢复物体的深度信息,包括:(1)对图像对进行立体匹配,得到对应点的视差图像;(2)根据对应点的视差与深度的关系计算出深度,从而将视差图像转化为深度图像。因此本发明实施例中,请参照图2B,可以运用如下公式计算像素点的深度:Z=B*f/(x-x’),其中,O和O’分别表示第一摄像头和第二摄像头,Z用于表示像素点的深度,B用于表示第一摄像机的光心与第二摄像机的光心的距离,f用于表示第一摄像机或第二摄像机的焦距,x和x’对应的是像素点和相机中心在图像平面上的投影点的距离,两者的差值即(x-x’)用于表示像素点的视差。After the device determines the target pixel in the matching area, since the target pixel is the pixel included in the first image and the second image, the device may determine the depth of the target pixel according to a preset algorithm, and the target pixel is The depth is taken as the depth of the point to be detected. Where depth refers to the distance from a point in the scene to the XY plane where the camera center is located. In practical applications, a depth map can be used to represent the depth information of each point in the scene, that is, each pixel in the depth map records the distance from a certain point in the scene to the XY plane where the camera center is located. In addition, there are various ways to determine the depth of a pixel. For example, a special hardware device can be used to actively acquire depth information of each pixel in the image, such as using an infrared pulse light source to transmit a signal to a scene, and then detecting by using an infrared sensor. The infrared light reflected back from the object in the scene to determine the distance from each pixel in the image to the camera; or, based on the traditional computer stereo vision method, by using two images of the same scene obtained at two different viewpoints The image or the plurality of viewpoint images are stereo-matched to restore the depth information of the object, including: (1) performing stereo matching on the image pair to obtain a parallax image of the corresponding point; and (2) calculating the depth according to the relationship between the parallax and the depth of the corresponding point. , thereby converting the parallax image into a depth image. Therefore, in the embodiment of the present invention, referring to FIG. 2B, the depth of the pixel point can be calculated by using the following formula: Z=B*f/(x-x'), where O and O' represent the first camera and the second camera, respectively. Z is used to indicate the depth of the pixel point, B is used to indicate the distance between the optical center of the first camera and the optical center of the second camera, and f is used to indicate the focal length of the first camera or the second camera, corresponding to x and x' It is the distance between the pixel point and the projection point of the camera center on the image plane, and the difference between them (x-x') is used to represent the parallax of the pixel point.
从以上技术方案可以看出,本发明实施例中,设备通过第一图像和第二图像对应的特征值集合以及待检测点的实际图像特征值在匹配区域中找出目标像素点,并以目标像素点的深度作为待检测点的深度,计算得到待检测点的深度,不需要额外增加摄像头,减少了不必要的硬件开支。As can be seen from the above technical solution, in the embodiment of the present invention, the device finds the target pixel point in the matching area by using the feature value set corresponding to the first image and the second image and the actual image feature value of the point to be detected, and The depth of the pixel is used as the depth of the point to be detected, and the depth of the point to be detected is calculated, and no additional camera is needed, which reduces unnecessary hardware expenditure.
为便于理解,下面将对本发明实施例的图像信息获取方法进行详细描述。请参阅图3A,图3A为本发明实施例中图像信息获取方法的另一实施例的流程图。For ease of understanding, the image information acquisition method of the embodiment of the present invention will be described in detail below. Referring to FIG. 3A, FIG. 3A is a flowchart of another embodiment of an image information acquiring method according to an embodiment of the present invention.
301、获取待检测点的实际图像特征值。301. Obtain an actual image feature value of the point to be detected.
302、获取第一图像和第二图像对应的特征值集合。302. Acquire a set of feature values corresponding to the first image and the second image.
本发明实施例中,图3A中的步骤301至步骤302与图2A中的步骤201至202类似,此处不再赘述。In the embodiment of the present invention, steps 301 to 302 in FIG. 3A are similar to steps 201 to 202 in FIG. 2A, and details are not described herein again.
303、判断特征值集合中是否存在目标实际图像特征值;若是,则执行步骤304;若否,则执行步骤306。303. Determine whether the target actual image feature value exists in the feature value set; if yes, execute step 304; if no, perform step 306.
在获得待检测图像的实际图像特征值和特征值集合后,判断特征值集合中是否存在目标实际图像特征值,其中该目标实际图像特征值与待检测点的实际图像特征值的差率小于第一预设差率值,即设备判断目标实际图像特征值与待检测的实际图像特征值相同或者误差范围在可接受范围内,若存在目标实际图像特征值,则执行步骤304;若不存在,则执行步骤306。After obtaining the actual image feature value and the feature value set of the image to be detected, determining whether the target actual image feature value exists in the feature value set, wherein the difference between the target actual image feature value and the actual image feature value of the to-be-detected point is smaller than the first a preset difference value, that is, the device determines that the target actual image feature value is the same as the actual image feature value to be detected or the error range is within an acceptable range. If the target actual image feature value exists, step 304 is performed; if not, Then step 306 is performed.
304、设备从目标实际图像特征值对应的像素点中选择一个像素点作为目标像素点。304. The device selects one pixel from the pixel corresponding to the target actual image feature value as the target pixel.
若设备判断存在特征值集合中存在目标实际图像特征值,可以理解的是, 目标实际图像特征值可以为一个或者多个,则其对应的像素点也为一个或者多个,故设备可以从对应的像素点中随机选择一个像素点作为目标像素点。If the device determines that there is a target actual image feature value in the feature value set, it can be understood that The target actual image feature value may be one or more, and the corresponding pixel point is also one or more, so the device may randomly select one pixel point from the corresponding pixel point as the target pixel point.
需要说明的是,实际应用中,选择目标像素点的方式有多种,例如还可以选择与待检测点的实际图像特征值的差率最小的像素点作为目标像素点,故目标像素点的选择方式具体此处不做限定。It should be noted that, in practical applications, there are various ways to select target pixel points. For example, a pixel point with the smallest difference between the actual image feature values of the point to be detected may be selected as the target pixel point, so the selection of the target pixel point. The method is not limited here.
305、将目标像素点的深度以待检测点的深度。305. The depth of the target pixel is the depth of the point to be detected.
本发明实施例中,图3A中的步骤305与图2A中的步骤204类似,此处不再赘述。In the embodiment of the present invention, the step 305 in FIG. 3A is similar to the step 204 in FIG. 2A, and details are not described herein again.
306、获取第一参考点的参考值。306. Obtain a reference value of the first reference point.
当确定特征值集合中不存在目标实际图像特征值时,设备从匹配区域中选取一个像素点,本发明实施例中可以称为第一参考点,并获得该第一参考点的参考值,其中参考值至少包括参考实际图像特征值和参考理论图像特征值,第一参考点的参考实际图像特征值可以通过图像检测技术得到,且获取第一参考点的参考实际图像特征值的方式与图2A所示的实施例中步骤201获取待检测点的实际图像特征值类似,此处不再赘述。另外,实际应用中,第一参考点的参考理论图像特征值可以通过预置的计算公式获得,例如,预置的计算公式可以为如下公式:C=d*F2/(2nU2*m),其中n表示摄像机的光圈值;C表示第一参考点的理论图像特征值,本公式中,理论图像特征值为理论模糊度值;U表示待检测点的深度;F表示摄像机的镜头焦距;d为当摄像机系统固定时的定值;m为景深,其中景深可以理解为在摄像机镜头或其他成像器前沿能够取得清晰成像所测定的被摄物体前后距离范围。When it is determined that the target actual image feature value does not exist in the feature value set, the device selects one pixel from the matching region, which may be referred to as a first reference point in the embodiment of the present invention, and obtains a reference value of the first reference point, where The reference value includes at least the reference actual image feature value and the reference theoretical image feature value, and the reference actual image feature value of the first reference point may be obtained by image detection technology, and the manner of obtaining the reference actual image feature value of the first reference point is compared with FIG. 2A Step 201 of the illustrated embodiment obtains similar actual image feature values of the point to be detected, and details are not described herein again. In addition, in practical applications, the reference theoretical image feature value of the first reference point can be obtained by a preset calculation formula. For example, the preset calculation formula can be the following formula: C=d*F 2 /(2nU 2 *m) Where n represents the aperture value of the camera; C represents the theoretical image feature value of the first reference point, in this formula, the theoretical image feature value is the theoretical ambiguity value; U represents the depth of the point to be detected; F represents the lens focal length of the camera; d is the fixed value when the camera system is fixed; m is the depth of field, where depth of field can be understood as the range of the distance between the front and back of the subject measured by the camera lens or other imager leading edge.
307、根据第一参考点的参考值和待检测点的实际图像特征值计算待检测点的理论图像特征值。307. Calculate a theoretical image feature value of the to-be-detected point according to the reference value of the first reference point and the actual image feature value of the point to be detected.
设备获得了第一参考点的参考值后,再根据检测得到的待检测点的实际图像特征值计算得到待检测点的理论图像特征值,该计算过程可以包括如下过程:设第一参考点的参考实际图像特征值为R1,第一参考点的参考理论图像特征值为M1,待检测点的实际图像特征值为R2,待检测点的理论图像特征值为M2,在实际应用中,可以认为R1/M1≈R2/M2,故通过该公式可以估算出待检测点的理论图像特征值。 After obtaining the reference value of the first reference point, the device calculates the theoretical image feature value of the point to be detected according to the detected actual image feature value of the to-be-detected point, and the calculation process may include the following process: setting the first reference point Referring to the actual image feature value R1, the reference theoretical image feature value of the first reference point is M1, the actual image feature value of the point to be detected is R2, and the theoretical image feature value of the point to be detected is M2. In practical applications, it can be considered R1/M1≈R2/M2, so the theoretical image feature value of the point to be detected can be estimated by this formula.
308、根据待检测点的理论图像特征值计算得到待检测点的深度。308. Calculate a depth of the point to be detected according to the theoretical image feature value of the point to be detected.
设备获得了待检测点的理论图像特征值后,根据预置的公式计算得到待检测点的深度,例如,可以按照如下的方式计算待检测点的深度:After the device obtains the theoretical image feature value of the point to be detected, the depth of the point to be detected is calculated according to a preset formula. For example, the depth of the point to be detected may be calculated as follows:
2ncU2/F2=d/m,其中n表示摄像机的光圈值;c表示待检测点的理论模糊度值;U表示待检测点的深度;F表示摄像机的镜头焦距;d为当摄像机系统固定时的定值;m为景深,其中景深可以理解为在摄像机镜头或其他成像器前沿能够取得清晰成像所测定的被摄物体前后距离范围。故n、c、F、d和m都已知,故可以计算得到待检测点的深度U。2ncU 2 /F 2 =d/m, where n represents the aperture value of the camera; c represents the theoretical ambiguity value of the point to be detected; U represents the depth of the point to be detected; F represents the focal length of the lens of the camera; d is fixed when the camera system is fixed The value of the time; m is the depth of field, where the depth of field can be understood as the distance between the front and back of the subject measured by the camera lens or other imager front edge. Therefore, n, c, F, d, and m are all known, so the depth U of the point to be detected can be calculated.
309、通过轮廓提取方法在第一区域和第二区域中分别对应找到第一闭合边和第二闭合边。309. Find a first closed edge and a second closed edge respectively in the first region and the second region by using a contour extraction method.
设备通过计算得到待检测点的深度后,为确保该待检测点的深度可信,对该待检测点的深度进行验证。在本发明实施例中,将匹配区域在第一图像中的区域称为第一区域,匹配区域在第二图像中的区域称为第二区域,设备可通过现有技术中的轮廓提取方法在第一区域中找出第一闭合边,在第二区域中找出第二闭合边,其中,轮廓提取方法的目的是为了获得图像的外围轮廓特征,轮廓提取方法的步骤可以包括先找到被提取图像轮廓上任意一点作为起始点,并且从这个起始点出发,沿着一个方向,对该起始点领域进行搜索,不断地找到被检测图像下一个轮廓边界点,最终得到完整的轮廓区域,并得到轮廓区域的闭合边。After the device obtains the depth of the point to be detected, the depth of the point to be detected is verified to ensure the depth of the point to be detected. In the embodiment of the present invention, the area of the matching area in the first image is referred to as a first area, and the area of the matching area in the second image is referred to as a second area, and the device may be used in the contour extraction method in the prior art. Finding a first closed edge in the first region and finding a second closed edge in the second region, wherein the purpose of the contour extraction method is to obtain a peripheral contour feature of the image, and the step of the contour extraction method may include first finding the extracted edge Any point on the contour of the image is used as the starting point, and from this starting point, the starting point field is searched in one direction, and the next contour boundary point of the detected image is continuously found, and finally the complete contour area is obtained, and the The closed edge of the outline area.
310、当第一闭合边和第二闭合边匹配时,根据第一闭合边或第二闭合边,和极平面找出第一世界点和第二世界点。310. When the first closed edge and the second closed edge match, the first world point and the second world point are found according to the first closed edge or the second closed edge, and the polar plane.
311、根据第一世界点和第二世界点确定目标交点。311. Determine a target intersection according to the first world point and the second world point.
312、将目标交点的深度作为目标值。312. The depth of the target intersection is taken as the target value.
设备在第一区域中找到第一闭合边、在第二区域中找到第二闭合边后,由于第一闭合边、第二闭合边的个数可以为一个或者多个,故需要确定与各第一闭合边匹配的第二闭合边。实际应用中,可以通过预置的匹配算法确定第一闭合边和第二闭合边匹配,具体可以包括将第一闭合边上的点,与第二闭合边上的点进行相关性计算,例如将第一闭合边上的每个点与各第二闭合边上的每个点的相关性值进行累加,得到累加值,在所述各第二闭合边中找出最大的累加 值所对应的第二闭合边即认为与该第一闭合边匹配。After the device finds the first closed edge in the first region and the second closed edge in the second region, since the number of the first closed edge and the second closed edge may be one or more, it is necessary to determine and A closed edge that matches the second closed edge. In a practical application, the first closed edge and the second closed edge may be matched by a preset matching algorithm, and specifically, the point on the first closed edge may be correlated with the point on the second closed edge, for example, The correlation value of each point on the first closed edge and each point on each second closed edge is accumulated to obtain an accumulated value, and the largest accumulated is found among the second closed edges The second closed edge corresponding to the value is considered to match the first closed edge.
假设第一摄像机和第二摄像机拍摄的目标上有一点P,它在第一摄像机成像平面上的投影点为P1,在第二摄像机成像平面上的投影点为P2,如图3B所示,其中,C1和C2分别为第一摄像机和第二摄像机的光心,即摄像机坐标系的原点。在极线几何中,称C1和C2的连线为基线。基线和第一摄像机成像平面的交点称为交点e1,该交点e1为第一摄像机的极点,同理,基线和第二摄像机成像平面的交点称为交点e2,该交点e2为第二摄像机的极点,它们分别为两个摄像机的光心C1和C2在对应的摄像机成像平面上的投影坐标。P、C1和C2组成的三角平面称为极平面π。π和两个摄像机成像平面的交线l1和l2称为极线,可以称l1为点P1对应的极线,l2为点P2对应的极线。Assuming that the first camera and the second camera have a point P on the target, the projection point on the imaging plane of the first camera is P 1 , and the projection point on the imaging plane of the second camera is P 2 , as shown in FIG. 3B. Wherein C 1 and C 2 are the optical centers of the first camera and the second camera, respectively, that is, the origin of the camera coordinate system. In the polar line geometry, the line connecting C 1 and C 2 is the baseline. The intersection of the baseline and the imaging plane of the first camera is called the intersection point e 1 , and the intersection point e 1 is the pole of the first camera. Similarly, the intersection of the baseline and the imaging plane of the second camera is called the intersection point e 2 , and the intersection point e 2 is the first point. The poles of the two cameras, which are the projection coordinates of the optical centers C 1 and C 2 of the two cameras on the corresponding camera imaging plane. The triangular plane composed of P, C 1 and C 2 is called the polar plane π. The intersection line π and the intersection planes of the two camera imaging planes l 1 and l 2 are called polar lines, and it can be said that l 1 is the polar line corresponding to the point P 1 , and l 2 is the polar line corresponding to the point P 2 .
当第一闭合边和第二闭合边匹配后,从第一闭合边上任取一个点M,确定点M、第一摄像机的光心和第二摄像机的光心组成的极平面,该极平面与第一摄像机成像平面的交线即为极线,在二维平面上,该极线与第一闭合边至少有两个交点,为方便描述,将该至少两个交点称为第一世界点和第二世界点,并在第一世界点、第二世界点、第一摄像机的光心和第二摄像机的光心四个点构成的四边形区域内,找出该四边形中对角线相交的点作为目标交点,并确定该目标交点的深度作为目标值以进行待检测点深度的验证。可以理解的是,实际应用中,也可以从第二闭合边任取一个点来与第一摄像机的光心和第二摄像机的光心组成极平面,具体此处不做限定。After the first closed edge and the second closed edge are matched, a point M is taken from the first closed edge, and the polar plane formed by the point M, the optical center of the first camera, and the optical center of the second camera is determined. The intersection line of the imaging plane of the first camera is an epipolar line. In the two-dimensional plane, the polar line has at least two intersection points with the first closed edge. For convenience of description, the at least two intersection points are referred to as a first world point and a second world point, and in a quadrilateral region composed of four points of the first world point, the second world point, the optical center of the first camera, and the optical center of the second camera, find a point where the diagonal intersects in the quadrilateral As the target intersection point, the depth of the target intersection point is determined as the target value to verify the depth of the point to be detected. It can be understood that, in practical applications, a point may be taken from the second closed edge to form a plane with the optical center of the first camera and the optical center of the second camera, which is not limited herein.
313、验证待检测点的深度是否大于目标值,若是,则执行步骤314;若否,则执行步骤315。313. Verify that the depth of the point to be detected is greater than the target value. If yes, go to step 314; if no, go to step 315.
设备获取到目标值后,根据该目标值验证待检测点的深度是否可信。通过比较该目标值和待检测点的深度,当待检测点的深度大于目标值时,则执行步骤314;当待检测点的深度不大于目标值时,则执行步骤315。After the device obtains the target value, it verifies whether the depth of the point to be detected is reliable according to the target value. When comparing the target value with the depth of the point to be detected, when the depth of the point to be detected is greater than the target value, step 314 is performed; when the depth of the point to be detected is not greater than the target value, step 315 is performed.
314、确认待检测点的深度通过验证。314. Confirm that the depth of the point to be detected passes verification.
当待检测点的深度大于目标值时,则设备确认该待检测点的深度可信,即确认该待检测点的深度通过验证。When the depth of the point to be detected is greater than the target value, the device confirms that the depth of the point to be detected is reliable, that is, confirms that the depth of the point to be detected passes the verification.
315、将与待检测点相邻的像素点的深度的平均值作为待检测点的深度。315. Use an average value of depths of pixel points adjacent to the point to be detected as the depth of the point to be detected.
当待检测点的深度不大于目标值时,设备将与待检测点相邻的像素点的深 度的平均值作为待检测点的深度,该相邻的像素点包括多个像素点,其中可以包括匹配区域中的像素点,且匹配区域中的像素点的深度已知,还可以包括不在匹配区域中的像素点,需要说明的是,该不在匹配区域中的像素点为已经估算出深度且其深度已经通过验证的点,即若待检测点相邻的像素点不在匹配区域中,且估算出的深度没有通过验证,则计算待检测点相邻的像素点的深度的平均值时,该点的深度不做考虑。When the depth of the point to be detected is not greater than the target value, the device will be deeper than the pixel point adjacent to the point to be detected The average value of the degree is taken as the depth of the point to be detected, and the adjacent pixel points include a plurality of pixel points, which may include pixel points in the matching area, and the depth of the pixel points in the matching area is known, and may also include not matching The pixel points in the area, it should be noted that the pixel points in the non-matching area are points that have been estimated to have depth and whose depth has been verified, that is, if the pixel points adjacent to the detected point are not in the matching area, and the estimation is performed. If the depth is not verified, the depth of the pixel adjacent to the detected point is calculated, and the depth of the point is not considered.
另外,可以理解的是,选择与待检测点相邻的像素点的方式有多种,例如,选择与待检测点直接相邻的像素点等,具体此处不做限定。In addition, it can be understood that there are various ways of selecting a pixel point adjacent to the point to be detected, for example, selecting a pixel point directly adjacent to the point to be detected, and the like, which is not limited herein.
316、将第一图像和第二图像合成深度图像。316. Combine the first image and the second image into a depth image.
设备得到待检测点的深度后,将第一图像和第二图像合成深度图像,即将两幅图像中匹配的区域重合所得到的图像,本发明实施例中,第一图像的第一区域与第二图像的第二区域匹配,则将第一图像的第一区域和第二图像的第二区域重合后,得到重合后的图像即深度图像,在该深度图像中包括匹配区域和遮挡区域,遮挡区域中包括第一图像中除去第一区域的区域,以及第二图像中除去第二区域的区域。After the device obtains the depth of the point to be detected, the first image and the second image are combined into a depth image, that is, the image obtained by the matching of the matching regions in the two images. In the embodiment of the present invention, the first region and the first image are After the second area of the two images is matched, the first area of the first image and the second area of the second image are overlapped to obtain a superimposed image, that is, a depth image, where the matching area and the occlusion area are included, and the occlusion is included. The area includes an area in the first image from which the first area is removed, and an area in the second image from which the second area is removed.
317、从子遮挡区域中选择目标遮挡点。317. Select a target occlusion point from the sub-occlusion area.
由于通过像素点的深度能确定灰度图像的每个像素点可能有的灰度级数,因此当遮挡区域中存在某个区域的深度分布与该区域在对应图像的灰度值分布不对应时,设备对该区域进行平滑预处理,本发明实施例中,为方便表述,将该区域称为子遮挡区域,即子遮挡区域的深度分布与子遮挡区域的灰度值分布不对应,例如,子遮挡区域的深度分布为递进式增大分布,而子遮挡区域的灰度值分布为先变小再变大分布,故设备需要对该子遮挡区域进行平滑预处理,并在该子遮挡区域中选择目标遮挡点,其中该目标遮挡点的深度与该目标遮挡点的相邻点的深度差大于预置差值。Since the number of gray levels that each pixel of the gray image may have is determined by the depth of the pixel, when the depth distribution of a certain region in the occlusion region does not correspond to the gray value distribution of the corresponding image The device performs smooth pre-processing on the area. In the embodiment of the present invention, for convenience of description, the area is referred to as a sub-occlusion area, that is, the depth distribution of the sub-occlusion area does not correspond to the gray value distribution of the sub-occlusion area, for example, The depth distribution of the sub-occlusion area is a progressively increasing distribution, and the gray value distribution of the sub-occlusion area is firstly smaller and then larger, so the device needs to perform smooth pre-processing on the sub-occlusion area, and the sub-occlusion is performed in the sub-occlusion area. A target occlusion point is selected in the region, wherein a depth difference between a depth of the target occlusion point and an adjacent point of the target occlusion point is greater than a preset difference.
318、将目标遮挡点的相邻点的深度平均值作为目标遮挡点的深度。318. Use a depth average of adjacent points of the target occlusion point as the depth of the target occlusion point.
设备确定了目标遮挡点后,获得该目标遮挡点的相邻点的深度,并将相邻点的深度做平均值运算,得到相邻点的深度平均值,设备可以将该深度平均值作为目标遮挡点的深度。After the device determines the target occlusion point, the depth of the adjacent point of the target occlusion point is obtained, and the depth of the adjacent point is averaged to obtain the depth average of the adjacent point, and the device can target the depth average The depth of the occlusion point.
其中,需要说明的是,设备通过步骤315至步骤317对第一图像和第二图 像做平滑预处理,实际应用中,该过程为可选步骤,具体此处不做限定。It should be noted that the device passes the first image and the second image through steps 315 to 317. For example, in the actual application, the process is an optional step, which is not limited herein.
本发明还提供了一种设备,请参阅图4,为本发明实施例中设备的装置图,其中所述设备40包括:存储器410、处理器420、传感器430。The present invention further provides an apparatus. Referring to FIG. 4, it is a device diagram of a device according to an embodiment of the present invention. The device 40 includes a memory 410, a processor 420, and a sensor 430.
其中,所述存储器410,用于存储操作指令;The memory 410 is configured to store an operation instruction;
处理器420通过调用存储器410存储的操作指令,用于执行如下步骤:The processor 420 is configured to perform the following steps by calling an operation instruction stored in the memory 410:
获取待检测点的实际图像特征值,所述待检测点为第一图像或者第二图像中的像素点,所述待检测点不包含在匹配区域内,所述匹配区域为所述第一图像和所述第二图像中都包括的区域,所述第一图像由第一摄像机拍摄,所述第二图像由第二摄像机拍摄,所述第一图像和所述第二图像为在不同角度拍摄同一目标所得到的图像;用于获取所述第一图像和所述第二图像对应的特征值集合,所述特征值集合包括所述匹配区域中各像素点的实际图像特征值;用于根据所述待检测点的实际图像特征值和所述特征值集合在所述匹配区域中找出目标像素点,所述目标像素点的实际图像特征值与所述待检测点的实际图像特征值的差率小于第一预置差率值;用于将所述目标像素点的深度作为所述待检测点的深度;Acquiring an actual image feature value of the to-be-detected point, where the to-be-detected point is a pixel in the first image or the second image, the to-be-detected point is not included in the matching area, and the matching area is the first image And an area included in the second image, the first image is captured by a first camera, the second image is captured by a second camera, and the first image and the second image are taken at different angles An image obtained by the same target; a set of feature values corresponding to the first image and the second image, the set of feature values comprising actual image feature values of each pixel in the matching region; The actual image feature value of the to-be-detected point and the feature value set find a target pixel point in the matching area, and the actual image feature value of the target pixel point and the actual image feature value of the to-be-detected point The difference is smaller than the first preset difference value; and the depth of the target pixel is used as the depth of the point to be detected;
传感器430,用于获取第一图像和第二图像。The sensor 430 is configured to acquire the first image and the second image.
需要说明的是,本实施例中,处理器420还可以称为中央处理单元(英文全称:Central Processing Unit,英文缩写:CPU)。It should be noted that, in this embodiment, the processor 420 may also be referred to as a central processing unit (English full name: Central Processing Unit, English abbreviation: CPU).
存储器410,用于存储操作指令和数据,以便处理器420调用上述操作指令实现相应操作,可以包括只读存储器和随机存取存储器。存储器410的一部分还可以包括非易失性随机存取存储器(英文全称:Non-Volatile Random Access Memory,英文缩写:NVRAM)。The memory 410 is configured to store operation instructions and data, so that the processor 420 invokes the above operation instructions to implement corresponding operations, and may include a read only memory and a random access memory. A portion of the memory 410 may also include a non-volatile random access memory (English name: Non-Volatile Random Access Memory, English abbreviation: NVRAM).
所述设备40还包括总线系统440,所述总线系统440将设备40的各个组件耦合在一起,上述各个组件包括传感器410、存储器420、处理器430,其中总线系统440除包括数据总线之外,还可以包括电源总线、控制总线和状态信号总线等。但是为了清楚说明起见,在图中将各种总线都标为总线系统440。The apparatus 40 also includes a bus system 440 that couples various components of the device 40, the various components including the sensor 410, the memory 420, and the processor 430, wherein the bus system 440 includes, in addition to the data bus, It can also include a power bus, a control bus, a status signal bus, and the like. However, for clarity of description, various buses are labeled as bus system 440 in the figure.
本实施例中,还需要说明的是,上述本发明实施例揭示的方法可以应用于处理器420中,或者由处理器420实现。处理器420可能是一种集成电路芯片,具有信号的处理能力。在实现过程中,上述方法的各步骤可以通过处理器420 中的硬件的集成逻辑电路或者软件形式的指令完成。上述的处理器420可以是通用处理器、数字信号处理器(英文全称:Digital Signal Processing,英文缩写:DSP)、专用集成电路(英文全称:Application Specific Integrated Circuit,英文缩写:ASIC)、现成可编程门阵列(英文全称:Field-Programmable Gate Array,英文缩写:FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件。可以实现或者执行本发明实施例中的公开的各方法、步骤及逻辑框图。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。结合本发明实施例所公开的方法的步骤可以直接体现为硬件译码处理器执行完成,或者用译码处理器中的硬件及软件模块组合执行完成。软件模块可以位于随机存储器,闪存、只读存储器,可编程只读存储器或者电可擦写可编程存储器、寄存器等本领域成熟的计算机存储介质中。该计算机存储介质位于存储器410,处理器420读取存储器410中的信息,结合其硬件完成上述方法的步骤。In this embodiment, it should be noted that the method disclosed in the foregoing embodiment of the present invention may be applied to the processor 420 or implemented by the processor 420. Processor 420 may be an integrated circuit chip with signal processing capabilities. In the implementation process, each step of the above method may pass through the processor 420. The integrated logic of the hardware or the instruction in the form of software is completed. The processor 420 may be a general-purpose processor, a digital signal processor (English name: Digital Signal Processing, English abbreviation: DSP), an application specific integrated circuit (English name: Application Specific Integrated Circuit, English abbreviation: ASIC), ready-made programmable Gate array (English name: Field-Programmable Gate Array, English abbreviation: FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components. The methods, steps, and logical block diagrams disclosed in the embodiments of the present invention may be implemented or carried out. The general purpose processor may be a microprocessor or the processor or any conventional processor or the like. The steps of the method disclosed in the embodiments of the present invention may be directly implemented by the hardware decoding processor, or may be performed by a combination of hardware and software modules in the decoding processor. The software modules can be located in a conventional computer storage medium of the art, such as random access memory, flash memory, read only memory, programmable read only memory or electrically erasable programmable memory, registers, and the like. The computer storage medium is located in memory 410, and processor 420 reads the information in memory 410 and, in conjunction with its hardware, performs the steps of the above method.
在上述实施例中处理器420根据第一实际图像特征值和特征值集合在匹配区域中找出目标像素点的具体实现可为:In the foregoing embodiment, the specific implementation of the processor 420 to find the target pixel in the matching area according to the first actual image feature value and the feature value set may be:
判断所述特征值集合中是否存在目标实际图像特征值,所述目标实际图像特征值与所述实际图像特征值的差率小于所述第一预置差率值;若存在,则从所述目标实际图像特征值对应的像素点中选择一个像素点作为所述目标像素点。Determining whether the target actual image feature value exists in the feature value set, the difference between the target actual image feature value and the actual image feature value is less than the first preset difference rate value; if yes, from the One pixel of the pixel corresponding to the target actual image feature value is selected as the target pixel.
在另一个可能的实施例中,处理器420还可以调用存储器410中的操作指令,执行如下步骤:In another possible embodiment, the processor 420 may also invoke an operation instruction in the memory 410 to perform the following steps:
当特征值集合中不存在目标实际图像特征值时,根据预置的计算公式获取第一参考点的参考理论图像特征值,第一参考点为匹配区域内的像素点;根据所述第一参考点的参考值和待检测点的实际图像特征值计算待检测点的理论图像特征值,第一参考点的参考值包括第一参考点的参考实际图像特征值和参考理论图像特征值;根据待检测点的理论图像特征值计算得到所述待检测点的深度。When the target actual image feature value does not exist in the feature value set, the reference theoretical image feature value of the first reference point is obtained according to a preset calculation formula, and the first reference point is a pixel point in the matching region; according to the first reference Calculating a theoretical image feature value of the point to be detected by the reference value of the point and the actual image feature value of the point to be detected, the reference value of the first reference point includes the reference actual image feature value of the first reference point and the reference theoretical image feature value; The theoretical image feature value of the detection point is calculated to obtain the depth of the point to be detected.
在另一个可能的实施例中,处理器420还可以调用存储器410中的操作指令,执行如下步骤: In another possible embodiment, the processor 420 may also invoke an operation instruction in the memory 410 to perform the following steps:
验证待检测点的深度是否大于目标值;Verify that the depth of the point to be detected is greater than the target value;
若是,则确认待检测点的深度通过验证;If yes, confirm that the depth of the point to be detected passes verification;
若否,则将与待检测点相邻的像素点的深度的平均值作为待检测点的深度。If not, the average value of the depths of the pixel points adjacent to the point to be detected is taken as the depth of the point to be detected.
在另一个可能的实施例中,处理器420还可以调用存储器410中的操作指令,执行如下步骤:In another possible embodiment, the processor 420 may also invoke an operation instruction in the memory 410 to perform the following steps:
通过轮廓提取方法在第一区域和第二区域中分别对应找到第一闭合边和第二闭合边,第一区域为匹配区域在第一图像中的区域,第二区域为匹配区域在所述第二图像中的区域;The first closed edge and the second closed edge are respectively correspondingly found in the first region and the second region by the contour extraction method, the first region is a region of the matching region in the first image, and the second region is a matching region in the first The area in the second image;
当第一闭合边和第二闭合边匹配时,根据第一闭合边或第二闭合边,和极平面找出第一世界点和第二世界点;When the first closed edge and the second closed edge match, the first world point and the second world point are found according to the first closed edge or the second closed edge, and the polar plane;
根据第一世界点和第二世界点确定目标交点;Determining the target intersection based on the first world point and the second world point;
将目标交点的深度作为目标值。The depth of the target intersection is taken as the target value.
在另一个可能的实施例中,处理器420还可以调用存储器410中的操作指令,执行如下步骤:In another possible embodiment, the processor 420 may also invoke an operation instruction in the memory 410 to perform the following steps:
将第一图像和第二图像合成深度图像,深度图像包括匹配区域和遮挡区域,遮挡区域包括第一图像中除去第一区域的区域和第二图像中除去第二区域的区域;Combining the first image and the second image into a depth image, the depth image including a matching area and an occlusion area, the occlusion area including an area in the first image from which the first area is removed and an area in the second image from which the second area is removed;
当子遮挡区域的深度分布与所述子遮挡区域在对应图像的灰度值分布不对应时,对遮挡区域进行预处理,子遮挡区域包含于遮挡区域。When the depth distribution of the sub-occlusion region does not correspond to the gray value distribution of the corresponding image in the sub-occlusion region, the occlusion region is preprocessed, and the sub-occlusion region is included in the occlusion region.
在上述实施例中处理器420对所述子遮挡区域进行预处理的具体实现可为:The specific implementation of the pre-processing of the sub-occlusion area by the processor 420 in the above embodiment may be:
从子遮挡区域中选择目标遮挡点,目标遮挡点与目标遮挡点的相邻点的深度差大于预置差值;Selecting a target occlusion point from the sub-occlusion area, and a depth difference between the target occlusion point and an adjacent point of the target occlusion point is greater than a preset difference;
将目标遮挡点的相邻点的深度平均值作为目标遮挡点的深度。The depth average of the adjacent points of the target occlusion point is taken as the depth of the target occlusion point.
以上实施例中,第一摄像机和第二摄像机分别在不同的角度对同一目标进行拍摄得到第一图像和第二图像,将仅存在于第一图像或者第二图像中的像素点称为待检测点,第一图像和第二图像均包括的区域称为匹配区域,并通过检测待检测点所在的图像得到该待检测点的实际图像特征值,通过检测第一图像 和第二图像得到匹配区域中各像素点的实际图像特征值以得到特征值集合,根据待检测点的实际图像特征值和特征值集合找出目标像素点,该目标像素点的实际图像特征值与待检测点的实际图像特征值的差率小于第一预设值,获得该目标像素点的深度并作为待检测点的深度,本实施例中,不需要额外增加摄像头来消除被遮挡的部分即只存在于第一图像或第二图像的部分,通过在将未被遮挡的部分即匹配区域中找出目标像素点,来还原被遮挡部分中的待检测点,减少了不必要的硬件开支。In the above embodiment, the first camera and the second camera respectively capture the same target at different angles to obtain the first image and the second image, and the pixels existing only in the first image or the second image are referred to as to be detected. a point, an area included in the first image and the second image is referred to as a matching area, and an actual image feature value of the point to be detected is obtained by detecting an image of the point to be detected, by detecting the first image Obtaining an actual image feature value of each pixel in the matching region with the second image to obtain a feature value set, and finding a target pixel point according to the actual image feature value and the feature value set of the to-be-detected point, the actual image feature value of the target pixel point The difference between the actual image feature value and the point to be detected is smaller than the first preset value, and the depth of the target pixel is obtained as the depth of the point to be detected. In this embodiment, the camera is not required to be added to eliminate the occluded portion. That is, only the portion of the first image or the second image exists, and the target pixel to be detected in the occluded portion is restored by finding the target pixel in the unoccluded portion, that is, the matching region, thereby reducing unnecessary hardware expenditure. .
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统,装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。A person skilled in the art can clearly understand that, for the convenience and brevity of the description, the specific working process of the system, the device and the unit described above can refer to the corresponding process in the foregoing method embodiment, and details are not described herein again.
在本申请所提供的几个实施例中,应该理解到,所揭露的系统,装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。In the several embodiments provided by the present application, it should be understood that the disclosed system, apparatus, and method may be implemented in other manners. For example, the device embodiments described above are merely illustrative. For example, the division of the unit is only a logical function division. In actual implementation, there may be another division manner, for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed. In addition, the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit. The above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储 介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本发明各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序代码的介质。The integrated unit, if implemented in the form of a software functional unit and sold or used as a standalone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present invention may contribute to the prior art or all or part of the technical solution may be embodied in the form of a software product stored in a storage. The medium includes instructions for causing a computer device (which may be a personal computer, server, or network device, etc.) to perform all or part of the steps of the methods described in various embodiments of the present invention. The foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like. .
以上所述,以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。 The above embodiments are only used to illustrate the technical solutions of the present invention, and are not intended to be limiting; although the present invention has been described in detail with reference to the foregoing embodiments, those skilled in the art will understand that The technical solutions described in the embodiments are modified, or the equivalents of the technical features are replaced by the equivalents of the technical solutions of the embodiments of the present invention.

Claims (15)

  1. 一种图像信息获取方法,其特征在于,所述方法应用于多目立体视觉技术,所述方法包括:An image information acquisition method, characterized in that the method is applied to a multi-view stereo vision technology, the method comprising:
    获取待检测点的实际图像特征值,所述待检测点为第一图像或者第二图像中的像素点,所述待检测点不包含在匹配区域内,所述匹配区域为所述第一图像和所述第二图像中都包括的区域,所述第一图像由第一摄像机拍摄,所述第二图像由第二摄像机拍摄,所述第一图像和所述第二图像为在不同角度拍摄同一目标所得到的图像;Acquiring an actual image feature value of the to-be-detected point, where the to-be-detected point is a pixel in the first image or the second image, the to-be-detected point is not included in the matching area, and the matching area is the first image And an area included in the second image, the first image is captured by a first camera, the second image is captured by a second camera, and the first image and the second image are taken at different angles Images obtained from the same target;
    获取所述第一图像和所述第二图像对应的特征值集合,所述特征值集合包括所述匹配区域中各像素点的实际图像特征值;Acquiring a set of feature values corresponding to the first image and the second image, where the set of feature values includes actual image feature values of each pixel in the matching region;
    根据所述待检测点的实际图像特征值和所述特征值集合在所述匹配区域中找出目标像素点,所述目标像素点的实际图像特征值与所述待检测点的实际图像特征值的差率小于第一预置差率值;Finding a target pixel point in the matching area according to the actual image feature value of the to-be-detected point and the feature value set, an actual image feature value of the target pixel point and an actual image feature value of the to-be-detected point The difference is less than the first preset difference value;
    将所述目标像素点的深度作为所述待检测点的深度。The depth of the target pixel is taken as the depth of the point to be detected.
  2. 根据权利要求1所述的图像信息获取方法,其特征在于,所述根据所述第一实际图像特征值和所述特征值集合在匹配区域中找出目标像素点包括:The image information acquiring method according to claim 1, wherein the finding the target pixel point in the matching area according to the first actual image feature value and the feature value set comprises:
    判断所述特征值集合中是否存在目标实际图像特征值,所述目标实际图像特征值与所述实际图像特征值的差率小于所述第一预置差率值;Determining whether there is a target actual image feature value in the set of feature values, and a difference between the target actual image feature value and the actual image feature value is smaller than the first preset difference value;
    若存在,则从所述目标实际图像特征值对应的像素点中选择一个像素点作为所述目标像素点。If so, one pixel point is selected from the pixel points corresponding to the target actual image feature value as the target pixel point.
  3. 根据权利要求2所述的图像信息获取方法,其特征在于,所述判断所述特征值集合中是否存在目标实际图像特征值之后,所述方法还包括:The image information obtaining method according to claim 2, wherein after the determining whether the target actual image feature value exists in the feature value set, the method further comprises:
    当判断所述特征值集合中不存在所述目标实际图像特征值时,根据预置的计算公式获取第一参考点的参考理论图像特征值,所述第一参考点为所述匹配区域内的像素点;When it is determined that the target actual image feature value does not exist in the feature value set, obtaining a reference theoretical image feature value of the first reference point according to a preset calculation formula, where the first reference point is within the matching area pixel;
    根据所述第一参考点的参考值和所述待检测点的实际图像特征值计算所述待检测点的理论图像特征值,所述第一参考点的参考值包括所述第一参考点的参考实际图像特征值和所述参考理论图像特征值;Calculating a theoretical image feature value of the to-be-detected point according to a reference value of the first reference point and an actual image feature value of the to-be-detected point, where the reference value of the first reference point includes the first reference point Referring to an actual image feature value and the reference theoretical image feature value;
    根据所述待检测点的理论图像特征值计算得到所述待检测点的深度。 Determining the depth of the point to be detected according to the theoretical image feature value of the point to be detected.
  4. 根据权利要求1所述的图像信息获取方法,其特征在于,所述将所述目标像素点的深度作为所述待检测点的深度之后,所述方法还包括:The image information obtaining method according to claim 1, wherein after the depth of the target pixel is used as the depth of the point to be detected, the method further includes:
    验证所述待检测点的深度是否大于目标值;Verifying whether the depth of the point to be detected is greater than a target value;
    若是,则确认所述待检测点的深度通过验证;If yes, confirm that the depth of the point to be detected passes verification;
    若否,则将与所述待检测点相邻的像素点的深度的平均值作为所述待检测点的深度。If not, the average value of the depths of the pixel points adjacent to the point to be detected is taken as the depth of the point to be detected.
  5. 根据权利要求4所述的图像信息获取方法,其特征在于,所述验证所述待检测点的深度是否大于目标值之前,所述方法还包括:The image information obtaining method according to claim 4, wherein the method further comprises: before the verifying whether the depth of the point to be detected is greater than a target value, the method further comprising:
    通过轮廓提取方法在第一区域和第二区域中分别对应找到第一闭合边和第二闭合边,所述第一区域为所述匹配区域在所述第一图像中的区域,所述第二区域为所述匹配区域在所述第二图像中的区域;Finding a first closed edge and a second closed edge respectively in the first area and the second area by a contour extraction method, the first area being an area of the matching area in the first image, the second An area is an area of the matching area in the second image;
    当所述第一闭合边和所述第二闭合边匹配时,根据所述第一闭合边或所述第二闭合边,和极平面找出第一世界点和第二世界点;When the first closed edge and the second closed edge match, the first world point and the second world point are found according to the first closed edge or the second closed edge, and the polar plane;
    根据所述第一世界点和所述第二世界点确定目标交点;Determining a target intersection according to the first world point and the second world point;
    将所述目标交点的深度作为所述目标值。The depth of the target intersection is taken as the target value.
  6. 根据权利要求5中任一项所述的图像信息获取方法,其特征在于,所述获得所述目标像素点的深度以作为所述待检测点的深度之后,所述方法还包括:The image information obtaining method according to any one of claims 5 to 5, wherein after the obtaining the depth of the target pixel point as the depth of the point to be detected, the method further comprises:
    将所述第一图像和所述第二图像合成深度图像,所述深度图像包括所述匹配区域和遮挡区域,所述遮挡区域包括所述第一图像中除去所述第一区域的区域和所述第二图像中除去所述第二区域的区域;Combining the first image and the second image into a depth image, the depth image including the matching area and an occlusion area, the occlusion area including an area and a portion of the first image from which the first area is removed Removing the region of the second region from the second image;
    当子遮挡区域的深度分布与所述子遮挡区域在对应图像的灰度值分布不对应时,对所述遮挡区域进行预处理,所述子遮挡区域包含于所述遮挡区域。When the depth distribution of the sub-occlusion region does not correspond to the gray value distribution of the corresponding image in the sub-occlusion region, the occlusion region is pre-processed, and the sub-occlusion region is included in the occlusion region.
  7. 根据权利要求6所述的图像信息获取方法,其特征在于,所述对所述子遮挡区域进行预处理包括:The image information obtaining method according to claim 6, wherein the preprocessing the sub-occlusion area comprises:
    从所述子遮挡区域中选择目标遮挡点,所述目标遮挡点与所述目标遮挡点的相邻点的深度差大于预置差值;Selecting a target occlusion point from the sub-occlusion area, wherein a depth difference between the target occlusion point and an adjacent point of the target occlusion point is greater than a preset difference;
    将所述目标遮挡点的相邻点的深度平均值作为所述目标遮挡点的深度。The depth average of the adjacent points of the target occlusion point is taken as the depth of the target occlusion point.
  8. 一种图像处理设备,其特征在于,所述图像处理设备包括: An image processing apparatus, characterized in that the image processing apparatus comprises:
    存储器、处理器和传感器;Memory, processor and sensor;
    所述存储器,用于存储操作指令;The memory is configured to store an operation instruction;
    所述处理器,用于获取待检测点的实际图像特征值,所述待检测点为第一图像或者第二图像中的像素点,所述待检测点不包含在匹配区域内,所述匹配区域为所述第一图像和所述第二图像中都包括的区域,所述第一图像由第一摄像机拍摄,所述第二图像由第二摄像机拍摄,所述第一图像和所述第二图像为在不同角度拍摄同一目标所得到的图像;用于获取所述第一图像和所述第二图像对应的特征值集合,所述特征值集合包括所述匹配区域中各像素点的实际图像特征值;用于根据所述待检测点的实际图像特征值和所述特征值集合在所述匹配区域中找出目标像素点,所述目标像素点的实际图像特征值与所述待检测点的实际图像特征值的差率小于第一预置差率值;用于将所述目标像素点的深度作为所述待检测点的深度;The processor is configured to acquire an actual image feature value of the to-be-detected point, where the to-be-detected point is a pixel in the first image or the second image, and the to-be-detected point is not included in the matching area, and the matching An area is an area included in the first image and the second image, the first image is captured by a first camera, and the second image is captured by a second camera, the first image and the first The image obtained by capturing the same target at different angles is used to obtain a set of feature values corresponding to the first image and the second image, where the set of feature values includes actual pixels of the matching region. An image feature value, configured to find a target pixel point in the matching area according to the actual image feature value of the to-be-detected point and the feature value set, and the actual image feature value of the target pixel point and the to-be-detected The difference between the actual image feature values of the points is smaller than the first preset difference rate value; and the depth of the target pixel points is used as the depth of the to-be-detected point;
    所述传感器,用于获取所述第一图像和所述第二图像。The sensor is configured to acquire the first image and the second image.
  9. 根据权利要求8所述的图像处理设备,其特征在于,所述处理器用于:The image processing device according to claim 8, wherein said processor is configured to:
    判断所述特征值集合中是否存在目标实际图像特征值,所述目标实际图像特征值与所述实际图像特征值的差率小于所述第一预置差率值;若存在,则从所述目标实际图像特征值对应的像素点中选择一个像素点作为所述目标像素点。Determining whether the target actual image feature value exists in the feature value set, the difference between the target actual image feature value and the actual image feature value is less than the first preset difference rate value; if yes, from the One pixel of the pixel corresponding to the target actual image feature value is selected as the target pixel.
  10. 根据权利要求9所述的图像处理设备,其特征在于,所述处理器还用于:The image processing device according to claim 9, wherein the processor is further configured to:
    当所述特征值集合中不存在目标实际图像特征值时,根据预置的计算公式获取第一参考点的参考理论图像特征值,所述第一参考点为所述匹配区域内的像素点;根据所述第一参考点的参考值和所述待检测点的实际图像特征值计算所述待检测点的理论图像特征值,所述第一参考点的参考值包括所述第一参考点的参考实际图像特征值和所述参考理论图像特征值;根据所述待检测点的理论图像特征值计算得到所述待检测点的深度。When the target actual image feature value does not exist in the feature value set, the reference theoretical image feature value of the first reference point is obtained according to a preset calculation formula, where the first reference point is a pixel point in the matching area; Calculating a theoretical image feature value of the to-be-detected point according to a reference value of the first reference point and an actual image feature value of the to-be-detected point, where the reference value of the first reference point includes the first reference point Referring to the actual image feature value and the reference theoretical image feature value; calculating the depth of the point to be detected according to the theoretical image feature value of the point to be detected.
  11. 根据权利要求9所述的图像处理设备,其特征在于,所述处理器还用于:The image processing device according to claim 9, wherein the processor is further configured to:
    验证所述待检测点的深度是否大于目标值;若是,则确认所述待检测点的 深度通过验证;若否,则将与所述待检测点相邻的像素点的深度的平均值作为所述待检测点的深度。Verifying whether the depth of the point to be detected is greater than a target value; if yes, confirming the point to be detected The depth passes the verification; if not, the average value of the depths of the pixel points adjacent to the point to be detected is taken as the depth of the point to be detected.
  12. 根据权利要求11所述的图像处理设备,其特征在于,所述处理器还用于:The image processing device according to claim 11, wherein the processor is further configured to:
    通过轮廓提取方法在第一区域和第二区域中分别对应找到第一闭合边和第二闭合边,所述第一区域为所述匹配区域在所述第一图像中的区域,所述第二区域为所述匹配区域在所述第二图像中的区域;确定所述第一闭合边和所述第二闭合边匹配;当所述第一闭合边和所述第二闭合边匹配时,根据所述第一闭合边或所述第二闭合边,和极平面找出第一世界点和第二世界点;根据所述第一世界点和所述第二世界点确定目标交点;将目标交点的深度作为所述目标值。Finding a first closed edge and a second closed edge respectively in the first area and the second area by a contour extraction method, the first area being an area of the matching area in the first image, the second a region is an area of the matching area in the second image; determining that the first closed edge and the second closed edge match; when the first closed edge and the second closed edge match, according to Determining a first world point and a second world point by the first closed edge or the second closed edge, and determining a target intersection point according to the first world point and the second world point; The depth is taken as the target value.
  13. 根据权利要求12中任一项所述的图像处理设备,其特征在于,所述处理器还用于:The image processing device according to any one of claims 12 to 12, wherein the processor is further configured to:
    将所述第一图像和所述第二图像合成深度图像,所述深度图像包括所述匹配区域和遮挡区域,所述遮挡区域包括所述第一图像中除去所述第一区域的区域和所述第二图像中除去所述第二区域的区域;当子遮挡区域的深度分布与所述子遮挡区域在对应图像的灰度值分布不对应时,对所述遮挡区域进行预处理,所述子遮挡区域包含于所述遮挡区域。Combining the first image and the second image into a depth image, the depth image including the matching area and an occlusion area, the occlusion area including an area and a portion of the first image from which the first area is removed Removing the region of the second region in the second image; pre-processing the occlusion region when the depth distribution of the sub-occlusion region does not correspond to the gray value distribution of the corresponding image in the sub-occlusion region, A sub-occlusion area is included in the occlusion area.
  14. 根据权利要求13所述的图像处理设备,其特征在于,所述处理器具体用于:The image processing device according to claim 13, wherein the processor is specifically configured to:
    从所述子遮挡区域中选择目标遮挡点,所述目标遮挡点与所述目标遮挡点的相邻点的深度差大于预置差值;将所述目标遮挡点的相邻点的深度平均值作为所述目标遮挡点的深度。Selecting a target occlusion point from the sub-occlusion area, a depth difference between the target occlusion point and an adjacent point of the target occlusion point is greater than a preset difference; and a depth average of adjacent points of the target occlusion point The depth of the target occlusion point.
  15. 一种计算机存储介质,包括指令,当其在计算机上运行时,使得计算机执行如权利要求1-7任一项所述的图像信息获取方法。 A computer storage medium comprising instructions which, when run on a computer, cause the computer to perform the image information acquisition method of any of claims 1-7.
PCT/CN2017/094932 2017-07-28 2017-07-28 Method for acquiring image information, image processing device, and computer storage medium WO2019019160A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2017/094932 WO2019019160A1 (en) 2017-07-28 2017-07-28 Method for acquiring image information, image processing device, and computer storage medium
CN201780092646.9A CN110800020B (en) 2017-07-28 2017-07-28 Image information acquisition method, image processing equipment and computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/094932 WO2019019160A1 (en) 2017-07-28 2017-07-28 Method for acquiring image information, image processing device, and computer storage medium

Publications (1)

Publication Number Publication Date
WO2019019160A1 true WO2019019160A1 (en) 2019-01-31

Family

ID=65039334

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/094932 WO2019019160A1 (en) 2017-07-28 2017-07-28 Method for acquiring image information, image processing device, and computer storage medium

Country Status (2)

Country Link
CN (1) CN110800020B (en)
WO (1) WO2019019160A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117575886A (en) * 2024-01-15 2024-02-20 之江实验室 Image edge detector, detection method, electronic equipment and medium

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111627061B (en) * 2020-06-03 2023-07-11 如你所视(北京)科技有限公司 Pose detection method and device, electronic equipment and storage medium
CN113484852B (en) * 2021-07-07 2023-11-07 烟台艾睿光电科技有限公司 Distance measurement method and system

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102204262A (en) * 2008-10-28 2011-09-28 皇家飞利浦电子股份有限公司 Generation of occlusion data for image properties
US20130135439A1 (en) * 2011-11-29 2013-05-30 Fujitsu Limited Stereoscopic image generating device and stereoscopic image generating method
CN103679739A (en) * 2013-12-26 2014-03-26 清华大学 Virtual view generating method based on shielding region detection
CN104063702A (en) * 2014-07-16 2014-09-24 中南大学 Three-dimensional gait recognition based on shielding recovery and partial similarity matching
CN104574331A (en) * 2013-10-22 2015-04-29 中兴通讯股份有限公司 Data processing method, device, computer storage medium and user terminal
CN105184780A (en) * 2015-08-26 2015-12-23 京东方科技集团股份有限公司 Prediction method and system for stereoscopic vision depth
CN105279786A (en) * 2014-07-03 2016-01-27 顾海松 Method and system for obtaining object three-dimensional model
CN106355570A (en) * 2016-10-21 2017-01-25 昆明理工大学 Binocular stereoscopic vision matching method combining depth characteristics

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102204262A (en) * 2008-10-28 2011-09-28 皇家飞利浦电子股份有限公司 Generation of occlusion data for image properties
US20130135439A1 (en) * 2011-11-29 2013-05-30 Fujitsu Limited Stereoscopic image generating device and stereoscopic image generating method
CN104574331A (en) * 2013-10-22 2015-04-29 中兴通讯股份有限公司 Data processing method, device, computer storage medium and user terminal
CN103679739A (en) * 2013-12-26 2014-03-26 清华大学 Virtual view generating method based on shielding region detection
CN105279786A (en) * 2014-07-03 2016-01-27 顾海松 Method and system for obtaining object three-dimensional model
CN104063702A (en) * 2014-07-16 2014-09-24 中南大学 Three-dimensional gait recognition based on shielding recovery and partial similarity matching
CN105184780A (en) * 2015-08-26 2015-12-23 京东方科技集团股份有限公司 Prediction method and system for stereoscopic vision depth
CN106355570A (en) * 2016-10-21 2017-01-25 昆明理工大学 Binocular stereoscopic vision matching method combining depth characteristics

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117575886A (en) * 2024-01-15 2024-02-20 之江实验室 Image edge detector, detection method, electronic equipment and medium
CN117575886B (en) * 2024-01-15 2024-04-05 之江实验室 Image edge detector, detection method, electronic equipment and medium

Also Published As

Publication number Publication date
CN110800020A (en) 2020-02-14
CN110800020B (en) 2021-07-09

Similar Documents

Publication Publication Date Title
US11830141B2 (en) Systems and methods for 3D facial modeling
US9424650B2 (en) Sensor fusion for depth estimation
EP2064675B1 (en) Method for determining a depth map from images, device for determining a depth map
WO2017076106A1 (en) Method and device for image splicing
JP6291469B2 (en) Obstacle detection device and obstacle detection method
JP2018534698A (en) Method and system for large scale determination of RGBD camera attitude
CN106981078B (en) Sight line correction method and device, intelligent conference terminal and storage medium
CN111160232B (en) Front face reconstruction method, device and system
WO2020147346A1 (en) Image recognition method, system and apparatus
KR20130104691A (en) Apparatus and method for image processing
WO2014180255A1 (en) Data processing method, apparatus, computer storage medium and user terminal
CN111882655A (en) Method, apparatus, system, computer device and storage medium for three-dimensional reconstruction
WO2019019160A1 (en) Method for acquiring image information, image processing device, and computer storage medium
CN111046845A (en) Living body detection method, device and system
CN106558038B (en) A kind of detection of sea-level and device
WO2023142352A1 (en) Depth image acquisition method and device, terminal, imaging system and medium
CN111160233B (en) Human face in-vivo detection method, medium and system based on three-dimensional imaging assistance
AU2022375768A1 (en) Methods, storage media, and systems for generating a three-dimensional line segment
JPH06160087A (en) Method and apparatus for measuring distance using image
KR101869226B1 (en) Disparity map generating method for mono cameras
CN111080689B (en) Method and device for determining face depth map
JP2958462B1 (en) Method and apparatus for correcting luminance of stereo image
CN117710467B (en) Unmanned plane positioning method, unmanned plane positioning equipment and aircraft
EP4423725A1 (en) Methods, storage media, and systems for generating a three-dimensional line segment
CN116721109A (en) Half global matching method for binocular vision images

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17919379

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17919379

Country of ref document: EP

Kind code of ref document: A1