WO2018233217A1 - Image processing method, device and augmented reality apparatus - Google Patents

Image processing method, device and augmented reality apparatus Download PDF

Info

Publication number
WO2018233217A1
WO2018233217A1 PCT/CN2017/113578 CN2017113578W WO2018233217A1 WO 2018233217 A1 WO2018233217 A1 WO 2018233217A1 CN 2017113578 W CN2017113578 W CN 2017113578W WO 2018233217 A1 WO2018233217 A1 WO 2018233217A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
matrix
camera
coordinate transformation
processing
Prior art date
Application number
PCT/CN2017/113578
Other languages
French (fr)
Chinese (zh)
Inventor
李祥艳
徐梁栋
Original Assignee
歌尔科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 歌尔科技有限公司 filed Critical 歌尔科技有限公司
Publication of WO2018233217A1 publication Critical patent/WO2018233217A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10048Infrared image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Definitions

  • the present invention relates to the field of image processing technologies, and in particular, to an image processing method, apparatus, and augmented reality device.
  • Augmented Reality (AR) technology is a real-time computing camera position and angle. It combines image processing technology to superimpose the scene of the virtual world into the real world scene and display it to the user. AR technology has real-time interactive, real-world and virtual world information integration and the ability to add positioning virtual objects in 3D scale space, bringing people a new visual experience.
  • the realistic scene in the AR scene is taken by the camera.
  • a certain type of camera is set in the AR device for collecting images of a real scene, such as a charge coupled device (CCD) camera.
  • CCD charge coupled device
  • the sharpness of the image taken at this time tends to be unsatisfactory, so that the image quality that the user finally sees is not good, affecting the user experience.
  • embodiments of the present invention provide an image processing method, apparatus, and augmented reality device, which improve image quality by performing image fusion processing on non-homologous images of the same scene.
  • an embodiment of the present invention provides an image processing method, including:
  • first image and the second image being non-homologous images obtained by capturing the same scene by the first camera and the second camera, respectively;
  • an embodiment of the present invention provides an image processing apparatus, including:
  • a receiving module configured to receive a first image and a second image, where the first image and the second image are non-homologous images obtained by capturing the same scene through the first camera and the second camera, respectively;
  • An acquiring module configured to acquire a coordinate transformation matrix corresponding to the first image, where the coordinate transformation matrix uses the second image as a reference image;
  • a transform module configured to perform coordinate transformation on the first image by using the coordinate transformation matrix
  • a fusion module configured to perform image fusion processing on the coordinate-converted first image and the second image.
  • an embodiment of the present invention provides an augmented reality device, including:
  • the memory is for storing one or more computer instructions, wherein the one or more computer instructions are executed by the processor to implement an image processing method as described above.
  • an embodiment of the present invention provides another augmented reality device, including:
  • the FPGA component includes functional logic that implements the image processing method as described above.
  • An image processing method and apparatus provided by an embodiment of the present invention, two different classes are set in an AR device
  • the first camera and the second camera are the first camera and the second camera, and the first camera and the second camera simultaneously capture the same scene to obtain the first image and the second image that are not homologous; the second image is used as the reference image, and the first image is acquired.
  • a coordinate transformation matrix corresponding to the image the coordinate transformation is performed on the first image by using the coordinate transformation matrix, so that the transformed first image corresponds to each pixel point in the second image; and further, the first image after the coordinate transformation Image fusion processing is performed with the second image. Since the first image and the second image are non-homologous images, the dominant features of the two are different. By combining the two, it is beneficial to fuse the superior features of the two and enhance the image quality after the fusion.
  • FIG. 1 is a flowchart of Embodiment 1 of an image processing method according to an embodiment of the present invention
  • Embodiment 2 is a flowchart of Embodiment 2 of an image processing method according to an embodiment of the present invention
  • FIG. 3 is a flowchart of Embodiment 3 of an image processing method according to an embodiment of the present disclosure
  • FIG. 4 is a schematic structural diagram of Embodiment 1 of an image processing apparatus according to an embodiment of the present disclosure
  • FIG. 5 is a schematic structural diagram of Embodiment 2 of an image processing apparatus according to an embodiment of the present disclosure
  • FIG. 6 is a schematic structural diagram of Embodiment 3 of an image processing apparatus according to an embodiment of the present disclosure.
  • FIG. 7 is a schematic structural diagram of Embodiment 1 of an augmented reality device according to an embodiment of the present disclosure.
  • FIG. 8 is a schematic structural diagram of Embodiment 2 of an augmented reality device according to an embodiment of the present disclosure.
  • FIG. 9 is a schematic structural diagram of a head mounted display device according to an embodiment of the present invention.
  • first, second, third, etc. may be used to describe XXX in embodiments of the invention, these XXX should not be limited to these terms. These terms are only used to distinguish XXX.
  • the first XXX may also be referred to as a second XXX without departing from the scope of the embodiments of the present invention.
  • the second XXX may also be referred to as a first XXX.
  • the words “if” and “if” as used herein may be interpreted to mean “when” or “when” or “in response to determining” or “in response to detecting.”
  • the phrase “if determined” or “if detected (conditions or events stated)” can be interpreted as “when determined” or “in response to determination” or “when detected (stated condition or event) “Time” or “in response to a test (condition or event stated)”.
  • FIG. 1 is a flowchart of Embodiment 1 of an image processing method according to an embodiment of the present invention.
  • the image processing method provided by this embodiment may be implemented by an image processing apparatus, and the image processing apparatus may be implemented as a field programmable gate array ( A combination of some hardware devices in the Field-Programmable Gate Array (FPGA) component, which can be integrated into the AR device.
  • the method includes the following steps:
  • different types of cameras such as the first camera and the second camera
  • the first camera and the second camera are disposed in the AR device, and the first camera and the second camera are respectively disposed on the left and right sides on a horizontal surface.
  • the first camera may be an infrared camera and the second camera may be a CCD camera.
  • the first image captured by the first camera is an infrared image
  • the second image captured by the second camera is a visible light image.
  • the following is an explanation of the image acquisition and image fusion of the same scene by using different types of cameras in combination with an infrared camera and a CCD camera.
  • the infrared camera and the CCD camera simultaneously take photos of the same scene and acquire the infrared image.
  • the fusion with the dominant feature information of the visible light image can finally obtain a fused image with obvious feature points and rich information.
  • the object when the temperature in the natural world is higher than the absolute zero, the object will produce infrared radiation, red.
  • External imaging uses an infrared camera to convert invisible infrared radiation into a visible temperature distribution image.
  • Infrared images are not susceptible to environmental influences, and temperature profiles of objects can be obtained in rain, snow, smoke, and dark environments.
  • the resolution of the infrared camera tends to be low, so that the obtained infrared image has poor definition, the scene detail information is not obvious, and the formed image does not conform to human visual habits.
  • the image acquired by the CCD camera is realized according to the energy of the reflected light of the object, and the image can better describe the detailed information of the scene, and the resolution is high, which is in line with the requirements of the human visual system.
  • CCD cameras also have some shortcomings: in bad weather, the image capture ability is poor, the useful information is lost, and the comprehensive and detailed image information in the description scene cannot be obtained.
  • the infrared image and the visible light image have their own advantages and disadvantages. Therefore, if the fusion algorithm is used to fuse the dominant features of the infrared image and the visible light image, the fused image contains rich feature point information, which is suitable for the human visual system. Will greatly enhance the visual experience of users watching.
  • the image processing method provided by the embodiment of the present invention can be implemented based on the hardware component of the FPGA component, that is, the fusion of the multi-source image is implemented based on the FPGA.
  • FPGAs have rich resources such as storage resources and faster computing speed.
  • the fused video is smoother, and the fused image can be output in real time to achieve a better visual experience.
  • an AR device that integrates the first camera and the second camera described above is often used to capture video images of a real scene. It can be understood that since the first camera and the second camera are used to capture the same scene, it is necessary to ensure the clock synchronization of the two, that is, at the same time, the two cameras are shooting the same object in the scene. However, since the shooting parameters such as the shooting position and shooting angle of the two cameras will be different, even if the same subject is shot, the captured image will often be different.
  • the first camera and the second camera simultaneously input the captured video image into the FPGA component through the video interface of the FPGA component, and after decoding the video decoder chip of the FPGA component, decoding into a YCbCr video image such as the BT.656 format.
  • the image processing method provided by the embodiment of the present invention performs fusion processing on two images corresponding to each time.
  • the description of the image fusion process is performed by taking only the first image and the second image corresponding to any one time as an example.
  • first image and the second image have differences in shooting parameters such as resolution and shooting angle
  • image registration of the first image and the second image is first required. Processing to establish a correspondence between each pixel point in the first image and the second image, so that the fusion of the first image and the second image can be performed based on the correspondence relationship of the pixel points.
  • the description of the image registration processing process is performed by taking the first image as an infrared image and the second image as a visible light image as an example.
  • Image registration processing includes scaling, rotation, and panning of the image. Since the visible light image has higher resolution and is more in line with the human eye visual habit than the infrared image, in this embodiment, the visible light image is used as the reference image, and the infrared image is used as the image to be registered for the infrared image. Zoom, rotate, and pan.
  • the scaling, rotation and translation operations of the infrared image are performed based on the obtained coordinate transformation matrix, that is, the coordinate transformation matrix contains the scaling parameters required for the scaling operation, and the rotation parameters required for the rotation operation, And the translation parameters required for the panning operation. And these scaling parameters, rotation parameters, and translation parameters may be obtained in advance, so that a coordinate transformation matrix can be generated based on these parameters obtained in advance.
  • the reason why the infrared image is transformed by using the coordinate transformation matrix is because the conversion is more efficient than the method of sequentially performing three transformations on the infrared image, because only the conversion is more efficient. It is necessary to use a matrix to perform three transformations for each pixel in the infrared image.
  • the coordinate-converted infrared image can be obtained based on matrix multiplication of the infrared image and the coordinate transformation matrix. Since the transformation parameters in the coordinate transformation matrix are based on the visible light image, the correspondence between the pixel points in the coordinate-converted infrared image and the pixel points in the visible light image can be obtained based on the transformation. Therefore, based on the correspondence, image fusion of the coordinate-converted infrared image and the visible light image can be performed.
  • a pixel-level fusion method is selected: a method of weighted average of gray values, that is, a weighted average calculation of gray values of corresponding pixel points is implemented. The fusion of two images.
  • two different types of cameras that is, a first camera and a second camera are disposed in the AR device, and the first camera and the second camera simultaneously capture the same scene to obtain a non-homologous first image and a second image; taking a second image as a reference image, acquiring a coordinate transformation matrix corresponding to the first image, and performing coordinate transformation on the first image by using the coordinate transformation matrix, so that the transformed first image and the second image are Corresponding to each pixel point; further, image fusion processing is performed on the coordinate-converted first image and the second image. Since the first image and the second image are non-homologous images, the dominant features of the two are different. By combining the two, it is beneficial to fuse the superior features of the two and enhance the image quality after the fusion.
  • FIG. 2 is a flowchart of Embodiment 2 of an image processing method according to an embodiment of the present invention. As shown in FIG. 2, on the basis of the embodiment shown in FIG. 1, after step 103, the following steps may be further included:
  • 201 Receive a first image and a second image, where the first image and the second image are non-homologous images obtained by capturing the same scene by the first camera and the second camera, respectively.
  • the first image and the second image are subjected to certain pre-processing.
  • the first image Take the first image as an infrared image and the second image as a visible light image as an example.
  • the infrared image is imaged according to the thermal radiation of the object, the brightness is too high and is not suitable for the human visual system.
  • the brightness of the infrared image is reduced by performing inverse processing of the gray value on the first image, and the feature points are highlighted.
  • the size of the infrared image Simage1 is M*N
  • the gray value of each pixel is 8 bits, that is, the gray level is divided into 28 or 256 gray levels, and the unit matrix E of M*N is constructed, and the inverse is performed.
  • the infrared image is Simage2
  • the visible light image is an image formed according to the principle of reflection of energy. Since the visible light image is acquired in a harsh environment with low illumination, the screen is dark and the prominent feature points are small. Therefore, it is necessary to perform image enhancement processing on the visible light image. Specifically, the gray value of the pixel of the visible light image may be threshold-divided, and the traditional three-stage image enhancement method is adopted to enhance the image image by stretching the transform coefficient for the pixel points in different threshold ranges. .
  • the coordinate transformation matrix corresponding to the preprocessed infrared image may be acquired based on the preprocessed infrared image and the visible light image.
  • the coordinate transformation matrix may be obtained according to the translation matrix A, the rotation matrix B, and the scaling matrix C, wherein the translation matrix A, the rotation matrix B, and the scaling matrix C are respectively used to represent a translation parameter, a rotation parameter, and a scaling parameter. Therefore, the translation matrix A, the rotation matrix B, and the scaling matrix C are respectively generated according to the translation parameter, the rotation parameter, and the scaling parameter.
  • the coordinate transformation matrix T corresponding to the preprocessed infrared image can be determined as: translation matrix A.
  • the scaling operation of the image is mainly performed for images of different resolutions. Since the resolution of the infrared image and the visible image are different, it is necessary to scale the preprocessed infrared image based on the preprocessed visible image to make the resolution and the resolution of the preprocessed visible image. Consistent.
  • the scaling factor in the X-axis direction is t x
  • the scaling factor in the Y-axis direction is t y , which is obtained after scaling transformation.
  • the scaling parameters t x and t y can be determined according to the resolution of the infrared camera and the CCD camera, that is, the ratio of the X-axis resolutions of the two can determine t x , and the ratio of the two Y-axis resolutions can determine t y . Therefore, when the infrared camera and the CCD camera in the AR device are set, the scaling parameters t x and t y can be determined, and the scaling parameters t x and t y can be pre-stored in the storage space of the FPGA component.
  • the rotation transformation of the image is mainly caused by the angle between the infrared image and the visible image due to human factors when shooting the infrared image and the visible light image, in order to make the corresponding feature points in the two images accurate.
  • the matching needs to be performed on the pre-processed visible light image as a reference, and the pre-processed infrared image is rotated in the two-dimensional space.
  • the pixel point P(x, y) is any pixel in the preprocessed infrared image. After the rotation transformation, the corresponding pixel point is P'(x', y'). If the matrix is used, P'(x) The rotation relationship between ',y') and P(x,y) is:
  • a Cartesian coordinate system is established centering on the origin in the preprocessed infrared image.
  • the angle between the connection of P(x, y) and the origin with the X axis is the first angle;
  • the angle between the connection of P'(x', y') and the origin with the X axis is the second angle, then the second angle
  • the difference from the first angle is ⁇ , which means the angle of deflection between P'(x', y') and P(x, y).
  • the rotation parameter ⁇ can be determined according to the setting of the infrared camera and the CCD camera in the AR device. Specifically, the angle between the lens center and the horizontal surface of the infrared camera and the lens center and the horizontal surface of the CCD camera can be measured. Angle, the angle difference between the two angles is the rotation parameter ⁇ . Therefore, when the infrared camera and the CCD camera in the AR device are set, the rotation parameter ⁇ can be determined, and the rotation parameter ⁇ can be pre-stored in the storage space of the FPGA component.
  • the translation parameters unlike the above rotation parameters and scaling parameters, when the translation transformation operation of the infrared image is required, the translation parameters need to be calculated based on the current infrared image and the visible light image. That is to say, the rotation parameter and the scaling parameter can be considered to be independent of the currently captured image, and do not depend on the currently captured image determination, but the translation parameter is related to the currently captured image and needs to be dependent on the current The captured image is determined.
  • the determination of the translation parameter needs to involve a responsible calculation process.
  • the image processing method provided by the embodiment of the present invention can be implemented based on the hardware component of the FPGA component. If the FPGA component is used to calculate the translation parameter, the limitation is limited. Therefore, optionally, the translation parameter can be calculated based on the image registration processing component, and the image registration processing component can be implemented as a software program, and the translation processing is performed by the image registration processing component to obtain a translation parameter and feedback to the FPGA component. So that the FPGA component generates the corresponding translation matrix A.
  • the image registration processing component mainly uses the preprocessed visible light image as the reference image, and the preprocessed infrared image is used as the image to be registered, and the image registration processing is performed on the preprocessed infrared image to obtain the translation parameter.
  • the pre-processed infrared image is also scaled and rotated. Therefore, the FPGA can take the preprocessed infrared image, the preprocessed visible image, and the locally stored rotation.
  • the parameter and the scaling parameter are sent to the image registration processing component, so that the image registration processing component takes the preprocessed visible light image as a reference image, and combines the rotation parameter and the scaling parameter to perform image registration processing on the preprocessed infrared image. Get pan parameter.
  • the pre-processed infrared image is subjected to scaling transformation and rotation transformation respectively; secondly, the transformed infrared image and the pre-processed visible light image common area are identified, and the feature points of the common area are identified; Then, a correspondence relationship between the transformed infrared image and the pre-processed visible light image common point feature point is established, and the translation parameter is determined based on the correspondence relationship.
  • the identification of the common area can be identified by, for example, a region of interest extraction (ROI) algorithm.
  • ROI region of interest extraction
  • the main idea is to define the contrast of the pixel in the color, brightness, direction, etc. as the significant value of the pixel (Saliency). ), the stronger the contrast, the greater the significant value of the pixel.
  • the significant values of all pixels constitute a significant picture.
  • the notable picture here is a grayscale image indicating the significance of each pixel of the image, and the brighter the greater the degree of saliency of the pixel.
  • the region of interest of the image can be obtained based on the salient map.
  • the respective regions of interest of the two images can be considered as common areas.
  • a differential Gaussian pyramid algorithm can be used to detect feature points.
  • the correspondence between the feature points in the two images is established. For example, suppose the coordinates of any feature point in the transformed infrared image are (x, y), and the coordinates of all detected feature points on the preprocessed visible light image are (X 1 , Y 1 ), (X) 2 , Y 2 )...(X N , Y N ), determining the cosine between (x, y) and (X 1 , Y 1 ), (X 2 , Y 2 ) (X N , Y N ), respectively Minimum value: min(arctan(xX 1 , yY 1 ), arctan(xX 2 , yY 2 )...arctan(xX N , yY N )), (X 1 , Y 1 ), (X 2 , the feature point corresponding to the minimum value in Y 2 )...(X N , Y N ) is the feature point corresponding to (x, y
  • the offset ⁇ x of (x, y) with respect to (X 1 , Y 1 ) in the X-axis direction can be determined according to the coordinate difference between x and X 1 , and the offset ⁇ y in the Y-axis direction can be based on y and Y 1 The coordinate difference is determined.
  • the mean of the offsets of the feature point pairs can be obtained to obtain the translation parameters (dx, dy).
  • the FPGA can generate the following translation matrix A:
  • the pixel point P(x, y) is any pixel in the preprocessed infrared image, and after the translation transformation, the corresponding pixel point is P'(x', y'), then :
  • the FPGA generates the translation matrix A, the rotation matrix B, and the scaling matrix C, and can calculate and obtain the coordinate transformation matrix T.
  • the FPGA component can multiply the preprocessed infrared image by the matrix T to obtain the coordinate transformed infrared image, and further, the coordinate transformed infrared image and the preprocessed image.
  • the visible light image is subjected to image fusion processing.
  • the image fusion processing process may include:
  • the FFT fusion process is performed on the coordinate-converted infrared image and the pre-processed visible light image according to the following formula to obtain the fused grayscale image:
  • g(x,y) w1(x,y)*f1(x,y)+w2(x,y)*f2(x,y), where f1(x,y) is the infrared image after coordinate transformation
  • f1(x,y) is the infrared image after coordinate transformation
  • the gray value of any pixel point (x, y), f2 (x, y) is the gray value of the corresponding pixel in the preprocessed visible light image, and g(x, y) is the gray image Gray value corresponding to the pixel
  • corresponding pixel points in the grayscale image are rendered with the chromaticity values of the respective pixels in the preprocessed visible light image to obtain a final fused image. Since the fused image is equivalent to a complementary fusion result of the dominant features of the infrared image and the visible image, the image quality is better.
  • FIG. 3 is a flowchart of Embodiment 3 of an image processing method according to an embodiment of the present invention. As shown in FIG. 3, on the basis of the embodiment shown in FIG. 1, before step 101, the following steps may be further included:
  • AR devices including different types of first camera and second camera are not only used in harsh environments, such as in low light environments, they are also used in normal environments. In a normal environment, if both the first camera and the second camera in the AR device work at the same time, it may not be necessary.
  • the embodiment also provides a scheme for controlling whether the first camera and the second camera operate based on the current environment.
  • the first camera as the infrared camera and the second camera as the CCD camera as an example.
  • the CCD camera In a normal environment, only the CCD camera can be operated, and in some harsh environments, the infrared camera and the CCD camera can work at the same time.
  • the identification of whether the current environment is a normal environment or a harsh environment can be determined by recognizing the pixel gradation value of the image captured by the CCD camera.
  • the CCD camera may be first controlled to randomly capture an image, that is, the third image.
  • An average gradation value is obtained by averaging the gradation values of all or part of the pixels in the third image.
  • the average gray value is compared with a preset gray threshold. If the gray threshold is greater than the gray threshold, the image resolution captured by the CCD camera can meet the viewing requirement, and the current environment is a normal environment. At this point, the CCD camera can be controlled to work alone. Conversely, if it is smaller than the gray threshold, it indicates that the resolution of the image captured by the CCD camera is insufficient to meet the viewing demand.
  • the current environment is a harsh environment. In this case, the infrared camera and the CCD camera need to be controlled to work simultaneously.
  • FIG. 4 is a schematic structural diagram of Embodiment 1 of an image processing apparatus according to an embodiment of the present invention. As shown in FIG. 5, the apparatus includes: a receiving module 11, an obtaining module 12, a transforming module 13, and a merging module 14.
  • the receiving module 11 is configured to receive the first image and the second image, where the first image and the second image are non-homologous images obtained by capturing the same scene by the first camera and the second camera, respectively.
  • the obtaining module 12 is configured to acquire a coordinate transformation matrix corresponding to the first image, where the coordinate transformation matrix uses the second image as a reference image.
  • the transform module 13 is configured to perform coordinate transformation on the first image by using the coordinate transformation matrix.
  • the fusion module 14 is configured to perform image fusion processing on the coordinate-converted first image and the second image.
  • the apparatus shown in FIG. 4 can perform the method of the embodiment shown in FIG. 1.
  • FIG. 5 is a schematic structural diagram of Embodiment 2 of an image processing apparatus according to an embodiment of the present invention. As shown in FIG. 5, on the basis of the embodiment shown in FIG. 4, the preprocessing module 21 is further included.
  • the pre-processing module 21 is configured to perform pre-processing on the first image and the second image, where the pre-processing comprises: performing gray level inversion processing on the first image, and performing the second image on the second image Image enhancement processing.
  • the obtaining module 12 includes: a generating unit 121 and a determining unit 122.
  • the generating unit 121 is configured to generate a translation matrix A, a rotation matrix B, and a scaling matrix C, respectively.
  • the determining unit 122 is configured to determine a coordinate transformation matrix T corresponding to the first image as a result of sequentially multiplying the translation matrix A, the rotation matrix B, and the scaling matrix C.
  • the generating unit 121 is specifically configured to:
  • the translation matrix A is generated according to the translation parameter fed back by the image registration processing component.
  • the fusion module 14 includes: a grayscale fusion unit 141 and a chroma rendering unit 142.
  • the gradation fusion unit 141 is configured to perform gradation fusion processing on the coordinate-converted first image and the second image according to the following formula to obtain a fused grayscale image:
  • the chroma rendering unit 142 is configured to render corresponding pixel points in the grayscale image with chroma values of respective pixels in the second image.
  • the apparatus shown in FIG. 5 can perform the method of the embodiment shown in FIG. 2.
  • FIG. 6 is a schematic structural diagram of Embodiment 3 of an image processing apparatus according to an embodiment of the present invention. As shown in FIG. 6, the receiving module 11 is further configured to receive the second camera. The third image.
  • the apparatus can also include a determination module 31.
  • the determining module 31 is configured to determine whether to trigger the first camera and the second camera to work simultaneously according to a comparison result between the average gray value of the third image and the preset gray threshold.
  • the apparatus shown in FIG. 6 can perform the method of the embodiment shown in FIG. 3.
  • the implementation process and technical effects of the technical solution refer to the description in the embodiment shown in FIG. 3, and details are not described herein again.
  • the device embodiments described above are merely illustrative, wherein the description as separate components
  • the units may or may not be physically separate, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the embodiment. Those of ordinary skill in the art can understand and implement without deliberate labor.
  • FIG. 7 is a schematic structural diagram of Embodiment 1 of an augmented reality device according to an embodiment of the present invention.
  • the AR device may include: a first camera 41, a second camera 42, a memory 43, and a processor 44;
  • the memory 43 is for storing one or more computer instructions, wherein the one or more computer instructions are executed by the processor 44 to implement an image processing method as provided by the embodiments shown in FIGS.
  • the setting of the first camera 41 and the second camera 42 in the AR device may be: setting a first camera 41 and a second camera 42 on the left and right sides on a horizontal plane, that is, the display of the first camera and the second camera from the user AR device The vertical distance of the screen is equal.
  • the first camera 41 is an infrared camera
  • the second camera 42 is a CCD camera.
  • FIG. 8 is a schematic structural diagram of Embodiment 2 of an augmented reality device according to an embodiment of the present invention. As shown in FIG. 8 , the AR device includes:
  • the FPGA component 53 includes functional logic for implementing the image processing method provided by the embodiment shown in FIGS. 1 to 3.
  • the FPGA component can be placed on the motherboard of the AR device.
  • the multi-source image fusion is realized. Because the storage resources of the FPGA are rich in resources, the computing speed will be faster. In the scene of video image acquisition and display, the fused video is smoother, and the fused image can be output in real time to achieve a better visual experience.
  • the electronic device may be an external head mounted display device or an integrated head mounted display device, wherein the external head mounted display device needs to be used in conjunction with an external processing system (eg, a computer processing system).
  • an external processing system eg, a computer processing system
  • FIG. 9 shows a schematic diagram of the internal configuration of the head mounted display device 900 in some embodiments.
  • the display unit 901 may include a display panel disposed on a side surface of the head mounted display device 900 facing the user's face, and may be a one-piece panel or left and right panels respectively corresponding to the left and right eyes of the user.
  • the display panel may be an electroluminescence (EL) element, a liquid crystal display or a microdisplay having a similar structure, or a laser-scanned display in which the retina may be directly displayed or similar.
  • EL electroluminescence
  • the virtual image optical unit 902 photographs the image displayed by the display unit 901 in an enlarged manner, and allows the user to observe the displayed image in the enlarged virtual image.
  • the display image outputted to the display unit 901 it may be an image of a virtual scene supplied from a content reproduction device (a Blu-ray disc or a DVD player) or a streaming server, or an image of a real scene photographed using an external camera 910.
  • virtual image optical unit 902 can include a lens unit, such as a spherical lens, an aspheric lens, a Fresnel lens, and the like.
  • the external camera 910 can specifically implement two cameras, that is, the first camera and the second camera mentioned in the foregoing embodiments.
  • the input operation unit 903 includes at least one operation member for performing an input operation, such as a button, a button, a switch, or other similarly functioned component, receives a user instruction through the operation member, and outputs an instruction to the control unit 907.
  • an input operation such as a button, a button, a switch, or other similarly functioned component
  • the status information acquisition unit 904 is configured to acquire status information of the user wearing the head mounted display device 900.
  • the status information acquisition unit 904 may include various types of sensors for detecting status information by itself, and may acquire status information from an external device such as a smartphone, a wristwatch, and other multi-function terminals worn by the user through the communication unit 905.
  • the status information acquisition unit 904 can acquire location information and/or posture information of the user's head.
  • the status information acquisition unit 904 may include one or more of a gyro sensor, an acceleration sensor, a global positioning system (GPS) sensor, a geomagnetic sensor, a Doppler effect sensor, an infrared sensor, and a radio frequency field intensity sensor.
  • GPS global positioning system
  • the state information acquisition unit 904 acquires state information of the user wearing the head-mounted display device 900, for example, acquires, for example, an operation state of the user (whether the user wears the head-mounted display device 900), an action state of the user (such as standing, walking, running) And the state of movement, such as the posture of the hand or fingertip, the open or closed state of the eye, Line direction, pupil size), mental state (whether the user is immersed in the image displayed by the observation and the like), and even the physiological state.
  • an operation state of the user whether the user wears the head-mounted display device 900
  • an action state of the user such as standing, walking, running
  • the state of movement such as the posture of the hand or fingertip, the open or closed state of the eye, Line direction, pupil size
  • mental state whether the user is immersed in the image displayed by the observation and the like
  • the communication unit 905 performs communication processing with the external device, modulation and demodulation processing, and encoding and decoding processing of the communication signal.
  • the control unit 907 can transmit transmission data from the communication unit 905 to an external device.
  • the communication method may be wired or wireless, such as mobile high-definition link (MHL) or universal serial bus (USB), high-definition multimedia interface (HDMI), wireless fidelity (Wi-Fi), Bluetooth communication, or low-power Bluetooth communication. And the mesh network of the IEEE802.11s standard.
  • communication unit 905 can be a cellular wireless transceiver that operates in accordance with Wideband Code Division Multiple Access (W-CDMA), Long Term Evolution (LTE), and the like.
  • W-CDMA Wideband Code Division Multiple Access
  • LTE Long Term Evolution
  • the head mounted display device 900 can also include a storage unit, the storage unit 906 being a mass storage device configured to have a solid state drive (SSD) or the like.
  • storage unit 906 can store applications or various types of data. For example, content viewed by the user using the head mounted display device 900 may be stored in the storage unit 906.
  • the head mounted display device 900 can also include a control unit, and the control unit 907 can include a computer processing unit (CPU) or other device having similar functionality.
  • control unit 907 can be used to execute an application stored by storage unit 906, or control unit 907 can also be used to perform the methods, functions, and operations disclosed in some embodiments of the present application.
  • the control unit 907 may further include a memory chip such as a ROM 9071, a RAM 9072, etc., for the control unit 907 to execute an application stored therein.
  • the image processing method provided by the foregoing embodiment can be implemented when the above application is executed.
  • the image processing unit 908 is for performing signal processing such as image quality correction related to the image signal output from the control unit 907, and converting its resolution into a resolution according to the screen of the display unit 901. Then, the display driving unit 909 sequentially selects each row of pixels of the display unit 901, and sequentially scans each row of pixels of the display unit 901 line by line, thereby providing pixel signals based on the signal-processed image signals.
  • the head mounted display device 900 can also include an external camera.
  • the external camera 910 may be disposed on the front surface of the body of the head mounted display device 900, and the external camera 910 may be one or more One.
  • the external camera 910 can acquire three-dimensional information and can also be used as a distance sensor. Additionally, a position sensitive detector (PSD) or other type of distance sensor that detects reflected signals from the object can be used with the external camera 910.
  • An external camera 910 and a distance sensor can be used to detect the body position, posture, and shape of the user wearing the head mounted display device 900. In addition, under certain conditions, the user can directly view or preview the real scene through the external camera 910.
  • the head mounted display device 900 may further include a sound processing unit 911 that may perform sound quality correction or sound amplification of a sound signal output from the control unit 907, signal processing of an input sound signal, and the like. Then, the sound input/output unit 912 outputs the sound to the outside and the sound from the microphone after the sound processing.
  • a sound processing unit 911 may perform sound quality correction or sound amplification of a sound signal output from the control unit 907, signal processing of an input sound signal, and the like. Then, the sound input/output unit 912 outputs the sound to the outside and the sound from the microphone after the sound processing.
  • the structure or component shown in bold in FIG. 1 may be independent of the head mounted display device 900, for example, may be disposed in an external processing system (eg, a computer system) for use with the head mounted display device 900; Alternatively, the structure or component shown in bold frame may be disposed inside or on the surface of the head mounted display device 900.
  • an external processing system eg, a computer system

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Processing (AREA)

Abstract

Embodiments of the present invention provide an image processing method, device and augmented reality apparatus. The method comprises: receiving a first image and a second image, wherein the first image and the second image are differently-sourced images acquired by capturing a same scene by a first camera and a second camera, respectively; acquiring a coordinate transformation matrix corresponding to the first image, wherein the coordinate transformation matrix uses the second image as a reference image; performing coordinate transformation on the first image by adopting the coordinate transformation matrix; and performing a fusion process on the first image having undergone coordinate transformation and the second image to fuse advantageous features of the two images and to enhance fused image quality.

Description

图像处理方法、装置和增强现实设备Image processing method, device and augmented reality device
交叉引用cross reference
本发明引用于2017年6月23日递交的名称为“图像处理方法、装置和增强现实设备”的第201710487271.8号中国专利申请,其通过引用被全部并入本发明。The present application is hereby incorporated by reference in its entirety in its entirety in its entirety in its entirety in the the the the the the the the the the the the
技术领域Technical field
本发明涉及图像处理技术领域,尤其涉及一种图像处理方法、装置和增强现实设备。The present invention relates to the field of image processing technologies, and in particular, to an image processing method, apparatus, and augmented reality device.
背景技术Background technique
增强现实(Augmented Reality,简称AR)技术是一种实时的计算摄像头摄像的位置和角度,结合图像处理技术将虚拟世界的场景叠加到现实世界的场景中并显示给用户。AR技术具有实时交互性、真实世界和虚拟世界的信息集成以及在三维尺度空间中增加定位虚拟物体的特点,为人们带来新的视觉体验。Augmented Reality (AR) technology is a real-time computing camera position and angle. It combines image processing technology to superimpose the scene of the virtual world into the real world scene and display it to the user. AR technology has real-time interactive, real-world and virtual world information integration and the ability to add positioning virtual objects in 3D scale space, bringing people a new visual experience.
AR场景中的现实场景是由摄像头拍得的。一般地,会在AR设备中设置某种类型的摄像头,以用于采集现实场景图像,比如通常设置电荷耦合元件(Charge coupled Device,简称CCD)摄像头。但是,如果拍摄现实场景的实际环境是处于较暗的低照度环境中,此时拍得的图像的清晰度往往会很不理想,使得用户最终看到的图像质量不佳,影响用户体验。 The realistic scene in the AR scene is taken by the camera. Generally, a certain type of camera is set in the AR device for collecting images of a real scene, such as a charge coupled device (CCD) camera. However, if the actual environment for shooting a real scene is in a dark, low-light environment, the sharpness of the image taken at this time tends to be unsatisfactory, so that the image quality that the user finally sees is not good, affecting the user experience.
发明内容Summary of the invention
有鉴于此,本发明实施例提供一种图像处理方法、装置和增强现实设备,通过对同一场景的非同源图像进行图像融合处理,提高图像质量。In view of this, embodiments of the present invention provide an image processing method, apparatus, and augmented reality device, which improve image quality by performing image fusion processing on non-homologous images of the same scene.
第一方面,本发明实施例提供一种图像处理方法,包括:In a first aspect, an embodiment of the present invention provides an image processing method, including:
接收第一图像和第二图像,所述第一图像和所述第二图像是分别通过第一摄像头和第二摄像头拍摄同一场景获得的非同源图像;Receiving a first image and a second image, the first image and the second image being non-homologous images obtained by capturing the same scene by the first camera and the second camera, respectively;
获取与所述第一图像对应的坐标变换矩阵,所述坐标变换矩阵以所述第二图像为基准图像;Obtaining a coordinate transformation matrix corresponding to the first image, where the coordinate transformation matrix uses the second image as a reference image;
采用所述坐标变换矩阵对所述第一图像进行坐标变换;Performing coordinate transformation on the first image by using the coordinate transformation matrix;
对坐标变换后的第一图像与所述第二图像进行图像融合处理。Performing image fusion processing on the coordinate-converted first image and the second image.
第二方面,本发明实施例提供一种图像处理装置,包括:In a second aspect, an embodiment of the present invention provides an image processing apparatus, including:
接收模块,用于接收第一图像和第二图像,所述第一图像和所述第二图像是分别通过第一摄像头和第二摄像头拍摄同一场景获得的非同源图像;a receiving module, configured to receive a first image and a second image, where the first image and the second image are non-homologous images obtained by capturing the same scene through the first camera and the second camera, respectively;
获取模块,用于获取与所述第一图像对应的坐标变换矩阵,所述坐标变换矩阵以所述第二图像为基准图像;An acquiring module, configured to acquire a coordinate transformation matrix corresponding to the first image, where the coordinate transformation matrix uses the second image as a reference image;
变换模块,用于采用所述坐标变换矩阵对所述第一图像进行坐标变换;a transform module, configured to perform coordinate transformation on the first image by using the coordinate transformation matrix;
融合模块,用于对坐标变换后的第一图像与所述第二图像进行图像融合处理。And a fusion module, configured to perform image fusion processing on the coordinate-converted first image and the second image.
第三方面,本发明实施例提供一种增强现实设备,包括:In a third aspect, an embodiment of the present invention provides an augmented reality device, including:
第一摄像头、第二摄像头、存储器和处理器;其中,a first camera, a second camera, a memory, and a processor; wherein
所述存储器用于存储一条或多条计算机指令,其中,所述一条或多条计算机指令被所述处理器执行时实现如上所述的图像处理方法。The memory is for storing one or more computer instructions, wherein the one or more computer instructions are executed by the processor to implement an image processing method as described above.
第四方面,本发明实施例提供另一种增强现实设备,包括:In a fourth aspect, an embodiment of the present invention provides another augmented reality device, including:
第一摄像头、第二摄像头、FPGA组件;其中,a first camera, a second camera, an FPGA component; wherein
所述FPGA组件中包含有实现如上所述的图像处理方法的功能逻辑。The FPGA component includes functional logic that implements the image processing method as described above.
本发明实施例提供的图像处理方法和装置,在AR设备中设置两种不同类 型的摄像头即第一摄像头和第二摄像头,通过第一摄像头和第二摄像头同时拍摄同一场景来获得非同源的第一图像和第二图像;以第二图像为基准图像,获取与第一图像对应的坐标变换矩阵,以采用该坐标变换矩阵对第一图像进行坐标变换,以使得变换后的第一图像与第二图像中的各像素点对应;进而,对坐标变换后的第一图像与第二图像进行图像融合处理。由于第一图像和第二图像是非同源图像,两者的优势特征不同,通过对两者进行融合,有利于融合两者的优势特征,增强融合后的图像质量。An image processing method and apparatus provided by an embodiment of the present invention, two different classes are set in an AR device The first camera and the second camera are the first camera and the second camera, and the first camera and the second camera simultaneously capture the same scene to obtain the first image and the second image that are not homologous; the second image is used as the reference image, and the first image is acquired. a coordinate transformation matrix corresponding to the image, the coordinate transformation is performed on the first image by using the coordinate transformation matrix, so that the transformed first image corresponds to each pixel point in the second image; and further, the first image after the coordinate transformation Image fusion processing is performed with the second image. Since the first image and the second image are non-homologous images, the dominant features of the two are different. By combining the two, it is beneficial to fuse the superior features of the two and enhance the image quality after the fusion.
附图说明DRAWINGS
为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作一简单地介绍,显而易见地,下面描述中的附图是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, a brief description of the drawings used in the embodiments or the prior art description will be briefly described below. Obviously, the drawings in the following description It is a certain embodiment of the present invention, and other drawings can be obtained from those skilled in the art without any creative work.
图1为本发明实施例提供的图像处理方法实施例一的流程图;FIG. 1 is a flowchart of Embodiment 1 of an image processing method according to an embodiment of the present invention;
图2为本发明实施例提供的图像处理方法实施例二的流程图;2 is a flowchart of Embodiment 2 of an image processing method according to an embodiment of the present invention;
图3为本发明实施例提供的图像处理方法实施例三的流程图;FIG. 3 is a flowchart of Embodiment 3 of an image processing method according to an embodiment of the present disclosure;
图4为本发明实施例提供的图像处理装置实施例一的结构示意图;FIG. 4 is a schematic structural diagram of Embodiment 1 of an image processing apparatus according to an embodiment of the present disclosure;
图5为本发明实施例提供的图像处理装置实施例二的结构示意图;FIG. 5 is a schematic structural diagram of Embodiment 2 of an image processing apparatus according to an embodiment of the present disclosure;
图6为本发明实施例提供的图像处理装置实施例三的结构示意图;FIG. 6 is a schematic structural diagram of Embodiment 3 of an image processing apparatus according to an embodiment of the present disclosure;
图7为本发明实施例提供的增强现实设备实施例一的结构示意图;FIG. 7 is a schematic structural diagram of Embodiment 1 of an augmented reality device according to an embodiment of the present disclosure;
图8为本发明实施例提供的增强现实设备实施例二的结构示意图FIG. 8 is a schematic structural diagram of Embodiment 2 of an augmented reality device according to an embodiment of the present disclosure;
图9为本发明实施例提供的头戴显示设备的结构示意图。 FIG. 9 is a schematic structural diagram of a head mounted display device according to an embodiment of the present invention.
具体实施方式Detailed ways
为使本发明实施例的目的、技术方案和优点更加清楚,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention will be clearly and completely described in conjunction with the drawings in the embodiments of the present invention. It is a partial embodiment of the invention, and not all of the embodiments. All other embodiments obtained by those skilled in the art based on the embodiments of the present invention without creative efforts are within the scope of the present invention.
在本发明实施例中使用的术语是仅仅出于描述特定实施例的目的,而非旨在限制本发明。在本发明实施例和所附权利要求书中所使用的单数形式的“一种”、“所述”和“该”也旨在包括多数形式,除非上下文清楚地表示其他含义,“多种”一般包含至少两种,但是不排除包含至少一种的情况。The terms used in the embodiments of the present invention are for the purpose of describing particular embodiments only and are not intended to limit the invention. The singular forms "a", "the", "the" and "the" Generally, at least two types are included, but the case of including at least one is not excluded.
应当理解,本文中使用的术语“和/或”仅仅是一种描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本文中字符“/”,一般表示前后关联对象是一种“或”的关系。It should be understood that the term "and/or" as used herein is merely an association describing the associated object, indicating that there may be three relationships, for example, A and/or B, which may indicate that A exists separately, while A and B, there are three cases of B alone. In addition, the character "/" in this article generally indicates that the contextual object is an "or" relationship.
应当理解,尽管在本发明实施例中可能采用术语第一、第二、第三等来描述XXX,但这些XXX不应限于这些术语。这些术语仅用来将XXX区分开。例如,在不脱离本发明实施例范围的情况下,第一XXX也可以被称为第二XXX,类似地,第二XXX也可以被称为第一XXX。It should be understood that although the terms first, second, third, etc. may be used to describe XXX in embodiments of the invention, these XXX should not be limited to these terms. These terms are only used to distinguish XXX. For example, the first XXX may also be referred to as a second XXX without departing from the scope of the embodiments of the present invention. Similarly, the second XXX may also be referred to as a first XXX.
取决于语境,如在此所使用的词语“如果”、“若”可以被解释成为“在……时”或“当……时”或“响应于确定”或“响应于检测”。类似地,取决于语境,短语“如果确定”或“如果检测(陈述的条件或事件)”可以被解释成为“当确定时”或“响应于确定”或“当检测(陈述的条件或事件)时”或“响应于检测(陈述的条件或事件)”。Depending on the context, the words "if" and "if" as used herein may be interpreted to mean "when" or "when" or "in response to determining" or "in response to detecting." Similarly, depending on the context, the phrase "if determined" or "if detected (conditions or events stated)" can be interpreted as "when determined" or "in response to determination" or "when detected (stated condition or event) "Time" or "in response to a test (condition or event stated)".
还需要说明的是,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的商品或者系统不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种商品或者系 统所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的商品或者系统中还存在另外的相同要素。It should also be noted that the terms "including", "comprising" or "comprising" or any other variations thereof are intended to encompass a non-exclusive inclusion, such that the item or system comprising a plurality of elements includes not only those elements but also Other elements, or are included for this commodity or department The elements inherent in the system. An element defined by the phrase "comprising a ..." does not exclude the presence of additional identical elements in the item or system including the element, without further limitation.
进一步值得说明的是,本发明各实施例中各步骤之间的顺序是可以调整的,不是必须按照以下举例的顺序执行。It is further noted that the order between the steps in the various embodiments of the present invention may be adjusted, and is not necessarily performed in the order illustrated below.
图1为本发明实施例提供的图像处理方法实施例一的流程图,本实施例提供的该图像处理方法可以由一图像处理装置来执行,该图像处理装置可以实现为现场可编程门阵列(Field-Programmable Gate Array,简称FPGA)组件中的部分硬件器件的组合,该FPGA组件可以集成设置在AR设备中。如图1所示,该方法包括如下步骤:FIG. 1 is a flowchart of Embodiment 1 of an image processing method according to an embodiment of the present invention. The image processing method provided by this embodiment may be implemented by an image processing apparatus, and the image processing apparatus may be implemented as a field programmable gate array ( A combination of some hardware devices in the Field-Programmable Gate Array (FPGA) component, which can be integrated into the AR device. As shown in FIG. 1, the method includes the following steps:
101、接收第一图像和第二图像,第一图像和第二图像是分别通过第一摄像头和第二摄像头拍摄同一场景获得的非同源图像。101. Receive a first image and a second image, where the first image and the second image are non-homologous images obtained by capturing the same scene by the first camera and the second camera, respectively.
102、获取与第一图像对应的坐标变换矩阵,坐标变换矩阵以第二图像为基准图像。102. Acquire a coordinate transformation matrix corresponding to the first image, where the coordinate transformation matrix uses the second image as a reference image.
103、采用坐标变换矩阵对第一图像进行坐标变换。103. Perform coordinate transformation on the first image by using a coordinate transformation matrix.
104、对坐标变换后的第一图像与第二图像进行图像融合处理。104. Perform image fusion processing on the first image and the second image after the coordinate transformation.
本发明实施例中,在同一AR设备中可以设置不同类型的摄像头,比如上述第一摄像头和第二摄像头,用于对同一场景进行拍摄。其中,第一摄像头和第二摄像头在AR设备中的设置可以是:在一水平面上,左右分别设置第一摄像头和第二摄像头。In the embodiment of the present invention, different types of cameras, such as the first camera and the second camera, may be disposed in the same AR device for capturing the same scene. The first camera and the second camera are disposed in the AR device, and the first camera and the second camera are respectively disposed on the left and right sides on a horizontal surface.
可选地,第一摄像头可以是红外摄像头,第二摄像头可以是CCD摄像头。相应地,第一摄像头拍得的第一图像为红外图像,第二摄像头拍得的第二图像为可见光图像。Alternatively, the first camera may be an infrared camera and the second camera may be a CCD camera. Correspondingly, the first image captured by the first camera is an infrared image, and the second image captured by the second camera is a visible light image.
下面结合红外摄像头和CCD摄像头来说明本发明实施例考虑采用不同类型的摄像头进行同一场景的图像采集以及图像融合的初衷:采用红外摄像头和CCD摄像头同时对同一场景进行拍照,并将获取的红外图像和可见光图像的优势特征信息进行融合,最终可以得到特征点明显,信息量丰富的融合图像。The following is an explanation of the image acquisition and image fusion of the same scene by using different types of cameras in combination with an infrared camera and a CCD camera. The infrared camera and the CCD camera simultaneously take photos of the same scene and acquire the infrared image. The fusion with the dominant feature information of the visible light image can finally obtain a fused image with obvious feature points and rich information.
具体来说,自然界中温度高于绝对零度时,物体就会产生红外辐射,红 外成像就是利用红外摄像头将不可见的红外辐射转换成可见的温度分布图像。红外图像不易受环境影响,在雨雪、烟雾以及黑暗的环境下,都可以获取物体的温度分布图。但红外摄像头的分辨率往往较低,使得获得的红外图像的清晰度差、场景细节信息不明显并且所成的图像不符合人的视觉习惯。反之,CCD摄像头获取的图像是根据物体反射光的能量来实现的,这种图像能够较好的刻画场景的细节信息,分辨率高,更是符合人的视觉系统的要求。但CCD摄像头同时也存在着一些不足之处:在恶劣天气下,图像的捕捉能力差,丢失有用信息,不能获取描述场景中全面细致的图像信息。由此可见,红外图像和可见光图像各有优点和缺点,因此,如果能够采用融合算法将红外图像和可见光图像的优势特征融合,使融合后的图像含有丰富的特征点信息,适合人眼视觉系统,将大大提高用户观看的视觉体验。Specifically, when the temperature in the natural world is higher than the absolute zero, the object will produce infrared radiation, red. External imaging uses an infrared camera to convert invisible infrared radiation into a visible temperature distribution image. Infrared images are not susceptible to environmental influences, and temperature profiles of objects can be obtained in rain, snow, smoke, and dark environments. However, the resolution of the infrared camera tends to be low, so that the obtained infrared image has poor definition, the scene detail information is not obvious, and the formed image does not conform to human visual habits. On the contrary, the image acquired by the CCD camera is realized according to the energy of the reflected light of the object, and the image can better describe the detailed information of the scene, and the resolution is high, which is in line with the requirements of the human visual system. However, CCD cameras also have some shortcomings: in bad weather, the image capture ability is poor, the useful information is lost, and the comprehensive and detailed image information in the description scene cannot be obtained. It can be seen that the infrared image and the visible light image have their own advantages and disadvantages. Therefore, if the fusion algorithm is used to fuse the dominant features of the infrared image and the visible light image, the fused image contains rich feature point information, which is suitable for the human visual system. Will greatly enhance the visual experience of users watching.
另外,本发明实施例提供的图像处理方法可以基于FPGA组件这一硬件平台实现,即基于FPGA实现多源图像的融合。相比于单纯的软件处理方法,FPGA的存储资源等资源丰富,运算速度更快。当在视频图像采集与展示的场景中,会使得融合后的视频流畅度更高,能够实时输出融合后的图像,达到更好的视觉体验效果。In addition, the image processing method provided by the embodiment of the present invention can be implemented based on the hardware component of the FPGA component, that is, the fusion of the multi-source image is implemented based on the FPGA. Compared with pure software processing methods, FPGAs have rich resources such as storage resources and faster computing speed. In the scene of video image acquisition and display, the fused video is smoother, and the fused image can be output in real time to achieve a better visual experience.
在某些情景中,集成上述第一摄像头和第二摄像头的AR设备往往用于采集现实场景的视频图像。可以理解的是,由于第一摄像头和第二摄像头用于对同一场景进行拍摄,因此,需要保证两者的时钟同步,即在同一时刻,这两个摄像头是对场景中的同一对象进行拍摄。但是,由于这两个摄像头的拍摄位置、拍摄角度等拍摄参数会有所差异,即使是对同一对象进行拍摄,拍得的图像也往往会有所不同。In some scenarios, an AR device that integrates the first camera and the second camera described above is often used to capture video images of a real scene. It can be understood that since the first camera and the second camera are used to capture the same scene, it is necessary to ensure the clock synchronization of the two, that is, at the same time, the two cameras are shooting the same object in the scene. However, since the shooting parameters such as the shooting position and shooting angle of the two cameras will be different, even if the same subject is shot, the captured image will often be different.
在上述情景中,第一摄像头和第二摄像头将拍得的视频图像通过FPGA组件的视频接口同时输入到FPGA组件,经过FPGA组件的视频解码芯片后,解码为比如BT.656格式的YCbCr视频图像。此情景中,本发明实施例提供的图像处理方法是对各时刻分别对应的两幅图像进行融合处理。为便于描述,仅以任一时刻对应的上述第一图像和第二图像为例进行图像融合过程的说明。 In the above scenario, the first camera and the second camera simultaneously input the captured video image into the FPGA component through the video interface of the FPGA component, and after decoding the video decoder chip of the FPGA component, decoding into a YCbCr video image such as the BT.656 format. . In this scenario, the image processing method provided by the embodiment of the present invention performs fusion processing on two images corresponding to each time. For convenience of description, the description of the image fusion process is performed by taking only the first image and the second image corresponding to any one time as an example.
由于第一图像和第二图像在分辨率、拍摄角度等拍摄参数方面存在差异,因此,要实现第一图像和第二图像的图像融合,首先需要对第一图像和第二图像进行图像配准处理,以建立第一图像和第二图像中各像素点的对应关系,从而,才能基于像素点的对应关系进行第一图像和第二图像的融合。Since the first image and the second image have differences in shooting parameters such as resolution and shooting angle, to achieve image fusion of the first image and the second image, image registration of the first image and the second image is first required. Processing to establish a correspondence between each pixel point in the first image and the second image, so that the fusion of the first image and the second image can be performed based on the correspondence relationship of the pixel points.
以第一图像为红外图像、第二图像为可见光图像为例进行图像配准处理过程的说明。The description of the image registration processing process is performed by taking the first image as an infrared image and the second image as a visible light image as an example.
图像配准处理包括了图像的缩放、旋转和平移操作。由于相比于红外图像来说,可见光图像具有更高的分辨率,更符合人眼视觉习惯,因此,本实施例中,以可见光图像为基准图像,红外图像作为待配准图像,针对红外图像进行缩放、旋转和平移操作。而对红外图像进行缩放、旋转和平移操作是基于获得的坐标变换矩阵进行的,也就是说,该坐标变换矩阵中包含了进行缩放操作所需的缩放参数,进行旋转操作所需的旋转参数,以及进行平移操作所需的平移参数。而这些缩放参数、旋转参数和平移参数可以是预先获得的,从而,基于预先获得的这些参数可以生成得到坐标变换矩阵。Image registration processing includes scaling, rotation, and panning of the image. Since the visible light image has higher resolution and is more in line with the human eye visual habit than the infrared image, in this embodiment, the visible light image is used as the reference image, and the infrared image is used as the image to be registered for the infrared image. Zoom, rotate, and pan. The scaling, rotation and translation operations of the infrared image are performed based on the obtained coordinate transformation matrix, that is, the coordinate transformation matrix contains the scaling parameters required for the scaling operation, and the rotation parameters required for the rotation operation, And the translation parameters required for the panning operation. And these scaling parameters, rotation parameters, and translation parameters may be obtained in advance, so that a coordinate transformation matrix can be generated based on these parameters obtained in advance.
值得说明的是,本实施例中,之所以采用坐标变换矩阵的方式对红外图像进行变换,是因为相比于对红外图像依次分别进行三种变换的方式,这样变换的效率更高,因为只需用一个矩阵就可以对红外图像中的各像素点一次进行了三种变换。It should be noted that, in this embodiment, the reason why the infrared image is transformed by using the coordinate transformation matrix is because the conversion is more efficient than the method of sequentially performing three transformations on the infrared image, because only the conversion is more efficient. It is necessary to use a matrix to perform three transformations for each pixel in the infrared image.
在获得上述坐标变换矩阵之后,可以基于红外图像与该坐标变换矩阵的矩阵乘法运算,得到坐标变换后的红外图像。由于上述坐标变换矩阵中的变换参数是以可见光图像为基准的,因此,基于该变换,可以得到坐标变换后的红外图像中的像素点与可见光图像中的像素点间的对应关系。从而,基于该对应关系,可以对坐标变换后的红外图像和可见光图像进行图像融合。After obtaining the coordinate transformation matrix described above, the coordinate-converted infrared image can be obtained based on matrix multiplication of the infrared image and the coordinate transformation matrix. Since the transformation parameters in the coordinate transformation matrix are based on the visible light image, the correspondence between the pixel points in the coordinate-converted infrared image and the pixel points in the visible light image can be obtained based on the transformation. Therefore, based on the correspondence, image fusion of the coordinate-converted infrared image and the visible light image can be performed.
在FPGA硬件平台实施图像的融合,要考虑到的硬件平台的资源,存储空间和处理速度等问题。为了得到较好的融合效果并充分利用FPGA的资源,本实施例中选择采用基于像素级的融合方法:灰度值加权平均的方法,即通过对应像素点的灰度值的加权平均计算,实现两幅图像的融合。 In the implementation of image fusion on the FPGA hardware platform, the resources of the hardware platform, storage space and processing speed should be considered. In order to obtain a better fusion effect and make full use of the resources of the FPGA, in this embodiment, a pixel-level fusion method is selected: a method of weighted average of gray values, that is, a weighted average calculation of gray values of corresponding pixel points is implemented. The fusion of two images.
综上,本实施例中,在AR设备中设置两种不同类型的摄像头即第一摄像头和第二摄像头,通过第一摄像头和第二摄像头同时拍摄同一场景来获得非同源的第一图像和第二图像;以第二图像为基准图像,获取与第一图像对应的坐标变换矩阵,以采用该坐标变换矩阵对第一图像进行坐标变换,以使得变换后的第一图像与第二图像中的各像素点对应;进而,对坐标变换后的第一图像与第二图像进行图像融合处理。由于第一图像和第二图像是非同源图像,两者的优势特征不同,通过对两者进行融合,有利于融合两者的优势特征,增强融合后的图像质量。In summary, in the embodiment, two different types of cameras, that is, a first camera and a second camera are disposed in the AR device, and the first camera and the second camera simultaneously capture the same scene to obtain a non-homologous first image and a second image; taking a second image as a reference image, acquiring a coordinate transformation matrix corresponding to the first image, and performing coordinate transformation on the first image by using the coordinate transformation matrix, so that the transformed first image and the second image are Corresponding to each pixel point; further, image fusion processing is performed on the coordinate-converted first image and the second image. Since the first image and the second image are non-homologous images, the dominant features of the two are different. By combining the two, it is beneficial to fuse the superior features of the two and enhance the image quality after the fusion.
图2为本发明实施例提供的图像处理方法实施例二的流程图,如图2所示,在图1所示实施例基础上,步骤103之后,还可以包括如下步骤:FIG. 2 is a flowchart of Embodiment 2 of an image processing method according to an embodiment of the present invention. As shown in FIG. 2, on the basis of the embodiment shown in FIG. 1, after step 103, the following steps may be further included:
201、接收第一图像和第二图像,第一图像和第二图像是分别通过第一摄像头和第二摄像头拍摄同一场景获得的非同源图像。201. Receive a first image and a second image, where the first image and the second image are non-homologous images obtained by capturing the same scene by the first camera and the second camera, respectively.
202、对第一图像和第二图像进行预处理,预处理包括:对第一图像进行灰度值取逆处理,对第二图像进行图像增强处理。202. Perform pre-processing on the first image and the second image, where the pre-processing comprises: performing inverse processing on the gray value of the first image, and performing image enhancement processing on the second image.
203、根据本地存储的旋转参数和缩放参数,分别生成旋转矩阵B和缩放矩阵C。203. Generate a rotation matrix B and a scaling matrix C according to locally stored rotation parameters and scaling parameters.
204、将预处理后的第一图像、预处理后的第二图像以及旋转参数和缩放参数发送至图像配准处理组件,以使图像配准处理组件以预处理后的第二图像为基准图像,结合旋转参数和缩放参数对预处理后的第一图像进行配准处理,以获得平移参数。204. Send the preprocessed first image, the preprocessed second image, and the rotation parameter and the scaling parameter to the image registration processing component, so that the image registration processing component uses the preprocessed second image as a reference image. The pre-processed first image is subjected to registration processing in combination with the rotation parameter and the scaling parameter to obtain a translation parameter.
205、根据图像配准处理组件反馈的平移参数,生成平移矩阵A。205. Generate a translation matrix A according to the translation parameter fed back by the image registration processing component.
206、确定与第一图像对应的坐标变换矩阵T为:平移矩阵A、旋转矩阵B和缩放矩阵C依次相乘的结果。206. Determine a coordinate transformation matrix T corresponding to the first image as a result of sequentially multiplying the translation matrix A, the rotation matrix B, and the scaling matrix C.
207、采用坐标变换矩阵对第一图像进行坐标变换。207. Perform coordinate transformation on the first image by using a coordinate transformation matrix.
208、对坐标变换后的第一图像与第二图像进行图像融合处理。208. Perform image fusion processing on the first image and the second image after the coordinate transformation.
本实施例中,为了保证后续的融合图像的质量,可选地,可以对接收到 的第一图像和第二图像进行一定的预处理。In this embodiment, in order to ensure the quality of the subsequent fused image, optionally, it may be received The first image and the second image are subjected to certain pre-processing.
以第一图像为红外图像、第二图像为可见光图像为例。Take the first image as an infrared image and the second image as a visible light image as an example.
由于红外图像根据物体的热辐射成像的,亮度太高,不适合人的视觉系统。本实施例中,通过对第一图像进行灰度值取逆处理来降低红外图像的亮度,突出特征点。Since the infrared image is imaged according to the thermal radiation of the object, the brightness is too high and is not suitable for the human visual system. In this embodiment, the brightness of the infrared image is reduced by performing inverse processing of the gray value on the first image, and the feature points are highlighted.
具体地,假设红外图像Simage1的大小为M*N,每个像素点的灰度值是8bits,即灰度等级划分为28即256个灰度级,构造M*N的单位矩阵E,取逆后的红外图像为Simage2,则根据如下公式确定取逆后的红外图像:Simage2=256*E-Simage1。Specifically, it is assumed that the size of the infrared image Simage1 is M*N, and the gray value of each pixel is 8 bits, that is, the gray level is divided into 28 or 256 gray levels, and the unit matrix E of M*N is constructed, and the inverse is performed. After the infrared image is Simage2, the inverse infrared image is determined according to the following formula: Simage2=256*E-Simage1.
可见光图像是根据能量的反射原理所成的像,由于是在低照度的恶劣环境下获取的可见光图像,画面比较暗,突出的特征点少,因此,需要对可见光图像进行图像增强处理。具体地,可以对可见光图像的像素点的灰度值进行阈值划分,采用传统的三段式图像增强的方法,对不同阈值范围内的像素点通过拉伸变换系数的方法,实现图像画面的增强。The visible light image is an image formed according to the principle of reflection of energy. Since the visible light image is acquired in a harsh environment with low illumination, the screen is dark and the prominent feature points are small. Therefore, it is necessary to perform image enhancement processing on the visible light image. Specifically, the gray value of the pixel of the visible light image may be threshold-divided, and the traditional three-stage image enhancement method is adopted to enhance the image image by stretching the transform coefficient for the pixel points in different threshold ranges. .
在对红外图像和可见光图像进行预处理之后,可以以预处理后的红外图像和可见光图像为基础,获取与预处理后的红外图像对应的坐标变换矩阵。After the infrared image and the visible light image are preprocessed, the coordinate transformation matrix corresponding to the preprocessed infrared image may be acquired based on the preprocessed infrared image and the visible light image.
具体地,该坐标变换矩阵可以根据平移矩阵A、旋转矩阵B和缩放矩阵C获得,其中,平移矩阵A、旋转矩阵B和缩放矩阵C分别是用于表现平移参数、旋转参数和缩放参数的。因此,需要先根据平移参数、旋转参数和缩放参数分别生成对应的平移矩阵A、旋转矩阵B和缩放矩阵C,进而,可以确定与预处理后的红外图像对应的坐标变换矩阵T为:平移矩阵A、旋转矩阵B和缩放矩阵C依次进行矩阵相乘的结果:T=ABC。Specifically, the coordinate transformation matrix may be obtained according to the translation matrix A, the rotation matrix B, and the scaling matrix C, wherein the translation matrix A, the rotation matrix B, and the scaling matrix C are respectively used to represent a translation parameter, a rotation parameter, and a scaling parameter. Therefore, the translation matrix A, the rotation matrix B, and the scaling matrix C are respectively generated according to the translation parameter, the rotation parameter, and the scaling parameter. Further, the coordinate transformation matrix T corresponding to the preprocessed infrared image can be determined as: translation matrix A. The rotation matrix B and the scaling matrix C sequentially perform matrix multiplication results: T=ABC.
针对缩放参数来说,图像的缩放变换操作主要是针对不同分辨率的图像进行的。由于红外图像与可见光图像的分辨率不同,因此,需要以预处理后的可见光图像为基准,对预处理后的红外图像进行缩放操作,以使得其分辨率与预处理后的可见光图像的分辨率一致。 For the scaling parameters, the scaling operation of the image is mainly performed for images of different resolutions. Since the resolution of the infrared image and the visible image are different, it is necessary to scale the preprocessed infrared image based on the preprocessed visible image to make the resolution and the resolution of the preprocessed visible image. Consistent.
假设像素点P(x,y)为预处理后的红外图像中的任一像素点,在X轴方向的缩放系数为tx,Y轴方向的缩放系数为ty,经过缩放变换后得到对应像素点为P'(x',y'),则有:x'=x*tx;y'=y*ty。若用矩阵表示,则为:Assume that the pixel point P(x, y) is any pixel in the preprocessed infrared image, the scaling factor in the X-axis direction is t x , and the scaling factor in the Y-axis direction is t y , which is obtained after scaling transformation. The pixel point is P'(x', y'), then there are: x'=x*t x ; y'=y*t y . If represented by a matrix, it is:
Figure PCTCN2017113578-appb-000001
Figure PCTCN2017113578-appb-000001
从而,缩放矩阵
Figure PCTCN2017113578-appb-000002
Thus, the scaling matrix
Figure PCTCN2017113578-appb-000002
由此可知,若要生成上述缩放矩阵C,需要获得缩放参数tx和ty。而缩放参数tx和ty可以根据红外摄像头和CCD摄像头的分辨率确定,即两者X轴分辨率的比值即可确定tx,两者Y轴分辨率的比值即可确定ty。因此,当AR设备中的红外摄像头和CCD摄像头设定后,即可确定出缩放参数tx和ty,该缩放参数tx和ty可以被预先存储在FPGA组件的存储空间中。From this, it can be seen that to generate the above-described scaling matrix C, it is necessary to obtain scaling parameters t x and t y . The scaling parameters t x and t y can be determined according to the resolution of the infrared camera and the CCD camera, that is, the ratio of the X-axis resolutions of the two can determine t x , and the ratio of the two Y-axis resolutions can determine t y . Therefore, when the infrared camera and the CCD camera in the AR device are set, the scaling parameters t x and t y can be determined, and the scaling parameters t x and t y can be pre-stored in the storage space of the FPGA component.
针对旋转参数来说,图像的旋转变换操作主要由于拍摄红外图像和可见光图像时,由于人为因素使得红外图像和可见光图像之间存在角度的偏移,为了使两幅图像中对应的特征点能够准确的匹配,需要以预处理后的可见光图像为基准,对预处理后的红外图像在二维空间内做图像的旋转变换。For the rotation parameters, the rotation transformation of the image is mainly caused by the angle between the infrared image and the visible image due to human factors when shooting the infrared image and the visible light image, in order to make the corresponding feature points in the two images accurate. The matching needs to be performed on the pre-processed visible light image as a reference, and the pre-processed infrared image is rotated in the two-dimensional space.
假设像素点P(x,y)为预处理后的红外图像中的任一像素点,经过旋转变换后得到对应像素点为P'(x',y'),若用矩阵表示P'(x',y')与P(x,y)的旋转关系,则为:Suppose the pixel point P(x, y) is any pixel in the preprocessed infrared image. After the rotation transformation, the corresponding pixel point is P'(x', y'). If the matrix is used, P'(x) The rotation relationship between ',y') and P(x,y) is:
Figure PCTCN2017113578-appb-000003
Figure PCTCN2017113578-appb-000003
其中,以预处理后的红外图像中的原点为中心,建立直角坐标系。假设P(x,y)与原点的连接与X轴的夹角为第一角度;P'(x',y')与原点的连接与X轴的夹角为第二角度,则第二角度与第一角度的差值即为θ,代表的含义是:P'(x',y')与P(x,y)之间的偏转角度。 Wherein, a Cartesian coordinate system is established centering on the origin in the preprocessed infrared image. Suppose that the angle between the connection of P(x, y) and the origin with the X axis is the first angle; the angle between the connection of P'(x', y') and the origin with the X axis is the second angle, then the second angle The difference from the first angle is θ, which means the angle of deflection between P'(x', y') and P(x, y).
从而,旋转矩阵
Figure PCTCN2017113578-appb-000004
Thus, the rotation matrix
Figure PCTCN2017113578-appb-000004
由此可知,若要生成上述旋转矩阵B,需要获得旋转参数θ。而旋转参数θ可以根据红外摄像头和CCD摄像头的在AR设备中的设置情况确定,具体来说,可以测量红外摄像头的镜头中心与水平面表面的夹角,以及CCD摄像头的镜头中心与水平面表面的夹角,两个夹角的角度差即为旋转参数θ。因此,当AR设备中的红外摄像头和CCD摄像头设定后,即可确定出旋转参数θ,该旋转参数θ可以被预先存储在FPGA组件的存储空间中。From this, it can be seen that to generate the above-described rotation matrix B, it is necessary to obtain the rotation parameter θ. The rotation parameter θ can be determined according to the setting of the infrared camera and the CCD camera in the AR device. Specifically, the angle between the lens center and the horizontal surface of the infrared camera and the lens center and the horizontal surface of the CCD camera can be measured. Angle, the angle difference between the two angles is the rotation parameter θ. Therefore, when the infrared camera and the CCD camera in the AR device are set, the rotation parameter θ can be determined, and the rotation parameter θ can be pre-stored in the storage space of the FPGA component.
针对平移参数来说,与上述旋转参数、缩放参数不同,当需要对红外图像进行平移变换操作时,平移参数需要基于当前的红外图像和可见光图像计算获得。也就是说,旋转参数和缩放参数可以认为是与当前拍得的图像无关的,不需依赖于当前拍得的图像确定,但是,平移参数是与当前拍得的图像有关的,需要依赖于当前拍得的图像确定。For the translation parameters, unlike the above rotation parameters and scaling parameters, when the translation transformation operation of the infrared image is required, the translation parameters need to be calculated based on the current infrared image and the visible light image. That is to say, the rotation parameter and the scaling parameter can be considered to be independent of the currently captured image, and do not depend on the currently captured image determination, but the translation parameter is related to the currently captured image and needs to be dependent on the current The captured image is determined.
平移参数的确定需要涉及到负责的计算过程,由于本发明实施例提供的图像处理方法是可基于FPGA组件这一硬件平台来实现的,若用FPGA组件来计算该平移参数,则比较受限,因此,可选地,可以基于图像配准处理组件来计算该平移参数,该图像配准处理组件可以实现为软件程序,经过该图像配准处理组件的计算处理,得到平移参数,反馈至FPGA组件,以便FPGA组件生成对应的平移矩阵A。The determination of the translation parameter needs to involve a responsible calculation process. The image processing method provided by the embodiment of the present invention can be implemented based on the hardware component of the FPGA component. If the FPGA component is used to calculate the translation parameter, the limitation is limited. Therefore, optionally, the translation parameter can be calculated based on the image registration processing component, and the image registration processing component can be implemented as a software program, and the translation processing is performed by the image registration processing component to obtain a translation parameter and feedback to the FPGA component. So that the FPGA component generates the corresponding translation matrix A.
而图像配准处理组件主要是以预处理后的可见光图像为基准图像,以预处理后的红外图像为待配准图像,对预处理后的红外图像进行图像配准处理后得到平移参数。而该图像配准处理过程中,也会涉及到对预处理后的红外图像的缩放、旋转变换操作,因此,FPGA可以将预处理后的红外图像、预处理后的可见光图像以及本地存储的旋转参数和缩放参数发送至图像配准处理组件,以使图像配准处理组件以预处理后的可见光图像为基准图像,结合该旋转参数和缩放参数对预处理后的红外图像进行图像配准处理,以获得平移 参数。The image registration processing component mainly uses the preprocessed visible light image as the reference image, and the preprocessed infrared image is used as the image to be registered, and the image registration processing is performed on the preprocessed infrared image to obtain the translation parameter. In the image registration process, the pre-processed infrared image is also scaled and rotated. Therefore, the FPGA can take the preprocessed infrared image, the preprocessed visible image, and the locally stored rotation. The parameter and the scaling parameter are sent to the image registration processing component, so that the image registration processing component takes the preprocessed visible light image as a reference image, and combines the rotation parameter and the scaling parameter to perform image registration processing on the preprocessed infrared image. Get pan parameter.
简单来说明下图像配准处理组件的图像配准过程:Simply explain the image registration process of the image registration processing component:
首先,基于缩放参数和旋转参数,分别对预处理后的红外图像进行缩放变换和旋转变换;其次,识别变换后的红外图像和预处理后的可见光图像的公共区域,识别公共区域的特征点;之后,建立变换后的红外图像和预处理后的可见光图像之间公共区域特征点的对应关系,以基于该对应关系确定出平移参数。First, based on the scaling parameter and the rotation parameter, the pre-processed infrared image is subjected to scaling transformation and rotation transformation respectively; secondly, the transformed infrared image and the pre-processed visible light image common area are identified, and the feature points of the common area are identified; Then, a correspondence relationship between the transformed infrared image and the pre-processed visible light image common point feature point is established, and the translation parameter is determined based on the correspondence relationship.
其中,对公共区域的识别可以通过比如感兴趣区域提取(ROI)算法进行识别,主要思路为:把像素点在颜色、亮度、方向等方面与背景的对比定义为该像素点的显著值(Saliency),对比越强,该像素点的显著值就越大。所有像素点的显著值构成一张显著图。这里显著图是一副表明图像各像素点显著性的灰度图像,越亮表明该像素点的显著度越大。基于该显著图可以获得图像的感兴趣区域。两幅图像各自的感兴趣区域可以认为是公共区域。The identification of the common area can be identified by, for example, a region of interest extraction (ROI) algorithm. The main idea is to define the contrast of the pixel in the color, brightness, direction, etc. as the significant value of the pixel (Saliency). ), the stronger the contrast, the greater the significant value of the pixel. The significant values of all pixels constitute a significant picture. The notable picture here is a grayscale image indicating the significance of each pixel of the image, and the brighter the greater the degree of saliency of the pixel. The region of interest of the image can be obtained based on the salient map. The respective regions of interest of the two images can be considered as common areas.
对于特征点来说,可以采用差分的高斯金字塔算法进行特征点的检测。For feature points, a differential Gaussian pyramid algorithm can be used to detect feature points.
在得到两幅图像中的特征点之后,建立两幅图像中的特征点之间的对应关系。举例来说,假设变换后的红外图像中任意一特征点的坐标为(x,y),预处理后的可见光图像上所有检测出的特征点的坐标为(X1,Y1)、(X2,Y2)…(XN,YN),确定(x,y)分别与(X1,Y1)、(X2,Y2)…(XN,YN)之间的余弦的最小值:即min(arctan(x-X1,y-Y1),arctan(x-X2,y-Y2)......arctan(x-XN,y-YN)),(X1,Y1)、(X2,Y2)…(XN,YN)中对应于该最小值的特征点即为与(x,y)对应的特征点,假设对应于该最小值的特征点为(X1,Y1)。则(x,y)相对于(X1,Y1)在X轴方向的偏移量Δx可以根据x与X1的坐标差确定,在Y轴方向的偏移量Δy可以根据y与Y1的坐标差确定。最后,对于所有特征点对,可以求取所述特征点对的偏移量的均值,以得到平移参数(dx,dy)。After the feature points in the two images are obtained, the correspondence between the feature points in the two images is established. For example, suppose the coordinates of any feature point in the transformed infrared image are (x, y), and the coordinates of all detected feature points on the preprocessed visible light image are (X 1 , Y 1 ), (X) 2 , Y 2 )...(X N , Y N ), determining the cosine between (x, y) and (X 1 , Y 1 ), (X 2 , Y 2 ) (X N , Y N ), respectively Minimum value: min(arctan(xX 1 , yY 1 ), arctan(xX 2 , yY 2 )...arctan(xX N , yY N )), (X 1 , Y 1 ), (X 2 , the feature point corresponding to the minimum value in Y 2 )...(X N , Y N ) is the feature point corresponding to (x, y), and it is assumed that the feature point corresponding to the minimum value is (X 1 , Y 1 ). Then, the offset Δx of (x, y) with respect to (X 1 , Y 1 ) in the X-axis direction can be determined according to the coordinate difference between x and X 1 , and the offset Δy in the Y-axis direction can be based on y and Y 1 The coordinate difference is determined. Finally, for all feature point pairs, the mean of the offsets of the feature point pairs can be obtained to obtain the translation parameters (dx, dy).
从而,FPGA基于该平移参数可以生成如下的平移矩阵A: Thus, based on the translation parameter, the FPGA can generate the following translation matrix A:
平移矩阵
Figure PCTCN2017113578-appb-000005
Translation matrix
Figure PCTCN2017113578-appb-000005
从而,在FPGA中,假设像素点P(x,y)为预处理后的红外图像中的任一像素点,经过平移变换后得到对应像素点为P'(x',y'),则有:Therefore, in the FPGA, it is assumed that the pixel point P(x, y) is any pixel in the preprocessed infrared image, and after the translation transformation, the corresponding pixel point is P'(x', y'), then :
Figure PCTCN2017113578-appb-000006
Figure PCTCN2017113578-appb-000006
FPGA在生成平移矩阵A、旋转矩阵B和缩放矩阵C,可以计算获得坐标变换矩阵T。The FPGA generates the translation matrix A, the rotation matrix B, and the scaling matrix C, and can calculate and obtain the coordinate transformation matrix T.
在基于前述过程得到坐标变换矩阵T后,FPGA组件可以将预处理后的红外图像与该矩阵T相乘,得到坐标变换后的红外图像,进而,将坐标变换后的红外图像与预处理后的可见光图像进行图像融合处理。After obtaining the coordinate transformation matrix T based on the foregoing process, the FPGA component can multiply the preprocessed infrared image by the matrix T to obtain the coordinate transformed infrared image, and further, the coordinate transformed infrared image and the preprocessed image. The visible light image is subjected to image fusion processing.
具体地,该图像融合处理过程可以包括:Specifically, the image fusion processing process may include:
根据如下公式对坐标变换后的红外图像与预处理后的可见光图像进行灰度融合处理,以获得融合后的灰度图像:The FFT fusion process is performed on the coordinate-converted infrared image and the pre-processed visible light image according to the following formula to obtain the fused grayscale image:
g(x,y)=w1(x,y)*f1(x,y)+w2(x,y)*f2(x,y),其中,f1(x,y)为坐标变换后的红外图像中的任一像素点(x,y)的灰度值,f2(x,y)为预处理后的可见光图像中的对应像素点的灰度值,g(x,y)为灰度图像中对应像素点的灰度值;w1(x,y)和w2(x,y)为加权系数,w1(x,y)+w2(x,y)=1;g(x,y)=w1(x,y)*f1(x,y)+w2(x,y)*f2(x,y), where f1(x,y) is the infrared image after coordinate transformation The gray value of any pixel point (x, y), f2 (x, y) is the gray value of the corresponding pixel in the preprocessed visible light image, and g(x, y) is the gray image Gray value corresponding to the pixel; w1 (x, y) and w2 (x, y) are weighting coefficients, w1 (x, y) + w2 (x, y) = 1;
进而,以预处理后的可见光图像中各像素点的色度值渲染灰度图像中的对应像素点,以获得最终的融合后图像。由于该融合后的图像相当于是红外图像和可见光图像的优势特征的互补融合结果,图像质量较佳。Further, corresponding pixel points in the grayscale image are rendered with the chromaticity values of the respective pixels in the preprocessed visible light image to obtain a final fused image. Since the fused image is equivalent to a complementary fusion result of the dominant features of the infrared image and the visible image, the image quality is better.
图3为本发明实施例提供的图像处理方法实施例三的流程图,如图3所示,在图1所示实施例基础上,步骤101之前,还可以包括如下步骤: FIG. 3 is a flowchart of Embodiment 3 of an image processing method according to an embodiment of the present invention. As shown in FIG. 3, on the basis of the embodiment shown in FIG. 1, before step 101, the following steps may be further included:
301、接收第二摄像头拍得的第三图像。301. Receive a third image captured by the second camera.
302、根据第三图像的平均灰度值与预设灰度阈值的比较结果,确定是否触发第一摄像头和第二摄像头同时工作,若是,则执行步骤101-104。302. Determine, according to a comparison result of the average gray value of the third image and the preset gray threshold, whether to trigger the first camera and the second camera to work simultaneously, and if yes, perform steps 101-104.
由于在实际应用中,包含了不同类型的第一摄像头和第二摄像头的AR设备不光仅用于恶劣环境中,比如光线较暗的环境中,还会被用于正常的环境中。而在正常的环境中,如果AR设备中的第一摄像头和第二摄像头都同时工作,可能是没有必要的。Since in practical applications, AR devices including different types of first camera and second camera are not only used in harsh environments, such as in low light environments, they are also used in normal environments. In a normal environment, if both the first camera and the second camera in the AR device work at the same time, it may not be necessary.
因此,本实施例还提供了基于当前环境的不同,控制第一摄像头和第二摄像头是否工作的方案。Therefore, the embodiment also provides a scheme for controlling whether the first camera and the second camera operate based on the current environment.
以第一摄像头为红外摄像头、第二摄像头为CCD摄像头为例。在正常环境中,可以仅让CCD摄像头工作,而在某些恶劣环境中,可以让红外摄像头和CCD摄像头同时工作。Take the first camera as the infrared camera and the second camera as the CCD camera as an example. In a normal environment, only the CCD camera can be operated, and in some harsh environments, the infrared camera and the CCD camera can work at the same time.
本实施例中,对于当前的环境是正常环境还是恶劣环境的识别,可以通过识别CCD摄像头拍得的图像的像素灰度值的情况来判定。In this embodiment, the identification of whether the current environment is a normal environment or a harsh environment can be determined by recognizing the pixel gradation value of the image captured by the CCD camera.
具体来说,当AR设备被启动时,可以先控制CCD摄像头随机拍得一幅图像,即上述第三图像。通过对该第三图像中全部或部分像素的灰度值进行求平均运算,得到平均灰度值。进而,将该平均灰度值与某预设的灰度阈值进行比较,如果大于该灰度阈值,说明此时CCD摄像头拍得的图像分辨率即可满足观看需求,当前的环境属于正常环境,此时控制CCD摄像头单独工作即可。相反地,如果小于该灰度阈值,说明此时CCD摄像头拍得的图像分辨率不足以满足观看需求,当前的环境属于恶劣环境,此时需控制红外摄像头和CCD摄像头同时工作。Specifically, when the AR device is activated, the CCD camera may be first controlled to randomly capture an image, that is, the third image. An average gradation value is obtained by averaging the gradation values of all or part of the pixels in the third image. Further, the average gray value is compared with a preset gray threshold. If the gray threshold is greater than the gray threshold, the image resolution captured by the CCD camera can meet the viewing requirement, and the current environment is a normal environment. At this point, the CCD camera can be controlled to work alone. Conversely, if it is smaller than the gray threshold, it indicates that the resolution of the image captured by the CCD camera is insufficient to meet the viewing demand. The current environment is a harsh environment. In this case, the infrared camera and the CCD camera need to be controlled to work simultaneously.
本实施例中,通过对当前环境是正常环境还是异常的恶劣环境的识别,对AR设备中设置的不同摄像头进行工作与否的控制,提高AR设备的智能化。In this embodiment, whether the current environment is a normal environment or an abnormally harsh environment is identified, and the operation of the different cameras set in the AR device is controlled to improve the intelligence of the AR device.
以下将详细描述本发明的一个或多个实施例的图像处理装置。本领域技术人员可以理解,这些图像处理装置均可使用市售的硬件组件通过本方案所 教导的步骤进行配置来构成。An image processing apparatus of one or more embodiments of the present invention will be described in detail below. Those skilled in the art will appreciate that these image processing devices can use commercially available hardware components through the solution. The steps taught are configured to be constructed.
图4为本发明实施例提供的图像处理装置实施例一的结构示意图,如图5所示,该装置包括:接收模块11、获取模块12、变换模块13、融合模块14。FIG. 4 is a schematic structural diagram of Embodiment 1 of an image processing apparatus according to an embodiment of the present invention. As shown in FIG. 5, the apparatus includes: a receiving module 11, an obtaining module 12, a transforming module 13, and a merging module 14.
接收模块11,用于接收第一图像和第二图像,所述第一图像和所述第二图像是分别通过第一摄像头和第二摄像头拍摄同一场景获得的非同源图像。The receiving module 11 is configured to receive the first image and the second image, where the first image and the second image are non-homologous images obtained by capturing the same scene by the first camera and the second camera, respectively.
获取模块12,用于获取与所述第一图像对应的坐标变换矩阵,所述坐标变换矩阵以所述第二图像为基准图像。The obtaining module 12 is configured to acquire a coordinate transformation matrix corresponding to the first image, where the coordinate transformation matrix uses the second image as a reference image.
变换模块13,用于采用所述坐标变换矩阵对所述第一图像进行坐标变换。The transform module 13 is configured to perform coordinate transformation on the first image by using the coordinate transformation matrix.
融合模块14,用于对坐标变换后的第一图像与所述第二图像进行图像融合处理。The fusion module 14 is configured to perform image fusion processing on the coordinate-converted first image and the second image.
图4所示装置可以执行图1所示实施例的方法,本实施例未详细描述的部分,可参考对图1所示实施例的相关说明。该技术方案的执行过程和技术效果参见图1所示实施例中的描述,在此不再赘述。The apparatus shown in FIG. 4 can perform the method of the embodiment shown in FIG. 1. For the parts not described in detail in this embodiment, reference may be made to the related description of the embodiment shown in FIG. 1. For the implementation process and technical effects of the technical solution, refer to the description in the embodiment shown in FIG. 1, and details are not described herein again.
图5为本发明实施例提供的图像处理装置实施例二的结构示意图,如图5所示,在图4所示实施例基础上,还包括:预处理模块21。FIG. 5 is a schematic structural diagram of Embodiment 2 of an image processing apparatus according to an embodiment of the present invention. As shown in FIG. 5, on the basis of the embodiment shown in FIG. 4, the preprocessing module 21 is further included.
预处理模块21,用于对所述第一图像和所述第二图像进行预处理,所述预处理包括:对所述第一图像进行灰度值取逆处理,对所述第二图像进行图像增强处理。The pre-processing module 21 is configured to perform pre-processing on the first image and the second image, where the pre-processing comprises: performing gray level inversion processing on the first image, and performing the second image on the second image Image enhancement processing.
可选地,所述获取模块12包括:生成单元121、确定单元122。Optionally, the obtaining module 12 includes: a generating unit 121 and a determining unit 122.
生成单元121,用于分别生成平移矩阵A、旋转矩阵B和缩放矩阵C。The generating unit 121 is configured to generate a translation matrix A, a rotation matrix B, and a scaling matrix C, respectively.
确定单元122,用于确定与所述第一图像对应的坐标变换矩阵T为:所述平移矩阵A、所述旋转矩阵B和所述缩放矩阵C依次相乘的结果。The determining unit 122 is configured to determine a coordinate transformation matrix T corresponding to the first image as a result of sequentially multiplying the translation matrix A, the rotation matrix B, and the scaling matrix C.
可选地,所述生成单元121具体用于:Optionally, the generating unit 121 is specifically configured to:
根据本地存储的旋转参数和缩放参数,分别生成所述旋转矩阵B和所述缩放矩阵C;Generating the rotation matrix B and the scaling matrix C according to locally stored rotation parameters and scaling parameters;
将所述第一图像、所述第二图像以及所述旋转参数和所述缩放参数发送 至图像配准处理组件,以使所述图像配准处理组件以所述第二图像为基准图像,结合所述旋转参数和所述缩放参数对所述第一图像进行配准处理,以获得平移参数;Transmitting the first image, the second image, and the rotation parameter and the scaling parameter And an image registration processing component, wherein the image registration processing component uses the second image as a reference image, and the first image is registered in combination with the rotation parameter and the scaling parameter to obtain a translation parameter;
根据所述图像配准处理组件反馈的所述平移参数,生成所述平移矩阵A。The translation matrix A is generated according to the translation parameter fed back by the image registration processing component.
可选地,所述融合模块14包括:灰度融合单元141、色度渲染单元142。Optionally, the fusion module 14 includes: a grayscale fusion unit 141 and a chroma rendering unit 142.
灰度融合单元141,用于根据如下公式对所述坐标变换后的第一图像与所述第二图像进行灰度融合处理,以获得融合后的灰度图像:The gradation fusion unit 141 is configured to perform gradation fusion processing on the coordinate-converted first image and the second image according to the following formula to obtain a fused grayscale image:
g(x,y)=w1(x,y)*f1(x,y)+w2(x,y)*f2(x,y),其中,f1(x,y)为所述坐标变换后的第一图像中的任一像素点(x,y)的灰度值,f2(x,y)为所述第二图像中的对应像素点的灰度值,g(x,y)为灰度图像中对应像素点的灰度值;w1(x,y)和w2(x,y)为加权系数,w1(x,y)+w2(x,y)=1;g(x,y)=w1(x,y)*f1(x,y)+w2(x,y)*f2(x,y), where f1(x,y) is the coordinate transformed a gray value of any pixel point (x, y) in the first image, f2 (x, y) is a gray value of a corresponding pixel point in the second image, and g(x, y) is a gray scale Gray value of the corresponding pixel in the image; w1 (x, y) and w2 (x, y) are weighting coefficients, w1 (x, y) + w2 (x, y) = 1;
色度渲染单元142,用于以所述第二图像中各像素点的色度值渲染所述灰度图像中的对应像素点。The chroma rendering unit 142 is configured to render corresponding pixel points in the grayscale image with chroma values of respective pixels in the second image.
图5所示装置可以执行图2所示实施例的方法,本实施例未详细描述的部分,可参考对图2所示实施例的相关说明。该技术方案的执行过程和技术效果参见图2所示实施例中的描述,在此不再赘述。The apparatus shown in FIG. 5 can perform the method of the embodiment shown in FIG. 2. For the parts not described in detail in this embodiment, reference may be made to the related description of the embodiment shown in FIG. 2. For the implementation process and technical effects of the technical solution, refer to the description in the embodiment shown in FIG. 2, and details are not described herein again.
图6为本发明实施例提供的图像处理装置实施例三的结构示意图,如图6所示,在前述实施例基础上,所述接收模块11,还用于接收所述第二摄像头拍得的第三图像。FIG. 6 is a schematic structural diagram of Embodiment 3 of an image processing apparatus according to an embodiment of the present invention. As shown in FIG. 6, the receiving module 11 is further configured to receive the second camera. The third image.
该装置还可以包括:确定模块31。The apparatus can also include a determination module 31.
确定模块31,用于根据所述第三图像的平均灰度值与预设灰度阈值的比较结果,确定是否触发所述第一摄像头和所述第二摄像头同时工作。The determining module 31 is configured to determine whether to trigger the first camera and the second camera to work simultaneously according to a comparison result between the average gray value of the third image and the preset gray threshold.
图6所示装置可以执行图3所示实施例的方法,本实施例未详细描述的部分,可参考对图3所示实施例的相关说明。该技术方案的执行过程和技术效果参见图3所示实施例中的描述,在此不再赘述。The apparatus shown in FIG. 6 can perform the method of the embodiment shown in FIG. 3. For the parts not described in detail in this embodiment, reference may be made to the related description of the embodiment shown in FIG. For the implementation process and technical effects of the technical solution, refer to the description in the embodiment shown in FIG. 3, and details are not described herein again.
以上所描述的装置实施例仅仅是示意性的,其中所述作为分离部件说明 的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部模块来实现本实施例方案的目的。本领域普通技术人员在不付出创造性的劳动的情况下,即可以理解并实施。The device embodiments described above are merely illustrative, wherein the description as separate components The units may or may not be physically separate, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the embodiment. Those of ordinary skill in the art can understand and implement without deliberate labor.
图7为本发明实施例提供的增强现实设备实施例一的结构示意图,如图7所示,该AR设备可以包括:第一摄像头41、第二摄像头42、存储器43和处理器44;其中,FIG. 7 is a schematic structural diagram of Embodiment 1 of an augmented reality device according to an embodiment of the present invention. As shown in FIG. 7, the AR device may include: a first camera 41, a second camera 42, a memory 43, and a processor 44;
所述存储器43用于存储一条或多条计算机指令,其中,所述一条或多条计算机指令被所述处理器44执行时实现如图1至图3所示实施例提供的图像处理方法。The memory 43 is for storing one or more computer instructions, wherein the one or more computer instructions are executed by the processor 44 to implement an image processing method as provided by the embodiments shown in FIGS.
第一摄像头41和第二摄像头42在AR设备中的设置可以是:在一水平面上,左右分别设置第一摄像头41和第二摄像头42,即第一摄像头和第二摄像头距离用户AR设备的显示屏的垂直距离相等。The setting of the first camera 41 and the second camera 42 in the AR device may be: setting a first camera 41 and a second camera 42 on the left and right sides on a horizontal plane, that is, the display of the first camera and the second camera from the user AR device The vertical distance of the screen is equal.
可选地,所述第一摄像头41为红外摄像头,所述第二摄像头42为CCD摄像头。Optionally, the first camera 41 is an infrared camera, and the second camera 42 is a CCD camera.
图8为本发明实施例提供的增强现实设备实施例二的结构示意图,如图8所示,该AR设备中包括:FIG. 8 is a schematic structural diagram of Embodiment 2 of an augmented reality device according to an embodiment of the present invention. As shown in FIG. 8 , the AR device includes:
第一摄像头51、第二摄像头52、FPGA组件53;其中,a first camera 51, a second camera 52, and an FPGA component 53; wherein
所述FPGA组件53中包含有实现如图1至图3所示实施例提供的图像处理方法的功能逻辑。FPGA组件可以设置在AR设备的主板上。The FPGA component 53 includes functional logic for implementing the image processing method provided by the embodiment shown in FIGS. 1 to 3. The FPGA component can be placed on the motherboard of the AR device.
基于FPGA组件这一平台实现多源图像的融合,由于FPGA的存储资源等资源丰富,运算速度会更快。在视频图像采集与展示的场景中,会使得融合后的视频流畅度更高,能够实时输出融合后的图像,达到更好的视觉体验效果。Based on the FPGA component platform, the multi-source image fusion is realized. Because the storage resources of the FPGA are rich in resources, the computing speed will be faster. In the scene of video image acquisition and display, the fused video is smoother, and the fused image can be output in real time to achieve a better visual experience.
本发明一些实施例提供的电子设备可以为外接式头戴显示设备或者一体式头戴显示设备,其中外接式头戴显示设备需要与外部处理系统(例如计算机处理系统)配合使用。 The electronic device provided by some embodiments of the present invention may be an external head mounted display device or an integrated head mounted display device, wherein the external head mounted display device needs to be used in conjunction with an external processing system (eg, a computer processing system).
图9示出了一些实施例中头戴显示设备900的内部配置结构示意图。FIG. 9 shows a schematic diagram of the internal configuration of the head mounted display device 900 in some embodiments.
显示单元901可以包括显示面板,显示面板设置在头戴显示设备900上面向用户面部的侧表面,可以为一整块面板、或者为分别对应用户左眼和右眼的左面板和右面板。显示面板可以为电致发光(EL)元件、液晶显示器或具有类似结构的微型显示器、或者视网膜可直接显示或类似的激光扫描式显示器。The display unit 901 may include a display panel disposed on a side surface of the head mounted display device 900 facing the user's face, and may be a one-piece panel or left and right panels respectively corresponding to the left and right eyes of the user. The display panel may be an electroluminescence (EL) element, a liquid crystal display or a microdisplay having a similar structure, or a laser-scanned display in which the retina may be directly displayed or similar.
虚拟图像光学单元902以放大方式拍摄显示单元901所显示的图像,并允许用户按放大的虚拟图像观察所显示的图像。作为输出到显示单元901上的显示图像,可以是从内容再现设备(蓝光光碟或DVD播放器)或流媒体服务器提供的虚拟场景的图像、或者使用外部相机910拍摄的现实场景的图像。一些实施例中,虚拟图像光学单元902可以包括透镜单元,例如球面透镜、非球面透镜、菲涅尔透镜等。可以理解的是,本发明实施例中,该外部相机910可以具体实现两个摄像头,即前述实施例中提及的第一摄像头和第二摄像头。The virtual image optical unit 902 photographs the image displayed by the display unit 901 in an enlarged manner, and allows the user to observe the displayed image in the enlarged virtual image. As the display image outputted to the display unit 901, it may be an image of a virtual scene supplied from a content reproduction device (a Blu-ray disc or a DVD player) or a streaming server, or an image of a real scene photographed using an external camera 910. In some embodiments, virtual image optical unit 902 can include a lens unit, such as a spherical lens, an aspheric lens, a Fresnel lens, and the like. It can be understood that, in the embodiment of the present invention, the external camera 910 can specifically implement two cameras, that is, the first camera and the second camera mentioned in the foregoing embodiments.
输入操作单元903包括至少一个用来执行输入操作的操作部件,例如按键、按钮、开关或者其他具有类似功能的部件,通过操作部件接收用户指令,并且向控制单元907输出指令。The input operation unit 903 includes at least one operation member for performing an input operation, such as a button, a button, a switch, or other similarly functioned component, receives a user instruction through the operation member, and outputs an instruction to the control unit 907.
状态信息获取单元904用于获取穿戴头戴显示设备900的用户的状态信息。状态信息获取单元904可以包括各种类型的传感器,用于自身检测状态信息,并可以通过通信单元905从外部设备(例如智能手机、腕表和用户穿戴的其它多功能终端)获取状态信息。状态信息获取单元904可以获取用户的头部的位置信息和/或姿态信息。状态信息获取单元904可以包括陀螺仪传感器、加速度传感器、全球定位系统(GPS)传感器、地磁传感器、多普勒效应传感器、红外传感器、射频场强度传感器中的一个或者多个。此外,状态信息获取单元904获取穿戴头戴显示设备900的用户的状态信息,例如获取例如用户的操作状态(用户是否穿戴头戴显示设备900)、用户的动作状态(诸如静止、行走、跑动和诸如此类的移动状态,手或指尖的姿势、眼睛的开或闭状态、视 线方向、瞳孔尺寸)、精神状态(用户是否沉浸在观察所显示的图像以及诸如此类的),甚至生理状态。The status information acquisition unit 904 is configured to acquire status information of the user wearing the head mounted display device 900. The status information acquisition unit 904 may include various types of sensors for detecting status information by itself, and may acquire status information from an external device such as a smartphone, a wristwatch, and other multi-function terminals worn by the user through the communication unit 905. The status information acquisition unit 904 can acquire location information and/or posture information of the user's head. The status information acquisition unit 904 may include one or more of a gyro sensor, an acceleration sensor, a global positioning system (GPS) sensor, a geomagnetic sensor, a Doppler effect sensor, an infrared sensor, and a radio frequency field intensity sensor. Further, the state information acquisition unit 904 acquires state information of the user wearing the head-mounted display device 900, for example, acquires, for example, an operation state of the user (whether the user wears the head-mounted display device 900), an action state of the user (such as standing, walking, running) And the state of movement, such as the posture of the hand or fingertip, the open or closed state of the eye, Line direction, pupil size), mental state (whether the user is immersed in the image displayed by the observation and the like), and even the physiological state.
通信单元905执行与外部装置的通信处理、调制和解调处理、以及通信信号的编码和解码处理。另外,控制单元907可以从通信单元905向外部装置发送传输数据。通信方式可以是有线或者无线形式,例如移动高清链接(MHL)或通用串行总线(USB)、高清多媒体接口(HDMI)、无线保真(Wi-Fi)、蓝牙通信或低功耗蓝牙通信,以及IEEE802.11s标准的网状网络等。另外,通信单元905可以是根据宽带码分多址(W-CDMA)、长期演进(LTE)和类似标准操作的蜂窝无线收发器。The communication unit 905 performs communication processing with the external device, modulation and demodulation processing, and encoding and decoding processing of the communication signal. In addition, the control unit 907 can transmit transmission data from the communication unit 905 to an external device. The communication method may be wired or wireless, such as mobile high-definition link (MHL) or universal serial bus (USB), high-definition multimedia interface (HDMI), wireless fidelity (Wi-Fi), Bluetooth communication, or low-power Bluetooth communication. And the mesh network of the IEEE802.11s standard. Additionally, communication unit 905 can be a cellular wireless transceiver that operates in accordance with Wideband Code Division Multiple Access (W-CDMA), Long Term Evolution (LTE), and the like.
一些实施例中,头戴显示设备900还可以包括存储单元,存储单元906是配置为具有固态驱动器(SSD)等的大容量存储设备。一些实施例中,存储单元906可以存储应用程序或各种类型的数据。例如,用户使用头戴显示设备900观看的内容可以存储在存储单元906中。In some embodiments, the head mounted display device 900 can also include a storage unit, the storage unit 906 being a mass storage device configured to have a solid state drive (SSD) or the like. In some embodiments, storage unit 906 can store applications or various types of data. For example, content viewed by the user using the head mounted display device 900 may be stored in the storage unit 906.
一些实施例中,头戴显示设备900还可以包括控制单元,控制单元907可以包括计算机处理单元(CPU)或者其他具有类似功能的设备。一些实施例中,控制单元907可以用于执行存储单元906存储的应用程序,或者控制单元907还可以用于执行本申请一些实施例公开的方法、功能和操作的电路。一些实施例中,控制单元907中还可以包括ROM9071、RAM9072等存储芯片,以供控制单元907可以执行其中存储的应用程序。上述应用程序被执行时可以实现前述实施例提供的图像处理方法。In some embodiments, the head mounted display device 900 can also include a control unit, and the control unit 907 can include a computer processing unit (CPU) or other device having similar functionality. In some embodiments, control unit 907 can be used to execute an application stored by storage unit 906, or control unit 907 can also be used to perform the methods, functions, and operations disclosed in some embodiments of the present application. In some embodiments, the control unit 907 may further include a memory chip such as a ROM 9071, a RAM 9072, etc., for the control unit 907 to execute an application stored therein. The image processing method provided by the foregoing embodiment can be implemented when the above application is executed.
图像处理单元908用于执行信号处理,比如与从控制单元907输出的图像信号相关的图像质量校正,以及将其分辨率转换为根据显示单元901的屏幕的分辨率。然后,显示驱动单元909依次选择显示单元901的每行像素,并逐行依次扫描显示单元901的每行像素,因而提供基于经信号处理的图像信号的像素信号。The image processing unit 908 is for performing signal processing such as image quality correction related to the image signal output from the control unit 907, and converting its resolution into a resolution according to the screen of the display unit 901. Then, the display driving unit 909 sequentially selects each row of pixels of the display unit 901, and sequentially scans each row of pixels of the display unit 901 line by line, thereby providing pixel signals based on the signal-processed image signals.
一些实施例中,头戴显示设备900还可以包括外部相机。外部相机910可以设置在头戴显示设备900主体前表面,外部相机910可以为一个或者多 个。外部相机910可以获取三维信息,并且也可以用作距离传感器。另外,探测来自物体的反射信号的位置灵敏探测器(PSD)或者其他类型的距离传感器可以与外部相机910一起使用。外部相机910和距离传感器可以用于检测穿戴头戴显示设备900的用户的身体位置、姿态和形状。另外,一定条件下用户可以通过外部相机910直接观看或者预览现实场景。In some embodiments, the head mounted display device 900 can also include an external camera. The external camera 910 may be disposed on the front surface of the body of the head mounted display device 900, and the external camera 910 may be one or more One. The external camera 910 can acquire three-dimensional information and can also be used as a distance sensor. Additionally, a position sensitive detector (PSD) or other type of distance sensor that detects reflected signals from the object can be used with the external camera 910. An external camera 910 and a distance sensor can be used to detect the body position, posture, and shape of the user wearing the head mounted display device 900. In addition, under certain conditions, the user can directly view or preview the real scene through the external camera 910.
一些实施例中,头戴显示设备900还可以包括声音处理单元,声音处理单元911可以执行从控制单元907输出的声音信号的声音质量校正或声音放大,以及输入声音信号的信号处理等。然后,声音输入/输出单元912在声音处理后向外部输出声音以及输入来自麦克风的声音。In some embodiments, the head mounted display device 900 may further include a sound processing unit 911 that may perform sound quality correction or sound amplification of a sound signal output from the control unit 907, signal processing of an input sound signal, and the like. Then, the sound input/output unit 912 outputs the sound to the outside and the sound from the microphone after the sound processing.
需要说明的是,图1中加粗框示出的结构或部件可以独立于头戴显示设备900之外,例如可以设置在外部处理系统(例如计算机系统)中与头戴显示设备900配合使用;或者,加粗框示出的结构或部件可以设置在头戴显示设备900内部或者表面上。It should be noted that the structure or component shown in bold in FIG. 1 may be independent of the head mounted display device 900, for example, may be disposed in an external processing system (eg, a computer system) for use with the head mounted display device 900; Alternatively, the structure or component shown in bold frame may be disposed inside or on the surface of the head mounted display device 900.
最后应说明的是:以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。 It should be noted that the above embodiments are only used to illustrate the technical solutions of the present invention, and are not limited thereto; although the present invention has been described in detail with reference to the foregoing embodiments, those skilled in the art should understand that The technical solutions described in the foregoing embodiments are modified, or the equivalents of the technical features are replaced. The modifications and substitutions do not depart from the spirit and scope of the technical solutions of the embodiments of the present invention.

Claims (17)

  1. 一种图像处理方法,其特征在于,包括:An image processing method, comprising:
    接收第一图像和第二图像,所述第一图像和所述第二图像是分别通过第一摄像头和第二摄像头拍摄同一场景获得的非同源图像;Receiving a first image and a second image, the first image and the second image being non-homologous images obtained by capturing the same scene by the first camera and the second camera, respectively;
    获取与所述第一图像对应的坐标变换矩阵,所述坐标变换矩阵以所述第二图像为基准图像;Obtaining a coordinate transformation matrix corresponding to the first image, where the coordinate transformation matrix uses the second image as a reference image;
    采用所述坐标变换矩阵对所述第一图像进行坐标变换;Performing coordinate transformation on the first image by using the coordinate transformation matrix;
    对坐标变换后的第一图像与所述第二图像进行图像融合处理。Performing image fusion processing on the coordinate-converted first image and the second image.
  2. 根据权利要求1所述的方法,其特征在于,所述获取与所述第一图像对应的坐标变换矩阵之前,还包括:The method according to claim 1, wherein before the acquiring the coordinate transformation matrix corresponding to the first image, the method further comprises:
    对所述第一图像和所述第二图像进行预处理,所述预处理包括:对所述第一图像进行灰度值取逆处理,对所述第二图像进行图像增强处理。Performing pre-processing on the first image and the second image, the pre-processing includes: performing gray level inversion processing on the first image, and performing image enhancement processing on the second image.
  3. 根据权利要求1所述的方法,其特征在于,所述获取与所述第一图像对应的坐标变换矩阵,包括:The method according to claim 1, wherein the acquiring a coordinate transformation matrix corresponding to the first image comprises:
    分别生成平移矩阵A、旋转矩阵B和缩放矩阵C;Generating a translation matrix A, a rotation matrix B, and a scaling matrix C, respectively;
    确定与所述第一图像对应的坐标变换矩阵T为:所述平移矩阵A、所述旋转矩阵B和所述缩放矩阵C依次相乘的结果。Determining a coordinate transformation matrix T corresponding to the first image is a result of sequentially multiplying the translation matrix A, the rotation matrix B, and the scaling matrix C.
  4. 根据权利要求3所述的方法,其特征在于,所述分别生成平移矩阵A、旋转矩阵B和缩放矩阵C,包括:The method according to claim 3, wherein the generating the translation matrix A, the rotation matrix B, and the scaling matrix C respectively comprises:
    根据本地存储的旋转参数和缩放参数,分别生成所述旋转矩阵B和所述缩放矩阵C;Generating the rotation matrix B and the scaling matrix C according to locally stored rotation parameters and scaling parameters;
    将所述第一图像、所述第二图像以及所述旋转参数和所述缩放参数发送至图像配准处理组件,以使所述图像配准处理组件以所述第二图像为基准图像,结合所述旋转参数和所述缩放参数对所述第一图像进行配准处理,以获得平移参数;Transmitting the first image, the second image, and the rotation parameter and the scaling parameter to an image registration processing component, so that the image registration processing component uses the second image as a reference image, combined The rotation parameter and the scaling parameter perform registration processing on the first image to obtain a translation parameter;
    根据所述图像配准处理组件反馈的所述平移参数,生成所述平移矩阵A。 The translation matrix A is generated according to the translation parameter fed back by the image registration processing component.
  5. 根据权利要求1所述的方法,其特征在于,所述对坐标变换后的第一图像与所述第二图像进行图像融合处理,包括:The method according to claim 1, wherein the performing image fusion processing on the coordinate-converted first image and the second image comprises:
    根据如下公式对所述坐标变换后的第一图像与所述第二图像进行灰度融合处理,以获得融合后的灰度图像:Performing gradation fusion processing on the coordinate-converted first image and the second image according to the following formula to obtain a fused grayscale image:
    g(x,y)=w1(x,y)*f1(x,y)+w2(x,y)*f2(x,y),其中,f1(x,y)为所述坐标变换后的第一图像中的任一像素点(x,y)的灰度值,f2(x,y)为所述第二图像中的对应像素点的灰度值,g(x,y)为灰度图像中对应像素点的灰度值;w1(x,y)和w2(x,y)为加权系数,w1(x,y)+w2(x,y)=1;g(x,y)=w1(x,y)*f1(x,y)+w2(x,y)*f2(x,y), where f1(x,y) is the coordinate transformed a gray value of any pixel point (x, y) in the first image, f2 (x, y) is a gray value of a corresponding pixel point in the second image, and g(x, y) is a gray scale Gray value of the corresponding pixel in the image; w1 (x, y) and w2 (x, y) are weighting coefficients, w1 (x, y) + w2 (x, y) = 1;
    以所述第二图像中各像素点的色度值渲染所述灰度图像中的对应像素点。Rendering corresponding pixel points in the grayscale image with chrominance values of respective pixel points in the second image.
  6. 根据权利要求1至5中任一项所述的方法,其特征在于,所述接收第一图像和第二图像之前,还包括:The method according to any one of claims 1 to 5, wherein before the receiving the first image and the second image, the method further comprises:
    接收所述第二摄像头拍得的第三图像;Receiving a third image captured by the second camera;
    根据所述第三图像的平均灰度值与预设灰度阈值的比较结果,确定是否触发所述第一摄像头和所述第二摄像头同时工作。And determining whether to trigger the first camera and the second camera to work simultaneously according to a comparison result of the average gray value of the third image and the preset gray threshold.
  7. 一种图像处理装置,其特征在于,包括:An image processing apparatus, comprising:
    接收模块,用于接收第一图像和第二图像,所述第一图像和所述第二图像是分别通过第一摄像头和第二摄像头拍摄同一场景获得的非同源图像;a receiving module, configured to receive a first image and a second image, where the first image and the second image are non-homologous images obtained by capturing the same scene through the first camera and the second camera, respectively;
    获取模块,用于获取与所述第一图像对应的坐标变换矩阵,所述坐标变换矩阵以所述第二图像为基准图像;An acquiring module, configured to acquire a coordinate transformation matrix corresponding to the first image, where the coordinate transformation matrix uses the second image as a reference image;
    变换模块,用于采用所述坐标变换矩阵对所述第一图像进行坐标变换;a transform module, configured to perform coordinate transformation on the first image by using the coordinate transformation matrix;
    融合模块,用于对坐标变换后的第一图像与所述第二图像进行图像融合处理。And a fusion module, configured to perform image fusion processing on the coordinate-converted first image and the second image.
  8. 根据权利要求7所述的装置,其特征在于,还包括:The device according to claim 7, further comprising:
    预处理模块,用于对所述第一图像和所述第二图像进行预处理,所述预处理包括:对所述第一图像进行灰度值取逆处理,对所述第二图像进行图像增强处理。 a pre-processing module, configured to perform pre-processing on the first image and the second image, where the pre-processing comprises: performing gray value inversion processing on the first image, and performing image on the second image Enhanced processing.
  9. 根据权利要求7所述的装置,其特征在于,所述获取模块包括:The device according to claim 7, wherein the obtaining module comprises:
    生成单元,用于分别生成平移矩阵A、旋转矩阵B和缩放矩阵C;Generating unit for respectively generating a translation matrix A, a rotation matrix B, and a scaling matrix C;
    确定单元,用于确定与所述第一图像对应的坐标变换矩阵T为:所述平移矩阵A、所述旋转矩阵B和所述缩放矩阵C依次相乘的结果。a determining unit, configured to determine a coordinate transformation matrix T corresponding to the first image as a result of sequentially multiplying the translation matrix A, the rotation matrix B, and the scaling matrix C.
  10. 根据权利要求9所述的装置,其特征在于,所述生成单元具体用于:The device according to claim 9, wherein the generating unit is specifically configured to:
    根据本地存储的旋转参数和缩放参数,分别生成所述旋转矩阵B和所述缩放矩阵C;Generating the rotation matrix B and the scaling matrix C according to locally stored rotation parameters and scaling parameters;
    将所述第一图像、所述第二图像以及所述旋转参数和所述缩放参数发送至图像配准处理组件,以使所述图像配准处理组件以所述第二图像为基准图像,结合所述旋转参数和所述缩放参数对所述第一图像进行配准处理,以获得平移参数;Transmitting the first image, the second image, and the rotation parameter and the scaling parameter to an image registration processing component, so that the image registration processing component uses the second image as a reference image, combined The rotation parameter and the scaling parameter perform registration processing on the first image to obtain a translation parameter;
    根据所述图像配准处理组件反馈的所述平移参数,生成所述平移矩阵A。The translation matrix A is generated according to the translation parameter fed back by the image registration processing component.
  11. 根据权利要求7所述的装置,其特征在于,所述融合模块包括:The device according to claim 7, wherein the fusion module comprises:
    灰度融合单元,用于根据如下公式对所述坐标变换后的第一图像与所述第二图像进行灰度融合处理,以获得融合后的灰度图像:a gradation fusion unit, configured to perform gradation fusion processing on the coordinate-transformed first image and the second image according to the following formula to obtain a fused grayscale image:
    g(x,y)=w1(x,y)*f1(x,y)+w2(x,y)*f2(x,y),其中,f1(x,y)为所述坐标变换后的第一图像中的任一像素点(x,y)的灰度值,f2(x,y)为所述第二图像中的对应像素点的灰度值,g(x,y)为灰度图像中对应像素点的灰度值;w1(x,y)和w2(x,y)为加权系数,w1(x,y)+w2(x,y)=1;g(x,y)=w1(x,y)*f1(x,y)+w2(x,y)*f2(x,y), where f1(x,y) is the coordinate transformed a gray value of any pixel point (x, y) in the first image, f2 (x, y) is a gray value of a corresponding pixel point in the second image, and g(x, y) is a gray scale Gray value of the corresponding pixel in the image; w1 (x, y) and w2 (x, y) are weighting coefficients, w1 (x, y) + w2 (x, y) = 1;
    色度渲染单元,用于以所述第二图像中各像素点的色度值渲染所述灰度图像中的对应像素点。And a chroma rendering unit, configured to render corresponding pixel points in the grayscale image with chroma values of respective pixels in the second image.
  12. 根据权利要求7至11中任一项所述的装置,其特征在于,所述接收模块,还用于接收所述第二摄像头拍得的第三图像;The device according to any one of claims 7 to 11, wherein the receiving module is further configured to receive a third image captured by the second camera;
    所述装置还包括:The device also includes:
    确定模块,用于根据所述第三图像的平均灰度值与预设灰度阈值的比较结果,确定是否触发所述第一摄像头和所述第二摄像头同时工作。And a determining module, configured to determine whether to trigger the first camera and the second camera to work simultaneously according to a comparison result between the average gray value of the third image and the preset gray threshold.
  13. 一种增强现实设备,其特征在于,包括: An augmented reality device, comprising:
    第一摄像头、第二摄像头、存储器和处理器;其中,a first camera, a second camera, a memory, and a processor; wherein
    所述第一摄像头,用于拍得第一图像;The first camera is configured to capture a first image;
    所述第二摄像头,用于拍得第二图像,所述第一图像和所述第二图像是分别拍摄同一场景获得的图像;The second camera is configured to capture a second image, where the first image and the second image are images obtained by respectively capturing the same scene;
    所述存储器用于存储一条或多条计算机指令,其中,所述一条或多条计算机指令被所述处理器执行时实现:The memory is for storing one or more computer instructions, wherein the one or more computer instructions are executed by the processor:
    获取与所述第一图像对应的坐标变换矩阵,所述坐标变换矩阵以所述第二图像为基准图像;Obtaining a coordinate transformation matrix corresponding to the first image, where the coordinate transformation matrix uses the second image as a reference image;
    采用所述坐标变换矩阵对所述第一图像进行坐标变换;Performing coordinate transformation on the first image by using the coordinate transformation matrix;
    对坐标变换后的第一图像与所述第二图像进行图像融合处理。Performing image fusion processing on the coordinate-converted first image and the second image.
  14. 根据权利要求13所述的设备,其特征在于,所述处理器还用于:对所述第一图像和所述第二图像进行预处理,所述预处理包括:对所述第一图像进行灰度值取逆处理,对所述第二图像进行图像增强处理。The device according to claim 13, wherein the processor is further configured to: perform pre-processing on the first image and the second image, the pre-processing comprising: performing the first image The gray value is inversely processed, and the second image is subjected to image enhancement processing.
  15. 根据权利要求13所述的设备,其特征在于,所述处理器在获取与所述第一图像对应的坐标变换矩阵时,具体用于:The device according to claim 13, wherein the processor is configured to: when acquiring a coordinate transformation matrix corresponding to the first image,
    分别生成平移矩阵A、旋转矩阵B和缩放矩阵C;Generating a translation matrix A, a rotation matrix B, and a scaling matrix C, respectively;
    确定与所述第一图像对应的坐标变换矩阵T为:所述平移矩阵A、所述旋转矩阵B和所述缩放矩阵C依次相乘的结果。Determining a coordinate transformation matrix T corresponding to the first image is a result of sequentially multiplying the translation matrix A, the rotation matrix B, and the scaling matrix C.
  16. 根据权利要求15所述的设备,其特征在于,所述处理器在分别生成平移矩阵A、旋转矩阵B和缩放矩阵C时,具体用于:The device according to claim 15, wherein when the processor generates the translation matrix A, the rotation matrix B, and the scaling matrix C, respectively, the processor is specifically configured to:
    根据本地存储的旋转参数和缩放参数,分别生成所述旋转矩阵B和所述缩放矩阵C;Generating the rotation matrix B and the scaling matrix C according to locally stored rotation parameters and scaling parameters;
    将所述第一图像、所述第二图像以及所述旋转参数和所述缩放参数发送至图像配准处理组件,以使所述图像配准处理组件以所述第二图像为基准图像,结合所述旋转参数和所述缩放参数对所述第一图像进行配准处理,以获得平移参数;Transmitting the first image, the second image, and the rotation parameter and the scaling parameter to an image registration processing component, so that the image registration processing component uses the second image as a reference image, combined The rotation parameter and the scaling parameter perform registration processing on the first image to obtain a translation parameter;
    根据所述图像配准处理组件反馈的所述平移参数,生成所述平移矩阵A。 The translation matrix A is generated according to the translation parameter fed back by the image registration processing component.
  17. 一种增强现实设备,其特征在于,包括:An augmented reality device, comprising:
    第一摄像头、第二摄像头、FPGA组件;其中,a first camera, a second camera, an FPGA component; wherein
    所述第一摄像头,用于拍得第一图像;The first camera is configured to capture a first image;
    所述第二摄像头,用于拍得第二图像,所述第一图像和所述第二图像是分别拍摄同一场景获得的图像;The second camera is configured to capture a second image, where the first image and the second image are images obtained by respectively capturing the same scene;
    所述FPGA组件中包含有实现如下步骤的功能逻辑:The FPGA component includes functional logic that implements the following steps:
    获取与所述第一图像对应的坐标变换矩阵,所述坐标变换矩阵以所述第二图像为基准图像;Obtaining a coordinate transformation matrix corresponding to the first image, where the coordinate transformation matrix uses the second image as a reference image;
    采用所述坐标变换矩阵对所述第一图像进行坐标变换;Performing coordinate transformation on the first image by using the coordinate transformation matrix;
    对坐标变换后的第一图像与所述第二图像进行图像融合处理。 Performing image fusion processing on the coordinate-converted first image and the second image.
PCT/CN2017/113578 2017-06-23 2017-11-29 Image processing method, device and augmented reality apparatus WO2018233217A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710487271.8A CN107230199A (en) 2017-06-23 2017-06-23 Image processing method, device and augmented reality equipment
CN201710487271.8 2017-06-23

Publications (1)

Publication Number Publication Date
WO2018233217A1 true WO2018233217A1 (en) 2018-12-27

Family

ID=59935307

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/113578 WO2018233217A1 (en) 2017-06-23 2017-11-29 Image processing method, device and augmented reality apparatus

Country Status (2)

Country Link
CN (1) CN107230199A (en)
WO (1) WO2018233217A1 (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107230199A (en) * 2017-06-23 2017-10-03 歌尔科技有限公司 Image processing method, device and augmented reality equipment
CN109727188A (en) * 2017-10-31 2019-05-07 比亚迪股份有限公司 Image processing method and its device, safe driving method and its device
CN111164962B (en) * 2018-09-26 2021-11-30 深圳市大疆创新科技有限公司 Image processing method, device, unmanned aerial vehicle, system and storage medium
CN109389630B (en) * 2018-09-30 2020-10-23 北京精密机电控制设备研究所 Method and device for determining and registering feature point set of visible light image and infrared image
CN111247558A (en) * 2018-12-04 2020-06-05 深圳市大疆创新科技有限公司 Image processing method, device, unmanned aerial vehicle, system and storage medium
CN109840881B (en) * 2018-12-12 2023-05-05 奥比中光科技集团股份有限公司 3D special effect image generation method, device and equipment
CN110160749B (en) * 2019-06-05 2022-12-06 歌尔光学科技有限公司 Calibration device and calibration method applied to augmented reality equipment
CN111127528A (en) * 2019-12-10 2020-05-08 Oppo广东移动通信有限公司 Image registration method, terminal and storage medium
CN113467601A (en) * 2020-03-31 2021-10-01 深圳光峰科技股份有限公司 Information display method, system and device based on augmented reality and projection equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1748167A (en) * 2003-02-06 2006-03-15 宝马股份公司 Method and device for visualizing a motor vehicle environment with environment-dependent fusion of an infrared image and a visual image
US20160093034A1 (en) * 2014-04-07 2016-03-31 Steven D. BECK Contrast Based Image Fusion
CN106296624A (en) * 2015-06-11 2017-01-04 联想(北京)有限公司 A kind of image interfusion method and device
CN107230199A (en) * 2017-06-23 2017-10-03 歌尔科技有限公司 Image processing method, device and augmented reality equipment

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102510436B (en) * 2011-10-17 2014-06-25 河海大学常州校区 Device and method for detecting high-speed tiny target online in real time by simulating fly vision
CN102982518A (en) * 2012-11-06 2013-03-20 扬州万方电子技术有限责任公司 Fusion method of infrared image and visible light dynamic image and fusion device of infrared image and visible light dynamic image
CN103606139A (en) * 2013-09-09 2014-02-26 上海大学 Sonar image splicing method
CN104535978A (en) * 2014-12-19 2015-04-22 西安工程大学 Three-dimensional InISAR image registration and fusion method based on mutual information
CN104680559B (en) * 2015-03-20 2017-08-04 青岛科技大学 The indoor pedestrian tracting method of various visual angles based on motor behavior pattern
CN105701828B (en) * 2016-01-14 2019-09-20 广州视睿电子科技有限公司 Image processing method and device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1748167A (en) * 2003-02-06 2006-03-15 宝马股份公司 Method and device for visualizing a motor vehicle environment with environment-dependent fusion of an infrared image and a visual image
US20160093034A1 (en) * 2014-04-07 2016-03-31 Steven D. BECK Contrast Based Image Fusion
CN106296624A (en) * 2015-06-11 2017-01-04 联想(北京)有限公司 A kind of image interfusion method and device
CN107230199A (en) * 2017-06-23 2017-10-03 歌尔科技有限公司 Image processing method, device and augmented reality equipment

Also Published As

Publication number Publication date
CN107230199A (en) 2017-10-03

Similar Documents

Publication Publication Date Title
WO2018233217A1 (en) Image processing method, device and augmented reality apparatus
US11182961B2 (en) Method and system for representing a virtual object in a view of a real environment
TWI675583B (en) Augmented reality system and color compensation method thereof
US10303973B2 (en) Image processing apparatus, image processing method, and storage medium for lighting processing on image using model data
WO2018103244A1 (en) Live streaming video processing method, device, and electronic apparatus
US9811910B1 (en) Cloud-based image improvement
US10021295B1 (en) Visual cues for managing image capture
CN106462937B (en) Image processing apparatus and image display apparatus
US20150206354A1 (en) Image processing apparatus and image display apparatus
WO2020237565A1 (en) Target tracking method and device, movable platform and storage medium
US11563889B2 (en) Electronic device and method for controlling camera using external electronic device
EP3629303A1 (en) Method and system for representing a virtual object in a view of a real environment
JP2020188448A (en) Imaging apparatus and imaging method
AU2016100369A4 (en) Method and system for providing position or movement information for controlling at least one function of a vehicle
CN113724140B (en) Image processing method, electronic device, medium and system
TWI736052B (en) Method and system for building environment map
WO2023124201A1 (en) Image processing method and electronic device
CN116437198B (en) Image processing method and electronic equipment
CN107426522B (en) Video method and system based on virtual reality equipment
KR101822169B1 (en) Electronic device for providing panorama image and control method thereof
US11636708B2 (en) Face detection in spherical images
WO2022179412A1 (en) Recognition method and electronic device
JP2023148897A (en) information processing system
JP2023148898A (en) information processing system
JP2023149009A (en) information processing system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17914571

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17914571

Country of ref document: EP

Kind code of ref document: A1