WO2021057582A1 - Image matching, 3d imaging and pose recognition method, device, and system - Google Patents
Image matching, 3d imaging and pose recognition method, device, and system Download PDFInfo
- Publication number
- WO2021057582A1 WO2021057582A1 PCT/CN2020/115736 CN2020115736W WO2021057582A1 WO 2021057582 A1 WO2021057582 A1 WO 2021057582A1 CN 2020115736 W CN2020115736 W CN 2020115736W WO 2021057582 A1 WO2021057582 A1 WO 2021057582A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- current
- deformed
- matching
- image block
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/33—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
Definitions
- the invention relates to the technical field of image matching, in particular to methods, devices and systems for image matching, 3D imaging and gesture recognition.
- the control unit matches a single image collected by a single image sensor with a reference image, or matches two images collected by the left and right image sensors, and then performs 3D imaging or performs posture recognition of the target object according to the matching result.
- the present invention provides a method, device and system for image matching, 3D imaging and gesture recognition.
- the first aspect of the present invention provides an image matching method.
- the image matching method includes:
- Detect the deformed area and/or non-deformed area in the initial image generate a matching result for the deformed area and/or generate a matching result for the non-deformed area.
- the detecting the deformed area and/or the non-deformed area in the initial image includes:
- the current image block is a current deformed image block or a current non-deformable image block; wherein the non-deformable area is a set of the current non-deformable image block; the deformed area Is the set of the current deformed image blocks.
- the detecting the deformed area and/or the non-deformed area in the initial image includes:
- the current image block is the current deformed image block; if yes, the current image block is the current non-deformed image block; wherein, the non-deformed area is a set of the current non-deformed image blocks; the deformed area Is the set of the current deformed image blocks.
- the generating a matching result for the deformed region includes:
- Matching is performed on the reference image to obtain a matching result for the deformed area.
- the converting the deformed area into a reference image includes:
- the current deformed image block is converted into the current reference image block.
- the converting the deformed area into a reference image includes:
- the deformed area is converted into a reference image based on Fourier transform.
- the converting the deformed area into a reference image includes:
- the current deformed image block is converted into a current reference image block.
- the converting the deformed area into a reference image includes:
- the generating a matching result for the deformed region includes:
- Matching is performed on the image group to obtain a matching result.
- the generating a matching result for the deformed region includes:
- Matching is performed on the template image group to obtain a matching result.
- the initial image is an image collected by an image sensor after an image is projected to the target object; wherein the projected image has a periodic gradual change pattern and is unique within a certain spatial range, or is within a certain spatial range It is unique.
- the second aspect of the present invention provides a gesture recognition method, the gesture recognition method includes:
- the posture information of the target object is generated.
- a third aspect of the present invention provides a 3D imaging method, the 3D imaging method includes:
- a fourth aspect of the present invention provides an image matching device, the image matching device includes:
- Image acquisition module to acquire the initial image
- the image matching module is used to detect the deformed area and/or non-deformed area in the initial image, generate a matching result for the deformed area and/or generate a matching result for the non-deformed area.
- a fifth aspect of the present invention provides a gesture recognition device, and the gesture recognition device includes:
- the posture generation module is used to generate posture information of the target object according to the matching result.
- a sixth aspect of the present invention provides a 3D imaging device, the 3D imaging device includes:
- the image generation module is used to generate a 3D image of the target object according to the matching result.
- a seventh aspect of the present invention provides a system, which includes: an image projector, an image sensor, and a control unit;
- the image projector is used to project an image to a target object
- the image sensor is used to collect the initial image of the target object after the image is projected
- the control unit is configured to implement the image matching method described in the first aspect; the gesture recognition method described in the second aspect; and/or the steps of the 3D imaging method described in the third aspect.
- An eighth aspect of the present invention provides a computer device that includes a memory, a processor, and a computer program that is stored in the memory and can run on the processor.
- the processor executes the computer program, Implement the steps of the image matching method described in the first aspect; the gesture recognition method described in the second aspect; and/or the 3D imaging method described in the third aspect.
- a ninth aspect of the present invention provides a computer-readable storage medium, the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the image matching method described in the first aspect is implemented; The gesture recognition method described above; and/or the steps of the 3D imaging method described in the third aspect.
- Fig. 1A is a first schematic structural diagram of an embodiment of a system provided by the present invention
- Fig. 1B is a second schematic structural diagram of an embodiment of a system provided by the present invention
- Fig. 1C is a third schematic structural diagram of an embodiment of the system provided by the present invention ;
- FIG. 2 is a schematic flowchart of an embodiment of an image matching method provided by the present invention.
- FIG. 3 is a schematic diagram of the first process of an embodiment of detecting a deformed area and/or a non-deformed area in an initial image provided by the present invention
- FIG. 4 is a schematic diagram of a second process of an embodiment of detecting a deformed area and/or a non-deformed area in an initial image provided by the present invention
- FIG. 5 is a schematic diagram of the first process of an embodiment of generating a matching result for a deformed area provided by the present invention
- FIG. 6 is a schematic diagram of a second process of an embodiment of generating a matching result for a deformed area provided by the present invention.
- FIG. 7 is a schematic diagram of a third process of an embodiment of generating a matching result for a deformed area provided by the present invention.
- FIG. 8 is a schematic diagram of the first process of an embodiment of detecting that a current image block is a current deformed image block or a current non-deformable image block based on image features according to the present invention
- FIG. 9 is a schematic diagram of a second process of an embodiment of detecting whether a current image block is a current deformed image block or a current non-deformable image block based on image characteristics according to the present invention.
- FIG. 10 is a schematic diagram of the first process of an embodiment of converting a deformed area into a reference image provided by the present invention
- FIG. 11 is a schematic diagram of the second process of an embodiment of converting a deformed area into a reference image provided by the present invention.
- FIG. 12 is a schematic diagram of the third process of an embodiment of converting a deformed area into a reference image provided by the present invention.
- FIG. 13 is a schematic flowchart of an embodiment of 3D imaging provided by the present invention.
- FIG. 14 is a schematic flowchart of an embodiment of a gesture recognition method provided by the present invention.
- FIG. 15 is a first structural block diagram of an embodiment of a matching device provided by the present invention.
- FIG. 16 is a second structural block diagram of an embodiment of a matching device provided by the present invention.
- FIG. 17 is a third structural block diagram of an embodiment of a matching device provided by the present invention.
- FIG. 18 is a fourth structural block diagram of an embodiment of a matching device provided by the present invention.
- FIG. 19 is a fifth structural block diagram of an embodiment of a matching device provided by the present invention.
- FIG. 20 is a sixth structural block diagram of an embodiment of a matching device provided by the present invention.
- 21 is a structural block diagram of an embodiment of a gesture recognition device provided by the present invention.
- FIG. 22 is a structural block diagram of an embodiment of a 3D imaging device provided by the present invention.
- Figure 23 is a structural block diagram of an embodiment of a computer device provided by the present invention.
- FIG. 24 is a schematic diagram of an embodiment of image conversion provided by the present invention.
- FIG. 25A is a first schematic diagram of an embodiment of an initial image provided by the present invention
- FIG. 25B is a second schematic diagram of an embodiment of an initial image provided by the present invention
- FIG. 25C is a third schematic diagram of an embodiment of an initial image provided by the present invention
- Figure 25D is a fourth schematic diagram of an embodiment of an initial image provided by the present invention
- Fig. 26A is a schematic diagram of an embodiment of a current preprocessed image block provided by the present invention
- Fig. 26B is a schematic diagram of an embodiment of a current preprocessed image block after conversion provided by the present invention
- FIG. 27 is a schematic diagram of intermediate results produced by the method for detecting deformed regions provided by the present invention.
- an embodiment of the present invention provides a system, which includes an image sensor 11, an image projector 12, and a control unit 13;
- the image projector 12 is used to project an image to a target object.
- the image projector 12 can be used to project a periodic gradual and unique image within a certain spatial range to the target object.
- the image projector 12 can be used to project a periodic gradual and unique image within a certain spatial range to the target object.
- a periodic gradual and unique image within a certain spatial range to the target object.
- the periodic gradual change rule may include, but is not limited to, periodically changing sine wave (or cosine wave, etc.) fringes.
- the sine wave does not necessarily fully meet the sine wave standard, and may also be fringes close to sine.
- incompletely regular sine fringes, or linearly changing sine fringes also called triangular waves.
- being unique within a certain spatial range means that an image frame of a certain specification moves in any area of the image, and the image in the corresponding area within the image frame is always unique. For example: sine wave stripes with random patterns.
- the image projector 12 can also be used to project a unique image within a certain spatial range, such as a scattered pattern, to the target object.
- the image projector 12 can also project any other images that can meet the requirements of subsequent matching.
- the image projector 12 may be a projector or a laser projector that can project the above-mentioned image that is currently or will be developed in the future.
- the image sensor 11 is used to collect the initial image of the target object after the projected image.
- the image sensor 11 sends the collected initial image to the control unit 13, the memory or the server, and so on.
- the image sensor 11 may be, but is not limited to, a camera, a video camera, a scanner, or other devices with related functions (mobile phones, computers, etc.), and so on.
- the image sensor 11 may be a group (as shown in FIG. 1A, 1B) or multiple groups (as shown in FIG. 1C) arranged around the target object; wherein, each group of image sensors may include one image sensor (as shown in FIG. 1A). (Shown), 2 image sensors (as shown in Figures 1B and 1C), or more than 2 image sensors (illustration omitted).
- the image sensor 11 can be fixedly arranged relative to the target object, or can be movably arranged relative to the target object, which is not limited in this specific embodiment.
- the control unit 13 respectively communicates with the image projector 11 and the image sensor 12 in a wired or wireless manner, and communicates with the image projector 11 and the image sensor 12.
- Wireless methods may include, but are not limited to: 3G/4G, WIFI, Bluetooth, WiMAX, Zigbee, UWB (ultra wideband), and other wireless connection methods that are currently or will be developed in the future.
- the control unit 13 may be a part of an independent computing device; it may also be a part of the image projector 11; and/or a part of the image sensor 12. In this specific embodiment, for convenience of description, the control unit 13 is a separate component Show.
- control unit 13 can be Programmable Logic Controller (PLC), Field-Programmable Gate Array (FPGA), Computer (Personal Computer, PC), Industrial Control Computer (Industrial Personal Computer, IPC) , Smart phone, tablet or server, etc.
- PLC Programmable Logic Controller
- FPGA Field-Programmable Gate Array
- PC Computer
- IPC Industrial Control Computer
- Smart phone Smart phone, tablet or server, etc.
- the control unit generates program instructions according to a pre-fixed program, combined with manually input information, parameters, or data collected by an external image sensor.
- the present invention provides an image matching method.
- the image matching method includes the following method steps:
- Step S110 acquiring an initial image
- Step S120 detects the deformed area and/or the non-deformed area in the initial image, and generates a matching result for the deformed area and/or generates a matching result for the non-deformed area.
- the deformed area and/or non-deformed area in the initial image are determined by detection, that is, the initial image may include both deformed areas and non-deformed areas at the same time, or all deformed areas, or all non-deformed areas.
- the image matching method described in the following embodiment can be used to directly match the non-deformed area, and for the deformed area, the deformed area needs to be processed (further detailed description will be given in the following embodiment). Then, perform matching according to the image matching method described in the following embodiment.
- Step S110 acquiring an initial image
- a group of image sensors can be set, each group of image sensors includes 1 image sensor, the initial image is one, and a series of different distances are stored in advance. Download the reference image of the target object after the projected image collected by the image sensor, and subsequently match the initial image with a series of reference images.
- a group of image sensors may be provided, each group of image sensors includes two image sensors 11, and each group of initial images includes a first initial image and a second initial image. Two images.
- N groups of image sensors are set according to the above description, where N is an integer greater than or equal to 2, and each group of images is composed of two image sensors 11, that is, each group of initial The image includes the first initial image and the second initial image.
- image splicing methods currently available or developed in the future can be used, such as: feature-based splicing and region-based splicing.
- the region-based splicing method can start from the gray value of the image to be spliced, and calculate the gray value of an area of the image to be spliced with the same size in the reference image using other methods such as least squares. The difference is compared to the difference to determine the similarity of the overlapping area of the image to be spliced, so as to obtain the range and position of the overlapping area of the spliced image, and then realize image splicing.
- the feature-based stitching is to derive the features of the image through pixels, and then use the image features as a standard to search and match the corresponding feature regions of the overlapping parts.
- each group of initial images varies according to the number of image sensors.
- the number of image sensors can be any number greater than or equal to 1. Since the imaging mode of more than 2 image sensors is also a combination of any two of them, the imaging method is The corresponding imaging methods under two image sensors are repeated. Therefore, this specific embodiment only uses a single group of image sensors, and the group of image sensors includes one image sensor or two image sensors as an example for description.
- Step S120 detects the deformed area and/or the non-deformed area in the initial image, and generates a matching result for the non-deformed area and/or generates a matching result for the deformed area, respectively.
- the deformed area refers to some areas in the initial image that are inconsistent with the reference image.
- the deformed area is due to the target object (1 image sensor as an example) or part of the surface on the target object (2 image sensors as an example) relative to the reference surface (taking 2 image sensors as an example, the reference surface can be parallel to the image sensor)
- the surface of the chip; taking an image sensor as an example, the reference surface can refer to the deflection, bending, etc.
- the inconsistency may be manifested as: the deformation area is deflected, stretched, compressed, and/or bent relative to the reference image.
- the sine wave image in the deformed area is deflected by 2 left (as shown in Fig. 25A), right deflection (as shown in Fig. 25B), and the sine wave in the deformed area Stretching (as shown in FIG. 25C), compression (as shown in FIG. 25D), or bending (illustration omitted), etc.
- the non-deformed area refers to the area in the initial image that is consistent with the reference image.
- the image matching method may include but is not limited to the following method steps:
- the matching is performed between the two initial images.
- the initial image includes a first initial image and a second initial image.
- a fixed-size image frame is usually set with the matched pixel as the center. Match the image blocks in the image frame.
- An n*n image block in the first initial image will be compared with the N image blocks of the same size in the second initial image along the epipolar direction of the two image sensors (N is one of the parallaxes of the two images Search scope).
- the method of comparison is to calculate the absolute value of the brightness difference of the corresponding pixels of the two image blocks, and then sum the absolute value to obtain a matching score.
- N matching scores can be obtained, and the minimum value of these N matching scores can be obtained, and the pixel point in the second initial image corresponding to this minimum value is the matched pixel point in the first initial image Correspondingly, and so on, so as to obtain the matching result of multiple pixels in the two images corresponding to each other.
- one image sensor is taken as an example.
- the above embodiment is to match between the first initial image and the second initial image, and in the embodiment of one image sensor, the initial The image is matched with the reference image, and the specific matching method can be referred to the above embodiment, which will not be repeated here.
- step S120 may include the following method steps:
- Step S121 acquiring the current image block in the initial image
- the current image block is the image block for current matching.
- Step S123 based on the image characteristics of the current image block, detect that the current image block is the current deformed image block or the current non-deformable image block;
- the initial image includes a deformed area and/or a non-deformed area.
- the deformed area refers to the current set of deformed image blocks;
- the non-deformed area refers to the current set of non-deformed image blocks.
- the deformed area can be detected by judging whether these characteristics are deformed, and then the deformed area can be converted to the image generated when the corresponding reference surface is generated.
- the image includes the characteristic part of the fringe with a sinusoidal gradual change.
- the sinusoidal fringe is parallel to the reference plane, the horizontal X is the repeated change of multiple cycles, and the vertical Y is aligned, which is Upright state.
- the period of the stripe in the horizontal X direction may be stretched or compressed, and/or the vertical Y direction may be inclined, and the current image block may be deformed according to the changes of these characteristic parts in the image.
- step S120 may include the following method steps:
- Step S122 obtains the current image block in the initial image
- Step S124 match the current image block; determine whether a matching result can be generated
- step S126 use the current image block as the current deformed image block
- step S128 use the current image block as the current non-deformable image block.
- an image frame of a preset size is sequentially moved by a unit amount (e.g., the first initial image) along the initial image (e.g., the first initial image).
- step S120 can be implemented through but not limited to the following method steps:
- step S120 may include the following method steps:
- Step S221 Obtain a deformed area
- the deformed area may be a collection of multiple current deformed image blocks, so the current deformed image block can be obtained, and the current deformed image block can be converted into the current reference image block in the following step S222.
- Step S222 Convert the deformed area into a reference image
- Step S223 performs matching on the reference image to obtain a matching result for the deformed area.
- the deformed area is replaced with a reference image, and the reference image is matched.
- the matching method please refer to the relevant description in the image matching method in the above embodiment, which will not be repeated here.
- step S120 may include the following method steps:
- Step S321 obtains a template image group in which the pre-generated deformation area has a unit deformation in sequence
- the template image group may be a pre-generated image template group obtained after the current deformed image block sequentially undergoes at least one unit deformation amount.
- a current template image group is generated in advance corresponding to each current image block.
- each template corresponds to the image block
- the template image can be generated in advance according to the deflection unit, combined with the spatial position conversion information of the image projector and the image sensor; or according to the method of fitting function in the embodiment shown in FIG. 11, the corresponding Template image.
- Step S322 performs matching on the template image group to obtain a matching result.
- the template image corresponding to the current image block in the first initial image is sequentially obtained (for example: deflection -60, -40, -20, 0, 20, 40, and 60 template images), the corresponding second initial image is located on the same epipolar line, and it can include multiple image blocks to be matched.
- Each template image and the second initial The sum of the absolute values of the gray-scale differences between the image blocks to be matched in the image corresponds to the second initial image with the smallest sum of the absolute values of the gray-scale differences and the current image block of the first initial image.
- the pixels corresponding to the two matching image blocks are matched pixels, and the matching result is obtained.
- the initial image and the reference image can be matched according to the method described in the above embodiment.
- the reference image can be regarded as the second initial image to generate the matching result.
- the template image group can be generated in advance according to the deflection unit, combined with the spatial position conversion information of the image projector and the image sensor; or according to the method of fitting function in the embodiment shown in FIG. 11, the corresponding Set of template images.
- step S120 may include the following method steps:
- Step S421 sequentially generates image groups after unit deformation of the deformed area occurs
- Step S422 performs matching on the image group to obtain a matching result.
- the deformed image block group formed after the current image block in the first initial image is deflected by a unit angle and the plurality of waiting images located on the same epipolar line in the second initial image are sequentially acquired.
- the matching image blocks are matched, and the sum of the absolute values of the gray-scale differences between each deformed image block and the image block to be matched in the second initial image is calculated, and then the second one with the smallest sum of the absolute values of the gray-scale differences is calculated.
- the to-be-matched image block of the initial image and the current image block of the first initial image are matched image blocks, so the pixels corresponding to the two matched image blocks are matched pixels, thereby obtaining the matching result.
- the image group after unit deformation can be generated according to the deflection unit, combined with the spatial position conversion information of the image projector and the image sensor; it can also be based on the fitting function in the embodiment shown in FIG. 9 below. Method to generate the corresponding image group.
- step S123 based on the image characteristics of the current image block, detecting that the current image block is the current deformed image block or the current non-deformable image block can be implemented by, but not limited to, the following method steps:
- step S123 may include the following method steps:
- the image frame can be moved along the epipolar direction of the image sensor on the initial image, and each time it moves one unit, the current image on the initial image corresponding to the image frame is the current image block.
- the current deformed image block when it is necessary to further convert the current deformed image block into a reference image based on the detection method, there may be a blank part on the edge of the current deformed image block after the conversion (as shown in FIG. 26B As shown), in order to ensure the integrity of the content of the current image block after conversion, it is usually necessary to obtain a current image block whose size is larger than the actual matching image block (as shown in Figure 26A), and then perform the conversion. The latter image is cropped to obtain the current deformed image block completely converted into the reference image (as shown in FIG. 26C).
- the peak or trough of each sine wave should be on the same reference line.
- the pixel point at the peak or valley is the maximum value pixel point, for example, the gray value is the highest or the lowest.
- S1235 fits the most value pixel points in each row to obtain the current fitting line of the current image block
- S1237 detects the current image shape variable of the current fitted line relative to the reference line
- a threshold can be preset. If the deformation is less than When it is equal to or less than the threshold (in one embodiment, the threshold may also be a theoretical value of zero), the current image block is considered to be the current non-deformed image block; if it is greater than or equal to a certain threshold, the current image block is determined Is the current deformed image block, and the shape variable of the fitted line relative to the reference line is the current image shape variable of the current image block.
- step S123 may include the following method steps:
- the current image block is fitted to obtain a fitting function.
- A is the amplitude of the sine function;
- 360/B is the period of the function's brightness change (unit: Pixels),
- C represents the inclination of the pixel in the Y direction.
- B pixels the inclination angle is arccotangent (C/B);
- D is the translation amount of the function in the lateral X direction;
- E is the translation amount of the function in the Z direction; where A, D and E are fixed values.
- C represents the inclination of the pixel in the Y direction.
- C is 0;
- arccotangent is the image shape variable of the current deformed image block.
- it can be determined whether the image deformation amount is greater than or equal to a certain threshold; if so, the current image block is a deformed image block; if not, the current image block is a non-deformable image block.
- step S222 converting the deformed area into a reference image can be achieved through but not limited to the following method steps:
- step S222 may include the following method steps:
- Step S2221 obtains the current deformed image block
- Step S2222 extracts the most value pixel points in each row of the current deformed image block
- Step S2223 fits the most value pixel points in each row to obtain the current fitted line of the current image block
- Step S2224 detects the current image shape variable of the current fitting line relative to the reference line
- step S2221 to step S2224 can refer to the description of step S1231 to step S1237 in step S123 in the above embodiment.
- steps S2221-step S2224 here can be omitted.
- Step S2225 converts the current deformed image block into the current reference image block based on the current image shape variable.
- step S222 may include the following method steps:
- Step S2231 obtains the current image block
- Step S2232 performs fitting on the current image block to obtain the current fitting function
- step S2231 to step S2232 can be found in the description of step S1232 to step S1234 in the above embodiment.
- steps S2231-step S2232 here can be omitted.
- Step S2233 converts the current deformed image block into the current reference image block based on the current fitting function
- the deformed area can be converted based on the fitting function, that is, the period of the deformed area is the reference period, and the C of the image is 0, so as to obtain the deformed area after conversion.
- the size of the image block to be acquired is (width W: 33, height H: 33), since there will be blanks at the edges of the converted image (as shown in Figure 26B), in order to ensure the integrity of the image block content, it is usually necessary to intercept An initial image block larger than the size of the image block, for example, as shown in Figure 26A, a wider area is selected as the initial image block (W: 63, H: 33), where the center of the initial image block Form a rectangular frame, the size of the rectangular frame corresponds to the size of the image block (W: 33, H: 33); in addition, the initial image block can also be any other size, as long as it is larger than the preset image block And ensure that the content in the converted image block is complete.
- the initial image block is fitted with a three-dimensional curve according to the above fitting function.
- the transformation method is as follows: Since the height of the image is H: 33, take the middle row (that is, the 17th row) without transformation.
- the middle row 17 is used as the 0 point reference, and the number of other rows is i, and the i-th row is translated along the positive direction of the X axis by -(i-17)*(C/B) pixels. After all rows are traversed, the transformation is completed. such as:
- the initial image is converted, and the converted initial image becomes as shown in FIG. 26B.
- the area within the rectangular frame (W: 33, H: 33) located in the middle of the deformed initial image is intercepted as the converted image block, and the converted image block after the interception is shown in FIG. 26C.
- step S222 may include the following method steps:
- Step S2241 obtains the current deformed image block
- Step S2242 generates the current deformation of the target object based on the current deformation image block
- Step S2243 based on the current deformation amount, converts the current deformed image block into the current reference image block;
- the actual spatial position coordinates of point A in the image corresponding to the image sensor coordinate system can be obtained according to the initial image; according to the deformation of point A relative to the reference plane L'(for example: the deformation can be implemented according to the above
- the image shape variable obtained in the example is converted based on the calibration result of the image sensor; or obtained based on the template image group described in the above embodiment)
- the point A corresponding to the direction of the projected image of the image projector 12 can be calculated
- the position coordinates of the virtual point A'projected on the reference plane O in the image sensor coordinate system According to the conversion between the first and second image sensor coordinate systems and the first and second image coordinate systems, the positions A'can be obtained respectively The coordinates in the first and second image coordinate systems, and so on, can obtain the corrected image of the deformed area.
- step S222 may be based on Fourier transform to convert the deformed area into a reference image.
- a 3D imaging method is further provided, and the 3D imaging method includes the matching method described in the above embodiment, and further includes the steps:
- S130 generates a 3D image of the target object according to the matching result.
- the posture information of the corresponding point of the matching point pair in the three-dimensional space can be obtained, based on the multiple matching point pairs included in the matching result , That is, you can draw a 3D image of the object in a three-dimensional space.
- the posture information can be 3d coordinates in a preset coordinate system for the target, and the motion of a rigid body in a 3-dimensional space can be described by 3d coordinates (a total of 6 degrees of freedom). Specifically, it can be divided into rotation and translation, each It is 3 degrees of freedom.
- the translation of a rigid body in a 3-dimensional space is an ordinary linear transformation, and a 3x1 vector can be used to describe the translation position; and the commonly used description methods of rotation pose include but are not limited to: rotation matrix, rotation vector, quaternion, Euler angle and Lie algebra.
- a gesture recognition method is also provided.
- the gesture recognition method includes the matching method described in the above embodiment, and further includes the steps:
- S140 generates a posture recognition result of the target object according to the matching result.
- each matching point pair is based on the triangulation algorithm or based on the corresponding reference image, and the posture information of the corresponding point in the three-dimensional space of the matching point pair can be obtained, based on the posture of one or more three-dimensional space points Information to obtain the result of the posture recognition of the target object.
- the result of the posture recognition of the target object may be the posture information representing the position and posture of the entire object or the posture information of a certain target position associated with the target object (for example, the target position may be located on the target object or on the including frame of the target object).
- the posture information of the target object can be obtained.
- the Linemod method refers to storing the corresponding point cloud images at multiple angles based on the 3D model of the target object (such as: CAD model) in advance, and matching the point cloud image obtained in the above embodiment with the image in the image library to determine The posture information of the target object.
- an image matching device is provided, and the image matching device includes:
- the image acquisition module 110 is used to acquire an initial image
- the result generating module 120 is configured to detect the deformed area and/or the non-deformed area in the initial image, generate a matching result for the deformed area and/or generate a matching result for the non-deformed area.
- the above result generation module 120 includes:
- the image acquisition unit 121 is configured to acquire the current image block in the initial image
- the image detection unit 123 is configured to detect whether the current image block is a current deformed image block or a current non-deformable image block based on the image characteristics of the current image block;
- the above result generation module 120 includes:
- the image acquisition unit 122 is configured to acquire the current image block in the initial image
- the image matching unit 124 is used to match the current image block, and the result judgment unit 126 is used to judge whether a matching result can be generated;
- the image determining unit 128 is configured to, if not, the current image block is the current deformed image block; if so, the current image block is the current non-deformable image block.
- the above result generation module 120 includes:
- the deformation acquiring unit 221 is configured to acquire the deformation area
- the deformation conversion unit 222 is configured to convert the deformation area into a reference image
- the reference matching unit 223 is configured to perform matching on the reference image to obtain a matching result for the deformed area.
- the above result generating module 120 includes:
- the template obtaining unit 321 is configured to obtain a template image group in which the pre-generated deformation area has a unit deformation in sequence;
- the template matching unit 322 is configured to perform matching on the template image group to obtain a matching result.
- the above result generation module 120 includes:
- the image generating unit 421 is configured to sequentially generate image groups after unit deformation of the deformed area occurs;
- the image matching unit 422 is configured to perform matching on the image group to obtain a matching result.
- the image detection unit 123 includes:
- the image acquisition unit 1231 is used to acquire the current image block
- the maximum value extraction unit 1233 is used to extract the maximum value pixel points in each row of the current image block
- the best value fitting unit 1235 is used to fit the best value pixel points in each row to obtain the current fitting line of the current image block;
- the deformation detection unit 1237 is used to detect the current image deformation of the current fitting line relative to the reference line;
- the image detection unit 1239 is configured to detect whether the current image block is the current deformed image block or the current non-deformable image block according to the current image deformation.
- the image detection unit 123 includes:
- the image acquisition unit 1232 acquires the current image block
- the function extraction unit 1234 fits the current image block to obtain a fitting function
- the image detection unit 1236 detects whether the current image block is the current deformed image block or the current non-deformed image block according to the fitting function.
- the deformation conversion unit 222 includes:
- the image acquiring unit 2221 is used to acquire the current deformed image block in the deformed area
- the maximum value extraction unit 2222 is configured to extract the maximum value pixel points in each row of the current deformed image block
- the best value fitting unit 2223 is configured to fit the best value pixel points in each row to obtain the current fitted line;
- the deformation calculation unit 2224 is used to calculate the current image deformation of the current fitting line relative to the reference line;
- the image conversion unit 2225 is configured to convert the current deformed image block into the current reference image block based on the current image shape variable.
- the deformation conversion unit 222 includes:
- the image acquiring unit 2231 is used to acquire the current deformed image block in the deformed area
- the function extraction part 2232 is used for fitting the current deformed image block to obtain the current fitting function
- the image conversion unit 2233 is configured to convert the current deformed image block into the current reference image block based on the current fitting function.
- the deformation conversion unit 222 includes:
- the image acquiring unit 2241 is used to acquire the current deformed image block of the deformed area
- the deformation generating unit 2242 is configured to generate the current deformation of the target object based on the current deformed image block;
- the image conversion unit 2243 is configured to convert the current deformed image block into the current reference image block based on the current deformation amount.
- the deformation conversion unit 222 includes:
- the image conversion unit 2251 is configured to convert the deformed area into a reference image based on Fourier transform.
- a gesture recognition device As shown in FIG. 21, in one embodiment, a gesture recognition device is provided, and the gesture recognition device includes:
- the posture generation module 130 is configured to generate posture information of the target object according to the matching result.
- a 3D imaging device is provided, and the 3D imaging device includes:
- the 3D imaging module 140 is configured to generate a 3D image of the target object according to the matching result.
- each module in each of the above-mentioned devices may be implemented in whole or in part by software, hardware, and a combination thereof.
- the above-mentioned modules may be embedded in the form of hardware or independent of the processor in the computer equipment, or may be stored in the memory of the computer equipment in the form of software, so that the processor can call and execute the operations corresponding to the above-mentioned modules.
- a computer-readable storage medium stores a computer program, and the computer program implements the image matching method described in the above embodiments when the computer program is executed by a processor, Steps of 3D imaging method and/or gesture recognition method.
- a computer device is also provided.
- the computer device includes a memory, a processor, and a computer program that is stored in the memory and can run on the processor.
- the processor executes the computer program, the steps of the image matching method, 3D imaging method, and/or gesture recognition method described in the above embodiments are implemented.
- Industrial control computers have important computer attributes and characteristics. Therefore, they all have computer central processing unit (CPU), hard disk, memory and other internal memories, as well as plug-in hard disks. External storage such as Smart Media Card (SMC), Secure Digital (SD) Card, Flash Card (Flash Card), etc., with operating system, control network and protocol, computing power, and friendly man-machine interface, It is to provide reliable, embedded, intelligent computers and industrial control computers for other structures/equipment/systems.
- SMC Smart Media Card
- SD Secure Digital
- Flash Card Flash Card
- the computer program may be divided into one or more modules/units, and the one or more modules/units are stored in the memory and executed by the processor to complete the present invention.
- the one or more modules/units may be a series of computer program instruction segments capable of completing specific functions, and the instruction segments are used to describe the execution process of the computer program in the control unit.
- the so-called processor can be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), off-the-shelf Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
- the memory may be a storage device built into the terminal, such as a hard disk or a memory.
- the memory may also be an external storage device of the control unit, such as a plug-in hard disk equipped on the control unit, a Smart Media Card (SMC), a Secure Digital (SD) card, and a flash memory. Card (Flash Card), etc.
- the memory may also include not only an internal storage unit of the control unit, but also an external storage device.
- the memory is used to store the computer program and other programs and data required by the terminal.
- the memory can also be used to temporarily store data that has been output or will be output.
- FIG. 23 is only an example of a computer device, and does not constitute a limitation on the computer device. It may include more or less components than shown in the figure, or a combination of certain components, or different components, such as
- the control unit may also include input and output devices, network access devices, buses, and the like.
- the disclosed devices and methods can be implemented in other ways.
- the embodiments of each device described above are only illustrative.
- the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, such as multiple units or Components can be combined or integrated into another system, or some features can be omitted or not implemented.
- the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
- the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
- the functional units in the various embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
- the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
- the integrated module/unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
- the present invention implements all or part of the processes in the above-mentioned embodiments and methods, and can also be completed by instructing relevant hardware through a computer program.
- the computer program can be stored in a computer-readable storage medium. When the program is executed by the processor, it can implement the steps of the foregoing method embodiments. .
- the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate forms.
- the computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signal, telecommunications signal, and software distribution media, etc.
- ROM Read-Only Memory
- RAM Random Access Memory
- electrical carrier signal telecommunications signal
- software distribution media etc.
- the content contained in the computer-readable medium can be appropriately added or deleted according to the requirements of the legislation and patent practice in the jurisdiction.
- the computer-readable medium Does not include electrical carrier signals and telecommunication signals.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (19)
- 一种图像匹配方法,其特征在于,所述图像匹配方法包括:An image matching method, characterized in that the image matching method includes:获取初始图像;Get the initial image;检测所述初始图像中的形变区域和/或非形变区域,生成针对所述形变区域的匹配结果和/或生成针对所述非形变区域的匹配结果。Detect the deformed area and/or non-deformed area in the initial image, generate a matching result for the deformed area and/or generate a matching result for the non-deformed area.
- 根据权利要求1所述的图像匹配方法,其特征在于,所述检测所述初始图像中的形变区域和/或非形变区域包括:The image matching method according to claim 1, wherein the detecting the deformed area and/or the non-deformed area in the initial image comprises:获取所述初始图像中的当前图像块;Acquiring the current image block in the initial image;基于所述当前图像块的图像特征,检测所述当前图像块为当前形变图像块或当前非形变图像块;其中,所述非形变区域为所述当前非形变图像块的集合;所述形变区域为所述当前形变图像块的集合。Based on the image characteristics of the current image block, it is detected that the current image block is a current deformed image block or a current non-deformable image block; wherein the non-deformable area is a set of the current non-deformable image block; the deformed area Is the set of the current deformed image blocks.
- 根据权利要求1所述的图像匹配方法,其特征在于,所述检测所述初始图像中的形变区域和/或非形变区域包括:The image matching method according to claim 1, wherein the detecting the deformed area and/or the non-deformed area in the initial image comprises:获取所述初始图像中的当前图像块;Acquiring the current image block in the initial image;匹配所述当前图像块,判断是否能够生成匹配结果;Match the current image block to determine whether a matching result can be generated;若否,所述当前图像块为当前形变图像块;若是,所述当前图像块为当前非形变图像块;其中,所述非形变区域为所述当前非形变图像块的集合;所述形变区域为所述当前形变图像块的集合。If not, the current image block is the current deformed image block; if yes, the current image block is the current non-deformed image block; wherein, the non-deformed area is a set of the current non-deformed image blocks; the deformed area Is the set of the current deformed image blocks.
- 根据权利要求1或2或3所述的图像匹配方法,其特征在于,所述生成针对所述形变区域的匹配结果包括:The image matching method according to claim 1 or 2 or 3, wherein the generating a matching result for the deformed area comprises:获取所述形变区域;Obtaining the deformation area;转换所述形变区域为基准图像;Converting the deformed area into a reference image;针对所述基准图像进行匹配,得到针对所述形变区域的匹配结果。Matching is performed on the reference image to obtain a matching result for the deformed area.
- 根据权利要求4所述的图像匹配方法,其特征在于,所述转换所述形变区域为基准图像包括:The image matching method according to claim 4, wherein said converting said deformed area into a reference image comprises:获取所述形变区域中的当前形变图像块;Acquiring the current deformed image block in the deformed area;提取所述当前形变图像块的每行中的最值像素点;Extracting the most value pixel point in each row of the current deformed image block;拟合所述每行中的所述最值像素点,得到拟合线;Fitting the most value pixel points in each row to obtain a fitting line;计算所述拟合线相对基准线的图像形变量;Calculating the image shape variable of the fitted line relative to the reference line;基于所述图像形变量,转换所述当前形变图像块为当前基准图像块。Based on the image deformation variable, the current deformed image block is converted into the current reference image block.
- 根据权利要求4所述的图像匹配方法,其特征在于,所述转换所述形变区域为基准图像包括:The image matching method according to claim 4, wherein said converting said deformed area into a reference image comprises:基于傅里叶变换转换所述形变区域为基准图像。The deformed area is converted into a reference image based on Fourier transform.
- 根据权利要求4所述的图像匹配方法,其特征在于,所述转换所述形变区域为基准图像包括:The image matching method according to claim 4, wherein said converting said deformed area into a reference image comprises:获取所述形变区域中的当前形变图像块;Acquiring the current deformed image block in the deformed area;对所述当前形变图像块进行拟合,得到拟合函数;Fitting the current deformed image block to obtain a fitting function;基于所述拟合函数,转换所述当前形变图像块为当前基准图像块。Based on the fitting function, the current deformed image block is converted into a current reference image block.
- 根据权利要求4所述的图像匹配方法,其特征在于,所述转换所述形变区域为基准图像包括:The image matching method according to claim 4, wherein said converting said deformed area into a reference image comprises:获取所述形变区域的当前形变图像块;Acquiring the current deformed image block of the deformed area;基于所述当前形变图像块,生成目标物的形变量;Based on the current deformed image block, generating a deformation variable of the target object;基于所述形变量,转换所述当前形变图像块为当前基准图像块。Based on the deformation variable, convert the current deformed image block into a current reference image block.
- 根据权利要求1或2或3所述的图像匹配方法,其特征在于,所述生成针对所述形变区域的匹配结果包括:The image matching method according to claim 1 or 2 or 3, wherein the generating a matching result for the deformed area comprises:依次生成所述形变区域发生单位形变量后的图像组;Sequentially generating image groups of the deformed area after unit deformation occurs;针对所述图像组进行匹配,得到匹配结果。Matching is performed on the image group to obtain a matching result.
- 根据权利要求1或2或3所述的图像匹配方法,其特征在于,所述生成针对所述形变区域的匹配结果包括:The image matching method according to claim 1 or 2 or 3, wherein the generating a matching result for the deformed area comprises:获取预生成的所述形变区域依次发生单位形变量后的模板图像组;Acquiring a pre-generated group of template images in which the deformation area sequentially undergoes unit deformation;针对所述模板图像组进行匹配,得到匹配结果。Matching is performed on the template image group to obtain a matching result.
- 根据权利要求1或2或3所述的图像匹配方法,其特征在于,所述初始图像为通过图像传感器采集的向目标物体投射图像后的图像;其中,被投射的所述图像成周期性渐变规律且在一定空间范围内具有唯一性,或在一定空间范围内具有唯一性。The image matching method according to claim 1 or 2 or 3, wherein the initial image is an image collected by an image sensor and projected to a target object; wherein the projected image becomes a periodic gradient Regular and unique within a certain space, or unique within a certain space.
- 一种姿态识别方法,其特征在于,所述姿态识别方法包括:A gesture recognition method, characterized in that the gesture recognition method includes:权利要求1-11任一项所述图像匹配方法;及The image matching method according to any one of claims 1-11; and根据所述匹配结果,生成目标物体的姿态信息。According to the matching result, the posture information of the target object is generated.
- 一种3D成像方法,其特征在于,所述3D成像方法包括:A 3D imaging method, characterized in that, the 3D imaging method includes:权利要求1-11任一项所述图像匹配方法;及The image matching method according to any one of claims 1-11; and根据所述匹配结果,生成目标物体的3D图像。According to the matching result, a 3D image of the target object is generated.
- 一种图像匹配装置,其特征在于,所述图像匹配装置包括:An image matching device, characterized in that the image matching device includes:图像获取模块,获取初始图像;Image acquisition module to acquire the initial image;图像匹配模块,用于检测所述初始图像中的形变区域和/或非形变区域,生成针对所述形变区域的匹配结果和/或生成针对所述非形变区域的匹配结果。The image matching module is used to detect the deformed area and/or non-deformed area in the initial image, generate a matching result for the deformed area and/or generate a matching result for the non-deformed area.
- 一种姿态识别装置,其特征在于,所述姿态识别装置包括:A gesture recognition device, characterized in that the gesture recognition device includes:权利要求14所述图像匹配装置;及The image matching device of claim 14; and姿态生成模块,用于根据所述匹配结果,生成目标物体的姿态信息。The posture generation module is used to generate posture information of the target object according to the matching result.
- 一种3D成像装置,其特征在于,所述3D成像装置包括:A 3D imaging device, characterized in that, the 3D imaging device comprises:权利要求14所述图像匹配装置;及The image matching device of claim 14; and图像生成模块,用于根据所述匹配结果,生成目标物体的3D图像。The image generation module is used to generate a 3D image of the target object according to the matching result.
- 一种系统,其特征在于,所述系统包括:图像投射器、图像传感器和控制单元;A system, characterized in that the system includes: an image projector, an image sensor, and a control unit;所述图像投射器,用于向目标物体投射图像;The image projector is used to project an image to a target object;所述图像传感器,用于采集被投射所述图像后的所述目标物体的初始图像;The image sensor is used to collect the initial image of the target object after the image is projected;所述控制单元,用于实现权利要求1-11任一项所述的图像匹配方法;权利要求12所述的姿态识别方法;和/或权利要求13所述的3D成像方法的步骤。The control unit is configured to implement the image matching method according to any one of claims 1-11; the gesture recognition method according to claim 12; and/or the steps of the 3D imaging method according to claim 13.
- 一种计算机设备,所述计算机设备包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现权利要求1-11任一项所述的图像匹配方法;权利要求12所述的姿态识别方法;和/或权利要求13所述的3D成像方法的步骤。A computer device, the computer device comprising a memory, a processor, and a computer program stored in the memory and running on the processor, wherein the processor implements the rights when the computer program is executed. The image matching method according to any one of claims 1-11; the gesture recognition method according to claim 12; and/or the steps of the 3D imaging method according to claim 13.
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现权利要求1-11任一项 所述的图像匹配方法;权利要求12所述的姿态识别方法;和/或权利要求13所述的3D成像方法的步骤。A computer-readable storage medium storing a computer program, wherein the computer program implements the image matching method according to any one of claims 1-11 when the computer program is executed by a processor; The gesture recognition method according to claim 12; and/or the steps of the 3D imaging method according to claim 13.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910926826.3 | 2019-09-27 | ||
CN201910926826.3A CN112581512A (en) | 2019-09-27 | 2019-09-27 | Image matching, 3D imaging and posture recognition method, device and system |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021057582A1 true WO2021057582A1 (en) | 2021-04-01 |
Family
ID=75110514
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/115736 WO2021057582A1 (en) | 2019-09-27 | 2020-09-17 | Image matching, 3d imaging and pose recognition method, device, and system |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN112581512A (en) |
WO (1) | WO2021057582A1 (en) |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08305854A (en) * | 1995-05-01 | 1996-11-22 | Nippon Telegr & Teleph Corp <Ntt> | Projection parameter extraction processing method |
CN1489382A (en) * | 2002-07-23 | 2004-04-14 | 日本电气视象技术株式会社 | Projecting apparatus |
CN102393398A (en) * | 2002-09-30 | 2012-03-28 | 应用材料以色列公司 | Illumination system for optical inspection |
CN108305218A (en) * | 2017-12-29 | 2018-07-20 | 努比亚技术有限公司 | Panoramic picture processing method, terminal and computer readable storage medium |
CN108391145A (en) * | 2017-02-20 | 2018-08-10 | 南安市耀森智能设备有限公司 | A kind of robot |
CN110139033A (en) * | 2019-05-13 | 2019-08-16 | Oppo广东移动通信有限公司 | Camera control method and Related product |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2787453B2 (en) * | 1988-10-25 | 1998-08-20 | 大日本印刷株式会社 | How to determine the pattern allocation area of the pattern film |
CN101198964A (en) * | 2005-01-07 | 2008-06-11 | 格斯图尔泰克股份有限公司 | Creating 3D images of objects by illuminating with infrared patterns |
JP6251489B2 (en) * | 2013-03-28 | 2017-12-20 | 株式会社 資生堂 | Image analysis apparatus, image analysis method, and image analysis program |
AU2015202937A1 (en) * | 2015-05-29 | 2016-12-15 | Canon Kabushiki Kaisha | Systems and methods for registration of images |
CN105956997B (en) * | 2016-04-27 | 2019-07-05 | 腾讯科技(深圳)有限公司 | The method and apparatus of image deformation processing |
WO2018050223A1 (en) * | 2016-09-14 | 2018-03-22 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Pattern detection |
CN108133492B (en) * | 2016-12-01 | 2022-04-26 | 京东方科技集团股份有限公司 | Image matching method, device and system |
CN109427046B (en) * | 2017-08-30 | 2021-07-20 | 深圳中科飞测科技股份有限公司 | Distortion correction method and device for three-dimensional measurement and computer readable storage medium |
-
2019
- 2019-09-27 CN CN201910926826.3A patent/CN112581512A/en active Pending
-
2020
- 2020-09-17 WO PCT/CN2020/115736 patent/WO2021057582A1/en active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08305854A (en) * | 1995-05-01 | 1996-11-22 | Nippon Telegr & Teleph Corp <Ntt> | Projection parameter extraction processing method |
CN1489382A (en) * | 2002-07-23 | 2004-04-14 | 日本电气视象技术株式会社 | Projecting apparatus |
CN102393398A (en) * | 2002-09-30 | 2012-03-28 | 应用材料以色列公司 | Illumination system for optical inspection |
CN108391145A (en) * | 2017-02-20 | 2018-08-10 | 南安市耀森智能设备有限公司 | A kind of robot |
CN108305218A (en) * | 2017-12-29 | 2018-07-20 | 努比亚技术有限公司 | Panoramic picture processing method, terminal and computer readable storage medium |
CN110139033A (en) * | 2019-05-13 | 2019-08-16 | Oppo广东移动通信有限公司 | Camera control method and Related product |
Also Published As
Publication number | Publication date |
---|---|
CN112581512A (en) | 2021-03-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3614340B1 (en) | Methods and devices for acquiring 3d face, and computer readable storage media | |
US10872227B2 (en) | Automatic object recognition method and system thereof, shopping device and storage medium | |
US10008005B2 (en) | Measurement system and method for measuring multi-dimensions | |
US9135710B2 (en) | Depth map stereo correspondence techniques | |
US10726580B2 (en) | Method and device for calibration | |
US10810718B2 (en) | Method and device for three-dimensional reconstruction | |
US10455219B2 (en) | Stereo correspondence and depth sensors | |
CN105335748B (en) | Image characteristic extracting method and system | |
EP3135033B1 (en) | Structured stereo | |
KR20200044676A (en) | Method and apparatus for active depth sensing and calibration method thereof | |
WO2012096747A1 (en) | Forming range maps using periodic illumination patterns | |
US11468609B2 (en) | Methods and apparatus for generating point cloud histograms | |
CN113362445B (en) | Method and device for reconstructing object based on point cloud data | |
CN116129037B (en) | Visual touch sensor, three-dimensional reconstruction method, system, equipment and storage medium thereof | |
CN109375833B (en) | Touch instruction generation method and device | |
Wei et al. | An accurate stereo matching method based on color segments and edges | |
Yin et al. | Estimation of the fundamental matrix from uncalibrated stereo hand images for 3D hand gesture recognition | |
CN112197708B (en) | Measuring method and device, electronic device and storage medium | |
JP7298687B2 (en) | Object recognition device and object recognition method | |
CN116912417A (en) | Texture mapping method, device, equipment and storage medium based on three-dimensional reconstruction of human face | |
WO2021057582A1 (en) | Image matching, 3d imaging and pose recognition method, device, and system | |
JP2021021577A (en) | Image processing device and image processing method | |
JP2017162251A (en) | Three-dimensional noncontact input device | |
CN106651940B (en) | Special processor for 3D interaction | |
Hu et al. | Active shape reconstruction using a novel visuotactile palm sensor |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20869735 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20869735 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 29.09.2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20869735 Country of ref document: EP Kind code of ref document: A1 |