CN106446883A - Scene reconstruction method based on light label - Google Patents

Scene reconstruction method based on light label Download PDF

Info

Publication number
CN106446883A
CN106446883A CN201610789231.4A CN201610789231A CN106446883A CN 106446883 A CN106446883 A CN 106446883A CN 201610789231 A CN201610789231 A CN 201610789231A CN 106446883 A CN106446883 A CN 106446883A
Authority
CN
China
Prior art keywords
optical label
image
scene
background image
label
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610789231.4A
Other languages
Chinese (zh)
Other versions
CN106446883B (en
Inventor
王晓东
方俊
李江亮
苏爱民
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Guangshi fusion Intelligent Technology Co.,Ltd.
Original Assignee
Xi'an Small Photon Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xi'an Small Photon Network Technology Co Ltd filed Critical Xi'an Small Photon Network Technology Co Ltd
Priority to CN201610789231.4A priority Critical patent/CN106446883B/en
Publication of CN106446883A publication Critical patent/CN106446883A/en
Application granted granted Critical
Publication of CN106446883B publication Critical patent/CN106446883B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/16Image acquisition using multiple overlapping images; Image stitching

Abstract

The invention relates to a scene reconstruction method based on a light label and can reinforce and highlight the transmission information, eliminate interference of the irrelevant information and guarantee information transmission accuracy and high effectiveness. The method comprises steps that 1, within the unit time, light label scanning is carried out through light label identification equipment, a light label in a scene is discovered, and a light label image is shot; the positioning identification coordinate position of the light label is mapped from a frame image to the light label image and then is taken as a positioning point; 2, a background image of the light label in the scene is acquired, and the background image comprises all the image regions of the scene except the positioning point; 3, image dimension proportion zooming, image information simplification, image edge enhancement processing and image rendering are sequentially carried out for the background image to acquire a reconstructed background image; and 4, the reconstructed background image and the light label image are fused and overlapped to acquire a reconstruction scene, and the reconstruction scene is sent to VR display equipment to accomplish scene reconstruction within the unit time.

Description

Scene reconstruction method based on optical label
Technical field
The present invention relates to the method for scene reconstruction, specially the scene reconstruction method based on optical label.
Background technology
Optical label, can be as VR (virtual reality) collecting device in entity space due to its unique information transmission characteristic In anchor point, locus is demarcated.Based on these optical label anchor points, soft hardware equipment just can be by entity space Object projects in VR space according to actual position relation, thus setting up the mapping of the two.Based on the mapping in both spaces, then More senior application can be implemented, and such as entity interaction accesses, space orientation and roaming, associate and calculate etc..However, only Anchor point sets up the mapping relations in space not enough, because in whole Virtual Space, in addition to anchor point, the blank between anchor point is also Needs are filled with, thus being more beneficial for user to the interpretation in VR space, understanding, the image filled also needs to reach prominent, strong Change the purpose of optical label.Herein, solving one of method that this problem can adopt is exactly that the background image of entity space is synchronous Collect, be filled into after process in VR scene.Because not all background image information is all related to the application of optical label, Therefore it is accomplished by carrying out proper treatment to these Backgrounds, emphasize including key character, render simplification with irrelevant information etc., then It is overlapped showing with optical label image, completes the reconstruct to scene, there is presently no the optical label scene reconstruction method of correlation.
Content of the invention
For problems of the prior art, the present invention provides a kind of scene reconstruction method based on optical label, can Optical label transmission information is strengthened and is projected, exclude the interference of other irrelevant informations it is ensured that the accuracy of information transmission and High efficiency.
The present invention is to be achieved through the following technical solutions:
Based on the scene reconstruction method of optical label, comprise the steps,
By optical label, step 1, within the unit interval, identifies that equipment carries out optical label scanning, finds the cursor in scene Sign, and pickup light label image;The positioning mark coordinate position of optical label is mapped to optical label image from two field picture, as Anchor point;
Step 2, collection optical label institute background image in the scene, described background image is included in scene except anchor point All image-regions in addition;
Step 3, carries out picture size proportional zoom successively, image information simplifies, at the enhancing of image border to background image Reason and image rendering, obtain the background image reconstructing;
Step 4, by the background image of reconstruct and optical label image carry out merge superposition, obtain reconstruct scene be sent to VR show Show the scene reconstruction completing in equipment in this unit interval.
Preferably, the comprising the following steps that of step 1,
Optical label identifies equipment continuous acquisition multiframe scene image, and shoots the figure that a width includes the optical label of anchor point Picture;Adjacent any two frames are made the difference and gets differential chart, differential chart finds the positioning mark of optical label, obtains optical label Coordinate position in optical label image;Thus positioning mark coordinate position luminous in optical label is mapped to cursor from two field picture Sign in image, obtain the coordinate position of optical label and the anchor point of optical label picture registration;Wherein, during collection between consecutive frame Between the time interval that flashes twice not less than the dynamic position identifiers of optical label.
Further, sit when the luminous positioning that mapping obtains identifies actual luminous positioning mark in coordinate and optical label image Mark when not being completely superposed;First, the position relationship according to optical label and luminous positioning mark, obtains optical label in optical label figure Position range in picture and scene areas;Secondly, in the scene areas comprising optical label, to the positioning mark in multiple image Symbol scans for identifying, respectively obtains positioning in multiple image and identifies corresponding coordinate, and then obtains the equal of all respective coordinates Value is as the positioning mark of optical label and the coincidence point of optical label image, i.e. anchor point.
Preferably, the comprising the following steps that of step 2, the time slot using optical label collection non-in optical label identification equipment is carried out The collection of background image.
Preferably, the comprising the following steps that of step 3,
Step 3.1, picture size proportional zoom;By the display in the background image displaying ratio of entity space and VR space Ratio is consistent, obtains scaling background image;
Step 3.2, image information simplifies;Scaling background image is carried out gray processing process, obtains gray scale pretreatment background Image;Using the reflection of RGB color component, heat, intensity of illumination or energy, remove unrelated figure in gray scale pretreatment background image As information realization simplifies, obtain gray scale background image;
Step 3.3, image border enhancing is processed;To gray scale background image I2Carry out high-pass filtering to obtain strengthening the pretreatment back of the body Scape image I2', then edge is obtained by following formula and strengthen background image I3,
I3=I2+k·I2’;
Wherein, k is adjustment factor;
Step 3.4, image rendering;Content obtaining is carried out to optical label information by optical label coding/decoding method, according to cursor Frame information content and the rgb value of storage and render the corresponding of striped and render feature in label information, with the position of optical label as the center of circle, Strengthen background image for radius to edge with adjustable pixel digit r to render, the background image after being reconstructed.
Further, in step 3.2, irrelevant information is removed simplify using RGB color component in gray scale background image When, using any one in following three kinds of methods,
Maximum value process, R=G=B=Max (R, G, B);
Mean value method, R=G=B=(R+G+B)/3;
Weighted average method, R=G=B=wr·R+wg·G+wb·B;
Wherein, R, G, B are the RGB color component value of any pixel of gray scale pretreatment background image;wr、wg、wbIt is respectively The weights of R, G, B, wr、wg、wbIt is in [0,1] interval.
Further, in step 3.4, coding/decoding method indicates coding by the algorithm flag bit in the signal element of optical label, and Obtained in the decoding algorithm database of standard by coding.
Compared with prior art, the present invention has following beneficial technique effect:
The present invention in the mapping carrying out the entity space based on optical label and VR space, using optical label as positioning " anchor Point ", the position forming above two space accurately corresponds to, and also needs to the image information of reality scene is especially located simultaneously Reason, reconstruct scene, form complete VR space the display of prominent optical label;By dividing to background image and cursor glyph From, and background image is reconstructed with process, solve enhancing optical label in the picture and show, propose and optical label correlation simultaneously Not high background information is so that user focuses more on the display of optical label.
Brief description
Optical label sample described in Fig. 1 present example.
Scene reconstruction schematic diagram described in Fig. 2 present example.
Image schematic diagram after Information Simplification described in Fig. 3 present example.
Virtual Space Scene after reconstruct described in Fig. 4 present example.
The scene reconstruction method flow diagram based on optical label for Fig. 5 present invention.
In figure:1 identifies equipment for optical label, and 2 is the real scene residing for optical label, and 3 is optical label, and 4 is in scene Background object, 5 is scene reconstruction processing server, and 6 is the reconstruct image in Virtual Space for the real scene.
Specific embodiment
With reference to specific embodiment, the present invention is described in further detail, described be explanation of the invention and It is not to limit.
The present invention in the mapping carrying out the entity space based on optical label and VR space, using optical label as positioning " anchor Point ", the position forming above two space accurately corresponds to, and also needs to the image information of reality scene is especially located simultaneously Reason, reconstruct scene, form complete VR space the display of prominent optical label.
Described optical label sample is as shown in Figure 1.Optical label includes signal element (cell) group (or referred to as " data Position ") and positioning mark (or referred to as " flag bit ") two parts, wherein position and be designated upper three larger rectangle frames (three of figure This rectangle frame individual is referred to as " one group of position identifiers "), under working condition, positioning mark is flashed with certain Frequency Synchronization, by figure The method that aberration divides can obtain quickly and be detected by picture pick-up device, and then can determine the position of signal element by positioning mark Put, thus carrying out data identification and reading;Signal element is the black and white rectangle between positioning mark, and multiple signal elements constitute one Group, usual signal element forms the array of 5 × 5 (being not limited to), and each signal element represents " 0 " or " 1 " of data signal, entirely (length of side of flag is the two of the data bit length of side to the digital signal sequences of matrix composition one frame of signal element group composition here Times, easily facilitate positioning), in order to increase the data space that signal element represents, under working condition, each signal element also can be according to Predetermined program is flashed, thus more signal contents are shown by multiframe.At this moment need to provide one in multiframe Start frame/end of identification frame, for demarcating the beginning/end position of one complete cycle of multiframe, this frame signal unit group sets Combine for a special data, such as:Not different with the information of the actual capabilities statement specific combination of full 0 or complete 1 or any.
The scene reconstruction based on optical label for the present invention is as shown in Figure 2.Scene reconstruction method based on optical label is with a list The position time is a cycle, and circulation execution, as shown in figure 5, its process is:
Step one:Carry out optical label scanning, find the optical label in scene;The method of above-mentioned optical label scanning is as follows:
Optical label identifies equipment 1 continuous acquisition multiframe label image, is designated as:f0, f1..., fm, collection between consecutive frame The time interval that time flashes twice not less than dynamic position identifiers;And then, shoot a width optical label image, be designated as p;Right Adjacent any two frames make the difference gets differential chart, find in differential chart optical label 3 positioning mark, obtain further its Coordinate position in optical label image;The position coordinates of luminescence unit is mapped to optical label image from two field picture;Due to can Can there is the reasons such as hand shaking, the luminous positioning that mapping obtains identifies actual luminous positioning mark in coordinate and optical label image p Coordinate is not completely superposed, and at this moment the position relationship according to optical label and luminous positioning mark, can get optical label in optical label Position range in image and scene areas, in the scene areas comprising optical label, enter to position identifiers in multiple image Line search identifies, respectively obtains the coordinate positioning mark in multiple image, the average calculating these coordinates is as final cursor The positioning mark signed and the coincidence point of optical label image p, i.e. anchor point.
Step 2:To all image-regions regard as being background image in addition to anchor point;Collection optical label institute is in the scene Background image, because optical label identifies that the shooting of equipment or photographing device are not all to be used for light in all of IMAQ time slot The data acquisition of label, therefore can be utilized the time slot of non-optical label collection to carry out the collection of background image;In unit interval T Interior, T=[t1,m1,t2,m2,t3,m3,t4,m4,t5,m5,t6...], wherein tiIt is optical label data acquisition time slot, i is positive integer, mjGather time slot for non-optical label, j is positive integer, then the background image obtaining in above-mentioned non-optical label data acquisition time slot represents For I, gather the shutter speed that shooting or photographing device need to be adopted as a upper time slot during I;
Step 3:Background image is carried out process reconstruct, the method for above-mentioned Image Reconstruction is as follows:
First dimension scale scaling is carried out to background image so that the background image displaying ratio of entity space and VR space In displaying ratio consistent, note source background image is I0, then the scaling background image after proportional zoom is I1, carry out image herein The method of proportional zoom can be arbitrary image scaling method;
Optical label scene information simplifies, and carries out gray processing process to Background herein, obtains gray scale pretreatment background image, Remove unrelated image information in gray scale pretreatment background image to realize simplifying, obtain gray scale background image I2, may be selected following Method one carries out this process, and three kinds of methods are as follows:
● maximum value process:R=G=B=Max (R, G, B);
● mean value method:R=G=B=(R+G+B)/3;
● weighted average method:R=G=B=wr·R+wg·G+wbB,
Herein, R, G, B are any pixel RGB color component value of gray scale pretreatment background image, wr、wg、wbRespectively R, The weights of G, B, wr、wg、wbIt is in [0,1] interval, Max is maximum value finding function, finally gives gray scale background image I2, this Place can also carry out simplification process according to other different quantizating index to image, such as:Heat, intensity of illumination, energy reflection etc., As Fig. 3, and it is not limited to RGB color component;
Optical label scene information is strengthened, herein to gray scale background image I2Carry out edge enhancing process, specifically to ash Degree background image I2Carry out high-pass filtering, obtain strengthening pretreatment background image I2', calculated according to the following formula:
I3=I2+k·I2’;
In above formula, k is adjustment factor, then obtain strengthening background image I through edge3
Carry out optical label scene information to render, content obtaining is carried out to optical label information by optical label coding/decoding method, should Coding/decoding method indicates coding by algorithm flag bit in the signal element of optical label, and by encoding the decoding algorithm data in standard Obtain in storehouse;If the information obtaining current cursor this frame of label is m, in colors list below, inquire about corresponding color:
The information content Rgb value Texturing patterns
m (R、G、B) T
Herein, T is texturing patterns, can be realized from third-party resource using any;This list is to preset;With light The position of label is the center of circle, according to frame information content above and rgb value and renders striped for radius with adjustable pixel digit r The corresponding feature that renders strengthens background image I to edge3Rendered, r>0;Through above-mentioned process, the background after being reconstructed Image I4, background image I after reconstruct4Eliminate the redundant image information unrelated with optical label and enhance around optical label Image shows;
Step 4:Background image after reconstruct is overlapped with optical label image, submits to VR display device, finish, Carry out the above-mentioned process of next time cycle if necessary, finally give the result of similar Fig. 4.
Present invention achieves strengthen optical label in the picture showing, the background letter not high with optical label correlation is proposed simultaneously Breath is so that user focuses more on the display of optical label.
In actually used, it is capable of following scene and function.User U is by optical label technology roaming access commercial street Road, so that highlighting in the VR equipment (Google glasses) that optical label uses at it, needs to scene reconstruction.Use first Family U uses image in face of image capture device acquisition;By there being two optical labels in face of optical label scanning discovery, identify one respectively Family restaurant and a hotel;After identification, this method is utilized respectively different time-gap and has collected optical label coding information and current field Scape information;Scene information is reconstructed, the optical label ambient background pointing to hotel is carried out lines enhancing, allows user see pure mellow wine The profile in shop and scale, and hotel's color rendering is become soft light green;The optical label ambient background pointing to restaurant is carried out Lines weaken, and allow user see the style within dining room clearly, and color rendering is become bright-coloured yellow;In above-mentioned background reconfiguration method With the help of, user U is more prone to be concerned about corresponding optical label, and enjoys the service that it is provided, and eliminates other unrelated The unduly interfering with of information.

Claims (7)

1. the scene reconstruction method based on optical label is it is characterised in that comprise the steps,
By optical label, step 1, within the unit interval, identifies that equipment carries out optical label scanning, finds the optical label in scene, and Pickup light label image;The positioning mark coordinate position of optical label is mapped to optical label image from two field picture, as positioning Point;
Step 2, collection optical label background image in the scene, described background image includes in scene in addition to anchor point All image-regions;
Step 3, background image is carried out successively picture size proportional zoom, image information simplify, image border strengthen process and Image rendering, obtains the background image reconstructing;
Step 4, the background image of reconstruct and optical label image are carried out merging superposition, obtain reconstructing scene being sent to VR and showing and set The scene reconstruction in this unit interval is completed in standby.
2. the scene reconstruction method based on optical label according to claim 1 is it is characterised in that the concrete steps of step 1 It is as follows,
Optical label identifies equipment continuous acquisition multiframe scene image, and shoots the image that a width includes the optical label of anchor point;Right Adjacent any two frames make the difference gets differential chart, finds the positioning mark of optical label, obtain optical label in light in differential chart Coordinate position in label image;Thus positioning mark coordinate position luminous in optical label is mapped to optical label figure from two field picture In picture, obtain the coordinate position of optical label and the anchor point of optical label picture registration;Wherein, the acquisition time between consecutive frame is not The time interval flashing twice less than the dynamic position identifiers of optical label.
3. the scene reconstruction method based on optical label according to claim 2 it is characterised in that when mapping obtain luminous When in positioning mark coordinate and optical label image, actual luminous positioning mark coordinate is not completely superposed;First, according to cursor Sign and the luminous position relationship positioning mark, obtain position range in optical label image for the optical label and scene areas;Secondly, In the scene areas comprising optical label, the position identifiers in multiple image is scanned for identify, respectively obtain multiframe figure Position in picture and identify corresponding coordinate, and then obtain the average of all respective coordinates and identify and optical label as the positioning of optical label The coincidence point of image, i.e. anchor point.
4. the scene reconstruction method based on optical label according to claim 1 is it is characterised in that the concrete steps of step 2 As follows, carry out the collection of background image using the time slot of optical label collection non-in optical label identification equipment.
5. the scene reconstruction method based on optical label according to claim 1 is it is characterised in that the concrete steps of step 3 It is as follows,
Step 3.1, picture size proportional zoom;By the displaying ratio in the background image displaying ratio of entity space and VR space Unanimously, obtain scaling background image;
Step 3.2, image information simplifies;Scaling background image is carried out gray processing process, obtains gray scale pretreatment background image; Using the reflection of RGB color component, heat, intensity of illumination or energy, remove unrelated image letter in gray scale pretreatment background image Breath is realized simplifying, and obtains gray scale background image;
Step 3.3, image border enhancing is processed;To gray scale background image I2Carry out high-pass filtering to obtain strengthening pretreatment Background As I2', then edge is obtained by following formula and strengthen background image I3,
I3=I2+k·I2’;
Wherein, k is adjustment factor;
Step 3.4, image rendering;Content obtaining is carried out to optical label information by optical label coding/decoding method, is believed according to optical label Frame information content and the rgb value of storage and render the corresponding of striped and render feature in breath, with the position of optical label as the center of circle, with can Pixel digit r adjusted is that radius renders to edge enhancing background image, the background image after being reconstructed.
6. the scene reconstruction method based on optical label according to claim 5 is it is characterised in that in step 3.2, using ash When in degree background image, RGB color component is removed to irrelevant information simplifying, using any one in following three kinds of methods,
Maximum value process, R=G=B=Max (R, G, B);
Mean value method, R=G=B=(R+G+B)/3;
Weighted average method, R=G=B=wr·R+wg·G+wb·B;
Wherein, R, G, B are the RGB color component value of any pixel of gray scale pretreatment background image;wr、wg、wbBe respectively R, G, The weights of B, wr、wg、wbIt is in [0,1] interval.
7. the scene reconstruction method based on optical label according to claim 5 is it is characterised in that in step 3.4, decoding side Method indicates coding by the algorithm flag bit in the signal element of optical label, and is obtained in the decoding algorithm database of standard by coding ?.
CN201610789231.4A 2016-08-30 2016-08-30 Scene reconstruction method based on optical label Active CN106446883B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610789231.4A CN106446883B (en) 2016-08-30 2016-08-30 Scene reconstruction method based on optical label

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610789231.4A CN106446883B (en) 2016-08-30 2016-08-30 Scene reconstruction method based on optical label

Publications (2)

Publication Number Publication Date
CN106446883A true CN106446883A (en) 2017-02-22
CN106446883B CN106446883B (en) 2019-06-18

Family

ID=58163637

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610789231.4A Active CN106446883B (en) 2016-08-30 2016-08-30 Scene reconstruction method based on optical label

Country Status (1)

Country Link
CN (1) CN106446883B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107886017A (en) * 2017-11-09 2018-04-06 陕西外号信息技术有限公司 A kind of read method and equipment of optical label sequence
CN109710198A (en) * 2018-12-29 2019-05-03 森大(深圳)技术有限公司 Method of printing, device and the equipment of local dynamic station variable image
CN110471580A (en) * 2018-05-09 2019-11-19 北京外号信息技术有限公司 Information equipment exchange method and system based on optical label
CN111754449A (en) * 2019-03-27 2020-10-09 北京外号信息技术有限公司 Scene reconstruction method based on optical communication device and corresponding electronic equipment
WO2021057887A1 (en) * 2019-09-26 2021-04-01 北京外号信息技术有限公司 Method and system for setting virtual object capable of being presented to target
TWI785332B (en) * 2020-05-14 2022-12-01 光時代科技有限公司 Three-dimensional reconstruction system based on optical label
WO2022257807A1 (en) * 2021-06-09 2022-12-15 丰疆智能科技股份有限公司 Positioning apparatus based on passive infrared tag

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101208723A (en) * 2005-02-23 2008-06-25 克雷格·萨默斯 Automatic scene modeling for the 3D camera and 3D video
CN101610421A (en) * 2008-06-17 2009-12-23 深圳华为通信技术有限公司 Video communication method, Apparatus and system
CN101742349A (en) * 2010-01-05 2010-06-16 浙江大学 Method for expressing three-dimensional scenes and television system thereof
CN103049728A (en) * 2012-12-30 2013-04-17 成都理想境界科技有限公司 Method, system and terminal for augmenting reality based on two-dimension code
CN103528571A (en) * 2013-10-12 2014-01-22 上海新跃仪表厂 Monocular stereo vision relative position/pose measuring method
CN103971079A (en) * 2013-01-28 2014-08-06 腾讯科技(深圳)有限公司 Augmented reality implementation method and device of two-dimensional code
US9294873B1 (en) * 2011-09-22 2016-03-22 Amazon Technologies, Inc. Enhanced guidance for electronic devices using objects within in a particular area

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101208723A (en) * 2005-02-23 2008-06-25 克雷格·萨默斯 Automatic scene modeling for the 3D camera and 3D video
CN101610421A (en) * 2008-06-17 2009-12-23 深圳华为通信技术有限公司 Video communication method, Apparatus and system
CN101742349A (en) * 2010-01-05 2010-06-16 浙江大学 Method for expressing three-dimensional scenes and television system thereof
US9294873B1 (en) * 2011-09-22 2016-03-22 Amazon Technologies, Inc. Enhanced guidance for electronic devices using objects within in a particular area
CN103049728A (en) * 2012-12-30 2013-04-17 成都理想境界科技有限公司 Method, system and terminal for augmenting reality based on two-dimension code
CN103971079A (en) * 2013-01-28 2014-08-06 腾讯科技(深圳)有限公司 Augmented reality implementation method and device of two-dimensional code
CN103528571A (en) * 2013-10-12 2014-01-22 上海新跃仪表厂 Monocular stereo vision relative position/pose measuring method

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107886017A (en) * 2017-11-09 2018-04-06 陕西外号信息技术有限公司 A kind of read method and equipment of optical label sequence
CN107886017B (en) * 2017-11-09 2021-02-19 陕西外号信息技术有限公司 Method and device for reading optical label sequence
CN110471580A (en) * 2018-05-09 2019-11-19 北京外号信息技术有限公司 Information equipment exchange method and system based on optical label
CN110471580B (en) * 2018-05-09 2021-06-15 北京外号信息技术有限公司 Information equipment interaction method and system based on optical labels
US11694055B2 (en) 2018-05-09 2023-07-04 Beijing Whyhow Information Technology Co., Ltd Optical tag based information apparatus interaction method and system
CN109710198A (en) * 2018-12-29 2019-05-03 森大(深圳)技术有限公司 Method of printing, device and the equipment of local dynamic station variable image
CN111754449A (en) * 2019-03-27 2020-10-09 北京外号信息技术有限公司 Scene reconstruction method based on optical communication device and corresponding electronic equipment
WO2021057887A1 (en) * 2019-09-26 2021-04-01 北京外号信息技术有限公司 Method and system for setting virtual object capable of being presented to target
TWI785332B (en) * 2020-05-14 2022-12-01 光時代科技有限公司 Three-dimensional reconstruction system based on optical label
WO2022257807A1 (en) * 2021-06-09 2022-12-15 丰疆智能科技股份有限公司 Positioning apparatus based on passive infrared tag

Also Published As

Publication number Publication date
CN106446883B (en) 2019-06-18

Similar Documents

Publication Publication Date Title
CN106446883B (en) Scene reconstruction method based on optical label
CN107945268B (en) A kind of high-precision three-dimensional method for reconstructing and system based on binary area-structure light
CN107370958B (en) Image blurs processing method, device and camera terminal
CN109920007B (en) Three-dimensional imaging device and method based on multispectral photometric stereo and laser scanning
CN108827316A (en) Mobile robot visual orientation method based on improved Apriltag label
CN111080724A (en) Infrared and visible light fusion method
CN109983501A (en) The method and relevant augmented reality method of a kind of hidden image or the object in video
CN104299220B (en) A kind of method that cavity in Kinect depth image carries out real-time filling
CN107154032B (en) A kind of image processing method and device
CN108573222A (en) The pedestrian image occlusion detection method for generating network is fought based on cycle
CN109360235A (en) A kind of interacting depth estimation method based on light field data
CN107392947A (en) 2D 3D rendering method for registering based on coplanar four point set of profile
CN106155299B (en) A kind of pair of smart machine carries out the method and device of gesture control
CN110443898A (en) A kind of AR intelligent terminal target identification system and method based on deep learning
CN108154514A (en) Image processing method, device and equipment
JPH10503627A (en) Improved chroma key system
CN107578437A (en) A kind of depth estimation method based on light-field camera, system and portable terminal
CN110188835A (en) Data based on production confrontation network model enhance pedestrian's recognition methods again
CN108592822A (en) A kind of measuring system and method based on binocular camera and structure light encoding and decoding
CN109889799B (en) Monocular structure light depth perception method and device based on RGBIR camera
CN110428439A (en) A kind of shadow detection method based on shadow region color saturation property
CN111292228B (en) Lens defect detection method
CN105158275B (en) Apple stem/calyx detection method based on near-infrared coding dot matrix structure light
CN107172386B (en) A kind of non-contact data transmission method based on computer vision
CN109360254A (en) Image processing method and device, electronic equipment, computer readable storage medium

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20211203

Address after: 201306 2nd floor, no.979 Yunhan Road, Lingang New Area, China (Shanghai) pilot Free Trade Zone, Pudong New Area, Shanghai

Patentee after: Shanghai Guangshi fusion Intelligent Technology Co.,Ltd.

Address before: 710075 Room 301, Block A, Innovation Information Building, Xi'an Software Park, No. 2 Science and Technology Road, Xi'an High-tech Zone, Shaanxi Province

Patentee before: XI'AN XIAOGUANGZI NETWORK TECHNOLOGY Co.,Ltd.

TR01 Transfer of patent right