CN106446883A - Scene reconstruction method based on light label - Google Patents
Scene reconstruction method based on light label Download PDFInfo
- Publication number
- CN106446883A CN106446883A CN201610789231.4A CN201610789231A CN106446883A CN 106446883 A CN106446883 A CN 106446883A CN 201610789231 A CN201610789231 A CN 201610789231A CN 106446883 A CN106446883 A CN 106446883A
- Authority
- CN
- China
- Prior art keywords
- optical label
- image
- scene
- background image
- label
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
- G06V10/16—Image acquisition using multiple overlapping images; Image stitching
Abstract
The invention relates to a scene reconstruction method based on a light label and can reinforce and highlight the transmission information, eliminate interference of the irrelevant information and guarantee information transmission accuracy and high effectiveness. The method comprises steps that 1, within the unit time, light label scanning is carried out through light label identification equipment, a light label in a scene is discovered, and a light label image is shot; the positioning identification coordinate position of the light label is mapped from a frame image to the light label image and then is taken as a positioning point; 2, a background image of the light label in the scene is acquired, and the background image comprises all the image regions of the scene except the positioning point; 3, image dimension proportion zooming, image information simplification, image edge enhancement processing and image rendering are sequentially carried out for the background image to acquire a reconstructed background image; and 4, the reconstructed background image and the light label image are fused and overlapped to acquire a reconstruction scene, and the reconstruction scene is sent to VR display equipment to accomplish scene reconstruction within the unit time.
Description
Technical field
The present invention relates to the method for scene reconstruction, specially the scene reconstruction method based on optical label.
Background technology
Optical label, can be as VR (virtual reality) collecting device in entity space due to its unique information transmission characteristic
In anchor point, locus is demarcated.Based on these optical label anchor points, soft hardware equipment just can be by entity space
Object projects in VR space according to actual position relation, thus setting up the mapping of the two.Based on the mapping in both spaces, then
More senior application can be implemented, and such as entity interaction accesses, space orientation and roaming, associate and calculate etc..However, only
Anchor point sets up the mapping relations in space not enough, because in whole Virtual Space, in addition to anchor point, the blank between anchor point is also
Needs are filled with, thus being more beneficial for user to the interpretation in VR space, understanding, the image filled also needs to reach prominent, strong
Change the purpose of optical label.Herein, solving one of method that this problem can adopt is exactly that the background image of entity space is synchronous
Collect, be filled into after process in VR scene.Because not all background image information is all related to the application of optical label,
Therefore it is accomplished by carrying out proper treatment to these Backgrounds, emphasize including key character, render simplification with irrelevant information etc., then
It is overlapped showing with optical label image, completes the reconstruct to scene, there is presently no the optical label scene reconstruction method of correlation.
Content of the invention
For problems of the prior art, the present invention provides a kind of scene reconstruction method based on optical label, can
Optical label transmission information is strengthened and is projected, exclude the interference of other irrelevant informations it is ensured that the accuracy of information transmission and
High efficiency.
The present invention is to be achieved through the following technical solutions:
Based on the scene reconstruction method of optical label, comprise the steps,
By optical label, step 1, within the unit interval, identifies that equipment carries out optical label scanning, finds the cursor in scene
Sign, and pickup light label image;The positioning mark coordinate position of optical label is mapped to optical label image from two field picture, as
Anchor point;
Step 2, collection optical label institute background image in the scene, described background image is included in scene except anchor point
All image-regions in addition;
Step 3, carries out picture size proportional zoom successively, image information simplifies, at the enhancing of image border to background image
Reason and image rendering, obtain the background image reconstructing;
Step 4, by the background image of reconstruct and optical label image carry out merge superposition, obtain reconstruct scene be sent to VR show
Show the scene reconstruction completing in equipment in this unit interval.
Preferably, the comprising the following steps that of step 1,
Optical label identifies equipment continuous acquisition multiframe scene image, and shoots the figure that a width includes the optical label of anchor point
Picture;Adjacent any two frames are made the difference and gets differential chart, differential chart finds the positioning mark of optical label, obtains optical label
Coordinate position in optical label image;Thus positioning mark coordinate position luminous in optical label is mapped to cursor from two field picture
Sign in image, obtain the coordinate position of optical label and the anchor point of optical label picture registration;Wherein, during collection between consecutive frame
Between the time interval that flashes twice not less than the dynamic position identifiers of optical label.
Further, sit when the luminous positioning that mapping obtains identifies actual luminous positioning mark in coordinate and optical label image
Mark when not being completely superposed;First, the position relationship according to optical label and luminous positioning mark, obtains optical label in optical label figure
Position range in picture and scene areas;Secondly, in the scene areas comprising optical label, to the positioning mark in multiple image
Symbol scans for identifying, respectively obtains positioning in multiple image and identifies corresponding coordinate, and then obtains the equal of all respective coordinates
Value is as the positioning mark of optical label and the coincidence point of optical label image, i.e. anchor point.
Preferably, the comprising the following steps that of step 2, the time slot using optical label collection non-in optical label identification equipment is carried out
The collection of background image.
Preferably, the comprising the following steps that of step 3,
Step 3.1, picture size proportional zoom;By the display in the background image displaying ratio of entity space and VR space
Ratio is consistent, obtains scaling background image;
Step 3.2, image information simplifies;Scaling background image is carried out gray processing process, obtains gray scale pretreatment background
Image;Using the reflection of RGB color component, heat, intensity of illumination or energy, remove unrelated figure in gray scale pretreatment background image
As information realization simplifies, obtain gray scale background image;
Step 3.3, image border enhancing is processed;To gray scale background image I2Carry out high-pass filtering to obtain strengthening the pretreatment back of the body
Scape image I2', then edge is obtained by following formula and strengthen background image I3,
I3=I2+k·I2’;
Wherein, k is adjustment factor;
Step 3.4, image rendering;Content obtaining is carried out to optical label information by optical label coding/decoding method, according to cursor
Frame information content and the rgb value of storage and render the corresponding of striped and render feature in label information, with the position of optical label as the center of circle,
Strengthen background image for radius to edge with adjustable pixel digit r to render, the background image after being reconstructed.
Further, in step 3.2, irrelevant information is removed simplify using RGB color component in gray scale background image
When, using any one in following three kinds of methods,
Maximum value process, R=G=B=Max (R, G, B);
Mean value method, R=G=B=(R+G+B)/3;
Weighted average method, R=G=B=wr·R+wg·G+wb·B;
Wherein, R, G, B are the RGB color component value of any pixel of gray scale pretreatment background image;wr、wg、wbIt is respectively
The weights of R, G, B, wr、wg、wbIt is in [0,1] interval.
Further, in step 3.4, coding/decoding method indicates coding by the algorithm flag bit in the signal element of optical label, and
Obtained in the decoding algorithm database of standard by coding.
Compared with prior art, the present invention has following beneficial technique effect:
The present invention in the mapping carrying out the entity space based on optical label and VR space, using optical label as positioning " anchor
Point ", the position forming above two space accurately corresponds to, and also needs to the image information of reality scene is especially located simultaneously
Reason, reconstruct scene, form complete VR space the display of prominent optical label;By dividing to background image and cursor glyph
From, and background image is reconstructed with process, solve enhancing optical label in the picture and show, propose and optical label correlation simultaneously
Not high background information is so that user focuses more on the display of optical label.
Brief description
Optical label sample described in Fig. 1 present example.
Scene reconstruction schematic diagram described in Fig. 2 present example.
Image schematic diagram after Information Simplification described in Fig. 3 present example.
Virtual Space Scene after reconstruct described in Fig. 4 present example.
The scene reconstruction method flow diagram based on optical label for Fig. 5 present invention.
In figure:1 identifies equipment for optical label, and 2 is the real scene residing for optical label, and 3 is optical label, and 4 is in scene
Background object, 5 is scene reconstruction processing server, and 6 is the reconstruct image in Virtual Space for the real scene.
Specific embodiment
With reference to specific embodiment, the present invention is described in further detail, described be explanation of the invention and
It is not to limit.
The present invention in the mapping carrying out the entity space based on optical label and VR space, using optical label as positioning " anchor
Point ", the position forming above two space accurately corresponds to, and also needs to the image information of reality scene is especially located simultaneously
Reason, reconstruct scene, form complete VR space the display of prominent optical label.
Described optical label sample is as shown in Figure 1.Optical label includes signal element (cell) group (or referred to as " data
Position ") and positioning mark (or referred to as " flag bit ") two parts, wherein position and be designated upper three larger rectangle frames (three of figure
This rectangle frame individual is referred to as " one group of position identifiers "), under working condition, positioning mark is flashed with certain Frequency Synchronization, by figure
The method that aberration divides can obtain quickly and be detected by picture pick-up device, and then can determine the position of signal element by positioning mark
Put, thus carrying out data identification and reading;Signal element is the black and white rectangle between positioning mark, and multiple signal elements constitute one
Group, usual signal element forms the array of 5 × 5 (being not limited to), and each signal element represents " 0 " or " 1 " of data signal, entirely
(length of side of flag is the two of the data bit length of side to the digital signal sequences of matrix composition one frame of signal element group composition here
Times, easily facilitate positioning), in order to increase the data space that signal element represents, under working condition, each signal element also can be according to
Predetermined program is flashed, thus more signal contents are shown by multiframe.At this moment need to provide one in multiframe
Start frame/end of identification frame, for demarcating the beginning/end position of one complete cycle of multiframe, this frame signal unit group sets
Combine for a special data, such as:Not different with the information of the actual capabilities statement specific combination of full 0 or complete 1 or any.
The scene reconstruction based on optical label for the present invention is as shown in Figure 2.Scene reconstruction method based on optical label is with a list
The position time is a cycle, and circulation execution, as shown in figure 5, its process is:
Step one:Carry out optical label scanning, find the optical label in scene;The method of above-mentioned optical label scanning is as follows:
Optical label identifies equipment 1 continuous acquisition multiframe label image, is designated as:f0, f1..., fm, collection between consecutive frame
The time interval that time flashes twice not less than dynamic position identifiers;And then, shoot a width optical label image, be designated as p;Right
Adjacent any two frames make the difference gets differential chart, find in differential chart optical label 3 positioning mark, obtain further its
Coordinate position in optical label image;The position coordinates of luminescence unit is mapped to optical label image from two field picture;Due to can
Can there is the reasons such as hand shaking, the luminous positioning that mapping obtains identifies actual luminous positioning mark in coordinate and optical label image p
Coordinate is not completely superposed, and at this moment the position relationship according to optical label and luminous positioning mark, can get optical label in optical label
Position range in image and scene areas, in the scene areas comprising optical label, enter to position identifiers in multiple image
Line search identifies, respectively obtains the coordinate positioning mark in multiple image, the average calculating these coordinates is as final cursor
The positioning mark signed and the coincidence point of optical label image p, i.e. anchor point.
Step 2:To all image-regions regard as being background image in addition to anchor point;Collection optical label institute is in the scene
Background image, because optical label identifies that the shooting of equipment or photographing device are not all to be used for light in all of IMAQ time slot
The data acquisition of label, therefore can be utilized the time slot of non-optical label collection to carry out the collection of background image;In unit interval T
Interior, T=[t1,m1,t2,m2,t3,m3,t4,m4,t5,m5,t6...], wherein tiIt is optical label data acquisition time slot, i is positive integer,
mjGather time slot for non-optical label, j is positive integer, then the background image obtaining in above-mentioned non-optical label data acquisition time slot represents
For I, gather the shutter speed that shooting or photographing device need to be adopted as a upper time slot during I;
Step 3:Background image is carried out process reconstruct, the method for above-mentioned Image Reconstruction is as follows:
First dimension scale scaling is carried out to background image so that the background image displaying ratio of entity space and VR space
In displaying ratio consistent, note source background image is I0, then the scaling background image after proportional zoom is I1, carry out image herein
The method of proportional zoom can be arbitrary image scaling method;
Optical label scene information simplifies, and carries out gray processing process to Background herein, obtains gray scale pretreatment background image,
Remove unrelated image information in gray scale pretreatment background image to realize simplifying, obtain gray scale background image I2, may be selected following
Method one carries out this process, and three kinds of methods are as follows:
● maximum value process:R=G=B=Max (R, G, B);
● mean value method:R=G=B=(R+G+B)/3;
● weighted average method:R=G=B=wr·R+wg·G+wbB,
Herein, R, G, B are any pixel RGB color component value of gray scale pretreatment background image, wr、wg、wbRespectively R,
The weights of G, B, wr、wg、wbIt is in [0,1] interval, Max is maximum value finding function, finally gives gray scale background image I2, this
Place can also carry out simplification process according to other different quantizating index to image, such as:Heat, intensity of illumination, energy reflection etc.,
As Fig. 3, and it is not limited to RGB color component;
Optical label scene information is strengthened, herein to gray scale background image I2Carry out edge enhancing process, specifically to ash
Degree background image I2Carry out high-pass filtering, obtain strengthening pretreatment background image I2', calculated according to the following formula:
I3=I2+k·I2’;
In above formula, k is adjustment factor, then obtain strengthening background image I through edge3;
Carry out optical label scene information to render, content obtaining is carried out to optical label information by optical label coding/decoding method, should
Coding/decoding method indicates coding by algorithm flag bit in the signal element of optical label, and by encoding the decoding algorithm data in standard
Obtain in storehouse;If the information obtaining current cursor this frame of label is m, in colors list below, inquire about corresponding color:
The information content | Rgb value | Texturing patterns |
m | (R、G、B) | T |
Herein, T is texturing patterns, can be realized from third-party resource using any;This list is to preset;With light
The position of label is the center of circle, according to frame information content above and rgb value and renders striped for radius with adjustable pixel digit r
The corresponding feature that renders strengthens background image I to edge3Rendered, r>0;Through above-mentioned process, the background after being reconstructed
Image I4, background image I after reconstruct4Eliminate the redundant image information unrelated with optical label and enhance around optical label
Image shows;
Step 4:Background image after reconstruct is overlapped with optical label image, submits to VR display device, finish,
Carry out the above-mentioned process of next time cycle if necessary, finally give the result of similar Fig. 4.
Present invention achieves strengthen optical label in the picture showing, the background letter not high with optical label correlation is proposed simultaneously
Breath is so that user focuses more on the display of optical label.
In actually used, it is capable of following scene and function.User U is by optical label technology roaming access commercial street
Road, so that highlighting in the VR equipment (Google glasses) that optical label uses at it, needs to scene reconstruction.Use first
Family U uses image in face of image capture device acquisition;By there being two optical labels in face of optical label scanning discovery, identify one respectively
Family restaurant and a hotel;After identification, this method is utilized respectively different time-gap and has collected optical label coding information and current field
Scape information;Scene information is reconstructed, the optical label ambient background pointing to hotel is carried out lines enhancing, allows user see pure mellow wine
The profile in shop and scale, and hotel's color rendering is become soft light green;The optical label ambient background pointing to restaurant is carried out
Lines weaken, and allow user see the style within dining room clearly, and color rendering is become bright-coloured yellow;In above-mentioned background reconfiguration method
With the help of, user U is more prone to be concerned about corresponding optical label, and enjoys the service that it is provided, and eliminates other unrelated
The unduly interfering with of information.
Claims (7)
1. the scene reconstruction method based on optical label is it is characterised in that comprise the steps,
By optical label, step 1, within the unit interval, identifies that equipment carries out optical label scanning, finds the optical label in scene, and
Pickup light label image;The positioning mark coordinate position of optical label is mapped to optical label image from two field picture, as positioning
Point;
Step 2, collection optical label background image in the scene, described background image includes in scene in addition to anchor point
All image-regions;
Step 3, background image is carried out successively picture size proportional zoom, image information simplify, image border strengthen process and
Image rendering, obtains the background image reconstructing;
Step 4, the background image of reconstruct and optical label image are carried out merging superposition, obtain reconstructing scene being sent to VR and showing and set
The scene reconstruction in this unit interval is completed in standby.
2. the scene reconstruction method based on optical label according to claim 1 is it is characterised in that the concrete steps of step 1
It is as follows,
Optical label identifies equipment continuous acquisition multiframe scene image, and shoots the image that a width includes the optical label of anchor point;Right
Adjacent any two frames make the difference gets differential chart, finds the positioning mark of optical label, obtain optical label in light in differential chart
Coordinate position in label image;Thus positioning mark coordinate position luminous in optical label is mapped to optical label figure from two field picture
In picture, obtain the coordinate position of optical label and the anchor point of optical label picture registration;Wherein, the acquisition time between consecutive frame is not
The time interval flashing twice less than the dynamic position identifiers of optical label.
3. the scene reconstruction method based on optical label according to claim 2 it is characterised in that when mapping obtain luminous
When in positioning mark coordinate and optical label image, actual luminous positioning mark coordinate is not completely superposed;First, according to cursor
Sign and the luminous position relationship positioning mark, obtain position range in optical label image for the optical label and scene areas;Secondly,
In the scene areas comprising optical label, the position identifiers in multiple image is scanned for identify, respectively obtain multiframe figure
Position in picture and identify corresponding coordinate, and then obtain the average of all respective coordinates and identify and optical label as the positioning of optical label
The coincidence point of image, i.e. anchor point.
4. the scene reconstruction method based on optical label according to claim 1 is it is characterised in that the concrete steps of step 2
As follows, carry out the collection of background image using the time slot of optical label collection non-in optical label identification equipment.
5. the scene reconstruction method based on optical label according to claim 1 is it is characterised in that the concrete steps of step 3
It is as follows,
Step 3.1, picture size proportional zoom;By the displaying ratio in the background image displaying ratio of entity space and VR space
Unanimously, obtain scaling background image;
Step 3.2, image information simplifies;Scaling background image is carried out gray processing process, obtains gray scale pretreatment background image;
Using the reflection of RGB color component, heat, intensity of illumination or energy, remove unrelated image letter in gray scale pretreatment background image
Breath is realized simplifying, and obtains gray scale background image;
Step 3.3, image border enhancing is processed;To gray scale background image I2Carry out high-pass filtering to obtain strengthening pretreatment Background
As I2', then edge is obtained by following formula and strengthen background image I3,
I3=I2+k·I2’;
Wherein, k is adjustment factor;
Step 3.4, image rendering;Content obtaining is carried out to optical label information by optical label coding/decoding method, is believed according to optical label
Frame information content and the rgb value of storage and render the corresponding of striped and render feature in breath, with the position of optical label as the center of circle, with can
Pixel digit r adjusted is that radius renders to edge enhancing background image, the background image after being reconstructed.
6. the scene reconstruction method based on optical label according to claim 5 is it is characterised in that in step 3.2, using ash
When in degree background image, RGB color component is removed to irrelevant information simplifying, using any one in following three kinds of methods,
Maximum value process, R=G=B=Max (R, G, B);
Mean value method, R=G=B=(R+G+B)/3;
Weighted average method, R=G=B=wr·R+wg·G+wb·B;
Wherein, R, G, B are the RGB color component value of any pixel of gray scale pretreatment background image;wr、wg、wbBe respectively R, G,
The weights of B, wr、wg、wbIt is in [0,1] interval.
7. the scene reconstruction method based on optical label according to claim 5 is it is characterised in that in step 3.4, decoding side
Method indicates coding by the algorithm flag bit in the signal element of optical label, and is obtained in the decoding algorithm database of standard by coding
?.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610789231.4A CN106446883B (en) | 2016-08-30 | 2016-08-30 | Scene reconstruction method based on optical label |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610789231.4A CN106446883B (en) | 2016-08-30 | 2016-08-30 | Scene reconstruction method based on optical label |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106446883A true CN106446883A (en) | 2017-02-22 |
CN106446883B CN106446883B (en) | 2019-06-18 |
Family
ID=58163637
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610789231.4A Active CN106446883B (en) | 2016-08-30 | 2016-08-30 | Scene reconstruction method based on optical label |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106446883B (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107886017A (en) * | 2017-11-09 | 2018-04-06 | 陕西外号信息技术有限公司 | A kind of read method and equipment of optical label sequence |
CN109710198A (en) * | 2018-12-29 | 2019-05-03 | 森大(深圳)技术有限公司 | Method of printing, device and the equipment of local dynamic station variable image |
CN110471580A (en) * | 2018-05-09 | 2019-11-19 | 北京外号信息技术有限公司 | Information equipment exchange method and system based on optical label |
CN111754449A (en) * | 2019-03-27 | 2020-10-09 | 北京外号信息技术有限公司 | Scene reconstruction method based on optical communication device and corresponding electronic equipment |
WO2021057887A1 (en) * | 2019-09-26 | 2021-04-01 | 北京外号信息技术有限公司 | Method and system for setting virtual object capable of being presented to target |
TWI785332B (en) * | 2020-05-14 | 2022-12-01 | 光時代科技有限公司 | Three-dimensional reconstruction system based on optical label |
WO2022257807A1 (en) * | 2021-06-09 | 2022-12-15 | 丰疆智能科技股份有限公司 | Positioning apparatus based on passive infrared tag |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101208723A (en) * | 2005-02-23 | 2008-06-25 | 克雷格·萨默斯 | Automatic scene modeling for the 3D camera and 3D video |
CN101610421A (en) * | 2008-06-17 | 2009-12-23 | 深圳华为通信技术有限公司 | Video communication method, Apparatus and system |
CN101742349A (en) * | 2010-01-05 | 2010-06-16 | 浙江大学 | Method for expressing three-dimensional scenes and television system thereof |
CN103049728A (en) * | 2012-12-30 | 2013-04-17 | 成都理想境界科技有限公司 | Method, system and terminal for augmenting reality based on two-dimension code |
CN103528571A (en) * | 2013-10-12 | 2014-01-22 | 上海新跃仪表厂 | Monocular stereo vision relative position/pose measuring method |
CN103971079A (en) * | 2013-01-28 | 2014-08-06 | 腾讯科技(深圳)有限公司 | Augmented reality implementation method and device of two-dimensional code |
US9294873B1 (en) * | 2011-09-22 | 2016-03-22 | Amazon Technologies, Inc. | Enhanced guidance for electronic devices using objects within in a particular area |
-
2016
- 2016-08-30 CN CN201610789231.4A patent/CN106446883B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101208723A (en) * | 2005-02-23 | 2008-06-25 | 克雷格·萨默斯 | Automatic scene modeling for the 3D camera and 3D video |
CN101610421A (en) * | 2008-06-17 | 2009-12-23 | 深圳华为通信技术有限公司 | Video communication method, Apparatus and system |
CN101742349A (en) * | 2010-01-05 | 2010-06-16 | 浙江大学 | Method for expressing three-dimensional scenes and television system thereof |
US9294873B1 (en) * | 2011-09-22 | 2016-03-22 | Amazon Technologies, Inc. | Enhanced guidance for electronic devices using objects within in a particular area |
CN103049728A (en) * | 2012-12-30 | 2013-04-17 | 成都理想境界科技有限公司 | Method, system and terminal for augmenting reality based on two-dimension code |
CN103971079A (en) * | 2013-01-28 | 2014-08-06 | 腾讯科技(深圳)有限公司 | Augmented reality implementation method and device of two-dimensional code |
CN103528571A (en) * | 2013-10-12 | 2014-01-22 | 上海新跃仪表厂 | Monocular stereo vision relative position/pose measuring method |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107886017A (en) * | 2017-11-09 | 2018-04-06 | 陕西外号信息技术有限公司 | A kind of read method and equipment of optical label sequence |
CN107886017B (en) * | 2017-11-09 | 2021-02-19 | 陕西外号信息技术有限公司 | Method and device for reading optical label sequence |
CN110471580A (en) * | 2018-05-09 | 2019-11-19 | 北京外号信息技术有限公司 | Information equipment exchange method and system based on optical label |
CN110471580B (en) * | 2018-05-09 | 2021-06-15 | 北京外号信息技术有限公司 | Information equipment interaction method and system based on optical labels |
US11694055B2 (en) | 2018-05-09 | 2023-07-04 | Beijing Whyhow Information Technology Co., Ltd | Optical tag based information apparatus interaction method and system |
CN109710198A (en) * | 2018-12-29 | 2019-05-03 | 森大(深圳)技术有限公司 | Method of printing, device and the equipment of local dynamic station variable image |
CN111754449A (en) * | 2019-03-27 | 2020-10-09 | 北京外号信息技术有限公司 | Scene reconstruction method based on optical communication device and corresponding electronic equipment |
WO2021057887A1 (en) * | 2019-09-26 | 2021-04-01 | 北京外号信息技术有限公司 | Method and system for setting virtual object capable of being presented to target |
TWI785332B (en) * | 2020-05-14 | 2022-12-01 | 光時代科技有限公司 | Three-dimensional reconstruction system based on optical label |
WO2022257807A1 (en) * | 2021-06-09 | 2022-12-15 | 丰疆智能科技股份有限公司 | Positioning apparatus based on passive infrared tag |
Also Published As
Publication number | Publication date |
---|---|
CN106446883B (en) | 2019-06-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106446883B (en) | Scene reconstruction method based on optical label | |
CN107945268B (en) | A kind of high-precision three-dimensional method for reconstructing and system based on binary area-structure light | |
CN107370958B (en) | Image blurs processing method, device and camera terminal | |
CN109920007B (en) | Three-dimensional imaging device and method based on multispectral photometric stereo and laser scanning | |
CN108827316A (en) | Mobile robot visual orientation method based on improved Apriltag label | |
CN111080724A (en) | Infrared and visible light fusion method | |
CN109983501A (en) | The method and relevant augmented reality method of a kind of hidden image or the object in video | |
CN104299220B (en) | A kind of method that cavity in Kinect depth image carries out real-time filling | |
CN107154032B (en) | A kind of image processing method and device | |
CN108573222A (en) | The pedestrian image occlusion detection method for generating network is fought based on cycle | |
CN109360235A (en) | A kind of interacting depth estimation method based on light field data | |
CN107392947A (en) | 2D 3D rendering method for registering based on coplanar four point set of profile | |
CN106155299B (en) | A kind of pair of smart machine carries out the method and device of gesture control | |
CN110443898A (en) | A kind of AR intelligent terminal target identification system and method based on deep learning | |
CN108154514A (en) | Image processing method, device and equipment | |
JPH10503627A (en) | Improved chroma key system | |
CN107578437A (en) | A kind of depth estimation method based on light-field camera, system and portable terminal | |
CN110188835A (en) | Data based on production confrontation network model enhance pedestrian's recognition methods again | |
CN108592822A (en) | A kind of measuring system and method based on binocular camera and structure light encoding and decoding | |
CN109889799B (en) | Monocular structure light depth perception method and device based on RGBIR camera | |
CN110428439A (en) | A kind of shadow detection method based on shadow region color saturation property | |
CN111292228B (en) | Lens defect detection method | |
CN105158275B (en) | Apple stem/calyx detection method based on near-infrared coding dot matrix structure light | |
CN107172386B (en) | A kind of non-contact data transmission method based on computer vision | |
CN109360254A (en) | Image processing method and device, electronic equipment, computer readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right |
Effective date of registration: 20211203 Address after: 201306 2nd floor, no.979 Yunhan Road, Lingang New Area, China (Shanghai) pilot Free Trade Zone, Pudong New Area, Shanghai Patentee after: Shanghai Guangshi fusion Intelligent Technology Co.,Ltd. Address before: 710075 Room 301, Block A, Innovation Information Building, Xi'an Software Park, No. 2 Science and Technology Road, Xi'an High-tech Zone, Shaanxi Province Patentee before: XI'AN XIAOGUANGZI NETWORK TECHNOLOGY Co.,Ltd. |
|
TR01 | Transfer of patent right |