CN106097454A - A kind of man-machine interactive system and exchange method - Google Patents
A kind of man-machine interactive system and exchange method Download PDFInfo
- Publication number
- CN106097454A CN106097454A CN201610394486.0A CN201610394486A CN106097454A CN 106097454 A CN106097454 A CN 106097454A CN 201610394486 A CN201610394486 A CN 201610394486A CN 106097454 A CN106097454 A CN 106097454A
- Authority
- CN
- China
- Prior art keywords
- physical object
- dummy object
- dummy
- model
- man
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
Landscapes
- Engineering & Computer Science (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Processing Or Creating Images (AREA)
Abstract
The present invention relates to virtual reality and field of human-computer interaction, particularly to a kind of man-machine interactive system and exchange method.The man-machine interactive system that the present invention provides includes receiving physical object and gathering the transmission terminal of this physical object characteristic information, the processing unit producing corresponding dummy object according to characteristic information and the display device of display dummy object;Can provide the user and put into physical object, demonstrate corresponding dummy object in a display device;Select dummy object in a display device, the true interactive experience of plan of corresponding physical object of taking from transmission terminal, can be widely applied to the field such as children education, electronic entertainment.Man-machine interactive system that the application present invention provides and man-machine interaction method can intend that very " UNICOM " virtual and real world, bring the sensation that user find everything fresh and new.
Description
Technical field
The present invention relates to virtual reality and field of human-computer interaction, particularly to a kind of man-machine interactive system and exchange method.
Background technology
Human-computer interaction technology always is that the prior development direction of computer realm, mouse, keyboard and display are current
Use history is the longest, the widest human-computer interaction device of range of application, but, along with the development of present multimedia technology, mouse, keyboard
And the matched combined of display can not meet the demand of user, now, touch screen technology, body-sensing technology, data glove the most completely
Technology is arisen at the historic moment, but, neozoic human-computer interaction technology still can not in settlement computer virtual world complete with true environment
Entirely isolating, the object in virtual world " cannot be walked " and go out screen, and the object in true environment can not " enter " to enter virtual world
Realistic problem, thus more real man-machine interaction experience cannot be provided the user.
Summary of the invention
Present invention aim to overcome that existing human-computer interaction technology, it is impossible to make the object in virtual world " walk " to go out screen
Curtain, and the object in true environment can not " enter " to enter virtual world, thus the most real man-machine friendship can not be provided the user
The problem experienced mutually;There is provided a kind of and object in true environment " is sent into " virtual world, the object of virtual world " is taken out "
The true man-machine interactive system of plan to real world.
It is an object of the invention to overcome the above-mentioned deficiency in the presence of prior art, it is provided that a kind of man-machine interactive system.
In order to realize foregoing invention purpose, the invention provides techniques below scheme:
A kind of man-machine interactive system, including,
Transmission terminal, for receiving the physical object that user puts into;
Feature extraction unit, carries out characteristic information collection, and the described characteristic information transmission extremely place that will obtain to this physical object
Reason unit;Described characteristic information includes the size of this physical object, shape, color, material;
Processing unit, for judge according to the characteristic information of this physical object received the implication of this physical object, title and/or
Form the dummy object corresponding with this physical object;
Display device, for showing described dummy object and/or the implication of dummy object, title.
Further, described man-machine interactive system also includes message input device, and described message input device is used for user
Input information;
Described processing unit is additionally operable to the information according to user's input and generates the dummy object of correspondence and/or containing of dummy object
Justice, title, and control display unit and show this dummy object and/or the implication of dummy object, title;
Described transmission terminal is additionally operable to the selection instruction according to user by this dummy object and/or the implication of dummy object, title
Corresponding physical object is sent;
Described physical object is for prestoring in systems or being printed by 3D printing equipment.
Concrete, described feature extraction unit includes graphics processing unit and at least one camera head, and described shooting fills
Put the image for shooting described physical object;Described graphics processing unit is for the spy according to physical object described in image zooming-out
Reference ceases.
Preferably, described processing unit includes model generation unit, model storage unit, Model Matching unit and animation wash with watercolours
Dye unit;Described model generation unit for producing the dummy model of physical object according to characteristic information;The storage of described model is single
Unit is for storing the model of model generation unit generation and being pre-stored in intrasystem model;Described Model Matching unit is used for basis
Characteristic information carries out Model Matching to dummy object;Described animation rendering unit renders for model is carried out animation.
In some embodiments, the described dummy object corresponding with physical object is two dimension and/or the three-dimensional of this physical object
Figure.
Present invention simultaneously provides a kind of man-machine interaction method, comprise the steps:
(1-1) obtain physical object, and this physical object is carried out the step of characteristic information collection;Described characteristic information includes this
The size of physical object, shape, color, material;
(1-2) dummy object corresponding with this physical object and/or dummy object are formed according to the described characteristic information collected
Implication, the step of title;
(1-3) this dummy object and/or the implication of dummy object, the step of title are shown.
Further, also comprise the steps:
(2-1) user's input information selects the step of dummy object;
(2-2) dummy object and/or the implication of dummy object, the step of title are generated according to the input information of user;
(2-3) dummy object and/or the implication of dummy object, the step of title of user's selection are shown;
(2-4) step that physical object corresponding for dummy object is taken out from system;Described physical object is for prestoring or by 3D
Printing equipment prints.
Preferably, in the step of described collection physical object characteristic information, including,
At least one camera head is used to obtain the step of image of this physical object;
Image according to the physical object obtained carries out image procossing and obtains the step of physical object characteristic information.
Preferably, generate dummy object and/or the implication of dummy object, title, the dummy object that display user selects
Before step, also include,
According to characteristic information newly-generated dummy object model or carry out dummy object Model Matching from the model prestored with life
Become the step of dummy object;
Show in the step of this dummy object and/or the implication of dummy object, title, including to the virtual object generated or match
Body Model carries out the step that animation renders.
Preferably, the described dummy object corresponding with physical object is two dimension or the 3-D graphic of this physical object.
Compared with prior art, beneficial effects of the present invention: the man-machine interactive system that the present invention provides includes receiving material object
Object also gathers the transmission terminal of these material object data, according to gathering data the extraction feature extraction unit of characteristic information, root in kind
Processing unit and the display device of display dummy object of corresponding dummy object is produced according to characteristic information;Can provide the user
Put into physical object, demonstrate corresponding dummy object in a display device;Select dummy object in a display device, from transmission eventually
Hold the true interactive experience of plan of corresponding physical object of taking, can be widely applied to the field such as children education, electronic entertainment.Apply this
The man-machine interactive system of bright offer and man-machine interaction method can intend that very " UNICOM " virtual and real world, bring user and find everything fresh and new
Sensation.
Accompanying drawing illustrates:
The structured flowchart of the man-machine interactive system that Fig. 1 provides for the present invention.
Physical object in the man-machine interaction method that Fig. 2 a provides for the present invention turns dummy object flow chart.
In the man-machine interaction method that Fig. 2 b provides for the present invention, dummy object turns physical object flow chart.
Detailed description of the invention
Below in conjunction with the accompanying drawings and specific embodiment the present invention is described in further detail.But this should be interpreted as this
The scope inventing above-mentioned theme is only limitted to below example, and all technology realized based on present invention belong to the present invention
Scope.
Embodiment 1: as it is shown in figure 1, the present embodiment provides a kind of man-machine interactive system, including,
Transmission terminal 1, for receiving the physical object (such as football, pencil, toy) that user puts into;
Feature extraction unit 2, carries out characteristic information collection, and the described characteristic information transmission extremely place that will obtain to this physical object
Reason unit 3;Described characteristic information includes the size of this physical object, shape, color, material;In actual application, characteristic information can
By in the form of images or transmit in the way of electromagnetic signal to processing unit 3;
Processing unit 3, for judge according to the characteristic information of this physical object received the implication of this physical object, title and/
Or form the dummy object corresponding with this physical object;Concrete, described processing unit 3 includes that model generation unit, model are deposited
Storage unit, Model Matching unit and animation rendering unit;Described model generation unit is for producing thing in kind according to characteristic information
The dummy model of body;Described model storage unit is for storing the model of model generation unit generation and the mould in systems that prestores
Type;Described Model Matching unit is for carrying out Model Matching according to characteristic information to dummy object;Described animation rendering unit is used
Render in model is carried out animation, so that this dummy object more like the physical object of its correspondence and has kinesthesia.Model is raw
Unit is become to use image recognition technology or electromagnetic signal identification technology that the characteristic information of physical object is identified, and by it
It is reduced to two dimensional image or the 3-D graphic of this physical object, i.e. forms the dummy object that this physical object is corresponding.Some are real
Executing in example, model generation unit needs to produce in real time the dummy object that this physical object is corresponding;And in other embodiment, mould
Type memory element prestores the dummy model of dummy object corresponding to common physical object;Meanwhile, described processing unit 3 includes
Model Matching unit, after characteristic information is identified by processing unit 3, described Model Matching unit is from the dummy object prestored
Model library in mate, then directly invoke this model show as the match is successful.
It should be noted that the model prestored in model library includes the characteristic information of object, such as the size of object, shape, face
Color, material, during coupling, preferentially mate according to the title of dummy object, now, and the reality that there may be in model library and obtain
The dummy object model that thing object names is identical, but individual other characteristic information differs (as color is different), and processing unit 3 is optional
After taking this dummy object model, its characteristic information is modified with the on all four virtual object of physical object realized with obtain
Body, newly-generated dummy object is stored in model library equally.
It should be noted that feature extraction unit 2 can be integrated at transmission terminal 1, it is also possible to be integrated into processing unit 3
Integrally, or, transmission terminal 1 becomes one with feature extraction unit 2, processing unit 3.
Display device 4, for showing described dummy object and/or the implication of this dummy object, title, it is thus well known that,
The implication title of this dummy object should be consistent with the implication of corresponding physical object, title.
Time specifically used, e.g., when user puts into a football, transmission terminal 1 gathers the characteristic information of this football, and will
The incoming processing unit of characteristic information 3 of football, processing unit 3 according to characteristic information identify its for after football, some embodiments
In, processing unit 3 directly generate the two dimension of this football or 3-D graphic and rendered after display, and other embodiment
In, the dummy object model library first prestored from system finds the model with or without football, if any, then directly invoke this model,
Thus form fantasy football, then show after it is rendered by animation rendering unit;It should be noted that animation rendering unit is basis
It is specifically rendered by the object color information in characteristic information;Animation renders, such as it may be that work as transmission terminal 1 by real
Thing football is hidden after receiving, and display device 4 coordinates the hide actions of transmission terminal 1, shows this fantasy football, and shows that this is virtual
Football " rolls into " animation effect in screen.
Further, described man-machine interactive system also includes that message input device 5, described message input device 5 are used for
Family input information, this input information may is that as concrete, and user can use input through keyboard object names or use mouse to select
Dummy object, as frame selects, clicks;In some embodiments, described input equipment can also be touch screen, thus user can be direct
From touch-screen input dummy object title, input dummy object characteristic information or directly to click, frame select virtual object image
Formal character dummy object;In other embodiment, problem (e.g., calculation question, question-and-answer problem can be generated in advance by display device 4
Deng), now, this input information it may be that as user input problem answers, under certain situation, message input device 5 is additionally operable to
User's input non-selection dummy object or the information answered a question, as opened, out code, page-turning instruction etc..
Described processing unit 3 controls display unit according to the selection of user and shows this dummy object;Concrete, user is from defeated
Entering after device selects dummy object in the way of input title, processing unit 3 is single from model storage according to this title or image
Corresponding dummy object model is found by unit, and this process is equally referred to as Model Matching;After the match is successful, animation render list
Model is carried out showing after animation renders by unit.
Described transmission terminal 1 is additionally operable to physical object corresponding for this dummy object is sent the selection according to user;Tool
The described physical object of body is for prestoring in systems or being printed by 3D printing equipment.
Concrete, equally as a example by football, user inputs football at message input device 5 or selects football, described
Processing unit 3 finds, from model storage unit, the dummy object model that football is corresponding, enters this model via animation rendering unit
Action picture shows through display device 4 after rendering.Animation renders, such as it may be that fantasy football be shown as in display device 4
The animation " rolled out " outside screen, after arriving certain position, fantasy football disappears, and then transmission terminal 1 coordinates animation to render
Effect, when fantasy football disappears, sends material object football, thus causing this material object football is the effect " rolled out " from screen
Really.
In some embodiments, described feature extraction unit 2 includes graphics processing unit 3 and at least one camera head, institute
State camera head for shooting the image of described physical object;Described graphics processing unit 3 is for according to real described in image zooming-out
The characteristic information of thing object;The quantity of camera head is arranged as required to, for the regular shape physical object that some are common, as
Football, basketball, ball, bowl, square block etc., can obtain its complete image information by a camera head.And some
There is the physical object of relatively complicated shape, as teacup, teapot etc. may need the camera head of different angles complete to obtain it
The image information of all angles;And in other detailed description of the invention, the material object that user puts into is probably the answer of standard configuration
Plate and the numeral write on Question answering board, picture, system the mode of putting into of the Question answering board to standard configuration carry out prespecified after, can
Can obtain the content on Question answering board only with a camera head, in this case, graphics processing unit 3 is believed according to feature
After breath judges that it is as the Question answering board of standard configuration, can only extract the contents such as the numeral on Question answering board, image, we are by the Question answering board of standard configuration
On numeral, image, word content be defined as the one of " implication of physical object ".
In actual application, this system can be childhood education system, and toy (such as ball, paintbrush) is put into transmission eventually by child
End 1, transmission terminal 1 is sent to specify position, feature extraction unit 2 characteristic information extraction after taking pictures it, by processing
Unit 3 is judged as ball (title of dummy object) according to this feature information, and processing unit 3 is first at the model that itself prestores
Storehouse (or data base) is found the model either with or without ball, if any, then select this model, and according to this ball in characteristic information
It is rendered and forms the virtual ball corresponding with practice ball by size, color, material, renders the step also including that animation renders,
To show that this ball " rolls to " animation in picture from picture;Increase the sense of reality of interactive experience.It should be noted that native system
Dummy object corresponding to quick physical object can be realized produce, thus realize " taking away " in 1 physical object of transmission terminal
While disappearance (or slightly time delay), i.e. show, in display device 4, the dummy object appearance that this physical object is corresponding, in virtual object
(as time delay 1-5s occurs) when body time delay occurs, display transition cartoon can be used during time delay (as flash of light animation, time shuttle back and forth
The effect animations such as animation) form carry out transition.
In the actual application of other, still as a example by childhood education system, system display problem, defeated from information by child
Entering device 5 to answer a question, after answering correctly, display module provides gift and selects picture, and child selects gift, system to show this void
Intend article from picture, " deliver to " animation outside picture, (or slightly time delay), institute while dummy object disappears in picture
State transmission terminal 1 physical object corresponding for this dummy object to be sent.
Embodiment 2: the present embodiment as shown in Fig. 2 a, Fig. 2 b provides a kind of man-machine interaction method, comprises the steps:
S110: obtain physical object, and this physical object is carried out characteristic information collection;Described characteristic information includes this material object thing
The size of body, shape, color, material;
Preferably, in the step of described collection physical object characteristic information, including,
S111: use at least one camera head to obtain the image of this physical object;The quantity of camera head is arranged as required to,
For the regular shape physical object that some are common, such as football, basketball, ball, bowl, square block etc., filled by a shooting
Put and can obtain its complete image information.And some have the physical object of relatively complicated shape, as teacup, teapot etc. may need
The camera head wanting different angles obtains the image information of its complete all angles.
S112: carry out image procossing according to the image of the physical object obtained and obtain physical object characteristic information.
S120: form the dummy object corresponding with this physical object and/or virtual according to the described characteristic information that collects
The implication of object, title;
Preferably, it includes,
S121: regenerate dummy object model and/or the implication of dummy object, title according to characteristic information, or from prestoring
Model in carry out dummy object Model Matching;The match is successful utilizes the dummy object model that the match is successful to form this virtual object afterwards
Body;
S122: the model matched is carried out animation and renders.
S130: show this dummy object and/or the implication of dummy object, title.
Further, also comprise the steps:
The step of S210: user's input information;User can use input through keyboard object names or characteristic information or, use mouse
Select dummy object, as frame selects, clicks;In some embodiments, described input equipment can also be touch screen, thus user is permissible
Directly from touch-screen input dummy object title, the characteristic information of input dummy object or directly to click, frame select dummy object figure
The formal character dummy object of picture;In other embodiment, can by display device 4 be generated in advance problem (e.g., calculation question, ask
Answer etc.), now, this input information is it may be that such as the problem answers of user's input, under certain situation, message input device 5 is also
For user's input non-selection dummy object or the information answered a question, as opened, out code, page-turning instruction etc..Should
It is noted that under individual cases, user also can directly input the implication of physical objects, title (corresponding virtual of this physical object
Object has consistent implication, title with it, therefore it is also contemplated that user is to have input the implication of dummy object, title).
S220: generate dummy object and/or the implication of dummy object, title according to the input information of user;Concrete, bag
Include:
S221: directly regenerate dummy object model and/or the implication of dummy object, title according to the input information of user,
Or characteristic information or implication, title according to dummy object carry out dummy object Model Matching from the model prestored, coupling
The dummy object model that the match is successful is utilized to form this dummy object after success;
S222: the dummy object model formed is carried out animation and renders.
S230: show this dummy object, it practice, after animation renders, the present embodiment can be shown that this dummy object
The cartoon picture " walked out " outside picture, to improve the sense of reality of man-machine interaction;This cartoon picture can be three-dimensional animation picture
Or 2 D animation picture.
S240: physical object corresponding for dummy object is taken out from system;Described physical object is for prestoring in systems
Or printed by 3D printing equipment.
Preferably, the described dummy object corresponding with physical object is two dimension or the 3-D graphic of this physical object.
Claims (10)
1. a man-machine interactive system, it is characterised in that include,
Transmission terminal, for receiving the physical object that user puts into;
Feature extraction unit, carries out characteristic information collection, and the described characteristic information transmission extremely place that will obtain to this physical object
Reason unit;Described characteristic information includes the size of this physical object, shape, color, material;
Processing unit, for judge according to the characteristic information of this physical object received the implication of this physical object, title and/or
Form the dummy object corresponding with this physical object;
Display device, for showing described dummy object and/or the implication of dummy object, title.
2. man-machine interactive system as claimed in claim 1, it is characterised in that described man-machine interactive system also includes that information inputs
Device, described message input device is used for user's input information;
Described processing unit is additionally operable to the information according to user's input and generates the dummy object of correspondence and/or containing of dummy object
Justice, title, and control display unit and show this dummy object and/or the implication of dummy object, title;
Described transmission terminal is additionally operable to the instruction according to user by this dummy object and/or the implication of dummy object, title correspondence
Physical object send;
Described physical object is for prestoring in systems or being printed by 3D printing equipment.
3. man-machine interactive system as claimed in claim 1, it is characterised in that described feature extraction unit includes image procossing list
Unit and at least one camera head, described camera head is for shooting the image of described physical object;Described graphics processing unit
For the characteristic information according to physical object described in image zooming-out.
4. man-machine interactive system as claimed in claim 1, it is characterised in that described processing unit include model generation unit,
Model storage unit, Model Matching unit and animation rendering unit;Described model generation unit is for producing according to characteristic information
The dummy model of physical object;Described model storage unit is for storing the model of model generation unit generation and being pre-stored in system
In model;Described Model Matching unit is for carrying out Model Matching according to characteristic information to dummy object;Described animation renders
Unit renders for model is carried out animation.
5. man-machine interactive system as claimed in claim 1, it is characterised in that the described dummy object corresponding with physical object is
The two dimension of this physical object and/or 3-D graphic.
6. a man-machine interaction method, it is characterised in that comprise the steps:
(1-1) obtain physical object, and this physical object is carried out the step of characteristic information collection;Described characteristic information includes this
The size of physical object, shape, color, material;
(1-2) dummy object corresponding with this physical object and/or dummy object are formed according to the described characteristic information collected
Implication, the step of title;
(1-3) this dummy object and/or the implication of dummy object, the step of title are shown.
7. man-machine interaction method as claimed in claim 6, it is characterised in that also comprise the steps:
(2-1) step of user's input information;
(2-2) dummy object and/or the implication of dummy object, the step of title are generated according to the input information of user;
(2-3) this dummy object and/or the implication of dummy object, the step of title are shown;
(2-4) step that physical object corresponding for dummy object is taken out from system;Described physical object is for prestoring or by 3D
Printing equipment prints.
8. man-machine interaction method as claimed in claim 6, it is characterised in that the step of described collection physical object characteristic information
In, including,
At least one camera head is used to obtain the step of image of this physical object;
Image according to the physical object obtained carries out image procossing and obtains the step of physical object characteristic information.
Man-machine interaction method the most as claimed in claims 6 or 7, it is characterised in that generate dummy object and/or dummy object
Implication, title include,
Newly-generated dummy object model or carry out dummy object Model Matching from the model prestored to generate dummy object;
Show in the step of this dummy object and/or the implication of dummy object, title, including to the virtual object generated or match
Body Model carries out the step that animation renders.
10. man-machine interaction method as claimed in claim 6, it is characterised in that the described dummy object corresponding with physical object
Two dimension or 3-D graphic for this physical object.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610394486.0A CN106097454A (en) | 2016-06-06 | 2016-06-06 | A kind of man-machine interactive system and exchange method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610394486.0A CN106097454A (en) | 2016-06-06 | 2016-06-06 | A kind of man-machine interactive system and exchange method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106097454A true CN106097454A (en) | 2016-11-09 |
Family
ID=57448064
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610394486.0A Pending CN106097454A (en) | 2016-06-06 | 2016-06-06 | A kind of man-machine interactive system and exchange method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106097454A (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106843473A (en) * | 2016-12-30 | 2017-06-13 | 武汉市马里欧网络有限公司 | Children based on AR tint system and method |
CN107291222A (en) * | 2017-05-16 | 2017-10-24 | 阿里巴巴集团控股有限公司 | Interaction processing method, device, system and the virtual reality device of virtual reality device |
CN112530219A (en) * | 2020-12-14 | 2021-03-19 | 北京高途云集教育科技有限公司 | Teaching information display method and device, computer equipment and storage medium |
CN115328316A (en) * | 2022-08-24 | 2022-11-11 | 中国科学院半导体研究所 | Meta-universe object material construction method and device based on VR technology |
WO2023182932A3 (en) * | 2022-03-25 | 2023-11-30 | 脸萌有限公司 | Target object identification method and apparatus, electronic device and storage medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101034307A (en) * | 2006-03-06 | 2007-09-12 | 乐金电子(中国)研究开发中心有限公司 | Remote tactile sensor and sensing method thereof |
CN101794459A (en) * | 2010-02-09 | 2010-08-04 | 北京邮电大学 | Seamless integration method of stereoscopic vision image and three-dimensional virtual object |
CN102323985A (en) * | 2011-09-08 | 2012-01-18 | 盛乐信息技术(上海)有限公司 | Real and virtuality conversion system and method |
US20130141419A1 (en) * | 2011-12-01 | 2013-06-06 | Brian Mount | Augmented reality with realistic occlusion |
US20140104274A1 (en) * | 2012-10-17 | 2014-04-17 | Microsoft Corporation | Grasping virtual objects in augmented reality |
CN105096382A (en) * | 2015-07-09 | 2015-11-25 | 浙江宇视科技有限公司 | Method and apparatus for associating actual object information in video monitoring image |
-
2016
- 2016-06-06 CN CN201610394486.0A patent/CN106097454A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101034307A (en) * | 2006-03-06 | 2007-09-12 | 乐金电子(中国)研究开发中心有限公司 | Remote tactile sensor and sensing method thereof |
CN101794459A (en) * | 2010-02-09 | 2010-08-04 | 北京邮电大学 | Seamless integration method of stereoscopic vision image and three-dimensional virtual object |
CN102323985A (en) * | 2011-09-08 | 2012-01-18 | 盛乐信息技术(上海)有限公司 | Real and virtuality conversion system and method |
US20130141419A1 (en) * | 2011-12-01 | 2013-06-06 | Brian Mount | Augmented reality with realistic occlusion |
US20140104274A1 (en) * | 2012-10-17 | 2014-04-17 | Microsoft Corporation | Grasping virtual objects in augmented reality |
CN105096382A (en) * | 2015-07-09 | 2015-11-25 | 浙江宇视科技有限公司 | Method and apparatus for associating actual object information in video monitoring image |
Non-Patent Citations (1)
Title |
---|
吴访升等: "家具3D虚拟展示与定制分布式原型系统", 《计算机应用》 * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106843473A (en) * | 2016-12-30 | 2017-06-13 | 武汉市马里欧网络有限公司 | Children based on AR tint system and method |
CN106843473B (en) * | 2016-12-30 | 2020-04-10 | 武汉市马里欧网络有限公司 | AR-based children painting system and method |
CN107291222A (en) * | 2017-05-16 | 2017-10-24 | 阿里巴巴集团控股有限公司 | Interaction processing method, device, system and the virtual reality device of virtual reality device |
CN107291222B (en) * | 2017-05-16 | 2020-08-04 | 阿里巴巴集团控股有限公司 | Interactive processing method, device and system of virtual reality equipment and virtual reality equipment |
CN112530219A (en) * | 2020-12-14 | 2021-03-19 | 北京高途云集教育科技有限公司 | Teaching information display method and device, computer equipment and storage medium |
WO2023182932A3 (en) * | 2022-03-25 | 2023-11-30 | 脸萌有限公司 | Target object identification method and apparatus, electronic device and storage medium |
CN115328316A (en) * | 2022-08-24 | 2022-11-11 | 中国科学院半导体研究所 | Meta-universe object material construction method and device based on VR technology |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106097454A (en) | A kind of man-machine interactive system and exchange method | |
CN107852573B (en) | Mixed reality social interactions | |
CN102270275B (en) | The method of selecting object and multimedia terminal in virtual environment | |
CN108305325A (en) | The display methods and device of virtual objects | |
KR101692335B1 (en) | System for augmented reality image display and method for augmented reality image display | |
CN106097417B (en) | Subject generating method, device, equipment | |
CN103366610A (en) | Augmented-reality-based three-dimensional interactive learning system and method | |
CN102496177B (en) | Method for producing three-dimensional water-and-ink animation | |
CN103530495A (en) | Augmented reality simulation continuum | |
CN103916621A (en) | Method and device for video communication | |
CN106200960A (en) | The content display method of electronic interactive product and device | |
CN106601043A (en) | Multimedia interaction education device and multimedia interaction education method based on augmented reality | |
CN103207677B (en) | System and method for realizing virtual-real somatosensory interaction of digital Zenghouyi bells | |
CN109254650A (en) | A kind of man-machine interaction method and device | |
US20130080976A1 (en) | Motion controlled list scrolling | |
CA3045132C (en) | Communication with augmented reality virtual agents | |
CN113382790A (en) | Toy system for augmented reality | |
CN111840999A (en) | Game education method based on three-dimensional object graph recognition | |
CN109395387A (en) | Display methods, device, storage medium and the electronic device of threedimensional model | |
TWI702504B (en) | System for splicing and converting images of chinese character into vocabularies and mobile terminal | |
CN205287569U (en) | Children's picture arragement based on augmented reality | |
CA3145029A1 (en) | User specific advertising in a virtual environment | |
CN101650822A (en) | Method and system for generating mutual information | |
WO2023066005A1 (en) | Method and apparatus for constructing virtual scenario, and electronic device, medium and product | |
CN116958487A (en) | Gesture prediction method, device, apparatus, storage medium, and computer program product |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C41 | Transfer of patent application or patent right or utility model | ||
TA01 | Transfer of patent application right |
Effective date of registration: 20170206 Address after: 610041 Chengdu province high tech Zone (West), West Park, the new route No. 4, No. Applicant after: Sichuan Lingling robot limited liability company Address before: The town street community is booming town of Chengdu Tianfu District of Sichuan province 610200 City 57 No. 2 unit 1 Building No. 1 Applicant before: Chengdu Tianfu create robot Co Ltd |
|
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20161109 |
|
RJ01 | Rejection of invention patent application after publication |