CN109085926A - A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings - Google Patents

A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings Download PDF

Info

Publication number
CN109085926A
CN109085926A CN201810954273.8A CN201810954273A CN109085926A CN 109085926 A CN109085926 A CN 109085926A CN 201810954273 A CN201810954273 A CN 201810954273A CN 109085926 A CN109085926 A CN 109085926A
Authority
CN
China
Prior art keywords
perception
augmented reality
module
modality imaging
auxiliary
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810954273.8A
Other languages
Chinese (zh)
Inventor
胡孟晗
李庆利
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
East China Normal University
Original Assignee
East China Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by East China Normal University filed Critical East China Normal University
Priority to CN201810954273.8A priority Critical patent/CN109085926A/en
Publication of CN109085926A publication Critical patent/CN109085926A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/016Input arrangements with force or tactile feedback as computer generated output to the user
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B2027/0178Eyeglass type
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/01Indexing scheme relating to G06F3/01
    • G06F2203/012Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment

Abstract

The invention discloses the augmented reality systems and its application of a kind of multi-modality imaging and more perception blendings, its main feature is that the system is by multi-modality imaging module, Multilayer Perception module, auxiliary sensing module and controlling terminal composition, the surrounding enviroment information input controlling terminal that the multi-modality imaging module and auxiliary sensing module will acquire, and treated image stream and data flow are fed back to the progress of Multilayer Perception module and accurately transmitted comprehensively by controlling terminal, and it is interacted by microphone and system, it is enforced the law using the augmented reality glasses of the augmented reality system architecture for intelligence, auxiliary visually impaired and driving auxiliary, realize comprehensive precisely reliable perception of the user to surrounding enviroment.There is the present invention multi-modality imaging to merge with the interaction of Multilayer Perception compared with prior art, realize the accurate acquisition and precisely transmitting of effective information under application scenarios, with the application demand under the varying environments such as the law enforcement of satisfaction intelligence, auxiliary visually impaired, auxiliary driving, easy to use, high reliablity.

Description

A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings
Technical field
The present invention relates to multi-modality imaging and Multilayer Perception technical field, especially a kind of multi-modality imaging and more perception are handed over The augmented reality system melted and its application.
Background technique
The application prospect of augmented reality glasses in real life is huge, since Google starts to issue Google glass, Many enterprises all start to set foot in augmented reality eyewear art in the world.Just existing augmented reality and its relevant eye on the market For mirror product, mainly have Google glass (G. Inc., Google Glass, 2012), eSight 3 (e. Co., eSight 3, 2017)、OrCam (O.T. Ltd., OrCam, 2015)、Enchroma (I. Enchroma, Enchroma, 2013), Intoer (L. Hangzhou KR-VISION Technology Co., Intoer, 2017) etc. Main product.Wherein most product is all that product is all single RGB camera as data acquisition equipment, then uses vision Or the feedback system of sound carries out the transmission of information.Mode existing for these products and the single problem of sensible form be it is current this Limited one of the maximum reason of a little products application scenes.
It with " augmented reality glasses " is that keyword is searched in the patent database in China, it is special there will be about 540 Benefit, 325 patents of invention (patent containing invention disclosed).Existing related patents are enumerated and analyzed below: as enhanced (application number: 201510009348.1) acquiring ambient condition information by camera in real time, and by virtual objects for real intelligent glasses It is superimposed reality with real picture, three-dimensional stereo effect is generated, is interacted by action sensor and sound transducer with system; Big field angle augmented reality glasses (application number: 201510573508.5) main that big field angle enhancing is realized by optics framework Reality glasses, to enhance the usage experience of user, which increases touch screen in another patent, real The interaction of user is showed, the actually perception of user and interaction is undivided, has perception, just there is interaction;Band augmented reality Intelligent glasses (application number: 201610035054.0) using visible light and depth camera collaboration by the way of enhance dysopia Perception of the user to extraneous things;Augmented reality glasses (application number: 201610270873.3) use double Visible Light Cameras and add The mode of flash lamp greatly improves aggregation capability and imaging image quality;Multiple field augmented reality intelligent glasses (application number: 201610316464.2) be used in combination filter glass, display eyeglass, functional lens optical design scheme realize the sense of multilayer Know;(application number: 201610049175.0) pass through communications protocol reality for realizing augmented reality interaction and the method shown, equipment The communication between disparate modules is showed, the Multilayer Perception of user is then carried out by way of vision, sound, vibration.
Existing patent and paper are all confined to single product or shape to the elaboration of augmented reality system and show form, Wherein the form of glasses is in the majority, and augmented reality glasses are only the one of the augmented reality system of multi-modality imaging and more perception blendings Kind concrete application.In actual application, according to specific task, augmented reality system should also include waistband, gloves, cap, The embodiment and displaying of the other forms such as walking stick, clothes, trousers, shoes, bracelet, wrist-watch, necklace, ring, foot ring, ear nail or packet. Now, there are not the related patents and paper that augmented reality system is described in a modular fashion also.
In conclusion the augmented reality system of the prior art there is mode it is single, perceptual model is deficient the problems such as, fail Will perception and preferably interacted and integrated, limit the large-scale practical application of augmented reality system, this be also it is current this Limited one of the maximum reason of a little products application scenes.
Summary of the invention
The purpose of the present invention is a kind of multi-modality imaging designed and more perception blendings in view of the deficiencies of the prior art Augmented reality system and its application are realized under complex application context effectively using the framework of multi-modality imaging fusion Multilayer Perception The accurate acquisition of information and precisely comprehensive transmitting, using the big visual angle augmented reality glasses of the augmented reality system architecture, or The auxiliary device that framework is applied as different scenes on walking stick, waistband, gloves, cap and clothes, to meet intelligence law enforcement, view Application demand under the varying environments such as barrier auxiliary, auxiliary driving, structure is simple, easy to use, high reliablity.
The object of the present invention is achieved like this: a kind of augmented reality system of multi-modality imaging and more perception blendings, Feature is that the augmented reality system is made of multi-modality imaging module, Multilayer Perception module, auxiliary sensing module and controlling terminal, The multi-modality imaging module is by visible light/near infrared camera, depth camera, far infrared thermal imager, near-infrared array light source, point Battle array laser and paliform laser are constituted;The Multilayer Perception module is occurred by sense of touch generator, gas ejector, vibration Device, bone conduction earphone and OLED display screen are constituted;The auxiliary sensing module is by gyroscope, ultrasonic distance-measuring sensor, infrared Distance measuring sensor and microphone are constituted;The controlling terminal is made of transmission module and microprocessor;The multi-modality imaging mould The surrounding enviroment information input controlling terminal that block and auxiliary sensing module will acquire, and will treated image stream by controlling terminal Feed back to Multilayer Perception module with data flow, by sense of touch generator, gas ejector, vibration machine, bone conduction earphone and OLED display screen progress is accurately transmitted comprehensively, and user makes corresponding behaviour decision making by the information fed back, and passes through wheat Gram wind is interacted with augmented reality system.
Visible light/the near infrared camera is realized the imaging of laser biology speckle by dot matrix laser and paliform laser Mode.
The near-infrared array light source, which carries out visible light/near infrared camera by change-over switch, optical mode (daytime) and nothing The switching of optical mode (night), and by far infrared thermal imager, ultrasonic distance-measuring sensor and infrared distance sensor to acquisition Object carries out thermometric, ranging and tests the speed in application scenarios.
The sense of touch generator is the heat or electro photoluminescence generated under pain threshold.
The gas ejector is to spray the application scenarios letter that the gas transmitting multi-modality imaging module of varying strength obtains Breath.
The vibration machine is to generate to beat stimulation under pain threshold.
The OLED display screen is that multi-modality images information is mapped to the full impregnated mode mirror in corresponding instant scene one by one Piece, and be equipped with the image information of eye movement or mechanical switch switching different modalities and its merge state and show.
The transmission module is that bluetooth, 4G or WIFI carry out the wireless transmission of each intermodule, or by the data of microprocessor It is transferred to cloud, and receives cloud treated data or decision information.
A kind of augmented reality systematic difference of multi-modality imaging and more perception blendings, feature utilize the augmented reality system Unite framework big visual angle augmented reality glasses for intelligence enforce the law, it is visually impaired auxiliary and driving auxiliary, or walking stick, waistband, gloves, The auxiliary device that framework is applied as different scenes on cap and clothes, the big visual angle augmented reality glasses be by it is multi-modal at As module, Multilayer Perception module, auxiliary sensing module and controlling terminal are separately positioned on the outer of frame, realization wearer Comprehensive precisely reliable perception to surrounding enviroment.
The present invention has the accurate acquisition of effective information under complex application context and precisely comprehensive compared with prior art Transmitting, with the application demand under the varying environments such as the law enforcement of satisfaction intelligence, auxiliary visually impaired, auxiliary driving, structure is simple, user Just, high reliablity provides a kind of resolving ideas for the acquisition and perception of information under high-precision complex environment, effectively integrates Visible Light Camera, near infrared camera, far infrared camera, laser biology speckle imaging technique and depth camera, and it is aided with near-infrared The different types of structure lights such as array light source, dot matrix laser and paliform laser, it is ensured that the information under complicated various environment Height precisely obtain, highly incorporate the cognition technology including the different human perception type hierarchy such as tactile, vision and sound, And heat/electric sense of touch generator, gas ejector, vibration machine, bone conduction earphone and OLED display screen etc. is various forms of Perception data, can largely improve the accuracy and robustness of system, thus meet intelligent law enforcement, it is visually impaired auxiliary, it is auxiliary The different application demands such as driving are helped, multi-modality imaging module and Multilayer Perception module is largely assisted to complete special scenes Under complex task, and related adjustment can be carried out according to actual requirement, better meet the application demand of different scenes.
Detailed description of the invention
Fig. 1 is schematic structural view of the invention;
Fig. 2 is multi-modality imaging modular structure schematic diagram;
Fig. 3 is dot matrix laser effect picture;
Fig. 4 is the effect picture of paliform laser;
Fig. 5 is Multilayer Perception modular structure schematic diagram;
Fig. 6 is auxiliary sensing module structural schematic diagram;
Fig. 7 is controlling terminal structural schematic diagram;
Fig. 8 is implementation example figure.
Specific embodiment
Refering to attached drawing 1, the present invention is whole by multi-modality imaging module 1, Multilayer Perception module 2, auxiliary sensing module 3 and control 4 composition of end, the surrounding enviroment information input controlling terminal 4 that the multi-modality imaging module 1 and auxiliary sensing module 3 will acquire, And treated image stream and data flow are fed back into the sense of touch generator 21 of Multilayer Perception module 2 by controlling terminal 4, gas sprays Emitter 22, vibration machine 23, bone conduction earphone 24 and OLED display screen 25 accurately transmit comprehensively, and user passes through anti- The information of feedback makes corresponding behaviour decision making, and is interacted by microphone 34 and augmented reality system.
Refering to attached drawing 2, the multi-modality imaging module 1 is by visible light/near infrared camera 11, depth camera 12, far infrared Thermal imaging system 13, near-infrared array light source 14, dot matrix laser 15 and paliform laser 16 are constituted, the visible light/near-infrared Camera 11 obtains the near-infrared image stream with structure optical information under the floor light of near-infrared array light source 14.
Refering to attached drawing 3 ~ 4, the visible light/near infrared camera 11 is realized by dot matrix laser 15 and paliform laser 16 The imaging pattern of laser biology speckle obtains the visible light/near-infrared image stream for having structure optical information.
Refering to attached drawing 5, the Multilayer Perception module 2 by sense of touch generator 21, gas ejector 22, vibration machine 23, Bone conduction earphone 24 and OLED display screen 25 are constituted;The sense of touch generator 21 is the heat or electricity thorn generated under pain threshold Swash;The gas ejector 22 is to spray the application scenarios information that the gas communicating multi-modality imaging module 1 of varying strength obtains; The vibration machine 23 is to generate to beat stimulation under pain threshold;The OLED display screen 25 is multi-modality images information The full impregnated mode eyeglass being mapped in corresponding instant scene one by one, and it is equipped with the figure of eye movement or mechanical switch switching different modalities As information and its fusion state are shown.
Refering to attached drawing 6, the auxiliary sensing module 3 is sensed by gyroscope 31, ultrasonic distance-measuring sensor 32, infrared distance measurement Device 33 and microphone 34 are constituted;In the ultrasonic distance-measuring sensor 32 and the application scenarios of 33 pairs of infrared distance sensor acquisitions Object carries out ranging and tests the speed.
Refering to attached drawing 7, the controlling terminal 4 is made of transmission module 41 and microprocessor 42;The transmission module 41 is Bluetooth, 4G or WIFI carry out the wireless transmission of each intermodule, or the data of microprocessor 42 are transferred to cloud, and receive cloud Data that treated or decision information.
Multi-modality imaging module 1, Multilayer Perception module 2, auxiliary sensing module 3 and controlling terminal 4 are respectively set below In the outer of frame, or be arranged in walking stick, waistband, gloves, on cap and clothes framework applied as different scenes it is auxiliary Device is helped, it is further detailed to realize that user makees the comprehensive utilization example precisely reliably perceived of surrounding enviroment to the present invention It describes in detail bright.
Embodiment 1
Refering to attached drawing 8, the big visual angle augmented reality glasses of framework of the present invention by conventional eyeglass 17 and OLED display screen 25 full impregnated Mode eyeglass composition, it will be seen that light/near infrared camera 11, near-infrared array light source 14, depth camera 12 and microprocessor 42 are set It sets in the side temple of mirror holder 20, the other side temple of mirror holder 20 is equipped with sense of touch generator 21, gas ejector 22, transmission Module 41, vibration machine 23 and bone conduction earphone 24, the nose frame of two eyeglasses of connection is equipped with dot matrix laser 15, paliform swashs Light device 16 and gyroscope 31, it is described routine eyeglass 17 above be equipped with infrared distance sensor 23;The full impregnated mode eyeglass It is equipped with ultrasonic distance-measuring sensor 32 and far infrared thermal imager 13 above OLED display screen 25, is equipped with microphone 34 below.
Embodiment 2
Application of the present invention in intelligently law enforcement, the big visual angle augmented reality glasses that staff wears framework of the present invention show When the law enforcement of field, there is unknown heat source on the stand of the discovery distant place retailer of far infrared thermal imager 13, and remind by vibration machine 23 Law enfrocement official, law enfrocement official pass through the information that bone conduction earphone 24 provides, and the data in OLED display screen 25 are switched to remote red The image of outer thermal imaging system 13 obtains the far infrared image by 34 acoustic control of microphone, and pass through transmission after finding unknown heat source The photo is sent to law enforcement general headquarters by module 41, meanwhile, law enfrocement official quickly goes at unknown heat source, prevents potential disaster in time Generation.
Embodiment 3
Application of the present invention in auxiliary visually impaired, when visually impaired people wears the big visual angle augmented reality glasses of framework of the present invention, depending on Barrier personage's pirouette circle is looked around, and determines whether according to the data of gyroscope 31 to be fully completed 360 ° of scanning.It is micro- Processor 42 is provided after data processing according to visible light/near infrared camera 11, depth camera 12 and far infrared thermal imager 13 Image data determines the direction of travel of next step, and the location data of gyroscope 31 is passed through vibration hair by bone conduction earphone 24 Raw device 23 or gas ejector 22 are directly perceived by the mechanics of visually impaired people's skin and specific provide direction of travel.In traveling In the process, it is seen that light/near infrared camera 11 determines 100 meters of Chu Youyijia convenience stores at a distance by image recognition technology, then leads to It crosses bone conduction earphone 24 and reminds this message of visually impaired people, this moment, visually impaired people just needs to buy relative article, and according to mentioning Show and go to target convenience store, however, ultrasonic distance-measuring sensor 32 and infrared distance sensor 33 are same since travel speed is too fast When perceive on the travel path of front and have a barrier, but the barrier could not be by other due to orientation, form etc. Multi-modality imaging detect, the danger that microprocessor 42 detects ultrasonic distance-measuring sensor 32 and infrared distance sensor 33 Dangerous information quickly reminds visually impaired people, vibration machine 23 and gas ejector by vibration machine 23 and gas ejector 22 22 vibration and jet frequency in case of emergency can all be added to acceptable maximum value.At this point, visually impaired people perceives front Danger and it is tentative advance, security risk is further excluded by the detection of other multi-modality imagings.After reaching convenience store, far Thermal infrared imager 13 perceives refrigerator, and occurs 21 by sense of touch and visually give visually impaired people's creeping chill, and visually impaired people goes to cold Cabinet picks up one piece of ice cream, and pass through the irradiation of dot matrix laser 15 under the auxiliary of visible light/near infrared camera 11, it is seen that Light/near infrared camera 11 collects laser biology speckle image, and informs that visually impaired people's ice cream is by bone conduction earphone 24 Paper material packaging.
Embodiment 4
Application of the present invention in auxiliary drives is driven when driver wears the big visual angle augmented reality glasses of framework of the present invention The vehicle for sailing personnel is not having form on the highway of street lamp, and sight is poor, and by bone conduction earphone, 24 driver recognizes The front obstacle information for assisting visible light/near infrared camera 11 to obtain by paliform laser 16, then, driver slows down Vehicle travel speed, and finally determine the barrier in front, then around other lanes.After long-duration driving, drive The personnel of sailing feel more fatigue, but its own is not discovered, and driver starts to doze off this moment, gyroscope 31, infrared distance measurement sensing Device 33 and ultrasonic distance-measuring sensor 32 can all show head pose and its variation abnormality, and meet the head pose of fatigue driving Driver is reminded by vibration machine 23, gas ejector 22, sense of touch generator 21 and bone conduction earphone 24 after mode Should not fatigue driving, drive with caution.At this point, sense of touch generator 21 mainly allows driver by way of electro photoluminescence It is able to maintain awake, the place that finding service area etc. can rest is rested, and fatigue driving is prevented.
Above only the present invention is further illustrated, and not to limit this patent, all is equivalence enforcement of the present invention, It is intended to be limited solely by within the scope of the claims of this patent.

Claims (9)

1. the augmented reality system of a kind of multi-modality imaging and more perception blendings, it is characterised in that the augmented reality system is by multimode State image-forming module, Multilayer Perception module, auxiliary sensing module and controlling terminal composition, the multi-modality imaging module from Light/near infrared camera, depth camera, far infrared thermal imager, near-infrared array light source, dot matrix laser and paliform laser structure At;The Multilayer Perception module is by sense of touch generator, gas ejector, vibration machine, bone conduction earphone and OLED display screen It constitutes;The auxiliary sensing module is made of gyroscope, ultrasonic distance-measuring sensor, infrared distance sensor and microphone;Institute Controlling terminal is stated to be made of transmission module and microprocessor;The week that the multi-modality imaging module and auxiliary sensing module will acquire Surrounding environment information input controlling terminal, and treated image stream and data flow are fed back into Multilayer Perception mould by controlling terminal Block is carried out comprehensively accurately by sense of touch generator, gas ejector, vibration machine, bone conduction earphone and OLED display screen Transmitting, user makes corresponding behaviour decision making by the information fed back, and is handed over by microphone and augmented reality system Mutually.
2. the augmented reality systems of multi-modality imaging and more perception blending according to claim 1, it is characterised in that it is described can Light-exposed/near infrared camera is realized the imaging pattern of laser biology speckle by dot matrix laser and paliform laser.
3. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that described close Infrared array light source, which carries out visible light/near infrared camera by change-over switch, is having optical mode and without the switching of optical mode, and by remote Thermal infrared imager, ultrasonic distance-measuring sensor and infrared distance sensor carry out thermometric to object in the application scenarios of acquisition, survey Away from test the speed.
4. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that the touching Sense generator is the heat or electro photoluminescence generated under pain threshold.
5. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that the gas Body injector is to spray the application scenarios information that the gas transmitting multi-modality imaging module of varying strength obtains.
6. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that the vibration Dynamic generator is to generate to beat stimulation under pain threshold.
7. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that described OLED display screen is that multi-modality images information is mapped to the full impregnated mode eyeglass in corresponding instant scene one by one, and is equipped with eye movement Or the image information and its fusion state of mechanical switch switching different modalities are shown.
8. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that the biography Defeated module is that bluetooth, 4G or WIFI carry out the wireless transmission of each intermodule, or the data of microprocessor are transferred to cloud, and connect Receive that treated in cloud data or decision information.
9. the augmented reality systematic difference of multi-modality imaging described in a kind of claim 1 and more perception blendings, it is characterised in that Using the big visual angle augmented reality glasses of the augmented reality system architecture for intelligence law enforcement, auxiliary visually impaired and driving auxiliary, or The auxiliary device that framework is applied as different scenes on walking stick, waistband, gloves, cap and clothes, the big visual angle enhancing are existing Real glasses are that frame is arranged in multi-modality imaging module, Multilayer Perception module, auxiliary sensing module and controlling terminal Comprehensive precisely reliable perception of the wearer to surrounding enviroment is realized in outer.
CN201810954273.8A 2018-08-21 2018-08-21 A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings Pending CN109085926A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810954273.8A CN109085926A (en) 2018-08-21 2018-08-21 A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810954273.8A CN109085926A (en) 2018-08-21 2018-08-21 A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings

Publications (1)

Publication Number Publication Date
CN109085926A true CN109085926A (en) 2018-12-25

Family

ID=64794082

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810954273.8A Pending CN109085926A (en) 2018-08-21 2018-08-21 A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings

Country Status (1)

Country Link
CN (1) CN109085926A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110246178A (en) * 2019-07-09 2019-09-17 华东师范大学 A kind of modularization intelligent auxiliary system visually impaired
CN112402194A (en) * 2019-08-23 2021-02-26 北京超维度计算科技有限公司 Auxiliary terminal system for visually impaired people
CN115079423A (en) * 2022-08-19 2022-09-20 歌尔股份有限公司 Intelligent glasses and ultrasonic tactile feedback method thereof

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011106797A1 (en) * 2010-02-28 2011-09-01 Osterhout Group, Inc. Projection triggering through an external marker in an augmented reality eyepiece
CN104306102A (en) * 2014-10-10 2015-01-28 上海交通大学 Head wearing type vision auxiliary system for patients with vision disorder
CN104473717A (en) * 2014-12-04 2015-04-01 上海交通大学 Wearable guide apparatus for totally blind people
CN105094417A (en) * 2014-05-19 2015-11-25 意美森公司 Non-collocated haptic cues in immersive environments
CN106580294A (en) * 2016-12-30 2017-04-26 上海交通大学 Physiological signal remote monitoring system based on multimodal imaging technique and application thereof
CN106817577A (en) * 2016-11-23 2017-06-09 杭州视氪科技有限公司 One kind is based on RGB D cameras and stereosonic visually impaired people's barrier early warning glasses
CN206920723U (en) * 2017-07-24 2018-01-23 深圳市科乐科技有限公司 A kind of augmented reality glasses
CN107884942A (en) * 2017-12-18 2018-04-06 国网浙江省电力公司培训中心 A kind of augmented reality display device
CN108319022A (en) * 2018-03-22 2018-07-24 巢湖永微信息科技有限公司 A kind of system of the intelligent glasses of the hot sensing function of band

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011106797A1 (en) * 2010-02-28 2011-09-01 Osterhout Group, Inc. Projection triggering through an external marker in an augmented reality eyepiece
CN105094417A (en) * 2014-05-19 2015-11-25 意美森公司 Non-collocated haptic cues in immersive environments
CN104306102A (en) * 2014-10-10 2015-01-28 上海交通大学 Head wearing type vision auxiliary system for patients with vision disorder
CN104473717A (en) * 2014-12-04 2015-04-01 上海交通大学 Wearable guide apparatus for totally blind people
CN106817577A (en) * 2016-11-23 2017-06-09 杭州视氪科技有限公司 One kind is based on RGB D cameras and stereosonic visually impaired people's barrier early warning glasses
CN106580294A (en) * 2016-12-30 2017-04-26 上海交通大学 Physiological signal remote monitoring system based on multimodal imaging technique and application thereof
CN206920723U (en) * 2017-07-24 2018-01-23 深圳市科乐科技有限公司 A kind of augmented reality glasses
CN107884942A (en) * 2017-12-18 2018-04-06 国网浙江省电力公司培训中心 A kind of augmented reality display device
CN108319022A (en) * 2018-03-22 2018-07-24 巢湖永微信息科技有限公司 A kind of system of the intelligent glasses of the hot sensing function of band

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110246178A (en) * 2019-07-09 2019-09-17 华东师范大学 A kind of modularization intelligent auxiliary system visually impaired
CN112402194A (en) * 2019-08-23 2021-02-26 北京超维度计算科技有限公司 Auxiliary terminal system for visually impaired people
CN115079423A (en) * 2022-08-19 2022-09-20 歌尔股份有限公司 Intelligent glasses and ultrasonic tactile feedback method thereof

Similar Documents

Publication Publication Date Title
EP3014338B1 (en) Tracking head movement when wearing mobile device
US9760167B2 (en) Visual stabilization system for head-mounted displays
US10656731B2 (en) Peripheral device for head-mounted display
WO2016184107A1 (en) Wearable apparatus for sight line focus positioning and method for sight line focus positioning
CN104956252B (en) Peripheral display for near-eye display device
CN104603673B (en) Head-mounted system and the method for being calculated using head-mounted system and rendering digital image stream
US10663729B2 (en) Peripheral device for head-mounted display
WO2017115618A1 (en) Information processing apparatus, information processing method, and program
CN109085926A (en) A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings
KR20180110051A (en) Systems and methods for augmented reality
US20130241805A1 (en) Using Convergence Angle to Select Among Different UI Elements
CN108919498A (en) A kind of augmented reality glasses based on multi-modality imaging and Multilayer Perception
CN104808340B (en) Head-mounted display device and control method thereof
CN109358754B (en) Mixed reality head-mounted display system
JPH0749744A (en) Head mounting type display input device
CN103731659B (en) head-mounted display apparatus
CN103748598A (en) Systems and methods for identifying gaze tracking scene reference locations
US11734898B2 (en) Program, information processing method, and information processing terminal
KR101203921B1 (en) Information providing apparatus using an eye tracking and local based service
WO2022095605A1 (en) Handle control tracker
US20120282585A1 (en) Interest-Attention Feedback System for Separating Cognitive Awareness into Different Left and Right Sensor Displays
CN108064447A (en) Method for displaying image, intelligent glasses and storage medium
CN207220273U (en) Visualize sea police's patrol helmet
WO2014128750A1 (en) Input/output device, input/output program, and input/output method
CN109445596B (en) Integrated mixed reality head-mounted display system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20181225