CN109085926A - A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings - Google Patents
A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings Download PDFInfo
- Publication number
- CN109085926A CN109085926A CN201810954273.8A CN201810954273A CN109085926A CN 109085926 A CN109085926 A CN 109085926A CN 201810954273 A CN201810954273 A CN 201810954273A CN 109085926 A CN109085926 A CN 109085926A
- Authority
- CN
- China
- Prior art keywords
- perception
- augmented reality
- module
- modality imaging
- auxiliary
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000003190 augmentative effect Effects 0.000 title claims abstract description 50
- 230000008447 perception Effects 0.000 title claims abstract description 47
- 238000003384 imaging method Methods 0.000 title claims abstract description 44
- 238000002156 mixing Methods 0.000 title claims abstract description 16
- 239000011521 glass Substances 0.000 claims abstract description 24
- 230000001771 impaired effect Effects 0.000 claims abstract description 16
- 239000000203 mixture Substances 0.000 claims abstract description 4
- 210000000988 bone and bone Anatomy 0.000 claims description 14
- 230000005540 biological transmission Effects 0.000 claims description 11
- 239000011159 matrix material Substances 0.000 claims description 9
- 230000000007 visual effect Effects 0.000 claims description 9
- 230000003287 optical effect Effects 0.000 claims description 7
- 208000000114 Pain Threshold Diseases 0.000 claims description 6
- 230000037040 pain threshold Effects 0.000 claims description 6
- 241001166076 Diapheromera femorata Species 0.000 claims description 5
- 239000007921 spray Substances 0.000 claims description 4
- 230000004424 eye movement Effects 0.000 claims description 3
- 230000004927 fusion Effects 0.000 claims description 3
- 238000005424 photoluminescence Methods 0.000 claims description 3
- 230000000638 stimulation Effects 0.000 claims description 3
- 238000012360 testing method Methods 0.000 claims description 3
- 230000002708 enhancing effect Effects 0.000 claims description 2
- 230000009897 systematic effect Effects 0.000 claims description 2
- 230000003993 interaction Effects 0.000 abstract description 5
- 230000004888 barrier function Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 238000000034 method Methods 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 235000015243 ice cream Nutrition 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 238000001931 thermography Methods 0.000 description 2
- 241000209140 Triticum Species 0.000 description 1
- 235000021307 Triticum Nutrition 0.000 description 1
- 206010047571 Visual impairment Diseases 0.000 description 1
- 230000005856 abnormality Effects 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000002776 aggregation Effects 0.000 description 1
- 238000004220 aggregation Methods 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000019771 cognition Effects 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 230000002950 deficient Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000005611 electricity Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000004806 packaging method and process Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/016—Input arrangements with force or tactile feedback as computer generated output to the user
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/012—Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment
Abstract
The invention discloses the augmented reality systems and its application of a kind of multi-modality imaging and more perception blendings, its main feature is that the system is by multi-modality imaging module, Multilayer Perception module, auxiliary sensing module and controlling terminal composition, the surrounding enviroment information input controlling terminal that the multi-modality imaging module and auxiliary sensing module will acquire, and treated image stream and data flow are fed back to the progress of Multilayer Perception module and accurately transmitted comprehensively by controlling terminal, and it is interacted by microphone and system, it is enforced the law using the augmented reality glasses of the augmented reality system architecture for intelligence, auxiliary visually impaired and driving auxiliary, realize comprehensive precisely reliable perception of the user to surrounding enviroment.There is the present invention multi-modality imaging to merge with the interaction of Multilayer Perception compared with prior art, realize the accurate acquisition and precisely transmitting of effective information under application scenarios, with the application demand under the varying environments such as the law enforcement of satisfaction intelligence, auxiliary visually impaired, auxiliary driving, easy to use, high reliablity.
Description
Technical field
The present invention relates to multi-modality imaging and Multilayer Perception technical field, especially a kind of multi-modality imaging and more perception are handed over
The augmented reality system melted and its application.
Background technique
The application prospect of augmented reality glasses in real life is huge, since Google starts to issue Google glass,
Many enterprises all start to set foot in augmented reality eyewear art in the world.Just existing augmented reality and its relevant eye on the market
For mirror product, mainly have Google glass (G. Inc., Google Glass, 2012), eSight 3 (e. Co.,
eSight 3, 2017)、OrCam (O.T. Ltd., OrCam, 2015)、Enchroma (I. Enchroma,
Enchroma, 2013), Intoer (L. Hangzhou KR-VISION Technology Co., Intoer, 2017) etc.
Main product.Wherein most product is all that product is all single RGB camera as data acquisition equipment, then uses vision
Or the feedback system of sound carries out the transmission of information.Mode existing for these products and the single problem of sensible form be it is current this
Limited one of the maximum reason of a little products application scenes.
It with " augmented reality glasses " is that keyword is searched in the patent database in China, it is special there will be about 540
Benefit, 325 patents of invention (patent containing invention disclosed).Existing related patents are enumerated and analyzed below: as enhanced
(application number: 201510009348.1) acquiring ambient condition information by camera in real time, and by virtual objects for real intelligent glasses
It is superimposed reality with real picture, three-dimensional stereo effect is generated, is interacted by action sensor and sound transducer with system;
Big field angle augmented reality glasses (application number: 201510573508.5) main that big field angle enhancing is realized by optics framework
Reality glasses, to enhance the usage experience of user, which increases touch screen in another patent, real
The interaction of user is showed, the actually perception of user and interaction is undivided, has perception, just there is interaction;Band augmented reality
Intelligent glasses (application number: 201610035054.0) using visible light and depth camera collaboration by the way of enhance dysopia
Perception of the user to extraneous things;Augmented reality glasses (application number: 201610270873.3) use double Visible Light Cameras and add
The mode of flash lamp greatly improves aggregation capability and imaging image quality;Multiple field augmented reality intelligent glasses (application number:
201610316464.2) be used in combination filter glass, display eyeglass, functional lens optical design scheme realize the sense of multilayer
Know;(application number: 201610049175.0) pass through communications protocol reality for realizing augmented reality interaction and the method shown, equipment
The communication between disparate modules is showed, the Multilayer Perception of user is then carried out by way of vision, sound, vibration.
Existing patent and paper are all confined to single product or shape to the elaboration of augmented reality system and show form,
Wherein the form of glasses is in the majority, and augmented reality glasses are only the one of the augmented reality system of multi-modality imaging and more perception blendings
Kind concrete application.In actual application, according to specific task, augmented reality system should also include waistband, gloves, cap,
The embodiment and displaying of the other forms such as walking stick, clothes, trousers, shoes, bracelet, wrist-watch, necklace, ring, foot ring, ear nail or packet.
Now, there are not the related patents and paper that augmented reality system is described in a modular fashion also.
In conclusion the augmented reality system of the prior art there is mode it is single, perceptual model is deficient the problems such as, fail
Will perception and preferably interacted and integrated, limit the large-scale practical application of augmented reality system, this be also it is current this
Limited one of the maximum reason of a little products application scenes.
Summary of the invention
The purpose of the present invention is a kind of multi-modality imaging designed and more perception blendings in view of the deficiencies of the prior art
Augmented reality system and its application are realized under complex application context effectively using the framework of multi-modality imaging fusion Multilayer Perception
The accurate acquisition of information and precisely comprehensive transmitting, using the big visual angle augmented reality glasses of the augmented reality system architecture, or
The auxiliary device that framework is applied as different scenes on walking stick, waistband, gloves, cap and clothes, to meet intelligence law enforcement, view
Application demand under the varying environments such as barrier auxiliary, auxiliary driving, structure is simple, easy to use, high reliablity.
The object of the present invention is achieved like this: a kind of augmented reality system of multi-modality imaging and more perception blendings,
Feature is that the augmented reality system is made of multi-modality imaging module, Multilayer Perception module, auxiliary sensing module and controlling terminal,
The multi-modality imaging module is by visible light/near infrared camera, depth camera, far infrared thermal imager, near-infrared array light source, point
Battle array laser and paliform laser are constituted;The Multilayer Perception module is occurred by sense of touch generator, gas ejector, vibration
Device, bone conduction earphone and OLED display screen are constituted;The auxiliary sensing module is by gyroscope, ultrasonic distance-measuring sensor, infrared
Distance measuring sensor and microphone are constituted;The controlling terminal is made of transmission module and microprocessor;The multi-modality imaging mould
The surrounding enviroment information input controlling terminal that block and auxiliary sensing module will acquire, and will treated image stream by controlling terminal
Feed back to Multilayer Perception module with data flow, by sense of touch generator, gas ejector, vibration machine, bone conduction earphone and
OLED display screen progress is accurately transmitted comprehensively, and user makes corresponding behaviour decision making by the information fed back, and passes through wheat
Gram wind is interacted with augmented reality system.
Visible light/the near infrared camera is realized the imaging of laser biology speckle by dot matrix laser and paliform laser
Mode.
The near-infrared array light source, which carries out visible light/near infrared camera by change-over switch, optical mode (daytime) and nothing
The switching of optical mode (night), and by far infrared thermal imager, ultrasonic distance-measuring sensor and infrared distance sensor to acquisition
Object carries out thermometric, ranging and tests the speed in application scenarios.
The sense of touch generator is the heat or electro photoluminescence generated under pain threshold.
The gas ejector is to spray the application scenarios letter that the gas transmitting multi-modality imaging module of varying strength obtains
Breath.
The vibration machine is to generate to beat stimulation under pain threshold.
The OLED display screen is that multi-modality images information is mapped to the full impregnated mode mirror in corresponding instant scene one by one
Piece, and be equipped with the image information of eye movement or mechanical switch switching different modalities and its merge state and show.
The transmission module is that bluetooth, 4G or WIFI carry out the wireless transmission of each intermodule, or by the data of microprocessor
It is transferred to cloud, and receives cloud treated data or decision information.
A kind of augmented reality systematic difference of multi-modality imaging and more perception blendings, feature utilize the augmented reality system
Unite framework big visual angle augmented reality glasses for intelligence enforce the law, it is visually impaired auxiliary and driving auxiliary, or walking stick, waistband, gloves,
The auxiliary device that framework is applied as different scenes on cap and clothes, the big visual angle augmented reality glasses be by it is multi-modal at
As module, Multilayer Perception module, auxiliary sensing module and controlling terminal are separately positioned on the outer of frame, realization wearer
Comprehensive precisely reliable perception to surrounding enviroment.
The present invention has the accurate acquisition of effective information under complex application context and precisely comprehensive compared with prior art
Transmitting, with the application demand under the varying environments such as the law enforcement of satisfaction intelligence, auxiliary visually impaired, auxiliary driving, structure is simple, user
Just, high reliablity provides a kind of resolving ideas for the acquisition and perception of information under high-precision complex environment, effectively integrates
Visible Light Camera, near infrared camera, far infrared camera, laser biology speckle imaging technique and depth camera, and it is aided with near-infrared
The different types of structure lights such as array light source, dot matrix laser and paliform laser, it is ensured that the information under complicated various environment
Height precisely obtain, highly incorporate the cognition technology including the different human perception type hierarchy such as tactile, vision and sound,
And heat/electric sense of touch generator, gas ejector, vibration machine, bone conduction earphone and OLED display screen etc. is various forms of
Perception data, can largely improve the accuracy and robustness of system, thus meet intelligent law enforcement, it is visually impaired auxiliary, it is auxiliary
The different application demands such as driving are helped, multi-modality imaging module and Multilayer Perception module is largely assisted to complete special scenes
Under complex task, and related adjustment can be carried out according to actual requirement, better meet the application demand of different scenes.
Detailed description of the invention
Fig. 1 is schematic structural view of the invention;
Fig. 2 is multi-modality imaging modular structure schematic diagram;
Fig. 3 is dot matrix laser effect picture;
Fig. 4 is the effect picture of paliform laser;
Fig. 5 is Multilayer Perception modular structure schematic diagram;
Fig. 6 is auxiliary sensing module structural schematic diagram;
Fig. 7 is controlling terminal structural schematic diagram;
Fig. 8 is implementation example figure.
Specific embodiment
Refering to attached drawing 1, the present invention is whole by multi-modality imaging module 1, Multilayer Perception module 2, auxiliary sensing module 3 and control
4 composition of end, the surrounding enviroment information input controlling terminal 4 that the multi-modality imaging module 1 and auxiliary sensing module 3 will acquire,
And treated image stream and data flow are fed back into the sense of touch generator 21 of Multilayer Perception module 2 by controlling terminal 4, gas sprays
Emitter 22, vibration machine 23, bone conduction earphone 24 and OLED display screen 25 accurately transmit comprehensively, and user passes through anti-
The information of feedback makes corresponding behaviour decision making, and is interacted by microphone 34 and augmented reality system.
Refering to attached drawing 2, the multi-modality imaging module 1 is by visible light/near infrared camera 11, depth camera 12, far infrared
Thermal imaging system 13, near-infrared array light source 14, dot matrix laser 15 and paliform laser 16 are constituted, the visible light/near-infrared
Camera 11 obtains the near-infrared image stream with structure optical information under the floor light of near-infrared array light source 14.
Refering to attached drawing 3 ~ 4, the visible light/near infrared camera 11 is realized by dot matrix laser 15 and paliform laser 16
The imaging pattern of laser biology speckle obtains the visible light/near-infrared image stream for having structure optical information.
Refering to attached drawing 5, the Multilayer Perception module 2 by sense of touch generator 21, gas ejector 22, vibration machine 23,
Bone conduction earphone 24 and OLED display screen 25 are constituted;The sense of touch generator 21 is the heat or electricity thorn generated under pain threshold
Swash;The gas ejector 22 is to spray the application scenarios information that the gas communicating multi-modality imaging module 1 of varying strength obtains;
The vibration machine 23 is to generate to beat stimulation under pain threshold;The OLED display screen 25 is multi-modality images information
The full impregnated mode eyeglass being mapped in corresponding instant scene one by one, and it is equipped with the figure of eye movement or mechanical switch switching different modalities
As information and its fusion state are shown.
Refering to attached drawing 6, the auxiliary sensing module 3 is sensed by gyroscope 31, ultrasonic distance-measuring sensor 32, infrared distance measurement
Device 33 and microphone 34 are constituted;In the ultrasonic distance-measuring sensor 32 and the application scenarios of 33 pairs of infrared distance sensor acquisitions
Object carries out ranging and tests the speed.
Refering to attached drawing 7, the controlling terminal 4 is made of transmission module 41 and microprocessor 42;The transmission module 41 is
Bluetooth, 4G or WIFI carry out the wireless transmission of each intermodule, or the data of microprocessor 42 are transferred to cloud, and receive cloud
Data that treated or decision information.
Multi-modality imaging module 1, Multilayer Perception module 2, auxiliary sensing module 3 and controlling terminal 4 are respectively set below
In the outer of frame, or be arranged in walking stick, waistband, gloves, on cap and clothes framework applied as different scenes it is auxiliary
Device is helped, it is further detailed to realize that user makees the comprehensive utilization example precisely reliably perceived of surrounding enviroment to the present invention
It describes in detail bright.
Embodiment 1
Refering to attached drawing 8, the big visual angle augmented reality glasses of framework of the present invention by conventional eyeglass 17 and OLED display screen 25 full impregnated
Mode eyeglass composition, it will be seen that light/near infrared camera 11, near-infrared array light source 14, depth camera 12 and microprocessor 42 are set
It sets in the side temple of mirror holder 20, the other side temple of mirror holder 20 is equipped with sense of touch generator 21, gas ejector 22, transmission
Module 41, vibration machine 23 and bone conduction earphone 24, the nose frame of two eyeglasses of connection is equipped with dot matrix laser 15, paliform swashs
Light device 16 and gyroscope 31, it is described routine eyeglass 17 above be equipped with infrared distance sensor 23;The full impregnated mode eyeglass
It is equipped with ultrasonic distance-measuring sensor 32 and far infrared thermal imager 13 above OLED display screen 25, is equipped with microphone 34 below.
Embodiment 2
Application of the present invention in intelligently law enforcement, the big visual angle augmented reality glasses that staff wears framework of the present invention show
When the law enforcement of field, there is unknown heat source on the stand of the discovery distant place retailer of far infrared thermal imager 13, and remind by vibration machine 23
Law enfrocement official, law enfrocement official pass through the information that bone conduction earphone 24 provides, and the data in OLED display screen 25 are switched to remote red
The image of outer thermal imaging system 13 obtains the far infrared image by 34 acoustic control of microphone, and pass through transmission after finding unknown heat source
The photo is sent to law enforcement general headquarters by module 41, meanwhile, law enfrocement official quickly goes at unknown heat source, prevents potential disaster in time
Generation.
Embodiment 3
Application of the present invention in auxiliary visually impaired, when visually impaired people wears the big visual angle augmented reality glasses of framework of the present invention, depending on
Barrier personage's pirouette circle is looked around, and determines whether according to the data of gyroscope 31 to be fully completed 360 ° of scanning.It is micro-
Processor 42 is provided after data processing according to visible light/near infrared camera 11, depth camera 12 and far infrared thermal imager 13
Image data determines the direction of travel of next step, and the location data of gyroscope 31 is passed through vibration hair by bone conduction earphone 24
Raw device 23 or gas ejector 22 are directly perceived by the mechanics of visually impaired people's skin and specific provide direction of travel.In traveling
In the process, it is seen that light/near infrared camera 11 determines 100 meters of Chu Youyijia convenience stores at a distance by image recognition technology, then leads to
It crosses bone conduction earphone 24 and reminds this message of visually impaired people, this moment, visually impaired people just needs to buy relative article, and according to mentioning
Show and go to target convenience store, however, ultrasonic distance-measuring sensor 32 and infrared distance sensor 33 are same since travel speed is too fast
When perceive on the travel path of front and have a barrier, but the barrier could not be by other due to orientation, form etc.
Multi-modality imaging detect, the danger that microprocessor 42 detects ultrasonic distance-measuring sensor 32 and infrared distance sensor 33
Dangerous information quickly reminds visually impaired people, vibration machine 23 and gas ejector by vibration machine 23 and gas ejector 22
22 vibration and jet frequency in case of emergency can all be added to acceptable maximum value.At this point, visually impaired people perceives front
Danger and it is tentative advance, security risk is further excluded by the detection of other multi-modality imagings.After reaching convenience store, far
Thermal infrared imager 13 perceives refrigerator, and occurs 21 by sense of touch and visually give visually impaired people's creeping chill, and visually impaired people goes to cold
Cabinet picks up one piece of ice cream, and pass through the irradiation of dot matrix laser 15 under the auxiliary of visible light/near infrared camera 11, it is seen that
Light/near infrared camera 11 collects laser biology speckle image, and informs that visually impaired people's ice cream is by bone conduction earphone 24
Paper material packaging.
Embodiment 4
Application of the present invention in auxiliary drives is driven when driver wears the big visual angle augmented reality glasses of framework of the present invention
The vehicle for sailing personnel is not having form on the highway of street lamp, and sight is poor, and by bone conduction earphone, 24 driver recognizes
The front obstacle information for assisting visible light/near infrared camera 11 to obtain by paliform laser 16, then, driver slows down
Vehicle travel speed, and finally determine the barrier in front, then around other lanes.After long-duration driving, drive
The personnel of sailing feel more fatigue, but its own is not discovered, and driver starts to doze off this moment, gyroscope 31, infrared distance measurement sensing
Device 33 and ultrasonic distance-measuring sensor 32 can all show head pose and its variation abnormality, and meet the head pose of fatigue driving
Driver is reminded by vibration machine 23, gas ejector 22, sense of touch generator 21 and bone conduction earphone 24 after mode
Should not fatigue driving, drive with caution.At this point, sense of touch generator 21 mainly allows driver by way of electro photoluminescence
It is able to maintain awake, the place that finding service area etc. can rest is rested, and fatigue driving is prevented.
Above only the present invention is further illustrated, and not to limit this patent, all is equivalence enforcement of the present invention,
It is intended to be limited solely by within the scope of the claims of this patent.
Claims (9)
1. the augmented reality system of a kind of multi-modality imaging and more perception blendings, it is characterised in that the augmented reality system is by multimode
State image-forming module, Multilayer Perception module, auxiliary sensing module and controlling terminal composition, the multi-modality imaging module from
Light/near infrared camera, depth camera, far infrared thermal imager, near-infrared array light source, dot matrix laser and paliform laser structure
At;The Multilayer Perception module is by sense of touch generator, gas ejector, vibration machine, bone conduction earphone and OLED display screen
It constitutes;The auxiliary sensing module is made of gyroscope, ultrasonic distance-measuring sensor, infrared distance sensor and microphone;Institute
Controlling terminal is stated to be made of transmission module and microprocessor;The week that the multi-modality imaging module and auxiliary sensing module will acquire
Surrounding environment information input controlling terminal, and treated image stream and data flow are fed back into Multilayer Perception mould by controlling terminal
Block is carried out comprehensively accurately by sense of touch generator, gas ejector, vibration machine, bone conduction earphone and OLED display screen
Transmitting, user makes corresponding behaviour decision making by the information fed back, and is handed over by microphone and augmented reality system
Mutually.
2. the augmented reality systems of multi-modality imaging and more perception blending according to claim 1, it is characterised in that it is described can
Light-exposed/near infrared camera is realized the imaging pattern of laser biology speckle by dot matrix laser and paliform laser.
3. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that described close
Infrared array light source, which carries out visible light/near infrared camera by change-over switch, is having optical mode and without the switching of optical mode, and by remote
Thermal infrared imager, ultrasonic distance-measuring sensor and infrared distance sensor carry out thermometric to object in the application scenarios of acquisition, survey
Away from test the speed.
4. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that the touching
Sense generator is the heat or electro photoluminescence generated under pain threshold.
5. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that the gas
Body injector is to spray the application scenarios information that the gas transmitting multi-modality imaging module of varying strength obtains.
6. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that the vibration
Dynamic generator is to generate to beat stimulation under pain threshold.
7. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that described
OLED display screen is that multi-modality images information is mapped to the full impregnated mode eyeglass in corresponding instant scene one by one, and is equipped with eye movement
Or the image information and its fusion state of mechanical switch switching different modalities are shown.
8. the augmented reality system of multi-modality imaging and more perception blendings according to claim 1, it is characterised in that the biography
Defeated module is that bluetooth, 4G or WIFI carry out the wireless transmission of each intermodule, or the data of microprocessor are transferred to cloud, and connect
Receive that treated in cloud data or decision information.
9. the augmented reality systematic difference of multi-modality imaging described in a kind of claim 1 and more perception blendings, it is characterised in that
Using the big visual angle augmented reality glasses of the augmented reality system architecture for intelligence law enforcement, auxiliary visually impaired and driving auxiliary, or
The auxiliary device that framework is applied as different scenes on walking stick, waistband, gloves, cap and clothes, the big visual angle enhancing are existing
Real glasses are that frame is arranged in multi-modality imaging module, Multilayer Perception module, auxiliary sensing module and controlling terminal
Comprehensive precisely reliable perception of the wearer to surrounding enviroment is realized in outer.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810954273.8A CN109085926A (en) | 2018-08-21 | 2018-08-21 | A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810954273.8A CN109085926A (en) | 2018-08-21 | 2018-08-21 | A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109085926A true CN109085926A (en) | 2018-12-25 |
Family
ID=64794082
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810954273.8A Pending CN109085926A (en) | 2018-08-21 | 2018-08-21 | A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109085926A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110246178A (en) * | 2019-07-09 | 2019-09-17 | 华东师范大学 | A kind of modularization intelligent auxiliary system visually impaired |
CN112402194A (en) * | 2019-08-23 | 2021-02-26 | 北京超维度计算科技有限公司 | Auxiliary terminal system for visually impaired people |
CN115079423A (en) * | 2022-08-19 | 2022-09-20 | 歌尔股份有限公司 | Intelligent glasses and ultrasonic tactile feedback method thereof |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2011106797A1 (en) * | 2010-02-28 | 2011-09-01 | Osterhout Group, Inc. | Projection triggering through an external marker in an augmented reality eyepiece |
CN104306102A (en) * | 2014-10-10 | 2015-01-28 | 上海交通大学 | Head wearing type vision auxiliary system for patients with vision disorder |
CN104473717A (en) * | 2014-12-04 | 2015-04-01 | 上海交通大学 | Wearable guide apparatus for totally blind people |
CN105094417A (en) * | 2014-05-19 | 2015-11-25 | 意美森公司 | Non-collocated haptic cues in immersive environments |
CN106580294A (en) * | 2016-12-30 | 2017-04-26 | 上海交通大学 | Physiological signal remote monitoring system based on multimodal imaging technique and application thereof |
CN106817577A (en) * | 2016-11-23 | 2017-06-09 | 杭州视氪科技有限公司 | One kind is based on RGB D cameras and stereosonic visually impaired people's barrier early warning glasses |
CN206920723U (en) * | 2017-07-24 | 2018-01-23 | 深圳市科乐科技有限公司 | A kind of augmented reality glasses |
CN107884942A (en) * | 2017-12-18 | 2018-04-06 | 国网浙江省电力公司培训中心 | A kind of augmented reality display device |
CN108319022A (en) * | 2018-03-22 | 2018-07-24 | 巢湖永微信息科技有限公司 | A kind of system of the intelligent glasses of the hot sensing function of band |
-
2018
- 2018-08-21 CN CN201810954273.8A patent/CN109085926A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2011106797A1 (en) * | 2010-02-28 | 2011-09-01 | Osterhout Group, Inc. | Projection triggering through an external marker in an augmented reality eyepiece |
CN105094417A (en) * | 2014-05-19 | 2015-11-25 | 意美森公司 | Non-collocated haptic cues in immersive environments |
CN104306102A (en) * | 2014-10-10 | 2015-01-28 | 上海交通大学 | Head wearing type vision auxiliary system for patients with vision disorder |
CN104473717A (en) * | 2014-12-04 | 2015-04-01 | 上海交通大学 | Wearable guide apparatus for totally blind people |
CN106817577A (en) * | 2016-11-23 | 2017-06-09 | 杭州视氪科技有限公司 | One kind is based on RGB D cameras and stereosonic visually impaired people's barrier early warning glasses |
CN106580294A (en) * | 2016-12-30 | 2017-04-26 | 上海交通大学 | Physiological signal remote monitoring system based on multimodal imaging technique and application thereof |
CN206920723U (en) * | 2017-07-24 | 2018-01-23 | 深圳市科乐科技有限公司 | A kind of augmented reality glasses |
CN107884942A (en) * | 2017-12-18 | 2018-04-06 | 国网浙江省电力公司培训中心 | A kind of augmented reality display device |
CN108319022A (en) * | 2018-03-22 | 2018-07-24 | 巢湖永微信息科技有限公司 | A kind of system of the intelligent glasses of the hot sensing function of band |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110246178A (en) * | 2019-07-09 | 2019-09-17 | 华东师范大学 | A kind of modularization intelligent auxiliary system visually impaired |
CN112402194A (en) * | 2019-08-23 | 2021-02-26 | 北京超维度计算科技有限公司 | Auxiliary terminal system for visually impaired people |
CN115079423A (en) * | 2022-08-19 | 2022-09-20 | 歌尔股份有限公司 | Intelligent glasses and ultrasonic tactile feedback method thereof |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3014338B1 (en) | Tracking head movement when wearing mobile device | |
US9760167B2 (en) | Visual stabilization system for head-mounted displays | |
US10656731B2 (en) | Peripheral device for head-mounted display | |
WO2016184107A1 (en) | Wearable apparatus for sight line focus positioning and method for sight line focus positioning | |
CN104956252B (en) | Peripheral display for near-eye display device | |
CN104603673B (en) | Head-mounted system and the method for being calculated using head-mounted system and rendering digital image stream | |
US10663729B2 (en) | Peripheral device for head-mounted display | |
WO2017115618A1 (en) | Information processing apparatus, information processing method, and program | |
CN109085926A (en) | A kind of the augmented reality system and its application of multi-modality imaging and more perception blendings | |
KR20180110051A (en) | Systems and methods for augmented reality | |
US20130241805A1 (en) | Using Convergence Angle to Select Among Different UI Elements | |
CN108919498A (en) | A kind of augmented reality glasses based on multi-modality imaging and Multilayer Perception | |
CN104808340B (en) | Head-mounted display device and control method thereof | |
CN109358754B (en) | Mixed reality head-mounted display system | |
JPH0749744A (en) | Head mounting type display input device | |
CN103731659B (en) | head-mounted display apparatus | |
CN103748598A (en) | Systems and methods for identifying gaze tracking scene reference locations | |
US11734898B2 (en) | Program, information processing method, and information processing terminal | |
KR101203921B1 (en) | Information providing apparatus using an eye tracking and local based service | |
WO2022095605A1 (en) | Handle control tracker | |
US20120282585A1 (en) | Interest-Attention Feedback System for Separating Cognitive Awareness into Different Left and Right Sensor Displays | |
CN108064447A (en) | Method for displaying image, intelligent glasses and storage medium | |
CN207220273U (en) | Visualize sea police's patrol helmet | |
WO2014128750A1 (en) | Input/output device, input/output program, and input/output method | |
CN109445596B (en) | Integrated mixed reality head-mounted display system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20181225 |