WO2015081694A1 - 智能眼镜及其控制方法 - Google Patents
智能眼镜及其控制方法 Download PDFInfo
- Publication number
- WO2015081694A1 WO2015081694A1 PCT/CN2014/081282 CN2014081282W WO2015081694A1 WO 2015081694 A1 WO2015081694 A1 WO 2015081694A1 CN 2014081282 W CN2014081282 W CN 2014081282W WO 2015081694 A1 WO2015081694 A1 WO 2015081694A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- signal
- display
- processor
- brain wave
- smart glasses
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 52
- 239000004984 smart glass Substances 0.000 title claims abstract description 52
- 210000004556 brain Anatomy 0.000 claims abstract description 73
- 238000004891 communication Methods 0.000 claims abstract description 18
- 238000012545 processing Methods 0.000 claims abstract description 10
- 238000006243 chemical reaction Methods 0.000 claims abstract description 6
- 230000005236 sound signal Effects 0.000 claims description 29
- 230000008569 process Effects 0.000 claims description 14
- 230000007613 environmental effect Effects 0.000 claims description 7
- 238000012790 confirmation Methods 0.000 claims description 5
- 239000011521 glass Substances 0.000 abstract description 9
- 208000032041 Hearing impaired Diseases 0.000 abstract description 5
- 230000006870 function Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 6
- 206010011878 Deafness Diseases 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 5
- 238000012546 transfer Methods 0.000 description 3
- 230000007547 defect Effects 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 210000003128 head Anatomy 0.000 description 1
- 208000016354 hearing loss disease Diseases 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B21/00—Teaching, or communicating with, the blind, deaf or mute
- G09B21/04—Devices for conversing with the deaf-blind
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B21/00—Teaching, or communicating with, the blind, deaf or mute
- G09B21/009—Teaching or communicating with deaf persons
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/015—Input arrangements based on nervous system activity detection, e.g. brain waves [EEG] detection, electromyograms [EMG] detection, electrodermal response detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B21/00—Teaching, or communicating with, the blind, deaf or mute
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0138—Head-up displays characterised by optical features comprising image capture systems, e.g. camera
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/06—Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
- G10L2021/065—Aids for the handicapped in understanding
Definitions
- Embodiments of the present invention relate to a smart eyewear and a control method thereof. Background technique
- Deaf-mute people do not have the ability to listen and/or speak because of their inherent physical defects, so they cannot know the thoughts of others, nor can they communicate and communicate with each other through language, which brings great inconvenience to daily life. Although most deaf-mute people can express what they want to say in sign language, they can't achieve effective communication and communication if they don't understand sign language. To overcome hearing defects, hearing impaired people can wear hearing aids. Although hearing aids help hearing impaired people to improve hearing impairment, they still have certain limitations. For example, there is no guarantee that all hearing impaired people will have the same hearing as ordinary people after wearing, so that some hearing impaired people still know what others are saying. There are difficulties. Summary of the invention
- a smart glasses including: a lens, a frame, and a temple; the lens includes a transparent display configured to perform double-sided display; the frame is provided with a camera and a pickup, respectively configured to Obtaining a gesture command and a voice signal and converting into a gesture signal and an audio signal; a brain wave identifier and a processor are disposed on the temple, the brain wave identifier is configured to acquire a brain wave signal of the wearer, and the processor is configured to receive the The gesture signal, the audio signal, and the brain wave signal are processed, and the processing result is sent to the transparent display for the two-sided display.
- the processor is configured to generate a processing result in the form of teletext information.
- the transparent display includes a display and two display surfaces disposed on the display, the two display surfaces being configured for front display and back display, respectively.
- the transparent display includes two displays, each display having a display surface, the two displays being configured for front display and back display, respectively.
- the transparent display is a flexible display.
- the smart glasses further include a parsing memory connected to the processor, and storing: a database including a correspondence relationship between the brain wave signals and the indication content thereof, a database including a correspondence relationship between the gesture signals and the indication contents thereof, and including A database of the correspondence of audio signals to their indicated content.
- the smart glasses further include a positioning system; the camera is further configured to acquire and transmit environmental information around the smart glasses to the processor; the processor is further configured to generate environmental information according to the camera And positioning the wearer in conjunction with the positioning system.
- the pickup is disposed at a nose pad position in the frame
- the electroencephalogram identifier is disposed in a middle portion of the temple
- the processor is disposed at a tail portion of the temple.
- the positioning system includes a memory pre-stored with location information.
- the positioning system includes a ranging device disposed on the temple, configured to sense a distance between a current position and a target position and to transmit a sensing result to the processor.
- the temple is also provided with a data transfer device configured to perform data transfer with the external device.
- the smart glasses further include a charging device configured to charge at least one of the transparent display, the camera, the pickup, the brain wave recognizer, and the processor.
- the charging device is a solar charging device integrated on the mirror surface of the lens.
- the ranging device is selected from the group consisting of: an ultrasonic range finder, an infrared range finder, and a laser range finder.
- control method for the above smart glasses comprising the following receiving method and expression method, and the receiving method and the expressing method can be performed independently, simultaneously or separately.
- Example: ⁇ , the connection ⁇ : method includes:
- Example: ⁇ the expression method includes:
- the step S1.2 includes: the processor receiving the received gesture signal and audio in a database storing a correspondence relationship between the gesture signal and the indication content thereof and a database storing the correspondence relationship between the audio signal and the indication content thereof.
- the signal is searched for matching, and the indication content is output in graphic form.
- the method may further include: prompting an error through the back display of the lens.
- the step S2.2 includes: the processor performing a search matching on the received first information in a database storing correspondence between the brain wave information and the indication content thereof, and outputting the indication content in a graphic form.
- the image information may be displayed on the back side, and the brain wave signal indicating the confirmation sent by the wearer is received, and then the graphic information is displayed on the front side.
- step S1.1 the audio signal is obtained by analog-to-digital conversion of the acquired voice signal by a pickup.
- step S1.1 the gesture signal and the audio signal are acquired, and the surrounding environment information is acquired by the camera, and the current position is located in combination with the positioning system.
- step S1.1 the environment information acquired by the camera is compared and calculated by the processor with location information stored in the positioning system.
- 1 is a schematic structural diagram of a smart glasses provided in an embodiment of the present invention
- 2 is a schematic diagram of a principle of identifying and judging a smart glasses in an expression mode according to an embodiment of the present invention
- FIG. 3 is a schematic diagram showing the principle of the matching judgment process of the smart glasses provided in the expression mode according to the embodiment of the present invention.
- FIG. 4 is a schematic diagram of the principle of the smart glasses provided in the receiving mode according to the embodiment of the present invention
- FIG. 5 is a flow chart of the steps of the method for controlling the smart glasses provided in the embodiment of the present invention. detailed description
- composition of the smart glasses according to an embodiment of the present invention is as shown in FIG. 1, which includes: a lens, a frame and a temple.
- the lens includes a transparent display 10 configured to perform double-sided display; the frame is provided with a camera 11 and a pickup 12, respectively configured to acquire gesture commands and voice signals, and if desired, convert them into gesture signals and audio signals ; the brain wave identifier 13 and the processor are disposed on the temple 14.
- the brain wave identifier 13 is configured to acquire a brainwave signal of the wearer, and the processor 14 is configured to receive the gesture signal, the audio signal, and the brain wave signal, perform processing, and transmit the processing result to the transparent display. Used for the double-sided display.
- the processor 14 is built into the tail of the temple on one side.
- a module having other functions may be built in the tail of the other side of the temple, for example.
- the transparent display 10 is configured to display the image information processed by the processor 14 on both sides.
- the transparent display 10 includes a display and two display surfaces disposed on the display, the two display surfaces being used for front display and back display, respectively.
- the transparent display 10 includes two displays, each having a display surface, so that there are also two display faces in total for the front display and the rear display, respectively.
- the front display and the back display in the embodiment of the present invention refer to two display modes.
- the display is transparent, during the display process, the opposite display can be observed on both sides of the lens regardless of the display mode.
- the second example above differs from the first example in that the number of displays used is different, as described below.
- each lens panel uses a display having two display surfaces, the display itself being transparent, and the two display surfaces of the transparent display correspond to the A and B sides of the lens, respectively, for back display and The two modes are displayed on the front side;
- each of the lenses is provided with two displays arranged opposite each other, that is, each lens is composed of two displays, and the two displays respectively correspond to the A of the lens
- the face and the B face are also used for the back display and the front display, respectively.
- the front display mode and the back display mode are respectively an expression mode and a reception mode: the expression mode is to display the wearer's thoughts or consciousness to the ordinary person in front of the glasses through the lens of the glasses, and the receiving mode is to obtain The ordinary people's voices, gestures and other information in front of them are finally presented to the wearer through the lenses.
- the front display of the smart glasses is used in the expression mode, and the so-called front side is along the direction of the wearer's line of sight, that is, the side of the ordinary person (assuming the A side).
- the receiving mode the rear side of the smart glasses is displayed.
- the so-called back is the opposite direction to the wearer's line of sight, that is, the side that is displayed to the wearer of the glasses (assuming the B side).
- the lens of the spectacles is spliced with a flexible display device to achieve a curved structural design of the lens.
- the display area on the frame is curved to extend to both sides of the user's head. The function is that when the lens of the glasses presents the display pattern, not only the user's front will appear a prompt screen, and the left and right sides thereof There is also a display area on the side and the graphic information is presented.
- the teletext information is information of one or more groups selected from the group consisting of: text information, picture information, and a combination of text information and picture information.
- the camera 11 is disposed on the frame for acquiring gesture commands, generating a gesture signal, and transmitting the gesture signal to the processor 14.
- the camera 11 is located in front of the smart glasses, on both sides of the left and right glasses, and performs an all-round detection around the wearer.
- the camera 11 is used for acquiring the gestures of other people in front of the wearer to obtain the idea and purpose thereof, and is also used for real-time collecting and detecting the environment and conditions surrounding the user.
- the information is sent to the processor 14 for processing, and is reasoned and calculated according to the internal database, and the position of the wearer is accurately positioned according to the result, in combination with, for example, a GPS positioning system.
- the positioning system of the smart glasses includes a GPS locator and a memory pre-stored with environmental information, which are respectively connected to the processor 14; at this time, the camera 11 is also configured to the surrounding environment of the smart glasses. Monitoring, obtaining surrounding environment information and transmitting it to the processor 14; the processor 14 compares and calculates the environmental information with the environmental information pre-stored in the memory database, and locates the current location to obtain positioning information.
- the positioning information obtained in this way is advantageous for other intelligent terminals to determine the position of the wearer according to the position information, thereby providing the wearer with the best route of the nearby place, for example, providing the wearer with the nearest subway station and from the current position to the The nearest route of the subway station, etc.; for example, other help information can be obtained according to the location information, the principle is the same as above, and details are not described herein again.
- the positioning system further includes a distance measuring device that senses a distance between the current position and the target position.
- the ranging device is built into the frame and functions to measure the distance of the wearer from the road sign at the location in real time to achieve accurate positioning.
- the distance measuring device measures the distance between the wearer and the communication object thereof, thereby setting the finger of the pickup 12 The distance to and from the sound.
- the ranging device is selected from the group consisting of: an ultrasonic range finder, an infrared range finder, and a laser range finder.
- the pickup 12 is disposed at the nose pad position of the frame, the brain wave identifier 13 is disposed at the center of the temple, and the processor 14 is disposed at the tail of the temple.
- the pickup 12 picks up a language signal (for example, an analog signal) in a specific range, converts it into a digital signal, and transmits it to the processor 14, and the processor 14 converts the signal into a graphic by voice recognition. The information is then transmitted to the transparent display 10 and then displayed through the back lens of the glasses.
- a language signal for example, an analog signal
- the electroencephalogram recognizer 13 is disposed in the middle of the temple and is very close to the wearer's brain when worn.
- the human brain generates an electroencephalogram signal when generating consciousness or an idea, and the electroencephalogram recognizer 13 is used to recognize the brain wave signal and read the brain wave information (that is, the operation instruction) when the wearer thinks.
- the first information is obtained by the decoding and encoding process and sent to the processor 14, and then the first information is parsed and processed by the processor 14, and displayed to the wearer's communication object on the A side in the form of the graphic information.
- the smart glasses include a parsing memory containing graphics storing brain wave information (first information) and a consciousness or idea (ie, content indicated by the signal) characterizing the signal transmitter A database of correspondences of information.
- the processor determines the indication content corresponding to the brain wave signal acquired at this time based on the received first information and performs a search comparison in the database, and displays the content.
- a data transmission device and a charging device are further disposed on the temple; the data transmission device is used for data transmission with the external device; the charging device is opposite to the transparent display 10, the camera 11, the pickup 12, and the brain. At least one of the electric wave identifier 13 and the processor 14 is charged to increase the endurance of the smart glasses.
- the charging device is a solar powered charging device integrated on the surface on either side of the lens.
- the data transfer device is built into the tail of the other side leg of the processor 14 that is not provided.
- the communication function of the smart glasses provided by the embodiment of the present invention is implemented by the data transmission device, that is, the smart glasses wearer provided in this embodiment is contacted by an RF (a radio frequency) system.
- RF radio frequency
- voice messages from other smart terminal users pass The corresponding information of the processing unit and the artificial intelligence system is processed and transmitted to the user through the display screen in the B-side display mode; the user's response is transmitted, recognized, and read by the brain wave, converted into graphic information, and displayed in the B-side display mode.
- the processor 14 receives the confirmation information of the wearer and transmits it to the communication object in the A-side display mode, or converts it into a voice signal by, for example, a processor, and transmits the data through the data.
- the device is sent out.
- the communication function is implemented by WIFI or Bluetooth or the like. Therefore, in one example, the smart glasses also have an entertainment function design, and the wearer can use brain waves to play games through the lens, use WIFI to access the Internet or transmit data with other devices.
- the working principle of the smart glasses in the expression mode is as follows.
- the wearer's brain sends a mind command to generate an electroencephalogram signal; the brain wave recognizer 13 recognizes the brain wave signal; if the brain wave signal fails to be recognized, the "recognition failure" is displayed on the B side of the transparent display 10, and the wearer passes the B side.
- the brainwave signal will be generated again after the feedback of "identification failure” is displayed.
- This process is a process of identifying and judging, but only feedbacks the error of reading information by the brain wave identifier.
- the schematic diagram is shown in Figure 2.
- the identified signal is compared with the brain wave database in the memory, for example, the identified signal is matched with the information code in the brain wave database, according to The matching degree judges whether the comparison passes; if the comparison passes, it indicates that the brain wave reading is successful, and the "first information" is obtained by the decoding and encoding process; the first information is sent to the processor 14 to continue parsing the information and The processing performs search matching based on, for example, a database including a correspondence relationship between the brain wave signal and the graphic information stored in the analysis memory, and outputs the matched graphic information, that is, displays through the A side of the transparent display 10.
- the graphic information obtained by matching the brain wave signal and the database may be displayed on the A side to the wearer for judgment;
- the graphic information conforms to the wearer's idea, and the wearer generates a confirmed brain wave signal.
- the processor 14 controls the transparent display to switch to the A-side display mode, and displays the matched graphic information to the wearer. If the graphic information displayed on the B side does not conform to the wearer's intention, the wearer's brain re-issues a new brain wave signal, and the brain wave recognizer receives the brain wave signal from the wearer's brain and repeats with the processor. The previous operation, until it meets the wearer's mind.
- the process is a matching judgment process, that is, whether the information output to the processor is desired by the wearer
- the expression of the idea is consistently judged, and the schematic diagram is shown in Figure 3.
- the graphic information display is performed on the A side of the glasses; however, if the processed information of the processor is different from the idea that the wearer wants to express , the wearer resends the mind command. Thereby, the wearer's idea can be expressed more accurately, and effective and accurate communication can be realized.
- the working principle of the smart glasses in the receiving mode in an embodiment of the present invention is as shown in FIG.
- the camera 11 and the pickup 12 respectively acquire gesture commands and voice signals, respectively convert the gesture commands and the voice signals to obtain gesture signals and audio signals, and then send the signals to the processor 14 for processing; the processor 14 stores the gesture signals stored in the analysis memory. Performing a search match in a database corresponding to the correspondence between the audio signal and the graphic information, determining an idea or intention expressed by the ordinary person of the gesture instruction and the voice signal to be expressed by the smart glasses wearer, and The form of the graphic information is displayed to the ordinary person in front of the wearer through the B side of the transparent display 10, thereby realizing the process of the wearer receiving the external information.
- the pickup 12 converts the speech signal into an audio signal by analog to digital conversion; and the camera 11 converts the gesture instruction into a gesture signal by a decoder and an encoder within the processor.
- the smart glasses provided by at least one embodiment of the present invention facilitate communication and communication between the deaf and the ordinary person wearing the smart mirror.
- the camera and the pickup obtain the gesture commands and voice signals of the ordinary person for the wearer, and after being recognized and processed by the processor, are displayed to the wearer through the back lens in the form of, for example, graphic information, so that the wearer knows the thoughts and intentions of the ordinary person;
- the brain wave identifier acquires the brain wave signal of the wearer, and after being parsed and processed by the processor, is displayed to the ordinary person through the front lens in the form of, for example, corresponding graphic information, so that the ordinary person knows the idea and intention of the wearer. This overcomes the obstacles that cannot be effectively communicated and communicated between deaf and ordinary people.
- the embodiment of the present invention further provides a control method based on any one of the above embodiments, comprising the steps S1-S2 for controlling the smart glasses to receive display and the steps S3 and S4 for controlling the smart glasses to display and display:
- the receive display control step is:
- S2 The audio signal and the gesture signal are respectively recognized by the processor to be converted into graphic information, and displayed on the back side through the lens.
- the expression display control step is:
- S4 Convert the first information into graphic information by using a processor, and perform front display through the lens. It should be noted that, in the receiving control, S1 and S2 are executed sequentially; in the expression control, S3 and S4 are executed sequentially; however, the two sets of steps can be executed simultaneously, sequentially, or separately.
- the smart glasses control method is as shown in Fig. 5, wherein the reception control is performed first, and then the expression control is executed (i.e., the steps of S1-S2-S3-S4 are performed).
- the expression control is executed (i.e., the steps of S1-S2-S3-S4 are performed).
- there may be only receiving control executing in steps S1-S2) or only expression control (executing in steps S3-S4) or performing expression control before performing reception control ( That is, it is performed in the order of S3-S4-S1-S2).
- step S1 includes: searching, by the processor, the received gesture signal and the audio signal in a database storing a correspondence relationship between the gesture signal and the indication content thereof and a database storing the correspondence relationship between the audio signal and the indication content thereof. Matches, the indication content is output in graphic form.
- the method further includes: prompting an error through the back display of the lens.
- step S4 includes: performing a search match on the received first information in a database storing correspondence between brain wave information and its indication content by the processor, and outputting the indication content in a graphic form.
- the method before the front lens display of the smart glasses in step S4, the method further includes: displaying the graphic information on the back lens, receiving the brain wave signal indicating the confirmation issued by the wearer, and then transmitting the graphic information through the lens. Perform a front display.
- step S1 the audio signal is obtained by analog-to-digital conversion of the acquired speech signal by the pickup.
- the gesture signal and the audio signal are acquired, and the surrounding environment information is acquired by the camera, and the current position is located in combination with the positioning system.
- step S1 the environment information acquired by the camera is compared and calculated by the processor with location information stored in the positioning system.
- the recognition, reading and matching of the gesture signal, the voice signal and the brain wave signal, the generation and output of the graphic information, the front and back display thereof, and the like The implementation of other auxiliary functions such as the positioning of the person and the data transmission of the external system can be referred to the related function description of the aforementioned smart glasses.
- the control method by wearing the smart glasses, people can communicate and communicate with ordinary people around them, mainly through a camera and a device having voice recognition and brain wave recognition or
- the module realizes the collection and conversion of gesture commands, voice signals and brain wave signals, and the obtained signals are processed by the processor and displayed in the form of graphic information. Since the smart eyes are transparent displays that can be displayed on both sides, it can be realized. Effective communication and communication between deaf and ordinary people, and improve the accuracy of existing methods such as sign language expression.
Abstract
Description
Claims
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/417,440 US20150379896A1 (en) | 2013-12-05 | 2014-06-30 | Intelligent eyewear and control method thereof |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201310652206.8A CN103646587B (zh) | 2013-12-05 | 2013-12-05 | 一种智能眼镜及其控制方法 |
CN201310652206.8 | 2013-12-05 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015081694A1 true WO2015081694A1 (zh) | 2015-06-11 |
Family
ID=50251793
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2014/081282 WO2015081694A1 (zh) | 2013-12-05 | 2014-06-30 | 智能眼镜及其控制方法 |
Country Status (3)
Country | Link |
---|---|
US (1) | US20150379896A1 (zh) |
CN (1) | CN103646587B (zh) |
WO (1) | WO2015081694A1 (zh) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106994689A (zh) * | 2016-01-23 | 2017-08-01 | 鸿富锦精密工业(武汉)有限公司 | 基于脑电信号控制的智能机器人系统和方法 |
CN105472256B (zh) * | 2016-01-05 | 2018-09-28 | 上海斐讯数据通信技术有限公司 | 拍摄和传输图像的方法、智能眼镜及系统 |
US11125998B2 (en) * | 2014-01-02 | 2021-09-21 | Nokia Technologies Oy | Apparatus or method for projecting light internally towards and away from an eye of a user |
CN114822172A (zh) * | 2022-06-23 | 2022-07-29 | 北京亮亮视野科技有限公司 | 基于ar眼镜的文字显示方法及装置 |
Families Citing this family (59)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103646587B (zh) * | 2013-12-05 | 2017-02-22 | 北京京东方光电科技有限公司 | 一种智能眼镜及其控制方法 |
CN104951259B (zh) * | 2014-03-28 | 2019-10-18 | 索尼公司 | 显示设备及其显示控制方法 |
CN104065388A (zh) * | 2014-07-09 | 2014-09-24 | 李永飞 | 人脑电台 |
CN104375641B (zh) * | 2014-10-27 | 2017-12-26 | 联想(北京)有限公司 | 一种控制方法及电子设备 |
TW201624469A (zh) * | 2014-12-26 | 2016-07-01 | Univ Chienkuo Technology | 聽障人士之電子智能溝通眼鏡 |
KR102311741B1 (ko) * | 2015-01-14 | 2021-10-12 | 삼성디스플레이 주식회사 | 머리 장착형 디스플레이 장치 |
CN106302974B (zh) * | 2015-06-12 | 2020-01-31 | 联想(北京)有限公司 | 一种信息处理的方法及电子设备 |
CN104966433A (zh) * | 2015-07-17 | 2015-10-07 | 江西洪都航空工业集团有限责任公司 | 一种辅助聋哑人对话的智能眼镜 |
DE102015214350A1 (de) * | 2015-07-29 | 2017-02-02 | Siemens Healthcare Gmbh | Verfahren zu einer Kommunikation zwischen einem medizinischen Netzwerk und einem medizinischen Bedienpersonal mittels einer mobilen Datenbrille, sowie eine mobile Datenbrille |
CN105137601B (zh) * | 2015-10-16 | 2017-11-14 | 上海斐讯数据通信技术有限公司 | 一种智能眼镜 |
CN105468140A (zh) * | 2015-11-05 | 2016-04-06 | 京东方科技集团股份有限公司 | 佩戴设备、应用设备系统 |
KR102450803B1 (ko) * | 2016-02-11 | 2022-10-05 | 한국전자통신연구원 | 양방향 수화 번역 장치 및 장치가 수행하는 양방향 수화 번역 방법 |
CN106157750A (zh) * | 2016-08-24 | 2016-11-23 | 深圳市铁格龙科技有限公司 | 一种智能聋哑人发音及交流学习眼镜 |
CN106205293A (zh) * | 2016-09-30 | 2016-12-07 | 广州音书科技有限公司 | 用于语音识别和手语识别的智能眼镜 |
CN106656352B (zh) * | 2016-12-27 | 2020-04-07 | 广东小天才科技有限公司 | 一种信息传递方法及装置、可穿戴设备 |
CN106601075A (zh) * | 2017-02-05 | 2017-04-26 | 苏州路之遥科技股份有限公司 | 脑电波输入训练器 |
US10854110B2 (en) | 2017-03-03 | 2020-12-01 | Microsoft Technology Licensing, Llc | Automated real time interpreter service |
US11861255B1 (en) | 2017-06-16 | 2024-01-02 | Apple Inc. | Wearable device for facilitating enhanced interaction |
CN109425983A (zh) * | 2017-08-27 | 2019-03-05 | 南京乐朋电子科技有限公司 | 一种脑电波成像投影眼镜 |
CN108106665A (zh) * | 2017-12-12 | 2018-06-01 | 深圳分云智能科技有限公司 | 一种具有玻璃监测功能的智能穿戴设备 |
US11435583B1 (en) * | 2018-01-17 | 2022-09-06 | Apple Inc. | Electronic device with back-to-back displays |
CN108198552B (zh) * | 2018-01-18 | 2021-02-02 | 深圳市大疆创新科技有限公司 | 一种语音控制方法及视频眼镜 |
CN110111651A (zh) * | 2018-02-01 | 2019-08-09 | 周玮 | 基于体态感知的智能语言交互系统 |
CN108509034B (zh) * | 2018-03-16 | 2021-05-11 | Oppo广东移动通信有限公司 | 电子装置、信息处理方法及相关产品 |
CN111954290B (zh) * | 2018-03-30 | 2023-04-18 | Oppo广东移动通信有限公司 | 电子装置、功率调整方法及相关产品 |
CN108711425A (zh) * | 2018-05-03 | 2018-10-26 | 华南理工大学 | 一种基于语音控制的视频输入听觉显示导盲装置及方法 |
CN108803871A (zh) * | 2018-05-07 | 2018-11-13 | 歌尔科技有限公司 | 头戴显示设备中数据内容的输出方法、装置及头戴显示设备 |
CN110058413A (zh) * | 2018-05-23 | 2019-07-26 | 王小峰 | 一种智能穿戴系统 |
US10908419B2 (en) | 2018-06-28 | 2021-02-02 | Lucyd Ltd. | Smartglasses and methods and systems for using artificial intelligence to control mobile devices used for displaying and presenting tasks and applications and enhancing presentation and display of augmented reality information |
CN109255314B (zh) * | 2018-08-30 | 2021-07-02 | Oppo广东移动通信有限公司 | 信息提示方法、装置、智能眼镜及存储介质 |
JP7283652B2 (ja) * | 2018-10-04 | 2023-05-30 | シーイヤー株式会社 | 聴覚サポートデバイス |
IT201800009607A1 (it) * | 2018-10-19 | 2020-04-19 | Andrea Previato | Sistema e metodo di ausilio ad utenti con disabilità comunicativa |
USD899500S1 (en) | 2019-03-22 | 2020-10-20 | Lucyd Ltd. | Smart glasses |
USD900920S1 (en) | 2019-03-22 | 2020-11-03 | Lucyd Ltd. | Smart glasses |
USD899493S1 (en) | 2019-03-22 | 2020-10-20 | Lucyd Ltd. | Smart glasses |
USD900206S1 (en) | 2019-03-22 | 2020-10-27 | Lucyd Ltd. | Smart glasses |
USD899499S1 (en) | 2019-03-22 | 2020-10-20 | Lucyd Ltd. | Smart glasses |
USD899498S1 (en) | 2019-03-22 | 2020-10-20 | Lucyd Ltd. | Smart glasses |
USD900205S1 (en) | 2019-03-22 | 2020-10-27 | Lucyd Ltd. | Smart glasses |
USD899497S1 (en) | 2019-03-22 | 2020-10-20 | Lucyd Ltd. | Smart glasses |
USD899494S1 (en) | 2019-03-22 | 2020-10-20 | Lucyd Ltd. | Smart glasses |
USD899496S1 (en) | 2019-03-22 | 2020-10-20 | Lucyd Ltd. | Smart glasses |
USD899495S1 (en) | 2019-03-22 | 2020-10-20 | Lucyd Ltd. | Smart glasses |
USD900204S1 (en) | 2019-03-22 | 2020-10-27 | Lucyd Ltd. | Smart glasses |
USD900203S1 (en) | 2019-03-22 | 2020-10-27 | Lucyd Ltd. | Smart glasses |
CN110351631A (zh) * | 2019-07-11 | 2019-10-18 | 京东方科技集团股份有限公司 | 聋哑人交流设备及其使用方法 |
CN112506335B (zh) * | 2019-09-16 | 2022-07-12 | Oppo广东移动通信有限公司 | 头戴式设备及其控制方法、装置、系统和存储介质 |
USD958234S1 (en) | 2019-12-12 | 2022-07-19 | Lucyd Ltd. | Round smartglasses having pivot connector hinges |
USD954135S1 (en) | 2019-12-12 | 2022-06-07 | Lucyd Ltd. | Round smartglasses having flat connector hinges |
USD955467S1 (en) | 2019-12-12 | 2022-06-21 | Lucyd Ltd. | Sport smartglasses having flat connector hinges |
USD954136S1 (en) | 2019-12-12 | 2022-06-07 | Lucyd Ltd. | Smartglasses having pivot connector hinges |
USD954137S1 (en) | 2019-12-19 | 2022-06-07 | Lucyd Ltd. | Flat connector hinges for smartglasses temples |
USD974456S1 (en) | 2019-12-19 | 2023-01-03 | Lucyd Ltd. | Pivot hinges and smartglasses temples |
CN111046854B (zh) * | 2020-01-10 | 2024-01-26 | 北京服装学院 | 一种脑电波外部识别方法、装置及系统 |
CN111258088A (zh) * | 2020-02-25 | 2020-06-09 | 厦门明睐科技有限公司 | 一种脑电波控制的智能眼镜设备及使用方法 |
US11282523B2 (en) * | 2020-03-25 | 2022-03-22 | Lucyd Ltd | Voice assistant management |
CN111751995A (zh) * | 2020-06-11 | 2020-10-09 | 重庆工业职业技术学院 | 一种声音视觉化的单目头戴式ar眼镜装置及其实现方法 |
CN111787264B (zh) * | 2020-07-21 | 2021-08-10 | 北京字节跳动网络技术有限公司 | 一种远程教学的提问方法、装置、提问终端和可读介质 |
CN115695620A (zh) * | 2021-07-22 | 2023-02-03 | 所乐思(深圳)科技有限公司 | 智能眼镜及其控制方法和系统 |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090128449A1 (en) * | 2007-11-15 | 2009-05-21 | International Business Machines Corporation | Augmenting Reality For A User |
CN101819334A (zh) * | 2010-04-01 | 2010-09-01 | 夏翔 | 多功能电子眼镜 |
CN202533867U (zh) * | 2012-04-17 | 2012-11-14 | 北京七鑫易维信息技术有限公司 | 一种头戴式眼控显示终端 |
US20120299950A1 (en) * | 2011-05-26 | 2012-11-29 | Nokia Corporation | Method and apparatus for providing input through an apparatus configured to provide for display of an image |
CN103211655A (zh) * | 2013-04-11 | 2013-07-24 | 深圳先进技术研究院 | 一种骨科手术导航系统及导航方法 |
CN103279232A (zh) * | 2012-06-29 | 2013-09-04 | 上海天马微电子有限公司 | 一种橱窗互动装置及其互动实施方法 |
CN103310683A (zh) * | 2013-05-06 | 2013-09-18 | 深圳先进技术研究院 | 智能眼镜及基于智能眼镜的语音交流系统及方法 |
CN103336579A (zh) * | 2013-07-05 | 2013-10-02 | 百度在线网络技术(北京)有限公司 | 穿戴式设备的输入方法和穿戴式设备 |
CN103646587A (zh) * | 2013-12-05 | 2014-03-19 | 北京京东方光电科技有限公司 | 一种智能眼镜及其控制方法 |
Family Cites Families (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4902120A (en) * | 1988-11-22 | 1990-02-20 | Weyer Frank M | Eyeglass headphones |
JP3289304B2 (ja) * | 1992-03-10 | 2002-06-04 | 株式会社日立製作所 | 手話変換装置および方法 |
US5610678A (en) * | 1993-12-30 | 1997-03-11 | Canon Kabushiki Kaisha | Camera including camera body and independent optical viewfinder |
US6433913B1 (en) * | 1996-03-15 | 2002-08-13 | Gentex Corporation | Electro-optic device incorporating a discrete photovoltaic device and method and apparatus for making same |
US6240392B1 (en) * | 1996-08-29 | 2001-05-29 | Hanan Butnaru | Communication device and method for deaf and mute persons |
EP1027627B1 (en) * | 1997-10-30 | 2009-02-11 | MYVU Corporation | Eyeglass interface system |
US6491394B1 (en) * | 1999-07-02 | 2002-12-10 | E-Vision, Llc | Method for refracting and dispensing electro-active spectacles |
US6510417B1 (en) * | 2000-03-21 | 2003-01-21 | America Online, Inc. | System and method for voice access to internet-based information |
US20020158816A1 (en) * | 2001-04-30 | 2002-10-31 | Snider Gregory S. | Translating eyeglasses |
US7023498B2 (en) * | 2001-11-19 | 2006-04-04 | Matsushita Electric Industrial Co. Ltd. | Remote-controlled apparatus, a remote control system, and a remote-controlled image-processing apparatus |
JP2004199027A (ja) * | 2002-10-24 | 2004-07-15 | Seiko Epson Corp | 表示装置、及び電子機器 |
JP2004144990A (ja) * | 2002-10-24 | 2004-05-20 | Alps Electric Co Ltd | 両面発光型液晶表示モジュール |
US7546158B2 (en) * | 2003-06-05 | 2009-06-09 | The Regents Of The University Of California | Communication methods based on brain computer interfaces |
US7120486B2 (en) * | 2003-12-12 | 2006-10-10 | Washington University | Brain computer interface |
US8965460B1 (en) * | 2004-01-30 | 2015-02-24 | Ip Holdings, Inc. | Image and augmented reality based networks using mobile devices and intelligent electronic glasses |
KR100594117B1 (ko) * | 2004-09-20 | 2006-06-28 | 삼성전자주식회사 | Hmd 정보 단말기에서 생체 신호를 이용하여 키를입력하는 장치 및 방법 |
US20060094974A1 (en) * | 2004-11-02 | 2006-05-04 | Cain Robert C | Systems and methods for detecting brain waves |
US8696113B2 (en) * | 2005-10-07 | 2014-04-15 | Percept Technologies Inc. | Enhanced optical and perceptual digital eyewear |
US11428937B2 (en) * | 2005-10-07 | 2022-08-30 | Percept Technologies | Enhanced optical and perceptual digital eyewear |
US20080144854A1 (en) * | 2006-12-13 | 2008-06-19 | Marcio Marc Abreu | Biologically fit wearable electronics apparatus and methods |
KR100866215B1 (ko) * | 2006-12-20 | 2008-10-30 | 삼성전자주식회사 | 뇌파를 이용한 단말의 구동 방법 및 장치 |
WO2010004698A1 (ja) * | 2008-07-11 | 2010-01-14 | パナソニック株式会社 | 脳波を用いた機器の制御方法および脳波インタフェースシステム |
CN100595635C (zh) * | 2009-01-14 | 2010-03-24 | 长春大学 | 智能盲人导航眼镜 |
US8964298B2 (en) * | 2010-02-28 | 2015-02-24 | Microsoft Corporation | Video display modification based on sensor input for a see-through near-to-eye display |
CN102236986A (zh) * | 2010-05-06 | 2011-11-09 | 鸿富锦精密工业(深圳)有限公司 | 手语翻译系统、手语翻译装置及手语翻译方法 |
US9994228B2 (en) * | 2010-05-14 | 2018-06-12 | Iarmourholdings, Inc. | Systems and methods for controlling a vehicle or device in response to a measured human response to a provocative environment |
US20110291918A1 (en) * | 2010-06-01 | 2011-12-01 | Raytheon Company | Enhancing Vision Using An Array Of Sensor Modules |
US8442626B2 (en) * | 2010-06-21 | 2013-05-14 | Aleksandrs Zavoronkovs | Systems and methods for communicating with a computer using brain activity patterns |
US20120078628A1 (en) * | 2010-09-28 | 2012-03-29 | Ghulman Mahmoud M | Head-mounted text display system and method for the hearing impaired |
GB201103200D0 (en) * | 2011-02-24 | 2011-04-13 | Isis Innovation | An optical device for the visually impaired |
US8593795B1 (en) * | 2011-08-09 | 2013-11-26 | Google Inc. | Weight distribution for wearable computing device |
KR20130045471A (ko) * | 2011-10-26 | 2013-05-06 | 삼성전자주식회사 | 전자장치 및 그 제어방법 |
US20170164878A1 (en) * | 2012-06-14 | 2017-06-15 | Medibotics Llc | Wearable Technology for Non-Invasive Glucose Monitoring |
TWI467539B (zh) * | 2012-07-20 | 2015-01-01 | Au Optronics Corp | 影像顯示的控制方法以及顯示系統 |
WO2014041871A1 (ja) * | 2012-09-12 | 2014-03-20 | ソニー株式会社 | 画像表示装置及び画像表示方法、並びに記録媒体 |
US9966075B2 (en) * | 2012-09-18 | 2018-05-08 | Qualcomm Incorporated | Leveraging head mounted displays to enable person-to-person interactions |
US10073201B2 (en) * | 2012-10-26 | 2018-09-11 | Qualcomm Incorporated | See through near-eye display |
JP6094190B2 (ja) * | 2012-12-10 | 2017-03-15 | ソニー株式会社 | 情報処理装置および記録媒体 |
US9240162B2 (en) * | 2012-12-31 | 2016-01-19 | Lg Display Co., Ltd. | Transparent display apparatus and method for controlling the same |
EP2972678A4 (en) * | 2013-03-15 | 2016-11-02 | Interaxon Inc | CLOTHING COMPUTER APPARATUS AND ASSOCIATED METHOD |
US9280972B2 (en) * | 2013-05-10 | 2016-03-08 | Microsoft Technology Licensing, Llc | Speech to text conversion |
CN106537233A (zh) * | 2014-04-22 | 2017-03-22 | 伊瓦·阿尔布佐夫 | 用于头戴式智能设备的热成像配件 |
CN105607253B (zh) * | 2014-11-17 | 2020-05-12 | 精工爱普生株式会社 | 头部佩戴型显示装置以及控制方法、显示系统 |
US9672760B1 (en) * | 2016-01-06 | 2017-06-06 | International Business Machines Corporation | Personalized EEG-based encryptor |
-
2013
- 2013-12-05 CN CN201310652206.8A patent/CN103646587B/zh active Active
-
2014
- 2014-06-30 WO PCT/CN2014/081282 patent/WO2015081694A1/zh active Application Filing
- 2014-06-30 US US14/417,440 patent/US20150379896A1/en not_active Abandoned
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090128449A1 (en) * | 2007-11-15 | 2009-05-21 | International Business Machines Corporation | Augmenting Reality For A User |
CN101819334A (zh) * | 2010-04-01 | 2010-09-01 | 夏翔 | 多功能电子眼镜 |
US20120299950A1 (en) * | 2011-05-26 | 2012-11-29 | Nokia Corporation | Method and apparatus for providing input through an apparatus configured to provide for display of an image |
CN202533867U (zh) * | 2012-04-17 | 2012-11-14 | 北京七鑫易维信息技术有限公司 | 一种头戴式眼控显示终端 |
CN103279232A (zh) * | 2012-06-29 | 2013-09-04 | 上海天马微电子有限公司 | 一种橱窗互动装置及其互动实施方法 |
CN103211655A (zh) * | 2013-04-11 | 2013-07-24 | 深圳先进技术研究院 | 一种骨科手术导航系统及导航方法 |
CN103310683A (zh) * | 2013-05-06 | 2013-09-18 | 深圳先进技术研究院 | 智能眼镜及基于智能眼镜的语音交流系统及方法 |
CN103336579A (zh) * | 2013-07-05 | 2013-10-02 | 百度在线网络技术(北京)有限公司 | 穿戴式设备的输入方法和穿戴式设备 |
CN103646587A (zh) * | 2013-12-05 | 2014-03-19 | 北京京东方光电科技有限公司 | 一种智能眼镜及其控制方法 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11125998B2 (en) * | 2014-01-02 | 2021-09-21 | Nokia Technologies Oy | Apparatus or method for projecting light internally towards and away from an eye of a user |
CN105472256B (zh) * | 2016-01-05 | 2018-09-28 | 上海斐讯数据通信技术有限公司 | 拍摄和传输图像的方法、智能眼镜及系统 |
CN106994689A (zh) * | 2016-01-23 | 2017-08-01 | 鸿富锦精密工业(武汉)有限公司 | 基于脑电信号控制的智能机器人系统和方法 |
CN114822172A (zh) * | 2022-06-23 | 2022-07-29 | 北京亮亮视野科技有限公司 | 基于ar眼镜的文字显示方法及装置 |
Also Published As
Publication number | Publication date |
---|---|
US20150379896A1 (en) | 2015-12-31 |
CN103646587A (zh) | 2014-03-19 |
CN103646587B (zh) | 2017-02-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2015081694A1 (zh) | 智能眼镜及其控制方法 | |
US9101459B2 (en) | Apparatus and method for hierarchical object identification using a camera on glasses | |
CN104983511A (zh) | 针对全盲视觉障碍者的语音帮助智能眼镜系统 | |
US10062302B2 (en) | Vision-assist systems for orientation and mobility training | |
KR20150144510A (ko) | 시각 장애인용 안경 시스템 | |
KR101835235B1 (ko) | 시각 장애인 보조장치 및 그 제어방법 | |
KR101684264B1 (ko) | 글라스형 웨어러블 디바이스의 버스도착 알림방법 및 이를 이용한 글라스형 웨어러블 디바이스용 프로그램 | |
CN109696748A (zh) | 一种用于同步翻译的增强现实字幕眼镜 | |
TWI652656B (zh) | 視覺輔助系統及具有該視覺輔助系統的可穿戴裝置 | |
KR101728707B1 (ko) | 글라스형 웨어러블 디바이스를 이용한 실내 전자기기 제어방법 및 제어프로그램 | |
CN112002186B (zh) | 一种基于增强现实技术的信息无障碍系统及方法 | |
KR101982848B1 (ko) | 분리형 웨어러블 디바이스 및 제어장치로 구성된 시력 취약계층용 시력 보조장치 | |
US10943117B2 (en) | Translation to braille | |
CN105824137A (zh) | 可视化智能眼镜 | |
KR20160024140A (ko) | 본 발명의 일실시예에 따른 글라스형 웨어러블 디바이스를 이용한 매장정보 제공서비스 시스템 및 방법 | |
CN210606226U (zh) | 一种双模式聋哑人交流设备 | |
Saha et al. | Vision maker: An audio visual and navigation aid for visually impaired person | |
CN218045797U (zh) | 一种盲人穿戴智慧云眼镜及系统 | |
KR102516155B1 (ko) | 홈 오토메이션과 연동되는 음성인식 기반의 웨어러블 디바이스 | |
KR102570418B1 (ko) | 사용자 행동을 분석하는 웨어러블 디바이스 및 이를 이용한 대상인식방법 | |
EP3882894A1 (en) | Seeing aid for a visually impaired individual | |
US20240105173A1 (en) | Method and apparatus for providing virtual space in which interaction with another entity is applied to entity | |
US20230104182A1 (en) | Smart Wearable Sensor-Based Bi-Directional Assistive Device | |
KR101661556B1 (ko) | 글라스형 웨어러블 디바이스를 이용한 신원 확인 방법 및 프로그램 | |
KR20230141395A (ko) | 정보를 제공하는 방법 및 이를 지원하는 전자 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 14417440 Country of ref document: US |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14868369 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205N DATED 17.11.2016) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 14868369 Country of ref document: EP Kind code of ref document: A1 |