WO2020236103A1 - A system and a method for speech tracking for hearing impaired individuals - Google Patents
A system and a method for speech tracking for hearing impaired individuals Download PDFInfo
- Publication number
- WO2020236103A1 WO2020236103A1 PCT/TR2020/050227 TR2020050227W WO2020236103A1 WO 2020236103 A1 WO2020236103 A1 WO 2020236103A1 TR 2020050227 W TR2020050227 W TR 2020050227W WO 2020236103 A1 WO2020236103 A1 WO 2020236103A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- processor
- speech
- audio signal
- vibration
- signal
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/06—Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B21/00—Teaching, or communicating with, the blind, deaf or mute
- G09B21/009—Teaching or communicating with deaf persons
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/06—Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
- G10L2021/065—Aids for the handicapped in understanding
Definitions
- the invention relates to a system and a method that enables hearing impaired individuals to follow conversations.
- the invention in particular relates to a system which allows the hearing impaired individual to follow conversations in real time by translating the conversations in the environment and projecting on any surface or transmitting as a vibration signal to the fingers, and a method for operation of this system.
- hearing aids can be used by individuals who are found to have hearing loss, so that individuals with hearing impairment can easily follow the conversations around them.
- the hearing aid is not effective in individuals who have lost hearing sensitivity to a large extent.
- the mentioned situation reveals the need for communication without the sense organs of these individuals.
- hearing impaired individuals communicate with other individuals by using sign language consisting of gestures and mimics or by writing.
- sign language is a language that needs to be learned and written communication results in time loss and is not suitable for all environments have made an innovation necessary in the relevant field.
- Patent application no. WO2018144896 has been encountered in a survey in the state of the art.
- the application relates to a device that can be attached to the wearer's clothing and can receive sounds from the environment and convert them to text and record them, then project them onto any smart device screen.
- the mentioned invention limits the use of the device by requiring an additional smart device screen to follow the converted text.
- the invention is inspired by existing situations and aims to solve the disadvantages mentioned above.
- the main aim of the invention is to enable hearing impaired individuals to follow conversations in the environment in real time, both in the field of education and in daily life.
- Another aim of the invention is to ensure that the written text of the conversations can be projected on any surface in real time so that hearing impaired individuals can follow them, without the need for computers, cell phones, tablets or any other additional devices.
- Another aim of the invention is to enable speech to be transmitted to the fingers of hearing impaired individuals using a defined vibrational alphabet, thus enabling hearing impaired individuals to follow speech more effectively.
- Another aim of the invention is to enable hearing impaired individuals who miss real-time speech tracking to be able to see the speech text.
- Figure 1 is a block diagram representation of the speech tracking system of the invention for individuals with hearing impairments.
- Figure 2 is a representation of the hand wearable device in the system of the invention.
- FIG. 3 is the flow chart of the signal processing steps in the speech tracking method of the invention for hearing impaired individuals. Description of Part References
- the invention a representation of which is given in Figure 1 ; is a system that enables individuals with hearing impairments to follow conversations in real time; comprising
- a processor (12) that receives the analog audio signal transmitted by the microphone (1 1 ) and converts sounds to speech text and speech signal using the software installed in it and transmits them wirelessly
- a projection lens (22) projecting the speech text received by the wireless module (21 ) to any surface to which the user is directed
- the wireless module (31 ) that receives and transmits the speech signal transmitted from the processor (12), o a vibration motor (32), which receives the speech signal transmitted from the wireless module (31 ) and converts it to a vibration signal in accordance with a specified vibration alphabet.
- a screen (33) can be introduced on the wearable device (3), and the speech signal transmitted from the processor (12) can be projected onto the screen (33) as text.
- the history of the last conversation can be displayed on the screen (33). In this way, the user can capture the missed parts of the conversation by looking at the screen (33).
- the working method of the invention is as follows:
- the processor (12) • by the processor (12), the classification of the digital audio signal by comparing it with the installed language model and acoustic model, and the conversion of the classified audio signal into speech text after the feature selection process and the generation of the vibration signal suitable for speech text.
- the vibration alphabet used in the invention is composed of 22 characters by omitting Vietnamese characters ( , G, I, J, O, ⁇ , LI) in order to avoid complexity. Each character has a specific finger and amount of vibration. The head and index fingers are used for frequently used letters, while the other letters were distributed to the remaining fingers in accordance with their frequency of use. For example, the word "Merhaba” is coded as:
- A Left hand index finger - 1 vibration.
Abstract
The invention relates to a system which converts the speech to text and projects them onto any surface at which the user is directed to, or transmits them to the users fingertips as a vibration signal according to a defined vibrational alphabet, thereby allowing hearing impaired individuals to follow conversations in real time more effectively, and a method of operation for this system.
Description
A system and a method for speech tracking for hearing impaired individuals Technical Field
The invention relates to a system and a method that enables hearing impaired individuals to follow conversations.
The invention in particular relates to a system which allows the hearing impaired individual to follow conversations in real time by translating the conversations in the environment and projecting on any surface or transmitting as a vibration signal to the fingers, and a method for operation of this system.
State of the Art
Today, hearing aids can be used by individuals who are found to have hearing loss, so that individuals with hearing impairment can easily follow the conversations around them. However, the hearing aid is not effective in individuals who have lost hearing sensitivity to a large extent. The mentioned situation reveals the need for communication without the sense organs of these individuals.
In the current technique, hearing impaired individuals communicate with other individuals by using sign language consisting of gestures and mimics or by writing. However, the fact that sign language is a language that needs to be learned and written communication results in time loss and is not suitable for all environments have made an innovation necessary in the relevant field.
Patent application no. WO2018144896 has been encountered in a survey in the state of the art. The application relates to a device that can be attached to the wearer's clothing and can receive sounds from the environment and convert them to text and record them, then project them onto any smart device screen. The mentioned invention limits the use of the device by requiring an additional smart device screen to follow the converted text.
Another study found in the state of the art is patent application no. US20140337023A1 . The mentioned application refers to a system that determines the target speaker using image processing methods, then takes the speaker's audio signal and converts it into text, reflecting the said text onto the inner surface of glasses worn by the user. Tracking the text
through the glasses by user creates difficulty in continuing conversations. In addition, the glasses in the system do not provide comfort due to obstructing the user's view.
As a result, due to the negativity described above and the inadequacy of the existing solutions on the subject, an improvement in the relevant technical area has been made necessary.
Object of the Invention
The invention is inspired by existing situations and aims to solve the disadvantages mentioned above.
The main aim of the invention is to enable hearing impaired individuals to follow conversations in the environment in real time, both in the field of education and in daily life.
Another aim of the invention is to ensure that the written text of the conversations can be projected on any surface in real time so that hearing impaired individuals can follow them, without the need for computers, cell phones, tablets or any other additional devices.
Another aim of the invention is to enable speech to be transmitted to the fingers of hearing impaired individuals using a defined vibrational alphabet, thus enabling hearing impaired individuals to follow speech more effectively.
Another aim of the invention is to enable hearing impaired individuals who miss real-time speech tracking to be able to see the speech text.
The structural and characteristic features and all advantages of the invention will be understood more clearly through the detailed explanation written with the following figures and references to these figures, and therefore assessment should be made by considering these figures and detailed explanation.
Brief Description of Drawings
Figure 1 is a block diagram representation of the speech tracking system of the invention for individuals with hearing impairments.
Figure 2 is a representation of the hand wearable device in the system of the invention.
Figure 3 is the flow chart of the signal processing steps in the speech tracking method of the invention for hearing impaired individuals.
Description of Part References
11. Microphone
12. Processor
2. Forehead piece
21. Wireless module
22. Projection lens
3. Hand wearable device
31. Wireless module
32. Vibration motor
33. Screen
Detailed Description of the Invention
In this detailed description, the preferred structures of speech tracking system and method of the invention for hearing impaired individuals are explained only for a better understanding of the subject.
The invention, a representation of which is given in Figure 1 ; is a system that enables individuals with hearing impairments to follow conversations in real time; comprising
• fastened to the wearer's clothing,
o a microphone (1 1 ) that receives and transmits ambient sounds as an analog audio signal,
o a processor (12) that receives the analog audio signal transmitted by the microphone (1 1 ) and converts sounds to speech text and speech signal using the software installed in it and transmits them wirelessly
• a forehead piece (2) worn on the wearer's forehead, comprising
o a wireless module (21 ) that receives and transmits speech text transmitted from the processor (12),
o a projection lens (22) projecting the speech text received by the wireless module (21 ) to any surface to which the user is directed
• at least one wearable device (3) worn on the user's hand, comprising
o the wireless module (31 ) that receives and transmits the speech signal transmitted from the processor (12),
o a vibration motor (32), which receives the speech signal transmitted from the wireless module (31 ) and converts it to a vibration signal in accordance with a specified vibration alphabet.
In an alternative embodiment of the invention, a screen (33) can be introduced on the wearable device (3), and the speech signal transmitted from the processor (12) can be projected onto the screen (33) as text. In addition, the history of the last conversation can be displayed on the screen (33). In this way, the user can capture the missed parts of the conversation by looking at the screen (33).
The working method of the invention is as follows:
• receiving the ambient audio signal by the microphone (1 1 ) and transmitting it to the processor (12),
• by the processor (12), sampling the analog audio signal using the software installed in it and converting it to digital audio signal, preprocessing the sampled audio signal to increase clarity, and extracting the cleaned audio signal from the important features for the audio signal through feature extraction,
• by the processor (12), the classification of the digital audio signal by comparing it with the installed language model and acoustic model, and the conversion of the classified audio signal into speech text after the feature selection process and the generation of the vibration signal suitable for speech text.
• receiving the speech text produced in the processor (12) by means of the wireless module (21 ) in the forehead piece (2) and transmitting it to the projection lens (22) in the forehead piece (2),
• projecting speech text by the projection lens (22) onto any surface towards which the user is directed to,
• receiving the vibration signal generated by the processor (12) through the wireless module (31 ) in the wearable device (3) and transmitting it to the vibration motor (32) in the wearable device (3),
• receiving the vibration signal transmitted from the wireless module (31 ) by the vibration motor (32) and transmitting it to the user's fingers by means of vibrations in accordance with a specified vibration alphabet.
The vibration alphabet used in the invention is composed of 22 characters by omitting Turkish characters ( , G, I, J, O, §, LI) in order to avoid complexity. Each character has a specific finger and amount of vibration. The head and index fingers are used for frequently
used letters, while the other letters were distributed to the remaining fingers in accordance with their frequency of use. For example, the word "Merhaba" is coded as:
• M: Right hand thumb - 1 vibration,
• E: Left hand index finger - 2 vibrations,
· R: Right hand little finger - 2 vibrations,
• H: Left hand ring finger - 1 vibration,
• A: Left hand index finger - 1 vibration,
• B: Right hand little finger - 1 vibration,
• A: Left hand index finger - 1 vibration. Thereby, the user will be able to make sense of the vibration signals that is transmitted to his hand, thus enabling speech tracking in a more effective way.
Claims
1. A system that allows hearing impaired individuals to track speech in real time, characterized by comprising;
o a microphone (1 1 ) that receives and transmits ambient sounds as an analog audio signal,
o a processor (12) that receives the analog audio signal transmitted by the microphone (1 1 ) and converts signal to speech text and speech signal using the software installed in it and transmitting them,
o a forehead piece (2) carrying a projection lens (22) projecting speech text transmitted from the processor (12) onto a surface towards which the user is directed.
2. A system according to claim 1 , characterized by comprising; the forehead piece (2) carrying a wireless module (21 ) which receives the speech text transmitted by the processor (12) and transmits it to the projection lens (22).
3. A system according to claim 1 , characterized by comprising; a wearable device (3) carrying a vibration motor (32) which receives the speech signal transmitted from the processor (12) and converts it to a vibration signal in accordance with a specified vibration alphabet.
4. A system according to claim 1 , characterized by comprising; the wearable device (3) carrying a wireless module (31 ) which receives the speech signal transmitted from the processor (12) and transmits it to the vibration motor (32).
5. A system according to claim 3 or 4, characterized by comprising; the wearable device (3) carrying a screen (33) which displays the speech text translated by the processor (12) to the user.
6. A system according to claim 1 , characterized by comprising; the projection lens (22) which has a moving head and is able to change orientation depending on the head movement of the user.
7. A method of operation for the system that allows hearing impaired individuals to track speech in real time, characterized by comprising; the process steps of
• receiving the ambient audio signal by the microphone (1 1 ) and transmitting it to the processor (12),
• by the processor (12), sampling the analog audio signal using the software installed in it and converting it to digital audio signal, preprocessing the sampled audio signal to increase clarity, and extracting the cleaned audio signal from the important features for the audio signal through feature extraction,
• by the processor (12), the classification of the digital audio signal by comparing it with the installed language model and acoustic model, and the conversion of the classified audio signal into speech text after the feature selection process and the generation of the vibration signal suitable for speech text.
• receiving the speech text produced in the processor (12) by means of the wireless module (21 ) in the forehead piece (2) and transmitting it to the projection lens (22) in the forehead piece (2),
• projecting speech text by the projection lens (22) onto any surface towards which the user is directed to.
8. A method according to Claim 7, characterized by comprising; the process steps of
• receiving the vibration signal generated by the processor (12) through the wireless module (31 ) in the wearable device (3) and transmitting it to the vibration motor (32) in the wearable device (3),
• receiving the vibration signal transmitted from the wireless module (31 ) by the vibration motor (32) and transmitting it to the user's fingers by means of vibrations in accordance with a specified vibration alphabet.
9. A method according to claim 8, characterized by comprising; the process step of projecting the speech text received from the processor (12) by the screen (33) introduced on the wearable device (3).
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
TR2019/07878A TR201907878A2 (en) | 2019-05-23 | 2019-05-23 | Speech tracking system and method for hearing impaired individuals |
TR2019/07878 | 2019-05-23 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020236103A1 true WO2020236103A1 (en) | 2020-11-26 |
Family
ID=67901480
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/TR2020/050227 WO2020236103A1 (en) | 2019-05-23 | 2020-03-20 | A system and a method for speech tracking for hearing impaired individuals |
Country Status (2)
Country | Link |
---|---|
TR (1) | TR201907878A2 (en) |
WO (1) | WO2020236103A1 (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6005536A (en) * | 1996-01-16 | 1999-12-21 | National Captioning Institute | Captioning glasses |
US20170287504A1 (en) * | 2016-04-01 | 2017-10-05 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus to assist speech training and/or hearing training after a cochlear implantation |
-
2019
- 2019-05-23 TR TR2019/07878A patent/TR201907878A2/en unknown
-
2020
- 2020-03-20 WO PCT/TR2020/050227 patent/WO2020236103A1/en active Application Filing
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6005536A (en) * | 1996-01-16 | 1999-12-21 | National Captioning Institute | Captioning glasses |
US20170287504A1 (en) * | 2016-04-01 | 2017-10-05 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus to assist speech training and/or hearing training after a cochlear implantation |
Non-Patent Citations (2)
Title |
---|
SHIVAKUMAR B.L., RAJASENATHIPATHI M.: "A New Approach for Hardware Control Procedure Used in Braille Glove Vibration System for Disabled Persons", RESEARCH JOURNAL OF APPLIED SCIENCES, ENGINEERING AND TECHNOLOGY, vol. 7, no. 9, 2014, pages 1863 - 1871, XP055762909 * |
YAGANOGLU, METE ET AL.: "Wearable vibration based computer interaction and communication system for deaf", APPLIED SCIENCES, vol. 7, no. 1296, 2017, pages 1 - 18, XP055762914 * |
Also Published As
Publication number | Publication date |
---|---|
TR201907878A2 (en) | 2019-07-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR102450803B1 (en) | Duplex sign language translation apparatus and the apparatus for performing the duplex sign language translation method | |
US9111545B2 (en) | Hand-held communication aid for individuals with auditory, speech and visual impairments | |
KR101777807B1 (en) | Sign language translator, system and method | |
WO2017142775A1 (en) | Hearing assistance with automated speech transcription | |
CN108427910B (en) | Deep neural network AR sign language translation learning method, client and server | |
TWI383377B (en) | Multi-sensory speech recognition system and method | |
US20170060850A1 (en) | Personal translator | |
US11481030B2 (en) | Methods and apparatus for gesture detection and classification | |
KR101830908B1 (en) | Smart glass system for hearing-impaired communication | |
US20020103649A1 (en) | Wearable display system with indicators of speakers | |
EP1083769A1 (en) | Speech converting device and method | |
CN107230476A (en) | A kind of natural man machine language's exchange method and system | |
US20170024380A1 (en) | System and method for the translation of sign languages into synthetic voices | |
CN109166409B (en) | Sign language conversion method and device | |
TW202044102A (en) | Real-time foreign language communication system having wearable translation device designed to be worn on the user's head to directly pick up the content of foreigners' speech for real-time translation | |
CN104361787A (en) | System and method for converting signals | |
JP2024516570A (en) | Data processing method, device, computer device and computer program | |
CN105118356A (en) | Sign language voice conversion method and device | |
WO2020236103A1 (en) | A system and a method for speech tracking for hearing impaired individuals | |
KR20200049404A (en) | System and Method for Providing Simultaneous Interpretation Service for Disabled Person | |
CN109254418A (en) | A kind of glasses for the crowd of becoming deaf | |
Wang et al. | HearASL: Your Smartphone Can Hear American Sign Language | |
CN210606226U (en) | Dual-mode communication equipment for deaf-mute | |
TWM596382U (en) | Sign language image recognition device | |
CN111326175A (en) | Prompting method for interlocutor and wearable device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20810532 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20810532 Country of ref document: EP Kind code of ref document: A1 |