WO2020236103A1 - A system and a method for speech tracking for hearing impaired individuals - Google Patents

A system and a method for speech tracking for hearing impaired individuals Download PDF

Info

Publication number
WO2020236103A1
WO2020236103A1 PCT/TR2020/050227 TR2020050227W WO2020236103A1 WO 2020236103 A1 WO2020236103 A1 WO 2020236103A1 TR 2020050227 W TR2020050227 W TR 2020050227W WO 2020236103 A1 WO2020236103 A1 WO 2020236103A1
Authority
WO
WIPO (PCT)
Prior art keywords
processor
speech
audio signal
vibration
signal
Prior art date
Application number
PCT/TR2020/050227
Other languages
French (fr)
Inventor
Mete YAĞANOĞLU
Original Assignee
Atatürk Üni̇versi̇tesi̇ Bi̇li̇msel Araştirma Projeleri̇ Bi̇ri̇mi̇
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Atatürk Üni̇versi̇tesi̇ Bi̇li̇msel Araştirma Projeleri̇ Bi̇ri̇mi̇ filed Critical Atatürk Üni̇versi̇tesi̇ Bi̇li̇msel Araştirma Projeleri̇ Bi̇ri̇mi̇
Publication of WO2020236103A1 publication Critical patent/WO2020236103A1/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/06Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B21/00Teaching, or communicating with, the blind, deaf or mute
    • G09B21/009Teaching or communicating with deaf persons
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/06Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
    • G10L2021/065Aids for the handicapped in understanding

Definitions

  • the invention relates to a system and a method that enables hearing impaired individuals to follow conversations.
  • the invention in particular relates to a system which allows the hearing impaired individual to follow conversations in real time by translating the conversations in the environment and projecting on any surface or transmitting as a vibration signal to the fingers, and a method for operation of this system.
  • hearing aids can be used by individuals who are found to have hearing loss, so that individuals with hearing impairment can easily follow the conversations around them.
  • the hearing aid is not effective in individuals who have lost hearing sensitivity to a large extent.
  • the mentioned situation reveals the need for communication without the sense organs of these individuals.
  • hearing impaired individuals communicate with other individuals by using sign language consisting of gestures and mimics or by writing.
  • sign language is a language that needs to be learned and written communication results in time loss and is not suitable for all environments have made an innovation necessary in the relevant field.
  • Patent application no. WO2018144896 has been encountered in a survey in the state of the art.
  • the application relates to a device that can be attached to the wearer's clothing and can receive sounds from the environment and convert them to text and record them, then project them onto any smart device screen.
  • the mentioned invention limits the use of the device by requiring an additional smart device screen to follow the converted text.
  • the invention is inspired by existing situations and aims to solve the disadvantages mentioned above.
  • the main aim of the invention is to enable hearing impaired individuals to follow conversations in the environment in real time, both in the field of education and in daily life.
  • Another aim of the invention is to ensure that the written text of the conversations can be projected on any surface in real time so that hearing impaired individuals can follow them, without the need for computers, cell phones, tablets or any other additional devices.
  • Another aim of the invention is to enable speech to be transmitted to the fingers of hearing impaired individuals using a defined vibrational alphabet, thus enabling hearing impaired individuals to follow speech more effectively.
  • Another aim of the invention is to enable hearing impaired individuals who miss real-time speech tracking to be able to see the speech text.
  • Figure 1 is a block diagram representation of the speech tracking system of the invention for individuals with hearing impairments.
  • Figure 2 is a representation of the hand wearable device in the system of the invention.
  • FIG. 3 is the flow chart of the signal processing steps in the speech tracking method of the invention for hearing impaired individuals. Description of Part References
  • the invention a representation of which is given in Figure 1 ; is a system that enables individuals with hearing impairments to follow conversations in real time; comprising
  • a processor (12) that receives the analog audio signal transmitted by the microphone (1 1 ) and converts sounds to speech text and speech signal using the software installed in it and transmits them wirelessly
  • a projection lens (22) projecting the speech text received by the wireless module (21 ) to any surface to which the user is directed
  • the wireless module (31 ) that receives and transmits the speech signal transmitted from the processor (12), o a vibration motor (32), which receives the speech signal transmitted from the wireless module (31 ) and converts it to a vibration signal in accordance with a specified vibration alphabet.
  • a screen (33) can be introduced on the wearable device (3), and the speech signal transmitted from the processor (12) can be projected onto the screen (33) as text.
  • the history of the last conversation can be displayed on the screen (33). In this way, the user can capture the missed parts of the conversation by looking at the screen (33).
  • the working method of the invention is as follows:
  • the processor (12) • by the processor (12), the classification of the digital audio signal by comparing it with the installed language model and acoustic model, and the conversion of the classified audio signal into speech text after the feature selection process and the generation of the vibration signal suitable for speech text.
  • the vibration alphabet used in the invention is composed of 22 characters by omitting Vietnamese characters ( , G, I, J, O, ⁇ , LI) in order to avoid complexity. Each character has a specific finger and amount of vibration. The head and index fingers are used for frequently used letters, while the other letters were distributed to the remaining fingers in accordance with their frequency of use. For example, the word "Merhaba” is coded as:
  • A Left hand index finger - 1 vibration.

Abstract

The invention relates to a system which converts the speech to text and projects them onto any surface at which the user is directed to, or transmits them to the users fingertips as a vibration signal according to a defined vibrational alphabet, thereby allowing hearing impaired individuals to follow conversations in real time more effectively, and a method of operation for this system.

Description

A system and a method for speech tracking for hearing impaired individuals Technical Field
The invention relates to a system and a method that enables hearing impaired individuals to follow conversations.
The invention in particular relates to a system which allows the hearing impaired individual to follow conversations in real time by translating the conversations in the environment and projecting on any surface or transmitting as a vibration signal to the fingers, and a method for operation of this system.
State of the Art
Today, hearing aids can be used by individuals who are found to have hearing loss, so that individuals with hearing impairment can easily follow the conversations around them. However, the hearing aid is not effective in individuals who have lost hearing sensitivity to a large extent. The mentioned situation reveals the need for communication without the sense organs of these individuals.
In the current technique, hearing impaired individuals communicate with other individuals by using sign language consisting of gestures and mimics or by writing. However, the fact that sign language is a language that needs to be learned and written communication results in time loss and is not suitable for all environments have made an innovation necessary in the relevant field.
Patent application no. WO2018144896 has been encountered in a survey in the state of the art. The application relates to a device that can be attached to the wearer's clothing and can receive sounds from the environment and convert them to text and record them, then project them onto any smart device screen. The mentioned invention limits the use of the device by requiring an additional smart device screen to follow the converted text.
Another study found in the state of the art is patent application no. US20140337023A1 . The mentioned application refers to a system that determines the target speaker using image processing methods, then takes the speaker's audio signal and converts it into text, reflecting the said text onto the inner surface of glasses worn by the user. Tracking the text through the glasses by user creates difficulty in continuing conversations. In addition, the glasses in the system do not provide comfort due to obstructing the user's view.
As a result, due to the negativity described above and the inadequacy of the existing solutions on the subject, an improvement in the relevant technical area has been made necessary.
Object of the Invention
The invention is inspired by existing situations and aims to solve the disadvantages mentioned above.
The main aim of the invention is to enable hearing impaired individuals to follow conversations in the environment in real time, both in the field of education and in daily life.
Another aim of the invention is to ensure that the written text of the conversations can be projected on any surface in real time so that hearing impaired individuals can follow them, without the need for computers, cell phones, tablets or any other additional devices.
Another aim of the invention is to enable speech to be transmitted to the fingers of hearing impaired individuals using a defined vibrational alphabet, thus enabling hearing impaired individuals to follow speech more effectively.
Another aim of the invention is to enable hearing impaired individuals who miss real-time speech tracking to be able to see the speech text.
The structural and characteristic features and all advantages of the invention will be understood more clearly through the detailed explanation written with the following figures and references to these figures, and therefore assessment should be made by considering these figures and detailed explanation.
Brief Description of Drawings
Figure 1 is a block diagram representation of the speech tracking system of the invention for individuals with hearing impairments.
Figure 2 is a representation of the hand wearable device in the system of the invention.
Figure 3 is the flow chart of the signal processing steps in the speech tracking method of the invention for hearing impaired individuals. Description of Part References
11. Microphone
12. Processor
2. Forehead piece
21. Wireless module
22. Projection lens
3. Hand wearable device
31. Wireless module
32. Vibration motor
33. Screen
Detailed Description of the Invention
In this detailed description, the preferred structures of speech tracking system and method of the invention for hearing impaired individuals are explained only for a better understanding of the subject.
The invention, a representation of which is given in Figure 1 ; is a system that enables individuals with hearing impairments to follow conversations in real time; comprising
• fastened to the wearer's clothing,
o a microphone (1 1 ) that receives and transmits ambient sounds as an analog audio signal,
o a processor (12) that receives the analog audio signal transmitted by the microphone (1 1 ) and converts sounds to speech text and speech signal using the software installed in it and transmits them wirelessly
• a forehead piece (2) worn on the wearer's forehead, comprising
o a wireless module (21 ) that receives and transmits speech text transmitted from the processor (12),
o a projection lens (22) projecting the speech text received by the wireless module (21 ) to any surface to which the user is directed
• at least one wearable device (3) worn on the user's hand, comprising
o the wireless module (31 ) that receives and transmits the speech signal transmitted from the processor (12), o a vibration motor (32), which receives the speech signal transmitted from the wireless module (31 ) and converts it to a vibration signal in accordance with a specified vibration alphabet.
In an alternative embodiment of the invention, a screen (33) can be introduced on the wearable device (3), and the speech signal transmitted from the processor (12) can be projected onto the screen (33) as text. In addition, the history of the last conversation can be displayed on the screen (33). In this way, the user can capture the missed parts of the conversation by looking at the screen (33).
The working method of the invention is as follows:
• receiving the ambient audio signal by the microphone (1 1 ) and transmitting it to the processor (12),
• by the processor (12), sampling the analog audio signal using the software installed in it and converting it to digital audio signal, preprocessing the sampled audio signal to increase clarity, and extracting the cleaned audio signal from the important features for the audio signal through feature extraction,
• by the processor (12), the classification of the digital audio signal by comparing it with the installed language model and acoustic model, and the conversion of the classified audio signal into speech text after the feature selection process and the generation of the vibration signal suitable for speech text.
• receiving the speech text produced in the processor (12) by means of the wireless module (21 ) in the forehead piece (2) and transmitting it to the projection lens (22) in the forehead piece (2),
• projecting speech text by the projection lens (22) onto any surface towards which the user is directed to,
• receiving the vibration signal generated by the processor (12) through the wireless module (31 ) in the wearable device (3) and transmitting it to the vibration motor (32) in the wearable device (3),
• receiving the vibration signal transmitted from the wireless module (31 ) by the vibration motor (32) and transmitting it to the user's fingers by means of vibrations in accordance with a specified vibration alphabet.
The vibration alphabet used in the invention is composed of 22 characters by omitting Turkish characters ( , G, I, J, O, §, LI) in order to avoid complexity. Each character has a specific finger and amount of vibration. The head and index fingers are used for frequently used letters, while the other letters were distributed to the remaining fingers in accordance with their frequency of use. For example, the word "Merhaba" is coded as:
• M: Right hand thumb - 1 vibration,
• E: Left hand index finger - 2 vibrations,
· R: Right hand little finger - 2 vibrations,
• H: Left hand ring finger - 1 vibration,
• A: Left hand index finger - 1 vibration,
• B: Right hand little finger - 1 vibration,
• A: Left hand index finger - 1 vibration. Thereby, the user will be able to make sense of the vibration signals that is transmitted to his hand, thus enabling speech tracking in a more effective way.

Claims

1. A system that allows hearing impaired individuals to track speech in real time, characterized by comprising;
o a microphone (1 1 ) that receives and transmits ambient sounds as an analog audio signal,
o a processor (12) that receives the analog audio signal transmitted by the microphone (1 1 ) and converts signal to speech text and speech signal using the software installed in it and transmitting them,
o a forehead piece (2) carrying a projection lens (22) projecting speech text transmitted from the processor (12) onto a surface towards which the user is directed.
2. A system according to claim 1 , characterized by comprising; the forehead piece (2) carrying a wireless module (21 ) which receives the speech text transmitted by the processor (12) and transmits it to the projection lens (22).
3. A system according to claim 1 , characterized by comprising; a wearable device (3) carrying a vibration motor (32) which receives the speech signal transmitted from the processor (12) and converts it to a vibration signal in accordance with a specified vibration alphabet.
4. A system according to claim 1 , characterized by comprising; the wearable device (3) carrying a wireless module (31 ) which receives the speech signal transmitted from the processor (12) and transmits it to the vibration motor (32).
5. A system according to claim 3 or 4, characterized by comprising; the wearable device (3) carrying a screen (33) which displays the speech text translated by the processor (12) to the user.
6. A system according to claim 1 , characterized by comprising; the projection lens (22) which has a moving head and is able to change orientation depending on the head movement of the user.
7. A method of operation for the system that allows hearing impaired individuals to track speech in real time, characterized by comprising; the process steps of
• receiving the ambient audio signal by the microphone (1 1 ) and transmitting it to the processor (12), • by the processor (12), sampling the analog audio signal using the software installed in it and converting it to digital audio signal, preprocessing the sampled audio signal to increase clarity, and extracting the cleaned audio signal from the important features for the audio signal through feature extraction,
• by the processor (12), the classification of the digital audio signal by comparing it with the installed language model and acoustic model, and the conversion of the classified audio signal into speech text after the feature selection process and the generation of the vibration signal suitable for speech text.
• receiving the speech text produced in the processor (12) by means of the wireless module (21 ) in the forehead piece (2) and transmitting it to the projection lens (22) in the forehead piece (2),
• projecting speech text by the projection lens (22) onto any surface towards which the user is directed to.
8. A method according to Claim 7, characterized by comprising; the process steps of
• receiving the vibration signal generated by the processor (12) through the wireless module (31 ) in the wearable device (3) and transmitting it to the vibration motor (32) in the wearable device (3),
• receiving the vibration signal transmitted from the wireless module (31 ) by the vibration motor (32) and transmitting it to the user's fingers by means of vibrations in accordance with a specified vibration alphabet.
9. A method according to claim 8, characterized by comprising; the process step of projecting the speech text received from the processor (12) by the screen (33) introduced on the wearable device (3).
PCT/TR2020/050227 2019-05-23 2020-03-20 A system and a method for speech tracking for hearing impaired individuals WO2020236103A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
TR2019/07878A TR201907878A2 (en) 2019-05-23 2019-05-23 Speech tracking system and method for hearing impaired individuals
TR2019/07878 2019-05-23

Publications (1)

Publication Number Publication Date
WO2020236103A1 true WO2020236103A1 (en) 2020-11-26

Family

ID=67901480

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/TR2020/050227 WO2020236103A1 (en) 2019-05-23 2020-03-20 A system and a method for speech tracking for hearing impaired individuals

Country Status (2)

Country Link
TR (1) TR201907878A2 (en)
WO (1) WO2020236103A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6005536A (en) * 1996-01-16 1999-12-21 National Captioning Institute Captioning glasses
US20170287504A1 (en) * 2016-04-01 2017-10-05 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus to assist speech training and/or hearing training after a cochlear implantation

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6005536A (en) * 1996-01-16 1999-12-21 National Captioning Institute Captioning glasses
US20170287504A1 (en) * 2016-04-01 2017-10-05 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus to assist speech training and/or hearing training after a cochlear implantation

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
SHIVAKUMAR B.L., RAJASENATHIPATHI M.: "A New Approach for Hardware Control Procedure Used in Braille Glove Vibration System for Disabled Persons", RESEARCH JOURNAL OF APPLIED SCIENCES, ENGINEERING AND TECHNOLOGY, vol. 7, no. 9, 2014, pages 1863 - 1871, XP055762909 *
YAGANOGLU, METE ET AL.: "Wearable vibration based computer interaction and communication system for deaf", APPLIED SCIENCES, vol. 7, no. 1296, 2017, pages 1 - 18, XP055762914 *

Also Published As

Publication number Publication date
TR201907878A2 (en) 2019-07-22

Similar Documents

Publication Publication Date Title
KR102450803B1 (en) Duplex sign language translation apparatus and the apparatus for performing the duplex sign language translation method
US9111545B2 (en) Hand-held communication aid for individuals with auditory, speech and visual impairments
KR101777807B1 (en) Sign language translator, system and method
WO2017142775A1 (en) Hearing assistance with automated speech transcription
CN108427910B (en) Deep neural network AR sign language translation learning method, client and server
TWI383377B (en) Multi-sensory speech recognition system and method
US20170060850A1 (en) Personal translator
US11481030B2 (en) Methods and apparatus for gesture detection and classification
KR101830908B1 (en) Smart glass system for hearing-impaired communication
US20020103649A1 (en) Wearable display system with indicators of speakers
EP1083769A1 (en) Speech converting device and method
CN107230476A (en) A kind of natural man machine language's exchange method and system
US20170024380A1 (en) System and method for the translation of sign languages into synthetic voices
CN109166409B (en) Sign language conversion method and device
TW202044102A (en) Real-time foreign language communication system having wearable translation device designed to be worn on the user's head to directly pick up the content of foreigners' speech for real-time translation
CN104361787A (en) System and method for converting signals
JP2024516570A (en) Data processing method, device, computer device and computer program
CN105118356A (en) Sign language voice conversion method and device
WO2020236103A1 (en) A system and a method for speech tracking for hearing impaired individuals
KR20200049404A (en) System and Method for Providing Simultaneous Interpretation Service for Disabled Person
CN109254418A (en) A kind of glasses for the crowd of becoming deaf
Wang et al. HearASL: Your Smartphone Can Hear American Sign Language
CN210606226U (en) Dual-mode communication equipment for deaf-mute
TWM596382U (en) Sign language image recognition device
CN111326175A (en) Prompting method for interlocutor and wearable device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20810532

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20810532

Country of ref document: EP

Kind code of ref document: A1