WO2018107489A1 - Method and apparatus for assisting people who have hearing and speech impairments and electronic device - Google Patents

Method and apparatus for assisting people who have hearing and speech impairments and electronic device Download PDF

Info

Publication number
WO2018107489A1
WO2018107489A1 PCT/CN2016/110475 CN2016110475W WO2018107489A1 WO 2018107489 A1 WO2018107489 A1 WO 2018107489A1 CN 2016110475 W CN2016110475 W CN 2016110475W WO 2018107489 A1 WO2018107489 A1 WO 2018107489A1
Authority
WO
WIPO (PCT)
Prior art keywords
sound
display signal
display
person
deaf
Prior art date
Application number
PCT/CN2016/110475
Other languages
French (fr)
Chinese (zh)
Inventor
廉士国
李强
Original Assignee
深圳前海达闼云端智能科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳前海达闼云端智能科技有限公司 filed Critical 深圳前海达闼云端智能科技有限公司
Priority to CN201680006924.XA priority Critical patent/CN107223277A/en
Priority to PCT/CN2016/110475 priority patent/WO2018107489A1/en
Publication of WO2018107489A1 publication Critical patent/WO2018107489A1/en

Links

Images

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B21/00Teaching, or communicating with, the blind, deaf or mute

Definitions

  • the present invention relates to the field of smart device technologies, and in particular, to a deaf-mute assist method, device, and electronic device.
  • Hearing is an important way for humans to perceive the world. Through hearing, human beings can realize the transmission and feedback of thoughts and feelings between people, and avoid dangerous situations in the environment.
  • the number of disabled people with hearing disabilities is the highest among the five disabled people, such as visual disability, disability, and disability.
  • the hearing language disabled group has many obstacles in life due to the shortcomings of listening language ability. Therefore, the hearing language disabled group is in need of help.
  • common hearing aids for deaf-mute people include hearing aids or cochlear implants. These devices are helpful for many deaf people, but they also have certain limitations.
  • the different degrees of disability have different requirements on the parameters of the hearing aid or the cochlear implant, and the user needs a complicated selection process when selecting the corresponding product.
  • the deaf-mute auxiliary equipment in the prior art has certain limitations. How to assist the deaf-mute person to perceive the sound conveniently and quickly is still a problem that is continuously studied by those skilled in the art.
  • Embodiments of the present invention provide a deaf-mute person assisting method, apparatus, and electronic device, which are mainly used to assist a deaf-mute person to perceive sound conveniently and quickly.
  • a method for assisting a deaf person comprising:
  • Display is performed under the driving of the display signal.
  • an AC assist device including:
  • a receiving unit configured to receive a sound
  • a converting unit configured to identify the sound and convert the sound into a display signal according to the recognition result
  • a display unit for displaying under the driving of the display signal.
  • an electronic device comprising: a sound collection device, a display device, a memory, and a processor, a sound collection device, a display device, and a memory coupled to the processor; the memory for storing a computer execution code,
  • the computer-executable code is for controlling the processor to perform the deaf-mute assist method of the first aspect.
  • a storage medium characterized in that the computer software instructions for storing the AC assist device of the second aspect are designed to perform the deaf-mute assist method described in the first aspect. code.
  • a computer program product can be directly loaded into an internal memory of a computer and includes software code, and the computer program can be loaded and executed by a computer to implement the deaf-mute assist method according to the first aspect. .
  • the deaf-mute assisting method provided by the embodiment of the present invention first receives a sound, then recognizes the received sound line and converts the sound into a display signal according to the recognition result, and finally displays the display signal, which is implemented by the present invention.
  • the deaf-mute assist method provided by the example can convert the received sound into a display signal and will display the signal
  • the display is driven to convert the received auditory signal into a visual signal, so that the deaf-mute person can visually see the display content corresponding to the sound. Therefore, the deaf-mute assisting method provided by the embodiment of the present invention can be assisted.
  • the dumb person perceives the sound.
  • the deaf-mute assisting method provided by the embodiment of the present invention does not require a complicated selection process and does not require language training, and thus the present invention is compared to the prior art.
  • the deaf-mute assist method provided can assist the deaf-mute person to perceive the sound conveniently and quickly.
  • FIG. 1 is a flow chart of steps of a deaf-mute assist method according to an embodiment of the present invention
  • FIG. 2 is a second flowchart of steps of a deaf-mute assist method according to an embodiment of the present invention
  • FIG. 3 is a third flowchart of the steps of the deaf-mute assisting method provided by the embodiment of the present invention.
  • FIG. 4 is a schematic diagram of a correspondence relationship between a sound orientation and a display position according to an embodiment of the present invention
  • FIG. 5 is a fourth flowchart of steps of a deaf-mute assist method according to an embodiment of the present invention.
  • FIG. 6 is a schematic structural diagram of a deaf-mute auxiliary device according to an embodiment of the present invention.
  • FIG. 7 is a second schematic structural diagram of a deaf-mute auxiliary device according to an embodiment of the present invention.
  • FIG. 8 is a third schematic structural diagram of a deaf-mute auxiliary device according to an embodiment of the present invention.
  • FIG. 9 is a schematic structural diagram of an electronic device according to an embodiment of the present invention.
  • the basic principle of the technical solution provided by the embodiment of the present invention is: identifying the received sound, converting the received sound into a display signal, and displaying the content corresponding to the sound driven by the display signal, thereby making the hoarseness People correspond to sound by watching The visual information perceives the sound.
  • the execution body of the deaf-mute assisting method provided by the embodiment of the present invention may be a deaf-mute auxiliary device or an electronic device that can be used to execute the deaf-mute assisting method.
  • the deaf-mute auxiliary device may be a combination of a central processing unit (CPU), a CPU and a memory in the electronic device, or may be another control unit or module in the electronic device.
  • the foregoing electronic device may be a mobile phone, augmented reality glasses (abbreviation: AR glasses), a personal computer (PC), a netbook, a personal digital assistant, which are assisted by the deaf-mute person by using the method provided by the embodiment of the present invention.
  • AR glasses augmented reality glasses
  • PC personal computer
  • netbook a personal digital assistant
  • server etc.
  • the above electronic device may be a PC, a server, etc., which is installed with a software client or software system or software application that can assist the deaf person, and a specific hardware implementation environment. It can be in the form of a general-purpose computer, or an ASIC, an FPGA, or a programmable extension platform such as Tensilica's Xtensa platform.
  • an embodiment of the present invention provides a deaf-mute assisting method.
  • the deaf-mute assisting method includes the following steps:
  • the sound in the above embodiment may be a voice that is emitted when another person communicates with the user, a voice that is broadcasted, or the like; or a voice in the environment, such as a sound of a car whistle, a barking voice, The sound of thunder and so on.
  • the sound can be received by a sound sensing device such as a microphone (English name: microphone, abbreviation: Mic) or a Mic array.
  • a sound sensing device such as a microphone (English name: microphone, abbreviation: Mic) or a Mic array.
  • the process of recognizing the sound in the above embodiment and converting the sound into the display signal according to the recognition result may be completed inside the deaf-mute auxiliary device, or may be assisted by the remote service device.
  • step S12 can be specifically realized by the following steps: a. Identifying the sound by the internal sound processing device, b The sound is converted into a corresponding display signal according to the recognition result of the sound processing device.
  • step S12 can be specifically implemented by: c. transmitting the sound to the remote server to make a sound with the end server. Identify and convert the sound into a display signal based on the recognition result. d. Receive the display signal sent by the remote server.
  • the remote service device can be a cloud server or the like.
  • the converting the sound into the display signal in the step S12 may be: converting the sound into at least one of a display signal for displaying characters, a display signal for displaying the logo, and a display signal for displaying the dynamic screen.
  • a display signal for displaying characters converting the sound into at least one of a display signal for displaying characters into at least one of a display signal for displaying characters, a display signal for displaying the logo, and a display signal for displaying the dynamic screen.
  • the received sound when the received sound is a voice uttered by a person who communicates face-to-face with the user, the received sound can be converted into a display signal for displaying the text.
  • the received sound when the received sound is a barking sound, the received sound can be converted into a display signal for displaying a logo such as a dog's cartoon drawing.
  • the received sound when the received sound is a sound emitted when the car is walking, the received sound can be converted into a display signal for displaying a dynamic picture of the car walking.
  • the received sound can be expressed more clearly by visual information by a combination of various ways.
  • the received sound when the received sound is a sound emitted when the car is walking, the received sound can be converted into a display signal for displaying a moving picture of the car and a car logo.
  • the received sound it is also possible to convert the received sound into other types of display signals based on the above embodiments, but these are all reasonable modifications of the embodiments of the present invention, and therefore should all belong to the protection of the embodiments of the present invention.
  • the received sound is a sound emitted when the car is walking
  • the received sound can be converted into a display signal for displaying a moving picture of the car and a car logo.
  • the specific display manner of displaying the visual information may be selected based on the execution subject of the deaf-mute assisting method provided by the embodiment of the present invention. For example, when the execution body of the deaf-mute assisting method provided by the above embodiment is a mobile phone, displaying under the driving of the display signal may be performed by driving the screen of the mobile phone through the display signal, for example: When the execution body of the deaf-mute assisting method provided by the above embodiment is an AR glasses, displaying the visual information may be performed by projecting the display content on the lens of the AR glasses by driving the projection display device by the display signal.
  • the deaf-mute assisting method provided by the embodiment of the present invention first receives a sound, then recognizes the received sound line and converts the sound into a display signal according to the recognition result, and finally displays the display signal, which is implemented by the present invention.
  • the deaf-mute assist method provided by the example can convert the received sound into a display signal and display it under the driving of the display signal, that is, the received auditory signal can be converted into a visual signal, thereby enabling the deaf person to pass the vision.
  • the display content corresponding to the sound is seen, so the deaf-mute assisting method provided by the embodiment of the present invention can assist the deaf-mute person to perceive the sound.
  • the deaf-mute assisting method provided by the embodiment of the present invention does not require a complicated selection process and does not require language training, and thus the present invention is compared to the prior art.
  • the deaf-mute assist method provided can assist the deaf-mute person to perceive the sound conveniently and quickly.
  • the sound is recognized in the above step S12, and the sound is converted into a display signal according to the recognition result, which can be specifically implemented by the following steps:
  • step S121 if it is determined that the sound is a speech sound by recognizing the type of the sound, step S122 is performed; and/or, if the sound is determined to be the ambient sound by recognizing the type of the sound, step S123 is executed. That is, step S122 and step S123 in the embodiment of the present invention may be performed either alternatively or alternatively.
  • the speaking voice in the embodiment of the present invention generally refers to a voice that is emitted by a human when talking, speaking, broadcasting news, and the like.
  • the voice may be received after being processed.
  • the voice of the speaker is amplified and output and received at the time of the lecture.
  • sounds are not spoken directly by humans, such sounds also belong to the voices of the embodiments of the present invention.
  • the ambient sound in the embodiment of the present invention is other sounds different from the talking sound, and the received sound is divided into a talking sound and an ambient sound.
  • the ring The sound of the environment can be: the sound of the car whistling, the sound of the dog barking, the sound of thunder, the noise in the environment, and so on.
  • the content of the voice recognition in the foregoing embodiment may be specifically implemented by: e. determining, by using a language type recognition technology, a language type of the received voice, for example, identifying the received voice as Chinese, English, French, etc. . f. Identify the spoken content according to the type of language in which the sound is received and the specific received voice. That is, when the received sound is a voice, the language type of the voice can be recognized first to identify the specific content of the voice.
  • the content of the speech is often complicated, it is difficult to clearly display the corresponding content by means of a logo, a dynamic picture, etc., so in the implementation of the present invention, when the sound is a voice, the sound is converted into a text according to the content of the voice. Thereby the content of the received speech is more clearly displayed.
  • the identifier in the above embodiment may specifically be: a cartoon drawing of a dog, a cartoon drawing of a car, a danger sign, a lightning sign, and the like.
  • the deaf-mute assisting method provided by the above embodiment can assist the deaf-mute person to perceive the speaking sound and various sounds in the environment, but when the user receives the speaking sound in a noisy environment, the received speech may be Contains noise from the environment, which may cause inaccurate content recognition of the voice.
  • the present invention provides a deaf-mute assisting method according to an embodiment of the present invention. Specifically, referring to FIG. 3, the deaf-mute assisting method provided by the embodiment of the present invention is provided on the basis of the deaf-mute assistant shown in FIG. It further includes:
  • the relative person is the person who makes the voice.
  • an image of a relative person may be acquired by one or more of a monocular camera, a binocular camera, a depth camera, an image sensor, and the like.
  • any image capturing device may be used to obtain an image of a relative person.
  • the manner of acquiring the image of the opposite person is not limited, so that the image of the opposite person can be obtained.
  • the exemplary, relative person's image may be a dynamic picture when speaking to a person.
  • step S122 converting the speech sound into a display signal for driving the display character according to the content of the speech sound can be realized by the implementation provided in step S33.
  • the image of the opposite person is recognized to obtain the lip motion of the opposite person, and then when the sound is the voice, the content of the voice is recognized, and according to the content of the voice and the relative person
  • the lip action converts the content of the speech into a display signal corresponding to the text, and since the lip language recognition technology can interpret the relative words spoken by the relative person's lip motion, the accuracy of the conversion can be improved.
  • deaf-mute assisting method provided by the foregoing embodiment further includes:
  • the display in the above step S13 under the driving of the display signal may be specifically implemented by displaying the position of the sound under the driving of the display signal at a corresponding position of the display interface.
  • the display content 41 corresponding to the sound is below the display interface 40; when the sound is located at the front of the user F2, the sound is The corresponding display content 42 is displayed above the display interface 40; when the sound is located on the left side of the user F3, the sound and the sound The corresponding display content 43 is displayed on the left side of the display interface 40; when the sound is located on the right side F4 of the user, the display content 44 corresponding to the sound is displayed on the right side of the display interface 40.
  • the display content is displayed on the corresponding position of the display interface, which can further enable the user to know the orientation of the sound, thereby helping the deaf-mute person to perceive the sound more comprehensively.
  • the deaf-mute assisting method includes:
  • detecting the user's hand motion may specifically: acquiring a dynamic picture of the user by one or more of a monocular camera, a binocular camera, a depth camera, an image sensor, etc., and then acquiring the user according to the dynamic picture of the user.
  • Hand movements In addition, detecting the user's hand movement can also detect the motion parameters such as the acceleration and the rotation angle of the user's hand through the hand wearing device, and acquire the user's hand motion according to the motion parameter.
  • the hand wearing device can be: a ring, a wristband, a data glove, and the like.
  • the process of recognizing the user's hand motion in the above embodiment and converting the user's hand motion into voice according to the recognition result may be completed inside the deaf-mute auxiliary device, or may be assisted by the remote service device.
  • step S52 can be specifically implemented by the following steps: A, assisted by the deaf person The image processing device inside the device recognizes the user's hand motion. B converts the user's hand motion into a corresponding voice according to the recognition result of the image processing apparatus.
  • the step S52 may be specifically implemented by: sending the image to the remote server, so that the remote server is the user.
  • the hand motion recognizes and converts the user's hand motion into voice based on the recognition result.
  • D. Receive the voice sent by the remote server.
  • the remote service device can be a cloud server or the like.
  • the sign language content of the gesture expression can be converted into voice by voice synthesis technology, and the voice is broadcasted through a speaker (English name: Speaker).
  • the sign language content can be converted into a voice and broadcasted, so that the person who does not understand the sign language can learn the content expressed by the sign language deaf and dumb person through the broadcasted voice, thereby further assisting the deaf and mute person to communicate.
  • FIG. 6 shows a possible structural diagram of the deaf-mute assisting device involved in the above embodiment.
  • the deaf-mute aid includes:
  • a receiving unit 61 configured to receive a sound
  • the converting unit 62 is configured to identify the sound and convert the sound into a display signal according to the recognition result;
  • the display unit 63 is configured to perform display under the driving of the display signal.
  • the deaf-mute auxiliary device provided by the embodiment of the present invention includes: a receiving unit, a converting unit, and a display unit, wherein the receiving unit is configured to receive a sound, and the converting unit is configured to identify the sound and convert the sound into a display signal according to the recognition result,
  • the display unit is configured to display under the driving of the display signal, so the deaf-mute auxiliary device provided by the embodiment of the present invention can convert the received auditory signal into a visual signal, thereby making the deaf person By visually seeing the display content corresponding to the sound, the deaf-mute assisting device provided by the embodiment of the present invention can assist the deaf-mute person to perceive the sound.
  • the deaf-mute auxiliary device provided by the embodiment of the present invention does not require a complicated selection process and does not need to perform language training, and thus is compared with the prior art embodiment of the present invention.
  • the deaf-mute aids are provided to assist the deaf and mute to sense sound easily and quickly.
  • the converting unit 62 is specifically configured to identify a type of the sound
  • the converting unit 62 is specifically configured to recognize the content of the talking sound when the sound is a talking sound, convert the talking sound into a display signal for driving the displayed text according to the content of the talking sound; and/or identify the environment when the sound is an ambient sound
  • the category of the sound which converts the ambient sound into a display signal for driving the display identifier according to the category of the ambient sound.
  • the receiving unit 61 is further configured to acquire an image of a relative person; wherein, the opposite person is a person who emits a voice;
  • the converting unit 62 is further configured to acquire a lip motion of the opposite person according to the image of the opposite person;
  • the converting unit 62 is specifically configured to convert the speaking sound into a display signal for driving the display text according to the content of the speaking sound and the lip motion of the opposite person.
  • the receiving unit 61 is further configured to acquire an orientation of the sound
  • the display unit 63 is further configured to perform display on a corresponding position of the display interface under the driving of the display signal according to the orientation of the sound.
  • the converting unit 62 includes: a sending module 71 and a receiving module 72;
  • the sending module 71 is configured to send the sound to the remote server to identify the sound with the end server and convert the sound into a display signal according to the recognition result;
  • the receiving module 72 is configured to receive a display signal sent by the remote server.
  • the deaf-mute assistant 600 further includes: voice broadcast Reporting unit 64;
  • the receiving unit 61 is further configured to detect a user's hand motion
  • the identification unit 62 is further configured to identify a user's hand motion and convert the user's hand motion into voice according to the recognition result;
  • the voice broadcast unit 64 is configured to broadcast the voice.
  • the receiving unit 61 is configured to implement the steps of receiving the sound, acquiring the image of the opposite person, and acquiring the orientation of the sound in the above-described deaf-mute assisting method;
  • the identifying unit 62 is configured to implement the sound recognition in the above-described deaf-mute assisting method And converting the sound into a display signal according to the recognition result, recognizing the type of the sound, recognizing the content of the speech sound, converting the speech sound into a display signal for driving the display text, and identifying the environment sound according to the content of the speech sound, according to The category of the ambient sound converts the ambient sound into a display signal for driving the display mark, acquires the lip motion of the opposite person according to the image of the opposite person, converts the content of the voice according to the content of the voice and the lip motion of the opposite person into the text Corresponding display signal and a step of recognizing the user's hand motion and converting the user's hand motion into voice according to the recognition result;
  • the sending module 71 is configured
  • the receiving unit 61 may be one or more of a Mic, a Mic array, a camera, an image sensor, an ultrasonic detecting device, an infrared camera, and the like.
  • the identification unit 62 may be a processor or a transceiver; the display unit 63 may be a display screen, a laser projection display device; the voice announcement unit 64 may be a speaker or the like.
  • the programs corresponding to the actions performed by the above-mentioned deaf-mute auxiliary device can be stored in software.
  • the deaf-mute auxiliary device is in the memory, so that the processor calls to perform the operations corresponding to the above respective units.
  • FIG. 9 shows a possible structural diagram of an electronic device including the deaf-mute aid device involved in the above embodiment.
  • the electronic device 900 includes a processor 91, a memory 92, a system bus 93, a communication interface 94, a sound collection device 95, and a display device 96.
  • the processor 91 may be a processor or a collective name of a plurality of processing elements.
  • the processor 91 can be a central processing unit (CPU).
  • the processor 91 can also be other general purpose processors, digital signal processing (DSP), application specific integrated circuits (ASICs), field-programmable gate arrays (FPGAs), or Other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, and the like, can implement or perform the various illustrative logical blocks, modules, and circuits described in connection with the present disclosure.
  • the general purpose processor may be a microprocessor or the processor or any conventional processor or the like.
  • the processor 91 may also be a dedicated processor, which may include at least one of a baseband processing chip, a radio frequency processing chip, and the like.
  • the processor can also be a combination of computing functions, for example, including one or more microprocessor combinations, a combination of a DSP and a microprocessor, and the like.
  • the dedicated processor may also include a chip having other specialized processing functions of the device.
  • the memory 92 is used to store computer execution code
  • the processor 91 is connected to the memory 92 through the system bus 93.
  • the processor 91 is configured to execute the computer execution code stored in the memory 92 to execute any of the embodiments provided by the embodiments of the present invention.
  • a deaf-mute assisting method for example, the processor 91 is configured to support the electronic device to perform step S12 shown in FIG. 1, steps S121, S122, and 123 shown in FIG. 2, and steps S32 and S33 shown in FIG. Step S52 shown in FIG. 5, and/or other processes for the techniques described herein, the specific deaf-mute assisting method may refer to the related descriptions above and in the drawings, and details are not described herein again.
  • System bus 93 can include a data bus, a power bus, a control bus, and a signal status bus. In the present embodiment, for the sake of clarity, various buses are shown in FIG. Means system bus 93.
  • Communication interface 94 may specifically be a transceiver on the device.
  • the transceiver can be a wireless transceiver.
  • the wireless transceiver can be an antenna or the like of the device.
  • the processor 91 communicates with other devices via the communication interface 94, for example, if the device is a module or component of the electronic device, the device is for data interaction with other modules in the electronic device.
  • the steps of the method described in connection with the present disclosure may be implemented in a hardware manner, or may be implemented by a processor executing software instructions.
  • the embodiment of the present invention further provides a storage medium for storing computer software instructions for use in the electronic device shown in FIG. 9, which includes program code designed to execute the deaf-mute assist method provided by any of the above embodiments.
  • the software instructions may be composed of corresponding software modules, and the software modules may be stored in a random access memory (English: random access memory, abbreviation: RAM), flash memory, read only memory (English: read only memory, abbreviation: ROM) , erasable programmable read-only memory (English: erasable programmable ROM, abbreviation: EPROM), electrically erasable programmable read-only memory (English: electrical EPROM, abbreviation: EEPROM), registers, hard disk, mobile hard disk, CD-ROM (CD-ROM) or any other form of storage medium known in the art.
  • An exemplary storage medium is coupled to the processor to enable the processor to read information from, and write information to, the storage medium.
  • the storage medium can also be an integral part of the processor.
  • the processor and the storage medium can be located in an ASIC.
  • the ASIC can be located in a core network interface device.
  • the processor and the storage medium may also exist as discrete components in the core network interface device.
  • the embodiment of the invention further provides a computer program product, which can be directly loaded into the internal memory of the computer and contains software code, and the computer program can be loaded and executed by the computer to implement the hoarseness provided by any of the above embodiments. Human assisted method.
  • the functions described herein can be implemented in hardware, software, firmware, or any combination thereof.
  • the functions may be stored in a computer readable medium or transmitted as one or more instructions or code on a computer readable medium.
  • Computer readable media includes both computer storage media and communication media including any medium that facilitates transfer of a computer program from one location to another.
  • a storage medium may be any available media that can be accessed by a general purpose or special purpose computer.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Business, Economics & Management (AREA)
  • Physics & Mathematics (AREA)
  • Educational Administration (AREA)
  • Educational Technology (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Controls And Circuits For Display Device (AREA)
  • Telephone Function (AREA)

Abstract

A method and apparatus for assisting people who have hearing and speech impairments and an electronic device, aimed at assisting people who have hearing and speech impairments in sensing sound in a quick and easy manner. The method comprises: receiving sound (S11); identifying the sound, and converting the sound into a display signal according to an identification result (S12); and performing display under drive of the display signal (S13).

Description

一种聋哑人辅助方法、装置以及电子设备Deaf-mute assistant method, device and electronic device 技术领域Technical field
本发明涉及智能装置技术领域,尤其涉及一种聋哑人辅助方法、装置以及电子设备。The present invention relates to the field of smart device technologies, and in particular, to a deaf-mute assist method, device, and electronic device.
背景技术Background technique
听觉是人类感知世界的一种重要方式。人类通过听觉可以实现人与人之间的思想与感情的传递和反馈、躲避环境中的危险状况等。Hearing is an important way for humans to perceive the world. Through hearing, human beings can realize the transmission and feedback of thoughts and feelings between people, and avoid dangerous situations in the environment.
调查显示,听力语言残疾人数居视力残疾、肢残、肢残等五大残疾人数之首,仅中国听力语言残疾人数就有2000万左右,并且其中包括许多七岁以下的儿童。听力语言残疾人群由于听力语言能力的缺陷,造成生活中存在非常多的障碍,因此听力语言残疾人群是亟需帮助的。目前,常见的听力常见的聋哑人辅助设备包括:助听器或人工耳蜗,这些设备对很多聋哑人都有帮助,但同时也存在一定的局限性。一方面,不同残疾程度对助听器或人工耳蜗的设备参数要求不同,用户在选择相应产品时需要复杂的甄选过程。另一方面,对于听力在语前已完全丧失且成年人,即使通过人工耳蜗植入等方式使其恢复听觉,也不是恢复听觉后就能听懂说话,而是要进行语言训练,而且由于错过了最佳语言学习期,语言训练效果往往不理想,导致交流时仍存在较大的障碍。综上,现有技术中的聋哑人辅助设备一定的局限性,如何辅助聋哑人方便、快捷的感知声音仍是本领域技术人员不断研究的问题。According to the survey, the number of disabled people with hearing disabilities is the highest among the five disabled people, such as visual disability, disability, and disability. There are only about 20 million people with hearing disabilities in China, including many children under the age of seven. The hearing language disabled group has many obstacles in life due to the shortcomings of listening language ability. Therefore, the hearing language disabled group is in need of help. At present, common hearing aids for deaf-mute people include hearing aids or cochlear implants. These devices are helpful for many deaf people, but they also have certain limitations. On the one hand, the different degrees of disability have different requirements on the parameters of the hearing aid or the cochlear implant, and the user needs a complicated selection process when selecting the corresponding product. On the other hand, if the hearing is completely lost before the language and the adult, even if it is restored to hearing by means of cochlear implantation, it is not possible to understand the speech after restoring the hearing, but to conduct language training, and because of the missed In the best language learning period, the language training effect is often not ideal, which leads to great obstacles in communication. In summary, the deaf-mute auxiliary equipment in the prior art has certain limitations. How to assist the deaf-mute person to perceive the sound conveniently and quickly is still a problem that is continuously studied by those skilled in the art.
发明内容Summary of the invention
本发明的实施例提供一种聋哑人辅助方法、装置以及电子设备,主要用于辅助聋哑人方便、快捷的感知声音。 Embodiments of the present invention provide a deaf-mute person assisting method, apparatus, and electronic device, which are mainly used to assist a deaf-mute person to perceive sound conveniently and quickly.
为达到上述目的,本发明的实施例采用如下技术方案:In order to achieve the above object, embodiments of the present invention adopt the following technical solutions:
第一方面,提供一种聋哑人辅助方法,包括:In a first aspect, a method for assisting a deaf person is provided, comprising:
接收声音;Receiving sound
对所述声音进行识别并根据识别结果将所述声音转换为显示信号;Identifying the sound and converting the sound into a display signal according to the recognition result;
在所述显示信号的驱动下进行显示。Display is performed under the driving of the display signal.
第二方面,提供一种交流辅助装置,包括:In a second aspect, an AC assist device is provided, including:
接收单元,用于接收声音;a receiving unit, configured to receive a sound;
转换单元,用于对所述声音进行识别并根据识别结果将所述声音转换为显示信号;a converting unit, configured to identify the sound and convert the sound into a display signal according to the recognition result;
显示单元,用于在所述显示信号的驱动下进行显示。a display unit for displaying under the driving of the display signal.
第三方面,提供一种电子设备,包括:声音采集设备、显示设备、存储器和处理器,声音采集设备、显示设备和存储器耦合至所述处理器;所述存储器用于存储计算机执行代码,所述计算机执行代码用于控制所述处理器执行第一方面所述的聋哑人辅助方法。In a third aspect, an electronic device is provided, comprising: a sound collection device, a display device, a memory, and a processor, a sound collection device, a display device, and a memory coupled to the processor; the memory for storing a computer execution code, The computer-executable code is for controlling the processor to perform the deaf-mute assist method of the first aspect.
第四方面,提供一种存储介质,其特征在于,用于储存为第二方面所述的交流辅助装置所用的计算机软件指令,其包含执行第一方面所述的聋哑人辅助方法所设计的程序代码。According to a fourth aspect, a storage medium is provided, characterized in that the computer software instructions for storing the AC assist device of the second aspect are designed to perform the deaf-mute assist method described in the first aspect. code.
第五方面,提供一种计算机程序产品,可直接加载到计算机的内部存储器中,并含有软件代码,所述计算机程序经由计算机载入并执行后能够实现第一方面所述的聋哑人辅助方法。In a fifth aspect, a computer program product is provided that can be directly loaded into an internal memory of a computer and includes software code, and the computer program can be loaded and executed by a computer to implement the deaf-mute assist method according to the first aspect. .
本发明的实施例提供的聋哑人辅助方法,首先接收声音,然后对接收到的声音行识别并根据识别结果将声音转换为显示信号,最后在显示信号的驱动下进行显示,由于本发明实施例提供的聋哑人辅助方法可以将接收到的声音转换为显示信号,并将在显示信号的 驱动下进行显示,即可以将接收到的听觉信号转换成视觉信号,进而使聋哑人通过视觉看到与声音对应的显示内容,因此通过本发明实施例提供的聋哑人辅助方法可以辅助聋哑人感知声音。此外,相比于现有技术中的聋哑人辅助设备,本发明实施例提供的聋哑人辅助方法无需复杂的甄选过程,也无需进行语言训练,因此相比于现有技术本发明实施例提供的聋哑人辅助方法可以辅助聋哑人方便、快捷的感知声音。The deaf-mute assisting method provided by the embodiment of the present invention first receives a sound, then recognizes the received sound line and converts the sound into a display signal according to the recognition result, and finally displays the display signal, which is implemented by the present invention. The deaf-mute assist method provided by the example can convert the received sound into a display signal and will display the signal The display is driven to convert the received auditory signal into a visual signal, so that the deaf-mute person can visually see the display content corresponding to the sound. Therefore, the deaf-mute assisting method provided by the embodiment of the present invention can be assisted. The dumb person perceives the sound. In addition, compared with the deaf-mute auxiliary device in the prior art, the deaf-mute assisting method provided by the embodiment of the present invention does not require a complicated selection process and does not require language training, and thus the present invention is compared to the prior art. The deaf-mute assist method provided can assist the deaf-mute person to perceive the sound conveniently and quickly.
附图说明DRAWINGS
为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the embodiments or the description of the prior art will be briefly described below. Obviously, the drawings in the following description are only It is a certain embodiment of the present invention, and other drawings can be obtained from those skilled in the art without any creative work.
图1为本发明的实施例提供的聋哑人辅助方法的步骤流程图之一;1 is a flow chart of steps of a deaf-mute assist method according to an embodiment of the present invention;
图2为本发明的实施例提供的聋哑人辅助方法的步骤流程图之二;2 is a second flowchart of steps of a deaf-mute assist method according to an embodiment of the present invention;
图3为本发明的实施例提供的聋哑人辅助方法的步骤流程图之三;3 is a third flowchart of the steps of the deaf-mute assisting method provided by the embodiment of the present invention;
图4为本发明的实施例提供的声音方位与显示位置的对应关系的示意图;4 is a schematic diagram of a correspondence relationship between a sound orientation and a display position according to an embodiment of the present invention;
图5为本发明的实施例提供的聋哑人辅助方法的步骤流程图之四;FIG. 5 is a fourth flowchart of steps of a deaf-mute assist method according to an embodiment of the present invention; FIG.
图6为本发明的实施例提供的聋哑人辅助装置的示意性结构图之一;FIG. 6 is a schematic structural diagram of a deaf-mute auxiliary device according to an embodiment of the present invention; FIG.
图7为本发明的实施例提供的聋哑人辅助装置的示意性结构图之二; FIG. 7 is a second schematic structural diagram of a deaf-mute auxiliary device according to an embodiment of the present invention; FIG.
图8为本发明的实施例提供的聋哑人辅助装置的示意性结构图之三;FIG. 8 is a third schematic structural diagram of a deaf-mute auxiliary device according to an embodiment of the present invention; FIG.
图9为本发明的实施例提供的电子设备的示意性结构图。FIG. 9 is a schematic structural diagram of an electronic device according to an embodiment of the present invention.
具体实施方式detailed description
本文中术语“和/或”,仅仅是一种描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本文中字符“/”,一般表示前后关联对象是一种“或”的关系。如果不加说明,本文中的“多个”是指两个或两个以上。The term "and/or" in this context is merely an association describing the associated object, indicating that there may be three relationships, for example, A and / or B, which may indicate that A exists separately, and both A and B exist, respectively. B these three situations. In addition, the character "/" in this article generally indicates that the contextual object is an "or" relationship. Unless otherwise stated, "multiple" in this document refers to two or more.
需要说明的是,本发明实施例中,“示例性的”或者“例如”等词用于表示作例子、例证或说明。本发明实施例中被描述为“示例性的”或者“例如”的任何实施例或设计方案不应被解释为比其它实施例或设计方案更优选或更具优势。确切而言,使用“示例性的”或者“例如”等词旨在以具体方式呈现相关概念。It should be noted that, in the embodiments of the present invention, the words "exemplary" or "such as" are used to mean an example, an illustration, or a description. Any embodiment or design described as "exemplary" or "for example" in the embodiments of the invention should not be construed as preferred or advantageous over other embodiments or designs. Rather, the use of the words "exemplary" or "such as" is intended to present the concepts in a particular manner.
需要说明的是,本发明实施例中,除非另有说明,“多个”的含义是指两个或两个以上。It should be noted that, in the embodiments of the present invention, the meaning of "a plurality" means two or more unless otherwise stated.
需要说明的是,本发明实施例中,“的(英文:of)”,“相应的(英文:corresponding,relevant)”和“对应的(英文:corresponding)”有时可以混用,应当指出的是,在不强调其区别时,其所要表达的含义是一致的。It should be noted that, in the embodiment of the present invention, "(English: of)", "corresponding (relevant)" and "corresponding" may sometimes be mixed. It should be noted that When the difference is not emphasized, the meaning to be expressed is the same.
下面将结合本发明实施例的说明书附图,对本发明实施例提供的技术方案进行说明。显然,所描述的是本发明的一部分实施例,而不是全部的实施例。需要说明的是,下文所提供的任意多个技术方案中的部分或全部技术特征在不冲突的情况下,可以结合使用,形成新的技术方案。The technical solutions provided by the embodiments of the present invention are described below in conjunction with the accompanying drawings of the embodiments of the present invention. It is apparent that some of the embodiments of the invention are described, rather than all of the embodiments. It should be noted that some or all of the technical features of any of the various technical solutions provided below may be combined to form a new technical solution without conflict.
本发明实施例所提供的技术方案的基本原理为:对接收到的声音进行识别,将接收到的声音转换为显示信号并在显示信号的驱动下将声音对应的内容显示出来,从而使聋哑人通过观看与声音对应 的视觉信息感知声音。The basic principle of the technical solution provided by the embodiment of the present invention is: identifying the received sound, converting the received sound into a display signal, and displaying the content corresponding to the sound driven by the display signal, thereby making the hoarseness People correspond to sound by watching The visual information perceives the sound.
本发明实施例提供的聋哑人辅助方法的执行主体可以为聋哑人辅助装置或者可以用于执行聋哑人辅助方法的电子设备。其中,聋哑人辅助装置可以为上述电子设备中的中央处理器(Central Processing Unit,CPU)、CPU与存储器等硬件的组合、或者可以为上述电子设备中的其他控制单元或者模块。The execution body of the deaf-mute assisting method provided by the embodiment of the present invention may be a deaf-mute auxiliary device or an electronic device that can be used to execute the deaf-mute assisting method. The deaf-mute auxiliary device may be a combination of a central processing unit (CPU), a CPU and a memory in the electronic device, or may be another control unit or module in the electronic device.
示例性的,上述电子设备可以为采用本发明实施例提供的方法进行聋哑人辅助的手机、增强现实眼镜(简称:AR眼镜)、个人计算机((personal computer,PC)、上网本、个人数字助理(英文:Personal Digital Assistant,简称:PDA)、服务器等,或者上述电子设备可以为安装有可以进行聋哑人辅助的软件客户端或软件系统或软件应用的PC、服务器等,具体的硬件实现环境可以通用计算机形式,或者是ASIC的方式,也可以是FPGA,或者是一些可编程的扩展平台例如Tensilica的Xtensa平台等等。Illustratively, the foregoing electronic device may be a mobile phone, augmented reality glasses (abbreviation: AR glasses), a personal computer (PC), a netbook, a personal digital assistant, which are assisted by the deaf-mute person by using the method provided by the embodiment of the present invention. (English: Personal Digital Assistant, PDA for short), server, etc., or the above electronic device may be a PC, a server, etc., which is installed with a software client or software system or software application that can assist the deaf person, and a specific hardware implementation environment. It can be in the form of a general-purpose computer, or an ASIC, an FPGA, or a programmable extension platform such as Tensilica's Xtensa platform.
基于上述内容,本发明的实施例提供一种聋哑人辅助方法,具体的参照图1所示,该聋哑人辅助方法包括如下步骤:Based on the above, an embodiment of the present invention provides a deaf-mute assisting method. Referring specifically to FIG. 1, the deaf-mute assisting method includes the following steps:
S11、接收声音。S11. Receive sound.
具体的,上述实施例中的声音可以为他人与用户进行交流时发出的说话声音、广播发出的说话声音等;也可以为环境中的声音,例如:汽车鸣笛的声音、狗叫的声音、打雷的声音等。Specifically, the sound in the above embodiment may be a voice that is emitted when another person communicates with the user, a voice that is broadcasted, or the like; or a voice in the environment, such as a sound of a car whistle, a barking voice, The sound of thunder and so on.
此外,上述步骤S11中具体可以通过麦克风(英文名称:microphone,简称:Mic)、Mic阵列等声音感应装置来对声音进行接收。In addition, in the above step S11, the sound can be received by a sound sensing device such as a microphone (English name: microphone, abbreviation: Mic) or a Mic array.
S12、对声音进行识别并根据识别结果将声音转换为显示信号。S12. Identify the sound and convert the sound into a display signal according to the recognition result.
具体的,上述实施例中对声音进行识别并根据识别结果将声音转换为显示信号的过程可以在聋哑人辅助装置内部完成,也可以通过远程服务设备协助完成。 Specifically, the process of recognizing the sound in the above embodiment and converting the sound into the display signal according to the recognition result may be completed inside the deaf-mute auxiliary device, or may be assisted by the remote service device.
当对声音进行识别并根据识别结果将声音转换为显示信号的过程在聋哑人辅助装置内部完成时,步骤S12具体可以通过如下步骤实现:a、通过内部的声音处理装置对声音进行识别,b根据声音处理装置的识别结果将声音转换为对应的显示信号。When the process of recognizing the sound and converting the sound into the display signal according to the recognition result is completed inside the deaf-mute auxiliary device, step S12 can be specifically realized by the following steps: a. Identifying the sound by the internal sound processing device, b The sound is converted into a corresponding display signal according to the recognition result of the sound processing device.
当对声音进行识别并根据识别结果将声音转换为显示信号的过程通过远程服务设备协助完成时,步骤S12具体可以通过如下步骤实现:c、将声音发送至远端服务器,以便与端服务器对声音进行识别并根据识别结果将声音转换为显示信号。d、接收远端服务器发送的显示信号。示例性的,远端服务设备可以为云端服务器等。When the process of recognizing the sound and converting the sound into the display signal according to the recognition result is completed by the remote service device, step S12 can be specifically implemented by: c. transmitting the sound to the remote server to make a sound with the end server. Identify and convert the sound into a display signal based on the recognition result. d. Receive the display signal sent by the remote server. Exemplarily, the remote service device can be a cloud server or the like.
可选的,上述步骤S12中将声音转换为显示信号具体可以为:将声音转换为用于显示文字的显示信号、用于显示标识的显示信号、用于显示动态画面的显示信号中的至少一种。Optionally, the converting the sound into the display signal in the step S12 may be: converting the sound into at least one of a display signal for displaying characters, a display signal for displaying the logo, and a display signal for displaying the dynamic screen. Kind.
示例性的,当接收到的声音为与用户面对面进行交流的人发出的语音时,可以将接收到的声音转换成用于显示文字的显示信号。再例如:当接收到的声音为狗叫声时,可以将接收到的声音转换成用于显示狗的卡通画等标识的显示信号。再例如:当接收到的声音为汽车行走时发出的声音时,可以将接收到的声音转换为用于显示汽车行走的动态画面的显示信号。此外,上述实施例中还可以通过多种方式的组合来更加清楚的通过视觉信息表现接收到的声音。例如:当接收到的声音为汽车行走时发出的声音时,可以将接收到的声音转换为用于显示汽车行走的动态画面以及汽车标识的显示信号。当然,在上述实施例基础上本领域技还可能想到将接收到的声音转换为其他类型的显示信号,但这都属于本发明实施例的合理变通方案,因此均应属于本发明实施例的保护范围之内。Illustratively, when the received sound is a voice uttered by a person who communicates face-to-face with the user, the received sound can be converted into a display signal for displaying the text. For another example, when the received sound is a barking sound, the received sound can be converted into a display signal for displaying a logo such as a dog's cartoon drawing. For another example, when the received sound is a sound emitted when the car is walking, the received sound can be converted into a display signal for displaying a dynamic picture of the car walking. In addition, in the above embodiments, the received sound can be expressed more clearly by visual information by a combination of various ways. For example, when the received sound is a sound emitted when the car is walking, the received sound can be converted into a display signal for displaying a moving picture of the car and a car logo. Of course, it is also possible to convert the received sound into other types of display signals based on the above embodiments, but these are all reasonable modifications of the embodiments of the present invention, and therefore should all belong to the protection of the embodiments of the present invention. Within the scope.
S13、在显示信号的驱动下进行显示。S13. Displayed under the driving of the display signal.
上述对视觉信息进行显示的具体显示方式可以基于本发明实施例提供的聋哑人辅助方法的执行主体来选择。例如:当上述实施例提供的聋哑人辅助方法的执行主体为手机时,在显示信号的驱动下进行显示可以为通过显示信号驱动手机屏幕进行显示,再例如: 当上述实施例提供的聋哑人辅助方法的执行主体为AR眼镜时,对视觉信息进行显示可以为通过显示信号驱动投影显示装置将显示内容投影在AR眼镜的镜片上进行显示。The specific display manner of displaying the visual information may be selected based on the execution subject of the deaf-mute assisting method provided by the embodiment of the present invention. For example, when the execution body of the deaf-mute assisting method provided by the above embodiment is a mobile phone, displaying under the driving of the display signal may be performed by driving the screen of the mobile phone through the display signal, for example: When the execution body of the deaf-mute assisting method provided by the above embodiment is an AR glasses, displaying the visual information may be performed by projecting the display content on the lens of the AR glasses by driving the projection display device by the display signal.
本发明的实施例提供的聋哑人辅助方法,首先接收声音,然后对接收到的声音行识别并根据识别结果将声音转换为显示信号,最后在显示信号的驱动下进行显示,由于本发明实施例提供的聋哑人辅助方法可以将接收到的声音转换为显示信号,并将在显示信号的驱动下进行显示,即可以将接收到的听觉信号转换成视觉信号,进而使聋哑人通过视觉看到与声音对应的显示内容,因此通过本发明实施例提供的聋哑人辅助方法可以辅助聋哑人感知声音。此外,相比于现有技术中的聋哑人辅助设备,本发明实施例提供的聋哑人辅助方法无需复杂的甄选过程,也无需进行语言训练,因此相比于现有技术本发明实施例提供的聋哑人辅助方法可以辅助聋哑人方便、快捷的感知声音。The deaf-mute assisting method provided by the embodiment of the present invention first receives a sound, then recognizes the received sound line and converts the sound into a display signal according to the recognition result, and finally displays the display signal, which is implemented by the present invention. The deaf-mute assist method provided by the example can convert the received sound into a display signal and display it under the driving of the display signal, that is, the received auditory signal can be converted into a visual signal, thereby enabling the deaf person to pass the vision. The display content corresponding to the sound is seen, so the deaf-mute assisting method provided by the embodiment of the present invention can assist the deaf-mute person to perceive the sound. In addition, compared with the deaf-mute auxiliary device in the prior art, the deaf-mute assisting method provided by the embodiment of the present invention does not require a complicated selection process and does not require language training, and thus the present invention is compared to the prior art. The deaf-mute assist method provided can assist the deaf-mute person to perceive the sound conveniently and quickly.
可选的,参照图2所示,上述步骤S12中对声音进行识别并根据识别结果将声音转换为显示信号,具体可以通过如下步骤实现:Optionally, referring to FIG. 2, the sound is recognized in the above step S12, and the sound is converted into a display signal according to the recognition result, which can be specifically implemented by the following steps:
S121、对声音的种类进行识别。S121. Identify the type of sound.
在步骤S121中,若通过对声音的种类进行识别后确定声音为说话声时,执行步骤S122;和/或,若通过对声音的种类进行识别后确定声音为环境声时,执行步骤S123。即,本发明实施例中的步骤S122和步骤S123可以均执行,也可以择一执行。In step S121, if it is determined that the sound is a speech sound by recognizing the type of the sound, step S122 is performed; and/or, if the sound is determined to be the ambient sound by recognizing the type of the sound, step S123 is executed. That is, step S122 and step S123 in the embodiment of the present invention may be performed either alternatively or alternatively.
需要说明的是,本发明实施例中的说话声一般是指人类在交谈、演讲、播报新闻等时发出的声音。此外,在一些情况下,说话声也可能会经过处理后被接收到,例如:在演讲时将演讲者发出的声音经过放大后输出并被接收到。虽然此类声音并不是人类直接发出的说话声,但此类声音也属于本发明实施例中的说话声。It should be noted that the speaking voice in the embodiment of the present invention generally refers to a voice that is emitted by a human when talking, speaking, broadcasting news, and the like. In addition, in some cases, the voice may be received after being processed. For example, the voice of the speaker is amplified and output and received at the time of the lecture. Although such sounds are not spoken directly by humans, such sounds also belong to the voices of the embodiments of the present invention.
还需要说明的是,本发明实施例中的环境声音是区别于说话声的其他声音,即将接收到的声音分为说话声和环境声。具体的,环 境声可以为:汽车鸣笛的声音、狗叫的声音、打雷的声音、环境中的噪声等。It should also be noted that the ambient sound in the embodiment of the present invention is other sounds different from the talking sound, and the received sound is divided into a talking sound and an ambient sound. Specifically, the ring The sound of the environment can be: the sound of the car whistling, the sound of the dog barking, the sound of thunder, the noise in the environment, and so on.
S122、识别说话声的内容,根据说话声的内容将说话声转换为用于驱动显示文字的显示信号。S122. Identify the content of the speech sound, and convert the speech sound into a display signal for driving the display text according to the content of the speech sound.
可选的,上述实施例中识别说话声的内容具体可以通过如下步骤实现:e、通过语种类型识别技术判断接收到声音的语种类型,例如:将接收到的声音识别为中文、英文、法文等。f、根据接收到声音的语种类型与具体接收到的声音识别说话内容。即,当接收到的声音为说话声时,可以先识别说话声的语种类型再识别说话的具体内容。Optionally, the content of the voice recognition in the foregoing embodiment may be specifically implemented by: e. determining, by using a language type recognition technology, a language type of the received voice, for example, identifying the received voice as Chinese, English, French, etc. . f. Identify the spoken content according to the type of language in which the sound is received and the specific received voice. That is, when the received sound is a voice, the language type of the voice can be recognized first to identify the specific content of the voice.
由于说话内容往往比较复杂,通过标识、动态画面等方法很难清楚的对应说话内容进行显示,因此在本发明实施中,当声音为说话声时,将根据说话声的内容将声音转为文字,从而更加清楚的显示出接收到的说话声的内容。Since the content of the speech is often complicated, it is difficult to clearly display the corresponding content by means of a logo, a dynamic picture, etc., so in the implementation of the present invention, when the sound is a voice, the sound is converted into a text according to the content of the voice. Thereby the content of the received speech is more clearly displayed.
S123、识别环境声的类别,根据环境声的类别将环境声转换为用于驱动显示标识的显示信号。S123. Identify a category of the ambient sound, and convert the ambient sound into a display signal for driving the display identifier according to the category of the ambient sound.
示例性的,上述实施例中的标识具体可以为:狗的卡通画、汽车的卡通画、危险标识、雷电标识等。Exemplarily, the identifier in the above embodiment may specifically be: a cartoon drawing of a dog, a cartoon drawing of a car, a danger sign, a lightning sign, and the like.
进一步的,上述实施例提供的聋哑人辅助方法可以辅助聋哑人感知说话声音以及环境中的各种声音,但是当用户处于嘈杂的环境中接受到说话声时,接收到的说话声中可能包含有环境中的噪音,进而可能会造成说话声的内容识别不准确。针对上述问题本发明实施例提供了一种聋哑人辅助方法,具体的,参照图3所示,在图2所示聋哑人辅助方的基础上本发明实施例提供的聋哑人辅助方法还进一步包括:Further, the deaf-mute assisting method provided by the above embodiment can assist the deaf-mute person to perceive the speaking sound and various sounds in the environment, but when the user receives the speaking sound in a noisy environment, the received speech may be Contains noise from the environment, which may cause inaccurate content recognition of the voice. The present invention provides a deaf-mute assisting method according to an embodiment of the present invention. Specifically, referring to FIG. 3, the deaf-mute assisting method provided by the embodiment of the present invention is provided on the basis of the deaf-mute assistant shown in FIG. It further includes:
S31、获取相对人的图像。S31. Obtain an image of a relative person.
其中,相对人为发出说话声的人。 Among them, the relative person is the person who makes the voice.
具体的,可以通过单目摄像头、双目摄像头、深度摄像头、图像传感器等的一种或多种来获取相对人的图像。本发明实施例中可以采用任一种图像采集装置还获取相对人的图像,本发明实施例中对获取相对人的图像的方式不作限定,以能够获取相对人的图像为准。此外,示例性的,相对人的图像可以为相对人说话时的动态画面。Specifically, an image of a relative person may be acquired by one or more of a monocular camera, a binocular camera, a depth camera, an image sensor, and the like. In the embodiment of the present invention, any image capturing device may be used to obtain an image of a relative person. In the embodiment of the present invention, the manner of acquiring the image of the opposite person is not limited, so that the image of the opposite person can be obtained. Moreover, the exemplary, relative person's image may be a dynamic picture when speaking to a person.
S32、根据相对人的图像获取相对人的嘴唇动作。S32. Acquire a lip motion of the opposite person according to the image of the relative person.
上述步骤S122中根据所述说话声的内容将所述说话声转换为用于驱动显示文字的显示信号,可以通过步骤S33所提供的实现方式来实现。In the above step S122, converting the speech sound into a display signal for driving the display character according to the content of the speech sound can be realized by the implementation provided in step S33.
S33、根据所述说话声的内容和所述相对人的嘴唇动作将所述说话声转换为用于驱动显示文字的显示信号。S33. Convert the speech sound into a display signal for driving the display text according to the content of the speech sound and the lip motion of the opposite person.
上述实施例中通过进一步获取获取相对人的图像,对相对人的图像进行识别获取相对人的嘴唇动作,然后在声音为说话声时,识别说话声的内容,以及根据说话声的内容和相对人的嘴唇动作将说话声的内容转换为与文字对应的显示信号,由于通过唇语识别技术可以根据相对人的嘴唇动作解读部分相对人说的话,因此可以提高转换的准确性。In the above embodiment, by acquiring the image of the relative person, the image of the opposite person is recognized to obtain the lip motion of the opposite person, and then when the sound is the voice, the content of the voice is recognized, and according to the content of the voice and the relative person The lip action converts the content of the speech into a display signal corresponding to the text, and since the lip language recognition technology can interpret the relative words spoken by the relative person's lip motion, the accuracy of the conversion can be improved.
进一步的,上述实施例提供的聋哑人辅助方法还包括:Further, the deaf-mute assisting method provided by the foregoing embodiment further includes:
获取声音的方位;Obtain the orientation of the sound;
上述步骤S13中在所述显示信号的驱动下进行显示具体可以通过如下实现方式来实现:根据所述声音的方位在所述显示信号的驱动下在显示界面的相应位置上进行显示。The display in the above step S13 under the driving of the display signal may be specifically implemented by displaying the position of the sound under the driving of the display signal at a corresponding position of the display interface.
示例性的,参照图4所示,当声音所在方为位于用户后方F1时,将与声音对应的显示内容41在显示界面40的下方;当声音所在方为位于用户前方F2时,将与声音对应的显示内容42显示在显示界面40的上方;当声音所在方为位于用户左侧F3时,将与声音 对应的显示内容43显示在显示界面40的左侧;当声音所在方为位于用户右侧F4时,将与声音对应的显示内容44显示在显示界面40的右侧。Exemplarily, referring to FIG. 4, when the sound is located at the rear F1 of the user, the display content 41 corresponding to the sound is below the display interface 40; when the sound is located at the front of the user F2, the sound is The corresponding display content 42 is displayed above the display interface 40; when the sound is located on the left side of the user F3, the sound and the sound The corresponding display content 43 is displayed on the left side of the display interface 40; when the sound is located on the right side F4 of the user, the display content 44 corresponding to the sound is displayed on the right side of the display interface 40.
根据声音的方位将显示内容在显示界面的相应位置上进行显示,可以进一步使用户了解到声音产生的方位,进而可以辅助聋哑人更加全面的对声音进行感知。According to the orientation of the sound, the display content is displayed on the corresponding position of the display interface, which can further enable the user to know the orientation of the sound, thereby helping the deaf-mute person to perceive the sound more comprehensively.
聋哑人可以通过手语与正常人进行交流、沟通,但当聋哑人面对一些对手语不了解的人时,交流、沟通将无法进行。针对上述问题,本发明实施例进一步提供了一种聋哑人辅助方法,具体的,参照图5所示,本发明实施例提供的聋哑人辅助方法包括:Deaf people can communicate and communicate with normal people through sign language, but when deaf people face people who do not understand their opponents, communication and communication will not be possible. The embodiment of the present invention further provides a deaf-mute assisting method. Specifically, referring to FIG. 5, the deaf-mute assisting method provided by the embodiment of the present invention includes:
S51、检测用户的手部动作。S51. Detecting a user's hand motion.
可选的,检测用户的手部动作具体可以为:通过单目摄像头、双目摄像头、深度摄像头、图像传感器等的一种或多种来获取用户的动态画面,进而根据用户的动态画面获取用户的手部动作。此外,检测用户的手部动作还可以通过手部穿戴设备,检测用户手部的加速度、旋转角度等运动参数,并根据运动参数来获取用户的手部动作。其中,手部穿戴设备可以为:指环、腕带、数据手套等。Optionally, detecting the user's hand motion may specifically: acquiring a dynamic picture of the user by one or more of a monocular camera, a binocular camera, a depth camera, an image sensor, etc., and then acquiring the user according to the dynamic picture of the user. Hand movements. In addition, detecting the user's hand movement can also detect the motion parameters such as the acceleration and the rotation angle of the user's hand through the hand wearing device, and acquire the user's hand motion according to the motion parameter. Among them, the hand wearing device can be: a ring, a wristband, a data glove, and the like.
S52、对用户的手部动作进行识别并根据识别结果将用户的手部动作转换为语音。S52. Identify the user's hand motion and convert the user's hand motion into voice according to the recognition result.
同样,上述实施例中对用户的手部动作进行识别并根据识别结果将用户的手部动作转换为语音的过程可以在聋哑人辅助装置内部完成,也可以通过远程服务设备协助完成。Similarly, the process of recognizing the user's hand motion in the above embodiment and converting the user's hand motion into voice according to the recognition result may be completed inside the deaf-mute auxiliary device, or may be assisted by the remote service device.
当对用户的手部动作进行识别并根据识别结果将用户的手部动作转换为语音的过程在聋哑人辅助装置内部完成时,步骤S52具体可以通过如下步骤实现:A、通过聋哑人辅助装置内部的图像处理装置对用户的手部动作进行识别。B根据图像处理装置的识别结果将用户的手部动作转换为对应的语音。 When the process of recognizing the user's hand motion and converting the user's hand motion into voice according to the recognition result is completed inside the deaf-mute auxiliary device, step S52 can be specifically implemented by the following steps: A, assisted by the deaf person The image processing device inside the device recognizes the user's hand motion. B converts the user's hand motion into a corresponding voice according to the recognition result of the image processing apparatus.
当对图像进行识别并根据识别结果将图像转换为声音信息的过程通过远程服务设备协助完成时,步骤S52具体可以通过如下步骤实现:C、将图像发送至远端服务器,以便远端服务器对用户的手部动作进行识别并根据识别结果将用户的手部动作转换为语音。D、接收远端服务器发送的语音。示例性的,远端服务设备可以为云端服务器等。When the process of recognizing the image and converting the image into the sound information according to the recognition result is completed by the remote service device, the step S52 may be specifically implemented by: sending the image to the remote server, so that the remote server is the user. The hand motion recognizes and converts the user's hand motion into voice based on the recognition result. D. Receive the voice sent by the remote server. Exemplarily, the remote service device can be a cloud server or the like.
S53、对语音进行播报。S53. Broadcast the voice.
具体的,可以通过语音合成技术将手势表达的手语内容转换为语音,并通过扬声器(英文名称:Speaker)将语音播报出来。Specifically, the sign language content of the gesture expression can be converted into voice by voice synthesis technology, and the voice is broadcasted through a speaker (English name: Speaker).
上述实施例中可以将手语内容转换成语音并播报出来,所以可以使不了解手语的人通过播报出来的语音获知手语聋哑人通手语所表达的内容,进而进一步辅助聋哑人进行沟通。In the above embodiment, the sign language content can be converted into a voice and broadcasted, so that the person who does not understand the sign language can learn the content expressed by the sign language deaf and dumb person through the broadcasted voice, thereby further assisting the deaf and mute person to communicate.
下面说明本发明实施例提供的与上文所提供的方法实施例相对应的装置实施例。需要说明的是,下述装置实施例中相关内容的解释,均可以参考上述方法实施例。The device embodiments corresponding to the method embodiments provided above are provided in the following description of the embodiments of the present invention. It should be noted that the explanation of the related content in the following device embodiments can refer to the foregoing method embodiments.
在采用对应各个功能划分各个功能模块的情况下,图6示出了上述实施例中所涉及的聋哑人辅助装置的一种可能的结构示意图。参照图6所示,聋哑人辅助装置包括:In the case where the respective functional modules are divided by corresponding functions, FIG. 6 shows a possible structural diagram of the deaf-mute assisting device involved in the above embodiment. Referring to Figure 6, the deaf-mute aid includes:
接收单元61,用于接收声音;a receiving unit 61, configured to receive a sound;
转换单元62,用于对声音进行识别并根据识别结果将声音转换为显示信号;The converting unit 62 is configured to identify the sound and convert the sound into a display signal according to the recognition result;
显示单元63,用于在显示信号的驱动下进行显示。The display unit 63 is configured to perform display under the driving of the display signal.
本发明实施例提供的聋哑人辅助装置包括:接收单元、转换单元以及显示单元,其中,接收单元用于接收声音,转换单元用于对声音进行识别并根据识别结果将声音转换为显示信号,显示单元用于在显示信号的驱动下进行显示,所以本发明实施例提供的聋哑人辅助装置可以将接收到的听觉信号转换成视觉信号,进而使聋哑人 通过视觉看到与声音对应的显示内容,因此通过本发明实施例提供的聋哑人辅助装置可以辅助聋哑人感知声音。此外,相比于现有技术中的聋哑人辅助设备,本发明实施例提供的聋哑人辅助装置无需复杂的甄选过程,也无需进行语言训练,因此相比于现有技术本发明实施例提供的聋哑人辅助装置可以辅助聋哑人方便、快捷的感知声音。The deaf-mute auxiliary device provided by the embodiment of the present invention includes: a receiving unit, a converting unit, and a display unit, wherein the receiving unit is configured to receive a sound, and the converting unit is configured to identify the sound and convert the sound into a display signal according to the recognition result, The display unit is configured to display under the driving of the display signal, so the deaf-mute auxiliary device provided by the embodiment of the present invention can convert the received auditory signal into a visual signal, thereby making the deaf person By visually seeing the display content corresponding to the sound, the deaf-mute assisting device provided by the embodiment of the present invention can assist the deaf-mute person to perceive the sound. In addition, the deaf-mute auxiliary device provided by the embodiment of the present invention does not require a complicated selection process and does not need to perform language training, and thus is compared with the prior art embodiment of the present invention. The deaf-mute aids are provided to assist the deaf and mute to sense sound easily and quickly.
可选的,转换单元62具体用于对声音的种类进行识别;Optionally, the converting unit 62 is specifically configured to identify a type of the sound;
转换单元62具体用于当声音为说话声时,识别说话声的内容,根据说话声的内容将说话声转换为用于驱动显示文字的显示信号;和/或当声音为环境声时,识别环境声的类别,根据环境声的类别将环境声转换为用于驱动显示标识的显示信号。The converting unit 62 is specifically configured to recognize the content of the talking sound when the sound is a talking sound, convert the talking sound into a display signal for driving the displayed text according to the content of the talking sound; and/or identify the environment when the sound is an ambient sound The category of the sound, which converts the ambient sound into a display signal for driving the display identifier according to the category of the ambient sound.
可选的,接收单元61还用于获取相对人的图像;其中,相对人为发出说话声的人;Optionally, the receiving unit 61 is further configured to acquire an image of a relative person; wherein, the opposite person is a person who emits a voice;
转换单元62还用于根据相对人的图像获取相对人的嘴唇动作;The converting unit 62 is further configured to acquire a lip motion of the opposite person according to the image of the opposite person;
所述转换单元62具体用于根据所述说话声的内容和所述相对人的嘴唇动作将所述说话声转换为用于驱动显示文字的显示信号。The converting unit 62 is specifically configured to convert the speaking sound into a display signal for driving the display text according to the content of the speaking sound and the lip motion of the opposite person.
可选的,接收单元61还用于获取声音的方位;Optionally, the receiving unit 61 is further configured to acquire an orientation of the sound;
所述显示单元63还用于根据所述声音的方位在所述显示信号的驱动下在显示界面的相应位置上进行显示。The display unit 63 is further configured to perform display on a corresponding position of the display interface under the driving of the display signal according to the orientation of the sound.
可选的,参照图7所示,转换单元62包括:发送模块71和接收模块72;Optionally, referring to FIG. 7, the converting unit 62 includes: a sending module 71 and a receiving module 72;
发送模块71用于将声音发送至远端服务器,以便与端服务器对声音进行识别并根据识别结果将声音转换为显示信号;The sending module 71 is configured to send the sound to the remote server to identify the sound with the end server and convert the sound into a display signal according to the recognition result;
接收模块72用于接收远端服务器发送的显示信号。The receiving module 72 is configured to receive a display signal sent by the remote server.
可选的,参照图8所示,聋哑人辅助装置600还包括:语音播 报单元64;Optionally, referring to FIG. 8, the deaf-mute assistant 600 further includes: voice broadcast Reporting unit 64;
接收单元61还用于检测用户的手部动作;The receiving unit 61 is further configured to detect a user's hand motion;
识别单元62还用于对用户的手部动作进行识别并根据识别结果将用户的手部动作转换为语音;The identification unit 62 is further configured to identify a user's hand motion and convert the user's hand motion into voice according to the recognition result;
语音播报单元64用于对语音进行播报。The voice broadcast unit 64 is configured to broadcast the voice.
即,接收单元61用于实现上述聋哑人辅助方法中的接收声音、获取相对人的图像以及获取声音的方位的步骤;识别单元62用于实现上述聋哑人辅助方法中的对声音进行识别并根据识别结果将声音转换为显示信号、对声音的种类进行识别、识别说话声的内容,根据说话声的内容将说话声转换为用于驱动显示文字的显示信号、识别环境声的类别,根据环境声的类别将环境声转换为用于驱动显示标识的显示信号、根据相对人的图像获取相对人的嘴唇动作、根据说话声的内容和相对人的嘴唇动作将说话声的内容转换为与文字对应的显示信号以及对用户的手部动作进行识别并根据识别结果将用户的手部动作转换为语音的步骤;发送模块71用于实现上述聋哑人辅助方法中的将声音发送至远端服务器的步骤;发送模块72用于实现上述聋哑人辅助方法中的接收远端服务器发送的显示信号的步骤;显示单元63用于实现上述聋哑人辅助方法中的在显示信号的驱动下进行显示和根据环境声的方位将标识显示在显示界面的相应位置的步骤,语音播报单元64用于实现上述聋哑人辅助方法中的对语音进行播报的步骤。That is, the receiving unit 61 is configured to implement the steps of receiving the sound, acquiring the image of the opposite person, and acquiring the orientation of the sound in the above-described deaf-mute assisting method; the identifying unit 62 is configured to implement the sound recognition in the above-described deaf-mute assisting method And converting the sound into a display signal according to the recognition result, recognizing the type of the sound, recognizing the content of the speech sound, converting the speech sound into a display signal for driving the display text, and identifying the environment sound according to the content of the speech sound, according to The category of the ambient sound converts the ambient sound into a display signal for driving the display mark, acquires the lip motion of the opposite person according to the image of the opposite person, converts the content of the voice according to the content of the voice and the lip motion of the opposite person into the text Corresponding display signal and a step of recognizing the user's hand motion and converting the user's hand motion into voice according to the recognition result; the sending module 71 is configured to implement the sound transmission to the remote server in the deaf-mute assist method The sending module 72 is configured to implement the receiving remote service in the above-mentioned deaf-mute assist method The step of displaying the display signal by the device; the display unit 63 is configured to implement the step of displaying the display under the driving of the display signal and displaying the identification on the corresponding position of the display interface according to the orientation of the ambient sound in the above-mentioned deaf-mute assisting method, the voice broadcast The unit 64 is configured to implement the step of broadcasting the voice in the above-described deaf-mute assist method.
还需说明的是,上述方法实施例涉及的各步骤的所有相关内容均可以援引到对应功能模块的功能描述,在此不再赘述。It should be noted that all the related content of the steps involved in the foregoing method embodiments may be referred to the functional description of the corresponding functional modules, and details are not described herein again.
在硬件实现上,上述的接收单元61可以是Mic、Mic阵列、摄像头、图像传感器、超声波检测装置、红外摄像头等中的一种或多种。识别单元62可以处理器或者收发机;显示单元63可以是显示屏、激光投影显示装置;语音播报单元64可以是扬声器等。上述聋哑人辅助装置所执行的动作所对应的程序均可以以软件形式存储于 聋哑人辅助装置的存储器中,以便于处理器调用执行以上各个单元对应的操作。In hardware implementation, the receiving unit 61 may be one or more of a Mic, a Mic array, a camera, an image sensor, an ultrasonic detecting device, an infrared camera, and the like. The identification unit 62 may be a processor or a transceiver; the display unit 63 may be a display screen, a laser projection display device; the voice announcement unit 64 may be a speaker or the like. The programs corresponding to the actions performed by the above-mentioned deaf-mute auxiliary device can be stored in software. The deaf-mute auxiliary device is in the memory, so that the processor calls to perform the operations corresponding to the above respective units.
在采用集成的单元的情况下,图9示出了包括上述实施例中所涉及的聋哑人辅助装置的电子设备的可能的结构示意图。电子设备900包括:处理器91、存储器92、系统总线93、通信接口94、声音采集设备95、显示设备96。In the case of employing an integrated unit, FIG. 9 shows a possible structural diagram of an electronic device including the deaf-mute aid device involved in the above embodiment. The electronic device 900 includes a processor 91, a memory 92, a system bus 93, a communication interface 94, a sound collection device 95, and a display device 96.
上述处理器91可以是一个处理器,也可以是多个处理元件的统称。例如,处理器91可以为中央处理器(central processing unit,CPU)。处理器91也可以为其他通用处理器、数字信号处理器(digital signal processing,DSP)、专用集成电路(application specific integrated circuit,ASIC)、现场可编程门阵列(field-programmable gate array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等,其可以实现或执行结合本发明公开内容所描述的各种示例性的逻辑方框,模块和电路。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。处理器91还可以为专用处理器,该专用处理器可以包括基带处理芯片、射频处理芯片等中的至少一个。处理器也可以是实现计算功能的组合,例如包含一个或多个微处理器组合,DSP和微处理器的组合等等。进一步地,该专用处理器还可以包括具有该装置其他专用处理功能的芯片。The processor 91 may be a processor or a collective name of a plurality of processing elements. For example, the processor 91 can be a central processing unit (CPU). The processor 91 can also be other general purpose processors, digital signal processing (DSP), application specific integrated circuits (ASICs), field-programmable gate arrays (FPGAs), or Other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, and the like, can implement or perform the various illustrative logical blocks, modules, and circuits described in connection with the present disclosure. The general purpose processor may be a microprocessor or the processor or any conventional processor or the like. The processor 91 may also be a dedicated processor, which may include at least one of a baseband processing chip, a radio frequency processing chip, and the like. The processor can also be a combination of computing functions, for example, including one or more microprocessor combinations, a combination of a DSP and a microprocessor, and the like. Further, the dedicated processor may also include a chip having other specialized processing functions of the device.
存储器92用于存储计算机执行代码,处理器91与存储器92通过系统总线93连接,当电子设备运行时,处理器91用于执行存储器92存储的计算机执行代码,以执行本发明实施例提供的任意一种聋哑人辅助方法,如,处理器91用于支持电子设备执行图1所示的步骤S12、图2所示的步骤S121、S122、123、图3所示的步骤S32、S33以及图5所示的步骤S52,和/或用于本文所描述的技术的其它过程,具体的聋哑人辅助方法可参考上文及附图中的相关描述,此处不再赘述。The memory 92 is used to store computer execution code, and the processor 91 is connected to the memory 92 through the system bus 93. When the electronic device is in operation, the processor 91 is configured to execute the computer execution code stored in the memory 92 to execute any of the embodiments provided by the embodiments of the present invention. A deaf-mute assisting method, for example, the processor 91 is configured to support the electronic device to perform step S12 shown in FIG. 1, steps S121, S122, and 123 shown in FIG. 2, and steps S32 and S33 shown in FIG. Step S52 shown in FIG. 5, and/or other processes for the techniques described herein, the specific deaf-mute assisting method may refer to the related descriptions above and in the drawings, and details are not described herein again.
系统总线93可以包括数据总线、电源总线、控制总线和信号状态总线等。本实施例中为了清楚说明,在图9中将各种总线都示 意为系统总线93。 System bus 93 can include a data bus, a power bus, a control bus, and a signal status bus. In the present embodiment, for the sake of clarity, various buses are shown in FIG. Means system bus 93.
通信接口94具体可以是该装置上的收发器。该收发器可以为无线收发器。例如,无线收发器可以是该装置的天线等。处理器91通过通信接口94与其他设备,例如,若该装置为该电子设备中的一个模块或组件时,该装置用于与该电子设备中的其他模块之间进行数据交互。 Communication interface 94 may specifically be a transceiver on the device. The transceiver can be a wireless transceiver. For example, the wireless transceiver can be an antenna or the like of the device. The processor 91 communicates with other devices via the communication interface 94, for example, if the device is a module or component of the electronic device, the device is for data interaction with other modules in the electronic device.
结合本发明公开内容所描述的方法的步骤可以硬件的方式来实现,也可以是由处理器执行软件指令的方式来实现。本发明实施例还提供一种存储介质,用于储存为图9所示的电子设备所用的计算机软件指令,其包含执行上述任一实施例提供的聋哑人辅助方法所设计的程序代码。其中,软件指令可以由相应的软件模块组成,软件模块可以被存放于随机存取存储器(英文:random access memory,缩写:RAM)、闪存、只读存储器(英文:read only memory,缩写:ROM)、可擦除可编程只读存储器(英文:erasable programmable ROM,缩写:EPROM)、电可擦可编程只读存储器(英文:electrically EPROM,缩写:EEPROM)、寄存器、硬盘、移动硬盘、只读光盘(CD-ROM)或者本领域熟知的任何其它形式的存储介质中。一种示例性的存储介质耦合至处理器,从而使处理器能够从该存储介质读取信息,且可向该存储介质写入信息。当然,存储介质也可以是处理器的组成部分。处理器和存储介质可以位于ASIC中。另外,该ASIC可以位于核心网接口设备中。当然,处理器和存储介质也可以作为分立组件存在于核心网接口设备中。The steps of the method described in connection with the present disclosure may be implemented in a hardware manner, or may be implemented by a processor executing software instructions. The embodiment of the present invention further provides a storage medium for storing computer software instructions for use in the electronic device shown in FIG. 9, which includes program code designed to execute the deaf-mute assist method provided by any of the above embodiments. The software instructions may be composed of corresponding software modules, and the software modules may be stored in a random access memory (English: random access memory, abbreviation: RAM), flash memory, read only memory (English: read only memory, abbreviation: ROM) , erasable programmable read-only memory (English: erasable programmable ROM, abbreviation: EPROM), electrically erasable programmable read-only memory (English: electrical EPROM, abbreviation: EEPROM), registers, hard disk, mobile hard disk, CD-ROM (CD-ROM) or any other form of storage medium known in the art. An exemplary storage medium is coupled to the processor to enable the processor to read information from, and write information to, the storage medium. Of course, the storage medium can also be an integral part of the processor. The processor and the storage medium can be located in an ASIC. Additionally, the ASIC can be located in a core network interface device. Of course, the processor and the storage medium may also exist as discrete components in the core network interface device.
本发明实施例还提供一种计算机程序产品,该计算机程序可直接加载到计算机的内部存储器中,并含有软件代码,计算机程序经由计算机载入并执行后能够实现上述任一实施例提供的聋哑人辅助方法。The embodiment of the invention further provides a computer program product, which can be directly loaded into the internal memory of the computer and contains software code, and the computer program can be loaded and executed by the computer to implement the hoarseness provided by any of the above embodiments. Human assisted method.
本领域技术人员应该可以意识到,在上述一个或多个示例中,本发明所描述的功能可以用硬件、软件、固件或它们的任意组合来实现。当使用软件实现时,可以将这些功能存储在计算机可读介质中或者作为计算机可读介质上的一个或多个指令或代码进行传输。 计算机可读介质包括计算机存储介质和通信介质,其中通信介质包括便于从一个地方向另一个地方传送计算机程序的任何介质。存储介质可以是通用或专用计算机能够存取的任何可用介质。Those skilled in the art will appreciate that in one or more examples described above, the functions described herein can be implemented in hardware, software, firmware, or any combination thereof. When implemented in software, the functions may be stored in a computer readable medium or transmitted as one or more instructions or code on a computer readable medium. Computer readable media includes both computer storage media and communication media including any medium that facilitates transfer of a computer program from one location to another. A storage medium may be any available media that can be accessed by a general purpose or special purpose computer.
以上所述,仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到的变化或替换,都应涵盖在本发明的保护范围之内。因此,本发明的保护范围应以权利要求的保护范围为准。 The above is only a specific embodiment of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily think of changes or substitutions within the technical scope of the present invention. All should be covered by the scope of the present invention. Therefore, the scope of protection of the present invention should be determined by the scope of the claims.

Claims (15)

  1. 一种聋哑人辅助方法,其特征在于,包括:A deaf-mute assisting method, characterized in that it comprises:
    接收声音;Receiving sound
    对所述声音进行识别并根据识别结果将所述声音转换为显示信号;Identifying the sound and converting the sound into a display signal according to the recognition result;
    在所述显示信号的驱动下进行显示。Display is performed under the driving of the display signal.
  2. 根据权利要求1所述的方法,其特征在于,所述对所述声音进行识别并根据识别结果将所述声音转换为显示信号,包括:The method according to claim 1, wherein said identifying said sound and converting said sound into a display signal according to the recognition result comprises:
    对所述声音的种类进行识别;Identifying the type of the sound;
    当所述声音为说话声时,识别所述说话声的内容,根据所述说话声的内容将所述说话声转换为用于驱动显示文字的显示信号;和/或,当所述声音为环境声时,识别所述环境声的类别,根据所述环境声的类别将所述环境声转换为用于驱动显示标识的显示信号。When the sound is a talking sound, identifying the content of the speaking sound, converting the speaking sound into a display signal for driving the displayed text according to the content of the speaking sound; and/or when the sound is an environment In the case of sound, the category of the ambient sound is identified, and the ambient sound is converted into a display signal for driving the display identifier according to the category of the ambient sound.
  3. 根据权利要求2所述的方法,其特征在于,所述方法还包括:The method of claim 2, wherein the method further comprises:
    获取相对人的图像;其中,所述相对人为发出说话声的人;Obtaining an image of a relative person; wherein the relative person is a person who makes a voice;
    根据所述相对人的图像获取所述相对人的嘴唇动作;Acquiring the lip motion of the opposite person according to the image of the opposite person;
    所述根据所述说话声的内容将所述说话声转换为用于驱动显示文字的显示信号,包括:Converting the voice into a display signal for driving the displayed text according to the content of the voice, including:
    根据所述说话声的内容和所述相对人的嘴唇动作将所述说话声转换为用于驱动显示文字的显示信号。The speech sound is converted into a display signal for driving the display text according to the content of the speech sound and the lip motion of the opposite person.
  4. 根据权利要求1所述的方法,其特征在于,所述方法还包括:The method of claim 1 further comprising:
    获取所述声音的方位;Obtaining the orientation of the sound;
    所述在所述显示信号的驱动下进行显示,包括: The displaying is performed under the driving of the display signal, including:
    根据所述声音的方位在所述显示信号的驱动下在显示界面的相应位置上进行显示。Displaying is performed at a corresponding position of the display interface under the driving of the display signal according to the orientation of the sound.
  5. 根据权利要求1所述的方法,其特征在于,所述对所述声音进行识别并根据识别结果将所述声音转换为显示信号包括:The method according to claim 1, wherein the recognizing the sound and converting the sound into a display signal according to the recognition result comprises:
    将所述声音发送至远端服务器,以便所述与端服务器对所述声音进行识别并根据识别结果将所述声音转换为显示信号;Transmitting the sound to a remote server, so that the end server identifies the sound and converts the sound into a display signal according to the recognition result;
    接收所述远端服务器发送的显示信号。Receiving a display signal sent by the remote server.
  6. 根据权利要求1所述的方法,其特征在于,所述方法还包括:The method of claim 1 further comprising:
    检测用户的手部动作;Detecting the user's hand movements;
    对所述用户的手部动作进行识别并根据识别结果将所述用户的手部动作转换为语音;Identifying the user's hand motion and converting the user's hand motion into voice according to the recognition result;
    对所述语音进行播报。The voice is broadcast.
  7. 一种聋哑人辅助装置,其特征在于,包括:A deaf-mute auxiliary device, comprising:
    接收单元,用于接收声音;a receiving unit, configured to receive a sound;
    转换单元,用于对所述声音进行识别并根据识别结果将所述声音转换为显示信号;a converting unit, configured to identify the sound and convert the sound into a display signal according to the recognition result;
    显示单元,用于在所述显示信号的驱动下进行显示。a display unit for displaying under the driving of the display signal.
  8. 根据权利要求7所述的装置,其特征在于,The device of claim 7 wherein:
    所述转换单元具体用于对所述声音的种类进行识别;The converting unit is specifically configured to identify the type of the sound;
    所述转换单元具体用于当所述声音为说话声时,识别所述说话声的内容,根据所述说话声的内容将所述说话声转换为用于驱动显示文字的显示信号;和/或,当所述声音为环境声时,识别所述环境声的类别,根据所述环境声的类别将所述环境声转换为用于驱动显示标识的显示信号。 The converting unit is specifically configured to: when the sound is a speaking sound, identify the content of the speaking sound, convert the speaking sound into a display signal for driving the displayed text according to the content of the speaking sound; and/or And when the sound is an ambient sound, identifying a category of the ambient sound, and converting the ambient sound into a display signal for driving the display identifier according to the category of the ambient sound.
  9. 根据权利要求8所述的装置,其特征在于,所述接收单元还用于获取相对人的图像;其中,所述相对人为发出说话声的人;The device according to claim 8, wherein the receiving unit is further configured to acquire an image of a relative person; wherein the relative person is a person who emits a voice;
    所述转换单元还用于根据所述相对人的图像获取所述相对人的嘴唇动作;The converting unit is further configured to acquire a lip motion of the opposite person according to the image of the opposite person;
    所述转换单元具体用于根据所述说话声的内容和所述相对人的嘴唇动作将所述说话声转换为用于驱动显示文字的显示信号。The converting unit is specifically configured to convert the speaking sound into a display signal for driving the display text according to the content of the speaking sound and the lip motion of the opposite person.
  10. 根据权利要求7所述的装置,其特征在于,The device of claim 7 wherein:
    所述接收单元还用于获取所述声音的方位;The receiving unit is further configured to acquire an orientation of the sound;
    所述显示单元还用于根据所述声音的方位在所述显示信号的驱动下在显示界面的相应位置上进行显示。The display unit is further configured to display at a corresponding position of the display interface under the driving of the display signal according to the orientation of the sound.
  11. 根据权利要求7所述的装置,其特征在于,所述转换单元包括:发送模块和接收模块;The apparatus according to claim 7, wherein the converting unit comprises: a transmitting module and a receiving module;
    所述发送模块用于将所述所述声音发送至远端服务器,以便所述与端服务器对所述声音进行识别并根据识别结果将所述声音转换为显示信号;The sending module is configured to send the sound to a remote server, so that the end server identifies the sound and converts the sound into a display signal according to the recognition result;
    所述接收模块用于接收所述远端服务器发送的显示信号。The receiving module is configured to receive a display signal sent by the remote server.
  12. 根据权利要求7所述的装置,其特征在于,所述装置还包括:语音播报单元;The device according to claim 7, wherein the device further comprises: a voice broadcast unit;
    所述接收单元还用于检测用户的手部动作;The receiving unit is further configured to detect a user's hand motion;
    所述识别单元还用于对所述用户的手部动作进行识别并根据识别结果将所述用户的手部动作转换为语音;The identification unit is further configured to identify the user's hand motion and convert the user's hand motion into voice according to the recognition result;
    所述语音播报单元用于对所述语音进行播报。The voice broadcast unit is configured to broadcast the voice.
  13. 一种电子设备,其特征在于,包括:声音采集设备、显示设备、存储器和处理器,声音采集设备、显示设备和存储器耦合至所述处理器;所述存储器用于存储计算机执行代码,所述计算机执 行代码用于控制所述处理器执行权利要求1-6任一项所述的聋哑人辅助方法。An electronic device, comprising: a sound collection device, a display device, a memory and a processor, a sound collection device, a display device and a memory coupled to the processor; the memory for storing computer execution code, Computer The line code is for controlling the processor to perform the deaf-mute assist method of any of claims 1-6.
  14. 一种存储介质,其特征在于,用于储存为权利要求7-12任一项所述的交流辅助装置所用的计算机软件指令,其包含执行权利要求1-6任一项所述的聋哑人辅助方法所设计的程序代码。A storage medium for storing the computer software instructions for use in the communication assisting device of any one of claims 7 to 12, comprising the deaf-mute person according to any one of claims 1-6 The program code designed by the helper method.
  15. 一种计算机程序产品,其特征在于,可直接加载到计算机的内部存储器中,并含有软件代码,所述计算机程序经由计算机载入并执行后能够实现权利要求1-6任一项所述的聋哑人辅助方法。 A computer program product, which can be directly loaded into an internal memory of a computer and contains software code, and the computer program can be loaded and executed by a computer to implement the device according to any one of claims 1 to 6. Dumb assist method.
PCT/CN2016/110475 2016-12-16 2016-12-16 Method and apparatus for assisting people who have hearing and speech impairments and electronic device WO2018107489A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201680006924.XA CN107223277A (en) 2016-12-16 2016-12-16 A kind of deaf-mute's householder method, device and electronic equipment
PCT/CN2016/110475 WO2018107489A1 (en) 2016-12-16 2016-12-16 Method and apparatus for assisting people who have hearing and speech impairments and electronic device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2016/110475 WO2018107489A1 (en) 2016-12-16 2016-12-16 Method and apparatus for assisting people who have hearing and speech impairments and electronic device

Publications (1)

Publication Number Publication Date
WO2018107489A1 true WO2018107489A1 (en) 2018-06-21

Family

ID=59928232

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/110475 WO2018107489A1 (en) 2016-12-16 2016-12-16 Method and apparatus for assisting people who have hearing and speech impairments and electronic device

Country Status (2)

Country Link
CN (1) CN107223277A (en)
WO (1) WO2018107489A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111128180A (en) * 2019-11-22 2020-05-08 北京理工大学 Auxiliary dialogue system for hearing-impaired people
CN113011245A (en) * 2021-01-28 2021-06-22 南京大学 Lip language recognition system and method based on ultrasonic sensing and knowledge distillation

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110111651A (en) * 2018-02-01 2019-08-09 周玮 Intelligent language interactive system based on posture perception
CN108510988A (en) * 2018-03-22 2018-09-07 深圳市迪比科电子科技有限公司 Language identification system and method for deaf-mutes
CN108596107A (en) 2018-04-26 2018-09-28 京东方科技集团股份有限公司 Lip reading recognition methods and its device, AR equipment based on AR equipment
CN108877407A (en) * 2018-06-11 2018-11-23 北京佳珥医学科技有限公司 Methods, devices and systems and augmented reality glasses for supplementary AC
CN111679745A (en) * 2019-03-11 2020-09-18 深圳市冠旭电子股份有限公司 Sound box control method, device, equipment, wearable equipment and readable storage medium
CN110020442A (en) * 2019-04-12 2019-07-16 上海电机学院 A kind of portable translating machine
CN110009973A (en) * 2019-04-15 2019-07-12 武汉灏存科技有限公司 Real-time inter-translation method, device, equipment and storage medium based on sign language
CN110351631A (en) * 2019-07-11 2019-10-18 京东方科技集团股份有限公司 Deaf-mute's alternating current equipment and its application method
TWI743624B (en) * 2019-12-16 2021-10-21 陳筱涵 Attention assist system
CN111343554A (en) * 2020-03-02 2020-06-26 开放智能机器(上海)有限公司 Hearing aid method and system combining vision and voice
CN112185415A (en) * 2020-09-10 2021-01-05 珠海格力电器股份有限公司 Sound visualization method and device, storage medium and MR mixed reality equipment
CN114267323A (en) * 2021-12-27 2022-04-01 深圳市研强物联技术有限公司 Voice hearing aid AR glasses for deaf-mutes and communication method thereof
CN114615609B (en) * 2022-03-15 2024-01-30 深圳市昂思科技有限公司 Hearing aid control method, hearing aid device, apparatus, device and computer medium

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020103649A1 (en) * 2001-01-31 2002-08-01 International Business Machines Corporation Wearable display system with indicators of speakers
CN101124617A (en) * 2005-01-21 2008-02-13 L·凯茨 Management and assistance system for the deaf
CN103946733A (en) * 2011-11-14 2014-07-23 谷歌公司 Displaying sound indications on a wearable computing system
CN104485104A (en) * 2014-12-16 2015-04-01 芜湖乐锐思信息咨询有限公司 Intelligent wearable equipment
CN104966433A (en) * 2015-07-17 2015-10-07 江西洪都航空工业集团有限责任公司 Intelligent glasses assisting deaf-mute conversation
CN105324811A (en) * 2013-05-10 2016-02-10 微软技术许可有限责任公司 Speech to text conversion
CN105529035A (en) * 2015-12-10 2016-04-27 安徽海聚信息科技有限责任公司 System for intelligent wearable equipment
CN105765486A (en) * 2013-09-24 2016-07-13 纽昂斯通讯公司 Wearable communication enhancement device

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020103649A1 (en) * 2001-01-31 2002-08-01 International Business Machines Corporation Wearable display system with indicators of speakers
CN101124617A (en) * 2005-01-21 2008-02-13 L·凯茨 Management and assistance system for the deaf
CN103946733A (en) * 2011-11-14 2014-07-23 谷歌公司 Displaying sound indications on a wearable computing system
CN105324811A (en) * 2013-05-10 2016-02-10 微软技术许可有限责任公司 Speech to text conversion
CN105765486A (en) * 2013-09-24 2016-07-13 纽昂斯通讯公司 Wearable communication enhancement device
CN104485104A (en) * 2014-12-16 2015-04-01 芜湖乐锐思信息咨询有限公司 Intelligent wearable equipment
CN104966433A (en) * 2015-07-17 2015-10-07 江西洪都航空工业集团有限责任公司 Intelligent glasses assisting deaf-mute conversation
CN105529035A (en) * 2015-12-10 2016-04-27 安徽海聚信息科技有限责任公司 System for intelligent wearable equipment

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111128180A (en) * 2019-11-22 2020-05-08 北京理工大学 Auxiliary dialogue system for hearing-impaired people
CN113011245A (en) * 2021-01-28 2021-06-22 南京大学 Lip language recognition system and method based on ultrasonic sensing and knowledge distillation
CN113011245B (en) * 2021-01-28 2023-12-12 南京大学 Lip language identification system and method based on ultrasonic sensing and knowledge distillation

Also Published As

Publication number Publication date
CN107223277A (en) 2017-09-29

Similar Documents

Publication Publication Date Title
WO2018107489A1 (en) Method and apparatus for assisting people who have hearing and speech impairments and electronic device
US9805619B2 (en) Intelligent glasses for the visually impaired
US11043231B2 (en) Speech enhancement method and apparatus for same
EP2842055B1 (en) Instant translation system
US10304452B2 (en) Voice interactive device and utterance control method
US20170243582A1 (en) Hearing assistance with automated speech transcription
US9307073B2 (en) Visual assistance systems and related methods
US20190019512A1 (en) Information processing device, method of information processing, and program
US20190138603A1 (en) Coordinating Translation Request Metadata between Devices
WO2015143114A1 (en) Sign language translation apparatus with smart glasses as display featuring a camera and optionally a microphone
US20170024380A1 (en) System and method for the translation of sign languages into synthetic voices
Salvi et al. Smart glass using IoT and machine learning technologies to aid the blind, dumb and deaf
US20180167745A1 (en) A head mounted audio acquisition module
CN113763940A (en) Voice information processing method and system for AR glasses
JP2021117371A (en) Information processor, information processing method and information processing program
JP6065768B2 (en) Information processing apparatus, information processing method, and program
CN113220912A (en) Interactive assistance method and device and computer readable storage medium
CN210072245U (en) Translation glasses
CN114115515A (en) Method and head-mounted unit for assisting a user
KR102000282B1 (en) Conversation support device for performing auditory function assistance
KR101410321B1 (en) Apparatus and method for silent voice recognition and speaking
JP7070402B2 (en) Information processing equipment
Sneha et al. AI-powered smart glasses for blind, deaf, and dumb
CN112154412B (en) Providing audio information with digital assistant
Laud et al. Indoor navigation aid system for the visually impaired

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16923641

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 05/11/2019)

122 Ep: pct application non-entry in european phase

Ref document number: 16923641

Country of ref document: EP

Kind code of ref document: A1