CN111800700B - Method and device for prompting object in environment, earphone equipment and storage medium - Google Patents

Method and device for prompting object in environment, earphone equipment and storage medium Download PDF

Info

Publication number
CN111800700B
CN111800700B CN202010718955.6A CN202010718955A CN111800700B CN 111800700 B CN111800700 B CN 111800700B CN 202010718955 A CN202010718955 A CN 202010718955A CN 111800700 B CN111800700 B CN 111800700B
Authority
CN
China
Prior art keywords
data
voiceprint
specific contact
earphone
environment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010718955.6A
Other languages
Chinese (zh)
Other versions
CN111800700A (en
Inventor
张峰
张斌
蒋兆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Zimi Electronic Technology Co Ltd
Original Assignee
Jiangsu Zimi Electronic Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu Zimi Electronic Technology Co Ltd filed Critical Jiangsu Zimi Electronic Technology Co Ltd
Priority to CN202010718955.6A priority Critical patent/CN111800700B/en
Publication of CN111800700A publication Critical patent/CN111800700A/en
Application granted granted Critical
Publication of CN111800700B publication Critical patent/CN111800700B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/10Earpieces; Attachments therefor ; Earphones; Monophonic headphones
    • H04R1/1083Reduction of ambient noise
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/10Earpieces; Attachments therefor ; Earphones; Monophonic headphones
    • H04R1/1091Details not provided for in groups H04R1/1008 - H04R1/1083
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2460/00Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
    • H04R2460/01Hearing devices using active noise cancellation

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Telephone Function (AREA)

Abstract

The invention discloses a method and a device for prompting an object in an environment, earphone equipment and a storage medium. The method comprises the following steps: acquiring sound data of an object in an environment; converting the sound data of the object into voiceprint data; matching the voiceprint data with voiceprint data of a specific contact person stored in the earphone; and when the voiceprint matching degree is greater than a preset value, generating prompt information based on the specific contact person to prompt the user. When the user can not know that the call of the specific contact person appears in the environment by using the earphone, the user is prompted to inform the user that the specific contact person appears in the environment, and the user can conveniently communicate with the specific contact person in time.

Description

Method and device for prompting object in environment, earphone equipment and storage medium
Technical Field
The embodiment of the invention relates to a sound signal processing technology, in particular to a method and a device for prompting an object in an environment, earphone equipment and a storage medium.
Background
Nowadays, new generation consumer electronics devices such as global smart phones and tablet computers are rapidly developed, and earphone products, especially wireless earphone products, are receiving more and more attention.
Among them, the Active Noise Cancellation (ANC) earphone can isolate external Noise and improve tone quality, and is gradually paid attention to by people. The active noise reduction function generates reverse sound waves equal to external noise through a noise reduction system, so that the noise is eliminated, and the noise reduction effect is realized. Most of the active noise reduction schemes in the market at present adopt elimination of external environment noise. In the prior art, the influence of different wearing modes and different auditory canal structures on a noise reduction system is reduced by actively adjusting the adaptability of the current filter coefficient of a feedforward filter, so that the noise reduction effect of an earphone is improved; or the proper active noise reduction coefficient is set for different types of earphones, so that noise reduction is realized for various types of earphones. These solutions eliminate the noise of the external environment and also eliminate some useful information, such as the conversation information of a specific user. When the user is called, the sound of the call is also eliminated as ambient noise, so that the user cannot know the called condition.
Disclosure of Invention
The invention provides a method and a device for prompting an object in an environment, earphone equipment and a storage medium, which can realize the effect of prompting a user when a specific contact person calls the user.
In a first aspect, an embodiment of the present invention provides an object prompting method in an environment, including:
acquiring sound data of an object in an environment;
converting the sound data of the object into voiceprint data;
matching the voiceprint data with voiceprint data of a specific contact person stored in the earphone;
and when the voiceprint matching degree is larger than a preset value, generating prompt information based on the specific contact person to prompt the user.
In a second aspect, an embodiment of the present invention further provides an apparatus for prompting an object in an environment, where the apparatus includes:
the first sound data acquisition module is used for acquiring sound data of an object in the environment;
the first voiceprint data acquisition module is used for converting the sound data of the object into voiceprint data;
the first voiceprint matching module is used for matching the voiceprint data with voiceprint data of a specific contact person stored in the earphone;
and the user prompting module generates prompting information based on the specific contact person to prompt the user when the voiceprint matching degree is greater than a preset value.
In a third aspect, an embodiment of the present invention further provides an earphone device, where the earphone device includes:
one or more processors;
a memory for storing one or more programs;
when executed by the one or more processors, cause the one or more processors to implement the in-environment object hinting method of any example.
In a fourth aspect, an embodiment of the present invention further provides a computer-readable storage medium, on which a computer program is stored, where the computer program is used, when executed by a processor, to execute the method for prompting an object in an environment according to any example.
According to the technical scheme of the method for prompting the object in the environment, the voice print data corresponding to the voice data of each object in the environment are identified, the identified voice print data are matched with the voice print data of the specific contact person pre-stored in the earphone, and when the matching is successful, the specific contact person in the environment is determined to be present, and the user is prompted. The user can quickly and accurately know the specific contact persons existing in the environment in the process of using the earphone, the interference of audio data in the earphone is avoided, and the user can conveniently communicate with the specific contact persons in time.
Drawings
FIG. 1 is a flowchart illustrating a method for prompting an object in an environment according to a first embodiment of the present invention;
FIG. 2 is a flowchart of an object hinting method in an environment according to a second embodiment of the present invention;
FIG. 3 is a flowchart of an object hinting method in an environment according to a third embodiment of the present invention;
FIG. 4 is a block diagram showing an arrangement of an object presentation apparatus in an environment according to a fourth embodiment of the present invention;
fig. 5 is a schematic structural diagram of an earphone device in the fifth embodiment of the present invention.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the invention and are not limiting of the invention. It should be further noted that, for the convenience of description, only some of the structures related to the present invention are shown in the drawings, not all of the structures.
Example one
Fig. 1 is a flowchart of an object prompting method in an environment according to an embodiment of the present invention, where this embodiment is applicable to a case where sound data of an object in an environment outside an earphone is collected, identified, and matched to prompt a user of the successfully matched object, and the method may be executed by an object prompting device in an environment according to an embodiment of the present invention. The device can be implemented in a software and/or hardware manner, and if the device is software, the device is stored in the earphone device provided by the embodiment of the invention, and if the device is hardware, the device is installed on the earphone device provided by the embodiment of the invention.
The method specifically comprises the following steps:
step 110, sound data of objects in the environment is obtained.
In particular, the object of the present invention refers to an unknown identity person in an environment that utters sound, and may include at least one object. Wherein, the sound data in the environment may include sound data of a person and sound data other than the person. The earphone collects sound data in the environment, wherein the sound data comprises human sound data and sound data except for human, and the collected sound data is classified by adopting modes of support vectors, K neighbor and the like to distinguish the human sound data.
The mode of collecting the sound data in the environment can be collected by adopting the sound collecting equipment originally arranged in the earphone or extra sound collecting equipment arranged on the earphone. Optionally, a feedback microphone of the headset is used to obtain the sound data. The collected sound data are classified, and the sound data of the object are distinguished, so that the effect of obtaining the sound data of the object in the environment is achieved. The period of collecting the sound data can be set according to actual conditions, for example, a sampling period of 1 second is set, the sampling period is short, the frequency of the obtained sound data is high, the collected sound data can be rapidly transmitted to the earphone, subsequent conversion work of the sound data and the sound pattern data is carried out, and the speed of sound identification is integrally increased.
Optionally, the headset of the present invention may have a headset that shields sound in the external environment, such as an active noise reduction headset. Or may be a normal earphone. When the earphone is an active noise reduction earphone, the active noise reduction function can be started or closed when the earphone starts to work.
Step 120, converting the sound data of the object into voiceprint data.
The voiceprint data refers to the voiceprint data of the object which is obtained by adopting a voiceprint recognition model to recognize and analyze the voice data of the object. Specifically, the sound data of the object is converted into voiceprint data, so that the subsequent recognition and analysis of the sound of the object are facilitated. When the environment includes multiple persons speaking simultaneously, analysis is performed on the acquisition of the sound data of each sampling period, and when the sound data of at least two persons are acquired in one sampling period, the acquired sound data can be classified firstly, and then voiceprint recognition is performed respectively. The sound data may be classified by using a neural network or the like, and a specific sound data classification method is not specifically limited herein.
And step 130, matching the voiceprint data with the voiceprint data of the specific contact person stored in the earphone.
And matching the voiceprint data of the object with the voiceprint data of the specific contact person stored in the earphone, and judging whether the object making sound in the environment is the specific contact person, wherein the specific contact person is the contact person stored by the user according to the actual situation of the user, such as leaders, family members and the like.
The voiceprint data of the specific contact is obtained by carrying out voiceprint recognition on the voice data of the specific contact in advance based on a trained voiceprint recognition model, wherein the voiceprint recognition model can be arranged in the electronic equipment. The electronic device can be a mobile phone, a tablet computer, a computer and the like, can be connected with a microphone to record sound, and has a storage function and a processing function.
Optionally, when storing the voiceprint data of the specific contact, first obtaining the voice data of the specific contact, performing voiceprint recognition on the voice data of the specific contact to obtain the voiceprint data of the specific contact, and storing the voiceprint data of the specific contact in the earphone.
The voiceprint recognition is that in the training stage, a section of training voice of a specific contact person is obtained, and is mapped into a voiceprint recognition model of the specific contact person through feature extraction and training of a voiceprint recognition model. When the voice of the object in the environment is recognized, the voice of the object with unknown identity in the environment is subjected to feature extraction to obtain voice features, the voice features are converted into voiceprint features, the similarity between the voiceprint features of the object in the environment and the voiceprint features of the trained specific contact in the voiceprint recognition model is calculated, and the confidence coefficient of the voiceprint of each object is obtained. And when the voice of the object with unknown identity is higher than a preset threshold value, judging that the voice of the object with unknown identity is the voice of a specific contact stored in advance, wherein the threshold value can be set to a certain value in [0.8-0.9 ].
Optionally, when storing the voiceprint data of the specific contact, the identity information of the specific contact and/or the classification of the specific contact may also be stored, for example, the identity information may be a code number or a name of the specific contact, and the classification of the specific contact may be set according to a user requirement, for example, a leader, a family, a colleague, a friend, and the like. And setting a corresponding prompting voice, prompting sound effect or vibration prompting mode according to the identity information of each specific contact person, so that subsequent prompt information is generated to carry out targeted prompting on the user.
The obtaining of the training voice of the specific contact may be obtaining of sound data of the specific contact for a preset text. Specifically, the voice data of the specific contact according to the specified text content is acquired, the voiceprint recognition model of each specific contact is established one by one, the voiceprint recognition model of the preset text is more accurate, of course, any voice data of the specific contact can be acquired, and the voiceprint recognition model of each specific contact is established one by one.
And 140, when the voiceprint matching degree is greater than a preset value, generating prompt information based on the specific contact person to prompt the user.
Specifically, a threshold of similarity is preset, when the similarity of voiceprint matching is greater than the preset threshold, it is determined that an object in the environment is a specific contact stored in the headset, and the specific contact is notified to the user in a prompt message manner. The prompting information can be a uniform prompting mode, that is, different specific contacts correspond to the same prompting mode, for example, the same vibration mode, the same sound effect (for example, dripping sound), or the same prompting voice (for example, a buddy calls the owner), and the like. The prompting information can also correspond to different prompting modes according to different specific contacts. The prompting mode corresponding to each specific contact person can be preset.
Optionally, the manner of prompting the user includes at least one of the following manners:
and playing prompt information in a voice form in the ear canal based on the earphone, wherein the voice of the prompt information comprises identification information of a specific contact.
And controlling the earphone to generate vibration based on the prompt information in the vibration form.
The prompt information based on the sound effect form is played in the ear canal based on the earphone, wherein the sound effect of the prompt information is correspondingly generated based on the specific contact.
Specifically, when it is determined that the object in the environment is a specific contact, the user is prompted in the form of voice, and specific information of the specific contact is notified to the user, for example, the object is XX leader or the object is XX colleague, and the like. Or in the form of vibration, the user is informed that the object in the environment is a specific contact, and different specific contacts correspond to different vibration modes, which can be distinguished by, for example, the vibration duration or the vibration frequency. Or playing the corresponding sound effect in the auditory canal to prompt the user according to different sound effects stored by different specific contacts through a sound effect prompting mode. The user can know the identity of the other side conveniently without seeing the object in the environment, and the prompting precision is improved.
Optionally, when the voiceprint matching degree is greater than a preset value, the method further includes: and converting the audio playing mode of the earphone into a transparent mode.
The sound of the audio data in the audio playing mode may be a sound during playing a video, a sound during playing music, a sound during playing a broadcast, or a sound during making a call. The pass-through mode is an audio playing mode in which the volume of currently played audio data is reduced to silence or reduced to a level that does not disturb the communication between the user and the objects in the environment. When a call is made, the through mode is a communication mode in which the volume is reduced to a level that does not disturb the communication between the user and the objects in the environment. When the earphone is in a transparent mode, the user can conveniently communicate with objects in the environment, and the interference of audio data is reduced.
Optionally, after the user finishes the conversation, the transparent mode may be converted into the audio playing mode by pressing a key, where the key may be a volume-down key or a volume-up key on the earphone or other keys, or may be a key on the user terminal. In some embodiments, when it is recognized that the voiceprint data of the user and the specific contact does not exist within the preset time period, the pass-through mode may be automatically switched to the audio playing mode, that is, the volume of the earphone is restored to the playing volume when the voiceprint data of the specific contact is not detected. Wherein the preset time period may be 30 seconds.
According to the technical scheme of the method for prompting the object in the environment, the voiceprint data corresponding to the sound data of each object in the environment are identified, the identified voiceprint data are matched with the voiceprint data of the specific contact person pre-stored in the earphone, when the matching is successful, the specific contact person in the environment is determined, and the user is prompted, so that the user can quickly and accurately know the specific contact person in the environment in the process of using the earphone, the interference of the audio data in the earphone is avoided, and the user can conveniently communicate with the specific contact person in time.
Example two
Fig. 2 is a flowchart of an object prompting method in an environment according to a second embodiment of the present invention, which is optimized based on the above embodiments, where explanations of terms that are the same as or correspond to the above embodiments are not repeated here, and the method specifically includes the following steps:
step 210, obtaining sound data of objects in the environment.
Step 220, converting the sound data of the object into text data, and identifying whether the text data has preset keywords. If yes, go to step 230, otherwise, go back to step 210.
Specifically, after sound data of an object in an environment is acquired, the sound data is converted into text data, one or more preset keywords are stored in the earphone, the preset keywords are matched in the text data, and when at least one preset keyword exists in the text data, the sound data of the object is converted into voiceprint data. On the contrary, when the text data corresponding to the sound data of the object does not have the preset keyword, the step of converting the sound data of the object into the voiceprint data is cancelled. The problem that when a specific contact person does not communicate with a user, the earphone obtains the sound data of the specific contact person to frequently prompt the user is solved, and the prompting accuracy is improved.
The preset keywords can be commonly used words for calling, and also can be codes or names of users, when the preset keywords exist in text data corresponding to the voice data of the object, the object is indicated to call the users, at the moment, the voice data of the object is converted into voiceprint data, the object is determined to be a certain specific contact person stored in the earphone, and the specific contact person is informed to the users in a prompt message mode.
Optionally, when the voiceprint data of the object does not match the voiceprint data of the specific contact stored in the headset, the object is also informed to the user in a prompt message manner. The user may be prompted that a person with unknown identity is calling the user. Because the voiceprint data and the identity information of the object are not stored in the earphone, when prompting, a mode different from the mode that the object is the prompting information of the specific contact person stored in the earphone can be adopted to prompt the user, for example, when voice prompting is carried out, a mode of 'an unknown identity person calls you' can be adopted, and a sound effect or vibration mode can also be adopted.
If the sound data of the object is converted into text data and the text data is identified to have the preset keywords, the following steps are carried out:
step 230, converting the sound data of the object into voiceprint data.
And step 240, matching the voiceprint data with the voiceprint data of the specific contact person stored in the earphone.
And step 250, when the voiceprint matching degree is larger than a preset value, generating prompt information based on the specific contact person to prompt the user.
According to the technical scheme of the embodiment, after sound data of an object in an environment are obtained, the sound data are converted into text data, whether preset keywords exist in the text data corresponding to the sound data of the object is determined, and whether the object calls a user is judged. And when the text data corresponding to the sound data of the object has preset keywords, carrying out voiceprint conversion on the sound data of the object and matching the voiceprint conversion with the voiceprint data of the specific contact person stored in the earphone. When the matching degree is greater than the preset value, the identity information of the specific contact corresponding to the object is informed to the user in a prompt message mode, when the matching degree is smaller than the preset value, the object in the environment is not the specific contact stored in the earphone, but the object in the environment calls the user, and at the moment, the object in the environment is also informed to the user in a prompt message mode different from the specific contact. The problem that the user is frequently prompted due to the fact that the earphone acquires the sound data of the specific contact when the specific contact does not communicate with the user is avoided, and prompting accuracy is improved.
EXAMPLE III
On the basis of any one of the above embodiments, the embodiment adds that "the voiceprint matching degree is smaller than the preset value, and when the volume reduction is detected within the preset time period, the prompt message for adding the specific contact person is generated. And when the feedback information of the prompt information is received and the addition is determined, storing the voiceprint data of the object in the local earphone. Wherein explanations of the same or corresponding terms as those of the above embodiments are omitted. Referring to fig. 3, the method for prompting an object in an environment provided by this embodiment includes:
step 310, sound data of objects in the environment is obtained.
Step 320, converting the sound data of the object into voiceprint data.
Step 330, matching the voiceprint data with the voiceprint data of the specific contact stored in the earphone.
And 340, when the voiceprint matching degree is larger than a preset value, generating prompt information based on the specific contact person to prompt the user.
And 350, when the voiceprint matching degree is smaller than a preset value and the audio volume is detected to be reduced in a preset time period, generating prompt information for adding a specific contact person.
Specifically, when the voiceprint matching degree is smaller than the preset value, it is indicated that the object in the environment is not the specific contact, but after the volume of the audio is detected to be reduced within the preset time by the earphone, it is indicated that the volume of the audio is reduced by the user, and the user may be communicating with the object in the environment at this time. And the earphone adds the voiceprint data of the object in the generated environment as prompt information of the specific contact person to prompt the user. So that when the object calls the user again, the headset can recognize the object as a particular contact and prompt the user for the object. The earphone does not store the event which is not prompted to the user and is caused by the fact that the object is a specific contact person when the user does not notice the object next time.
And step 360, when the feedback information of the received prompt information is determined to be added, storing the voiceprint data of the object in the earphone body.
Specifically, when the user confirms that the object is added as a specific contact, the voiceprint data of the object is stored in the headphone body. The method for the user to confirm may be a key on the earphone or a key on the user terminal, which is not limited specifically here. And after the conversation between the user and the object is finished, the identity information of the object is added on the user terminal by the user and is stored on the earphone body.
According to the technical scheme of the embodiment, after the sound data of the object in the environment is acquired and converted into the voiceprint data, when the matching degree of the voiceprint data and the voiceprint data of the specific contact person stored in the earphone is judged, when the matching degree is larger than a preset value, namely the object is a certain specific contact person stored in the earphone, the specific contact person is informed to the user in a prompt message mode. When the matching degree is smaller than the preset value, but the sound volume of the audio detected by the earphone is reduced in the preset time, the user finds that the object calls the user. The user operates the earphone or the user terminal to reduce the audio volume, and the earphone sends prompt information whether to store the voiceprint data of the object in the earphone body to the user at the moment. When receiving the positive prompt message of the user, the earphone stores the voiceprint data of the object to the earphone body, and when the object calls the user again, the user can be informed in time.
Example four
Fig. 4 is a block diagram of an object presentation apparatus 40 in an environment according to an embodiment of the present invention. The device is used for executing the method for prompting the object in the environment provided by any embodiment, and the device can be implemented by software or hardware. The apparatus 40 comprises:
the first sound data acquisition module 410 acquires sound data of an object in an environment.
The first voiceprint data acquisition module 420 converts the sound data of the object into voiceprint data.
The first voiceprint matching module 430 matches the voiceprint data with the voiceprint data of the specific contact stored in the headset.
And the user prompting module 440 generates a prompting message based on the specific contact person to prompt the user when the voiceprint matching degree is greater than the preset value.
Optionally, the user prompting module 440 further includes:
and converting the audio playing mode of the earphone into a transparent mode.
Optionally, the apparatus 40 further comprises:
and the keyword judgment module is used for converting the sound data of the object into text data and identifying whether the text data has preset keywords.
And the conversion canceling module cancels the step of converting the sound data of the object into the voiceprint data if the preset key words do not exist in the text data.
Optionally, the apparatus 40 further comprises:
and the second sound acquisition module acquires the sound data of the specific contact.
And the second voiceprint acquisition module is used for carrying out voiceprint recognition on the voice data of the specific contact person so as to acquire the voiceprint data of the specific contact person.
The first storage module stores the voiceprint data of the specific contact in the earphone.
Optionally, the second sound obtaining module includes:
and acquiring sound data of the specific contact for the preset text.
Optionally, the apparatus 40 further comprises:
and the second big voiceprint matching module generates prompt information for adding a specific contact when the voiceprint matching degree is smaller than a preset value and the volume reduction is detected within a preset time period.
And the second storage module is used for storing the voiceprint data of the object to the earphone body when the feedback information of the prompt information is received and is determined to be added.
Optionally, the user prompting module 440 includes at least one sub-module of:
and the voice playing sub-module plays the prompt information in a voice form in the ear canal based on the earphone, wherein the voice of the prompt information comprises the identification information of the specific contact.
And the vibration submodule controls the earphone to generate vibration based on the prompt information of the vibration form.
And the sound effect playing sub-module plays the sound in the auditory canal based on the earphone in the prompt information of the sound effect form, wherein the sound effect of the prompt information is correspondingly generated based on the specific contact.
The object prompting device in the environment provided by the embodiment of the invention comprises: the method comprises the steps of obtaining sound data of an object in the environment, converting the sound data into voiceprint data, matching the voiceprint data of the object with voiceprint data of a specific contact person stored in an earphone, when the matching degree is larger than a preset value, indicating that the object making sound in the environment is the specific contact person stored in the earphone, and generating prompt information based on the specific contact person to prompt a user. When the user can not know that the call of the specific contact person appears in the environment by using the earphone, the user is prompted to inform the user that the specific contact person appears in the environment, and the user can conveniently communicate with the specific contact person in time.
EXAMPLE five
Fig. 5 is a schematic structural diagram of an earphone device according to a fifth embodiment of the present invention, as shown in fig. 5, the earphone device includes a processor 510, a memory 520, an input device 530, and an output device 540. The number of processors 510 in the earphone device may be one or more, and one processor 510 is taken as an example in fig. 5. The processor 510, the memory 520, the input device 530 and the output device 540 in the headset apparatus may be connected by a bus or other means, which is exemplified in fig. 5.
The memory 520 is a computer-readable storage medium, and can be used for storing software programs, computer-executable programs, and modules, such as program instructions/modules corresponding to the object prompting method in the environment in the embodiment of the present invention (for example, the first sound data obtaining module 410, the first voiceprint data obtaining module 420, and the voiceprint matching module 430 in the object prompting device in the environment). The processor 510 executes various functional applications and data processing of the headphone device by executing software programs, instructions, and modules stored in the memory 520, that is, implements the object prompting method in the above-described environment.
The memory 520 may mainly include a program storage area and a data storage area, wherein the program storage area may store an operating system, an application program required for at least one function, and the like. The storage data area may store data created according to the use of the terminal, and the like. Further, the memory 520 may include high speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other non-volatile solid state storage device. In some examples, the memory 520 may further include memory located remotely from the processor 510, which may be connected to the headset device via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The input means 530 may be used to receive input numeric or character information and generate key signal inputs related to user settings and function control of the headset device. The output device 540 may include a display device such as a display screen.
EXAMPLE six
An embodiment of the present invention further provides a storage medium containing computer-executable instructions, which when executed by a computer processor, perform a method for object hinting in an environment, the method including:
sound data of an object in an environment is acquired.
Sound data of the object is converted into voiceprint data.
And matching the voiceprint data with the voiceprint data of the specific contact person stored in the earphone.
And when the voiceprint matching degree is greater than a preset value, generating prompt information based on the specific contact person to prompt the user.
Of course, the storage medium provided by the embodiment of the present invention contains computer-executable instructions, and the computer-executable instructions are not limited to the operations of the method described above, and may also perform related operations in the object prompting method in the environment provided by any embodiment of the present invention.
From the above description of the embodiments, it is obvious for those skilled in the art that the present invention can be implemented by software and necessary general hardware, and certainly, can also be implemented by hardware, but the former is a better embodiment in many cases. Based on such understanding, the technical solutions of the present invention may be embodied in the form of a software product, which can be stored in a computer-readable storage medium, such as a floppy disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a FLASH Memory (FLASH), a hard disk or an optical disk of a computer, and includes several instructions for enabling a computer device (which may be a personal computer, a server, or a network device) to execute the methods according to the embodiments of the present invention.
It should be noted that, in the embodiment of the object prompting apparatus in the above environment, the included modules and sub-modules are only divided according to functional logic, but are not limited to the above division, as long as the corresponding functions can be implemented. In addition, the specific names of the functional modules are only for convenience of distinguishing from each other and are not used for limiting the protection scope of the present invention.
It is to be noted that the foregoing is only illustrative of the preferred embodiments of the present invention and the technical principles employed. It will be understood by those skilled in the art that the present invention is not limited to the particular embodiments described herein, but is capable of various obvious changes, rearrangements and substitutions as will now become apparent to those skilled in the art without departing from the scope of the invention. Therefore, although the present invention has been described in greater detail by the above embodiments, the present invention is not limited to the above embodiments, and may include other equivalent embodiments without departing from the spirit of the present invention, and the scope of the present invention is determined by the scope of the appended claims.

Claims (10)

1. An object prompting method in an environment is characterized by comprising the following steps:
acquiring sound data of an object in an environment;
converting the sound data of the object into voiceprint data;
matching the voiceprint data with voiceprint data of a specific contact person stored in the earphone;
when the voiceprint matching degree is larger than a preset value, generating prompt information based on the specific contact person to prompt a user;
when the voiceprint matching degree is smaller than a preset value and the audio volume is detected to be reduced within a preset time period, generating prompt information for adding a specific contact person;
and when the feedback information of the prompt information is received and is determined to be added, storing the voiceprint data of the object in the earphone body.
2. The method of claim 1, wherein when the voiceprint matching degree is greater than a preset value, the method further comprises:
and converting the audio playing mode of the earphone into a transparent mode.
3. The method of claim 1, further comprising, prior to converting the sound data of the object to voiceprint data:
converting the sound data of the object into text data, and identifying whether preset keywords exist in the text data;
and if the preset keywords do not exist in the text data, canceling the step of converting the sound data of the object into the voiceprint data.
4. The method of claim 1, further comprising:
acquiring sound data of the specific contact;
carrying out voiceprint recognition on the voice data of the specific contact person to obtain the voiceprint data of the specific contact person;
storing the voiceprint data for the particular contact in the headset.
5. The method of claim 4, wherein the obtaining voice data of the specific contact comprises:
and acquiring sound data of the specific contact for a preset text.
6. The method of any of claims 1-5, wherein the prompting the user comprises at least one of:
playing the prompt information in a voice form in an ear canal based on the earphone, wherein the voice of the prompt information comprises identification information of the specific contact;
controlling the earphone to generate vibration based on the prompt information in the form of vibration;
the prompt information based on the sound effect form is played in the ear canal based on the earphone, wherein the sound effect of the prompt information is generated correspondingly based on the specific contact.
7. The method of any of claims 1-5, wherein the headset is an active noise reduction headset.
8. An apparatus for prompting an object in an environment, comprising:
the first sound data acquisition module is used for acquiring sound data of an object in the environment;
the first voiceprint data acquisition module is used for converting the sound data of the object into voiceprint data;
the first voiceprint matching module is used for matching the voiceprint data with voiceprint data of a specific contact person stored in the earphone;
the user prompting module is used for generating prompting information based on the specific contact person to prompt a user when the voiceprint matching degree is larger than a preset value;
the second big voiceprint matching module is used for generating prompt information for adding a specific contact when the voiceprint matching degree is smaller than a preset value and the volume reduction is detected within a preset time period;
and the second storage module is used for storing the voiceprint data of the object to the earphone body when the feedback information of the prompt information is received and is determined to be added.
9. An earphone device, characterized in that the earphone device comprises:
one or more processors;
a memory for storing one or more programs;
when executed by the one or more processors, cause the one or more processors to implement the in-environment object hinting method of any one of claims 1-5.
10. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, carries out a method for object hinting in an environment according to any one of claims 1-5.
CN202010718955.6A 2020-07-23 2020-07-23 Method and device for prompting object in environment, earphone equipment and storage medium Active CN111800700B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010718955.6A CN111800700B (en) 2020-07-23 2020-07-23 Method and device for prompting object in environment, earphone equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010718955.6A CN111800700B (en) 2020-07-23 2020-07-23 Method and device for prompting object in environment, earphone equipment and storage medium

Publications (2)

Publication Number Publication Date
CN111800700A CN111800700A (en) 2020-10-20
CN111800700B true CN111800700B (en) 2022-04-22

Family

ID=72828347

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010718955.6A Active CN111800700B (en) 2020-07-23 2020-07-23 Method and device for prompting object in environment, earphone equipment and storage medium

Country Status (1)

Country Link
CN (1) CN111800700B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP4207805A4 (en) 2021-02-23 2024-04-03 Samsung Electronics Co., Ltd. Electronic device and control method thereof
CN113938785A (en) * 2021-11-24 2022-01-14 英华达(上海)科技有限公司 Noise reduction processing method, device, equipment, earphone and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105657185A (en) * 2016-02-29 2016-06-08 宇龙计算机通信科技(深圳)有限公司 Method and device for controlling mobile terminal
CN107799117A (en) * 2017-10-18 2018-03-13 倬韵科技(深圳)有限公司 Key message is identified to control the method, apparatus of audio output and audio frequency apparatus
CN108391206A (en) * 2018-03-30 2018-08-10 广东欧珀移动通信有限公司 Signal processing method, device, terminal, earphone and readable storage medium storing program for executing
CN110691300A (en) * 2019-09-12 2020-01-14 连尚(新昌)网络科技有限公司 Audio playing device and method for providing information

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105657185A (en) * 2016-02-29 2016-06-08 宇龙计算机通信科技(深圳)有限公司 Method and device for controlling mobile terminal
CN107799117A (en) * 2017-10-18 2018-03-13 倬韵科技(深圳)有限公司 Key message is identified to control the method, apparatus of audio output and audio frequency apparatus
CN108391206A (en) * 2018-03-30 2018-08-10 广东欧珀移动通信有限公司 Signal processing method, device, terminal, earphone and readable storage medium storing program for executing
CN110691300A (en) * 2019-09-12 2020-01-14 连尚(新昌)网络科技有限公司 Audio playing device and method for providing information

Also Published As

Publication number Publication date
CN111800700A (en) 2020-10-20

Similar Documents

Publication Publication Date Title
CN107995360B (en) Call processing method and related product
US6233556B1 (en) Voice processing and verification system
CN110493678B (en) Earphone control method and device, earphone and storage medium
CN112585676A (en) Biometric authentication
CN107172256B (en) Earphone call self-adaptive adjustment method and device, mobile terminal and storage medium
CN109686367B (en) Earphone noise reduction method, device and equipment and readable storage medium
US20180152163A1 (en) Noise control method and device
CN102484461A (en) A system and a method for providing sound signals
US20200219530A1 (en) Adaptive spatial vad and time-frequency mask estimation for highly non-stationary noise sources
CN111800700B (en) Method and device for prompting object in environment, earphone equipment and storage medium
CN112770214A (en) Earphone control method and device and earphone
CN112911441A (en) Noise reduction method, apparatus, audio device, and computer-readable storage medium
WO2019228329A1 (en) Personal hearing device, external sound processing device, and related computer program product
CN111491236A (en) Active noise reduction earphone, awakening method and device thereof and readable storage medium
CN111199751B (en) Microphone shielding method and device and electronic equipment
CN112383855A (en) Bluetooth headset charging box, recording method and computer readable storage medium
CN104851423B (en) Sound information processing method and device
JP2019184809A (en) Voice recognition device and voice recognition method
CN115482830A (en) Speech enhancement method and related equipment
CN110197663B (en) Control method and device and electronic equipment
CN106549954A (en) Method of speech processing and device
US11940896B2 (en) Information processing device, information processing method, and program
US20190304457A1 (en) Interaction device and program
CN114333817A (en) Remote controller and remote controller voice recognition method
JP2023531417A (en) LIFELOGGER USING AUDIO RECOGNITION AND METHOD THEREOF

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant