WO2023245390A1 - Smart earphone control method and apparatus, electronic device and storage medium - Google Patents

Smart earphone control method and apparatus, electronic device and storage medium Download PDF

Info

Publication number
WO2023245390A1
WO2023245390A1 PCT/CN2022/099966 CN2022099966W WO2023245390A1 WO 2023245390 A1 WO2023245390 A1 WO 2023245390A1 CN 2022099966 W CN2022099966 W CN 2022099966W WO 2023245390 A1 WO2023245390 A1 WO 2023245390A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
voice
smart
data
voice data
Prior art date
Application number
PCT/CN2022/099966
Other languages
French (fr)
Chinese (zh)
Inventor
彭聪
Original Assignee
北京小米移动软件有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京小米移动软件有限公司 filed Critical 北京小米移动软件有限公司
Priority to PCT/CN2022/099966 priority Critical patent/WO2023245390A1/en
Priority to CN202280004138.1A priority patent/CN117751585A/en
Publication of WO2023245390A1 publication Critical patent/WO2023245390A1/en

Links

Images

Definitions

  • the present application relates to the field of computer application technology, and in particular to control methods, devices, electronic devices and storage media for smart earphones.
  • Terminal devices are equipped with smart headphones during use, which brings convenience to people's daily use.
  • Smart headphones with noise reduction function can isolate the surrounding noise, and their current use is becoming more and more popular.
  • This application proposes a control method, device, electronic device and storage medium for smart earphones to improve the convenience of using smart earphones.
  • an embodiment of this application proposes a control method for smart headphones, including:
  • the audio data is identified to determine that the terminal device is in a target working scenario; the microphone of the smart headset is in a voice collection state under the target working scenario.
  • the smart earphone is controlled to switch between the at least two voice modes.
  • Another embodiment of the present application provides a control device for smart headphones, including:
  • An acquisition module used to acquire audio data played on the terminal device
  • the first determination module is used to identify the audio data to determine that the terminal device is in a target working scenario; in the target working scenario, the microphone of the smart headset is in a voice collection state;
  • the acquisition module is also used to acquire the mode switching instruction of the smart headset in the target working scenario
  • a control module configured to control the smart earphone to switch between the at least two voice modes according to the mode switching instruction.
  • Another aspect of the present application provides an electronic device, including a memory, a processor, and a computer program stored in the memory and executable on the processor.
  • the processor executes the program, the aforementioned aspect is implemented. the method described.
  • Another embodiment of the present application proposes a non-transitory computer-readable storage medium on which a computer program is stored.
  • the computer program is executed by a processor, the method as described in the foregoing aspect is implemented.
  • Another aspect of the present application provides a computer program product on which a computer program is stored.
  • the program is executed by a processor, the method as described in the foregoing aspect is implemented.
  • the control method, device, electronic equipment and storage medium of the smart headset proposed in this application acquire the audio data played on the terminal device and identify the audio data to determine that the terminal device is in the target working scene and the microphone of the smart headset in the target working scene.
  • the voice collection state in the target work scenario, obtain the mode switching instruction of the smart headset, and control the smart headset to switch between at least two voice modes according to the mode switching instruction, by controlling the smart headset to switch between at least two voice modes in the target work scenario. Switching between modes can help users avoid taking off or putting on smart headphones frequently when using smart headphones, which improves the convenience of using smart headphones.
  • Figure 1 is a schematic flowchart of a smart headset control method provided by an embodiment of the present application
  • FIG. 2 is a schematic flow chart of another smart headset control method provided by an embodiment of the present application.
  • FIG. 3 is a schematic flow chart of another smart headset control method provided by an embodiment of the present application.
  • Figure 4 is a schematic flow chart of another smart headset control method provided by an embodiment of the present application.
  • FIG. 5 is a schematic structural diagram of a control device for a smart headset provided by an embodiment of the present application.
  • Figure 6 is a block diagram of an electronic device provided by an embodiment of the present application.
  • FIG. 1 is a schematic flowchart of a method for controlling a smart headset provided by an embodiment of the present application.
  • the execution subject of the smart earphone control method in the embodiment of the present application is a control device of the smart earphone.
  • the device can be installed in an electronic device.
  • the electronic device can be a smart earphone.
  • the smart earphone is a noise-cancelling earphone with a noise reduction function.
  • the method may include the following steps:
  • Step 101 Obtain audio data played on the terminal device.
  • the terminal device may be a smartphone, a handheld computer, a smart wearable device, a computer, etc., which are not limited in this embodiment.
  • the smart earphone is a Bluetooth earphone with a noise reduction function.
  • the smart earphone and the terminal device are connected through Bluetooth. After the connection is established, the terminal device sends the played audio data to the smart earphone according to the set frequency. , for example, sent in real time, or sent every 200ms, where the audio data can be human voice data, audio-visual entertainment data, etc.
  • Step 102 Recognize the audio data to determine that the terminal device is in the target working scene.
  • the microphone of the smart headset in the target working scenario, is in the voice collection state, which means that when the terminal device is in the target working scenario, the microphone of the smart headset is turned on and can be used to collect voice data of the environment.
  • the audio data is identified to identify whether the audio data belongs to audio-visual entertainment data or human voice data.
  • audio-visual entertainment data such as music, movies, short videos, etc.
  • audio-visual entertainment data such as music, movies, short videos, etc.
  • voice data recognition is voice data, for example, the voice data of people in a conference call scenario, or the voice data of people using instant chat software. Voice data, etc.
  • the voice data of the user using the smart headset needs to be transmitted to users of other terminal devices participating in conference calls or instant chats. Therefore, the microphone on the smart headset needs to be turned on to collect the users using the smart headset.
  • the terminal device it is determined that the terminal device is in the target working scene; if the audio data is identified as audio-visual entertainment data, for example, music listened to on the application, video watched, etc., in this scenario, the smart headset is used to use the terminal device
  • the played audio data is transmitted to the human ear through the earphones, and the microphone on the smart earphones is not turned on. There is no need to collect the sound data or environmental data of the user using the smart earphones.
  • audio data of a set duration is obtained, the audio data is identified, the voice segment included in the audio data is determined, and the response is that the voice segment includes the first target voice segment and does not include the second target voice segment.
  • the target voice band and the third target voice band determine that the terminal device is in the target working scenario.
  • the lower frequency limit of the first target speech band is greater than the upper frequency limit of the second target speech band, and the upper frequency limit of the first target speech band is less than the lower frequency limit of the third target speech band.
  • the speech band includes a low frequency band, a medium frequency band and a high frequency band
  • the first target speech band is the medium frequency band
  • the second target speech band is the low frequency band
  • the third target speech band is the high frequency band.
  • the acquired user voice data contains fewer low-frequency and high-frequency bands, which means that the frequency of the user's voice data mainly belongs to the mid-frequency band.
  • the audio-visual entertainment data contains music data
  • the music data contains more low-frequency bands. data and high-frequency band data. Therefore, by identifying the frequency band contained in the audio data, it can be determined whether the audio data is human voice data or audio-visual entertainment data. When the audio data is recognized as human voice data, it is determined that the terminal device is in Target work scenario.
  • the audio data is considered to be the user's voice data, that is, it is determined that the terminal device is in the target working scene.
  • audio data of multiple set durations can be collected for identification.
  • voice segments contained in the audio data of multiple set durations meet the above requirements, it is determined that the terminal device is in Target work scenario.
  • the source information carried by the audio data is identified, the source of the audio data is determined, and in response to the source of the audio data being the target source, it is determined that the terminal device is in the target working scene.
  • the source information carried by the audio data can be identified, where the source information indicates the application software to which the audio data belongs. If the application to which the audio data belongs is identified as the target application, it is determined that the terminal device is in the target working scene. For example, if the source information indicates that the pinyin data belongs to an audio-visual entertainment application, not a call application, for example, a conference call application, then it is determined that the terminal device does not belong to the target work scenario, otherwise it is determined that the terminal device belongs to the target work scenario.
  • Step 103 Obtain the mode switching instruction of the smart headset in the target working scenario.
  • the smart headset has at least two voice modes in the target working scenario, which is a scenario where multiple users communicate, for example, a phone call. Meeting scenarios, voice chat scenarios, etc.
  • the functional mode corresponding to the setting button of the smart headset can be changed.
  • the functional mode includes play and pause functions, and in the target working mode, the functional mode is switched to include at least two voice modes. Therefore, in the target working mode, in response to the user pressing the setting button of the smart earphone, the mode switching instruction of the smart earphone can be obtained, and the mode switching instruction is used to switch the smart earphone to the voice mode indicated by the mode switching instruction.
  • smart headphones include two voice modes, namely noise reduction mode and communication mode.
  • the external sound will be noise reduction processing, that is, after the noise reduction processing, the external sound is less than the set volume; in the communication mode, The external sound will not be denoised, but the original sound can be maintained or the original sound can be amplified so that the external sound is greater than the set volume. If it is detected that the user presses the setting button, it is determined which voice mode the smart headset was in before the user pressed the setting button. If the smart headset is in the noise reduction mode, it responds to the operation of the button and generates a communication mode switching command. Similarly, If the smart headset is in the communication mode, it will generate a noise reduction mode switching instruction in response to the operation of the button.
  • the mode switching instruction corresponding to the obtained voice recognition can be used to change the corresponding functional mode of the smart headset.
  • the acquired voice signal can be identified to identify the keywords contained in the voice signal, and the corresponding mode switching instruction is determined based on the keywords, where the mode switching instruction is used to switch the smart headset to the mode indicated by the mode switching instruction.
  • Voice mode Regarding the voice mode of the smart headset, please refer to the description in the previous implementation, and will not be described again here.
  • Step 104 Control the smart headset to switch between at least two voice modes according to the mode switching instruction.
  • the smart headset is controlled to switch between at least two voice modes, so that the smart headset can be switched between at least two voice modes while wearing the smart headset, which improves the efficiency of voice mode switching and increases the number of voice modes. User convenience in using smart headphones.
  • This embodiment takes two voice modes, namely the noise reduction mode and the communication mode, as an example to illustrate.
  • the smart headset is in the noise reduction mode, that is to say, the user using the smart headset makes a voice call through the client program. Then according to the mode switching instruction, switch to the communication mode.
  • the communication mode the sound of the voice call will be reduced to less than the set volume, so that when the user wears the smart headset, he will not be affected by the audio played by the device on the terminal.
  • the sound of data affects the acquisition of external sounds; in another scenario, the smart earphones are in communication mode, and after detecting the mode switching command, the smart earphones will switch from the communication mode to the noise reduction mode.
  • the external sounds It will be processed by noise reduction, so that the volume of external sounds heard by the user is lower than the set volume.
  • the audio data played on the terminal device is obtained, and the audio data is identified to determine that the terminal device is in the target working scene.
  • the microphone of the smart headset is in the voice collection state.
  • obtain the mode switching instruction of the smart headset and control the smart headset to switch between at least two voice modes according to the mode switching instruction.
  • FIG. 2 is a schematic flow chart of another smart earphone control method provided by an embodiment of the present application.
  • the voice mode including the noise reduction mode and the communication mode is used as an example for explanation.
  • the smart headset After the smart headset switches to the noise reduction mode, the smart headset performs noise reduction processing on the acquired environmental voice data to avoid the impact of human voice data on the target mode.
  • the method includes the following steps:
  • Step 201 Obtain audio data played on the terminal device.
  • Step 202 Recognize the audio data to determine that the terminal device is in the target working scene.
  • the microphone of the smart headset is in the voice collection state in the target work scenario.
  • Step 203 Obtain the mode switching instruction of the smart headset in the target working scenario.
  • Step 204 Control the smart headset to switch between at least two voice modes according to the mode switching instruction.
  • Step 205 In response to controlling the smart headset to switch to the noise reduction mode, obtain the audio data played on the terminal device.
  • steps 205 to 208 after step 204 is only an example, that is to say, the execution timing of steps 205 to 208 is not limited.
  • audio data played on the terminal device is obtained.
  • the audio data is voice data collected by other terminal devices. Users of other terminal devices and the terminal device They are all in the target work scenario.
  • the voice data contains human voice data. For the convenience of distinction, it is called the first user's voice data.
  • the target work scenario is a conference call scenario.
  • the conference call scenario there are three users participating in the conference call, namely User 1, User 2 and User 3.
  • User 1 is the user corresponding to the terminal device
  • User 2 and User 3 are participating in the conference call.
  • 3 corresponds to other terminal devices
  • the obtained audio data played on the terminal device includes the sound data of the two first users, that is, the sound data of user 2 and user 3.
  • Step 206 Recognize the audio data to identify the first user's voice data included in the audio data.
  • the audio data is identified to identify the sound segments corresponding to different voices in the audio data. Since the voices of different people have different frequencies, therefore, the segments corresponding to the voices of different people can be identified. Different sound frequency bands or different sound spectrum characteristics can be used to identify the first user's sound data included in the audio data.
  • the identified first user's sound data can be one or more, which are related to the target work scene. Next, it is related to the number of participants. The more participants, the more voice data of the first user will be recognized.
  • feature recognition such as timbre and color can be performed on the audio data based on the trained speech recognition model to identify the first user's voice data contained in the audio data.
  • the audio data obtained after identification includes the voice data of the two first users, which can be called the voice data of the first user A. and the voice data of the first user B.
  • Step 207 Obtain the first environmental voice data collected by the microphone on the smart headset.
  • the microphone on the smart headset is in the activated state and can collect environmental voice data in real time. In order to facilitate differentiation from the environmental voice data collected in other modes, it is called the first environmental voice data and is sent to the smart headset.
  • Step 208 Perform noise reduction processing on the first environmental voice data according to the first user's voice data.
  • the first environmental voice data can be recognized to obtain the human voice data contained in the first environmental voice data, wherein the first environmental voice data can be recognized to obtain the human voice data.
  • the method please refer to the explanations in the previous steps. The principles are the same and will not be repeated here.
  • the volume of the voice segments in the first environmental voice data that do not belong to the first user's voice data is reduced or directly deleted, so as to improve the first user's voice data.
  • the ambient voice data undergoes noise reduction processing to reduce the impact of the ambient voice data on the voice data played on the terminal device.
  • the voice data of the second user using the smart headset is obtained, and based on the voice data of the first user and the voice data of the second user, the voice data of the first user and the voice data of the second user in the first environmental voice data are compared.
  • Sound data other than the user's sound data is subjected to noise reduction processing.
  • the noise reduction level of the noise reduction processing can be set according to needs.
  • the noise reduction can be set to the highest to achieve the acquisition of the third sound.
  • the voice data other than the first user's voice data and the second user's voice data are filtered out in the first ambient voice data to maximize the effect of calls or communication in the target work scenario, and achieve the goal of filtering out the first ambient voice data that is not included in the first user's voice data.
  • the audio data played on the terminal device in response to controlling the smart headset to switch to the noise reduction mode, is obtained, and the audio data is identified to identify the first user's voice included in the audio data.
  • the audio data obtain the ambient voice data collected by the microphone on the smart headset, and perform noise reduction processing on the ambient voice data based on the first user's voice data to reduce the impact of the ambient voice data on the voice data played on the terminal device and improve the intelligence Ease of use of headphones.
  • FIG. 3 is a schematic flow chart of another smart earphone control method provided by an embodiment of the present application.
  • the voice mode including the noise reduction mode and the communication mode is used as an example for explanation.
  • the smart headset switches to communication mode, how does the smart headset obtain the sound data of a third user outside the second user using the smart headset in the ambient voice data, so as to achieve communication without taking off the smart headset.
  • the method consists of the following steps:
  • Step 301 Obtain audio data played on the terminal device.
  • Step 302 Recognize the audio data to determine that the terminal device is in the target working scene.
  • the microphone of the smart headset is in the voice collection state in the target work scenario.
  • Step 303 Obtain the mode switching instruction of the smart headset in the target working scenario.
  • Step 304 Control the smart headset to switch between at least two voice modes according to the mode switching instruction.
  • Step 305 In response to controlling the smart earphone to switch to the communication mode according to the mode switching instruction, obtain the first sound direction of the second user using the smart earphone.
  • steps 305 to 307 after step 304 is only an example. Steps 305 to 307 can also be executed before step 304. That is to say, the execution timing of steps 305 to 307 in this implementation There are no restrictions in the example.
  • the second user using the smart earphone when the smart earphone is in the communication mode, the second user using the smart earphone will communicate with surrounding users.
  • a voice prompt can be sent to the second user using the smart earphone through the smart earphone, prompting the second user to speak the setting sentence, thereby determining to collect the setting of the second user's voice The first sound position corresponding to the sentence.
  • the first sound direction of the second user using the smart earphone stored by the smart earphone can be obtained, because the second user usually wears the smart earphone when using the smart earphone. is fixed, so the first sound direction of the second user's sound data collected by the smart earphone is also fixed, so the first sound direction of the second user can be obtained from the storage unit of the smart earphone.
  • Step 306 Control the microphone to collect the sound data of the third user in the environment at a second sound direction other than the first sound direction of the second user.
  • the third user may be a user who makes a sound in the environment, may be a user communicating with the second user, or may be another user who is not communicating with the second user but is talking.
  • the second sound direction is the direction from which the third user makes the sound in the environment.
  • the microphone is a microphone array
  • the first microphone in the microphone array used to collect the first sound direction of the second user is controlled to stop collecting the second user's sound data
  • the first microphone in the microphone array is controlled to stop collecting the second user's sound data.
  • a second microphone other than the microphone collects the sound data of the third user in the environment at a second sound direction other than the first sound direction of the second user, so that the sound data collected by the microphone does not include the sound data of the second user, improving Improves the accuracy of sound collection in communication mode.
  • the direction with the strongest sound intensity can be used as the second sound direction.
  • Step 307 Use smart headphones to play the third user's voice data.
  • the voice data of the third user collected by the microphone is sent to the smart earphone.
  • the smart earphone amplifies the voice data of the third user, it plays the voice data of the third user, which improves the accuracy of the voice data of the third user.
  • the volume improves the reliability of the second user's acquisition of the communicating user's voice data, allowing the second user to clearly hear the communicating third user's voice without taking off the earphones, thereby improving the communication effect.
  • the noise reduction mode in response to controlling the smart headset to switch to the communication mode, is turned off to avoid noise reduction processing on the environmental voice data collected by the microphone.
  • the target work scenario is, for example, a conference call scenario, that is, to prevent other customers participating in the conference call from listening. to the second user and the third user's communication voice.
  • the user when the smart headset is switched to the communication mode according to the mode switching instruction, the user can be prompted to trigger the prohibition instruction in the interactive interface of the terminal device, so that the terminal device passes the prohibition instruction through Bluetooth is sent to the smart headset.
  • the smart headset prohibits data transmission between the microphone and the terminal device based on the obtained prohibition instruction sent by the terminal device. Specifically, it can prohibit the transmission of the third user's voice data, thereby avoiding the second user from being connected to the surroundings.
  • the voice content communicated by the third user is transmitted to the terminal device, thereby preventing the terminal device from playing the communicated voice content in the target work scenario.
  • the target work scenario is, for example, a conference call scenario, that is, to prevent other customers participating in the conference call from hearing the third user.
  • Using smart earphones to play the voice data of a third user does not collect the voice data of the second user using the smart earphones, but collects the voice data of the third user communicating with the second user, so that the user does not take off the smart earphones. It can communicate with surrounding users smoothly, improving the reliability of communication while wearing smart headphones.
  • FIG. 4 is a schematic flow chart of another smart headset control method provided by an embodiment of the present application. It specifically illustrates how the smart headset obtains environmental voice data using the smart headset after the smart headset switches to the communication mode. The voice data of the third user outside the second user enables communication without taking off the smart earphones.
  • the method consists of the following steps:
  • Step 401 Obtain audio data played on the terminal device.
  • Step 402 Recognize the audio data to determine that the terminal device is in the target working scene.
  • the microphone of the smart headset is in the voice collection state in the target work scenario.
  • Step 403 Obtain the mode switching instruction of the smart headset in the target working scenario.
  • Step 404 Control the smart headset to switch between at least two voice modes according to the mode switching instruction.
  • Step 405 In response to controlling the smart headset to switch to the communication mode, obtain the second environmental voice data collected by the microphone on the smart headset.
  • steps 405 to 407 after step 404 is only an example, and the execution timing of steps 405 to 407 is not limited in this embodiment.
  • the content of the second environmental voice data collected by the microphone on the smart headset may be the same as or different from the first environmental voice data in the previous embodiment. It is only to distinguish the environmental voice data collected in different modes. Please refer to the aforementioned The explanation of the first environmental voice data in the embodiment will not be described again here.
  • Step 406 Determine the voice data of the third user in the second environmental voice data based on the voice data of the second user using the smart headset.
  • the third user and the second user are different users.
  • the third user is the user communicating with the second user.
  • Step 407 Use smart headphones to play the third user's voice data.
  • the second user's voice data is identified to obtain the voice segment or timbre feature information of the second user's voice data, and the second user's voice data is identified based on the voice segment or timbre feature information of the second user's voice data.
  • the noise data that does not belong to the human voice in the third user's voice data can be filtered out and amplified and then played through the smart earphones to ensure that the third user wearing the smart earphones The second user can clearly obtain the voice of the third user.
  • the voice data of the third user may be the voice data of one user or the voice data of multiple users.
  • the method of identifying the second user's voice data may refer to the explanations in the foregoing embodiments. The principles are the same and will not be described again here.
  • the noise reduction mode in response to controlling the smart headset to switch to the communication mode, is turned off to avoid noise reduction processing on the environmental voice data collected by the microphone.
  • the second user using the headset communicates with the surrounding third user
  • data transmission between the microphone and the terminal device is prohibited, which means that the microphone is prohibited from transmitting data to the surrounding third user.
  • the collected second environment voice data is sent to the terminal device, thereby preventing the terminal device from playing the second environment voice data containing the third user's voice data in the target work scenario.
  • the target work scenario is, for example, a conference call scene, that is, to avoid participating in Other users of the conference call hear the exchanges between the second user and the third user.
  • the user when the smart headset is switched to the communication mode according to the mode switching instruction, the user can be prompted to trigger the prohibition instruction in the interactive interface of the terminal device, so that the terminal device passes the prohibition instruction through Bluetooth is sent to the smart headset.
  • the smart headset prohibits data transmission between the microphone and the terminal device according to the prohibition instruction sent by the terminal device. Specifically, it can prohibit the transmission of the second environment voice data collected by the microphone, thereby avoiding the second user.
  • the voice content communicated with the surrounding third users is transmitted to the terminal device, thereby preventing the terminal device from playing the communicated voice content in the target work scenario.
  • the target work scenario is, for example, a conference call scenario, that is, to prevent other customers participating in the conference call from listening. Speech content of the exchange to the second user and the third user.
  • the second environmental voice data collected by the microphone on the smart headset is obtained, and the second environmental voice data collected by the microphone on the smart headset is determined based on the voice data of the second user using the smart headset.
  • the third user's voice data in the second ambient voice data is used to play the third user's voice data using smart headphones.
  • embodiments of the present application also provide a control device for smart earphones.
  • FIG. 5 is a schematic structural diagram of a control device for a smart headset provided by an embodiment of the present application.
  • the device may include:
  • the acquisition module 51 is used to acquire the audio data played on the terminal device.
  • the first determination module 52 is used to identify the audio data to determine that the terminal device is in a target working scenario; in the target working scenario, the microphone of the smart headset is in a voice collection state.
  • the obtaining module 51 is also used to obtain the mode switching instruction of the smart headset in the target working scenario.
  • the control module 53 is configured to control the smart earphone to switch between the at least two voice modes according to the mode switching instruction.
  • At least two voice modes include a noise reduction mode
  • the device further includes: a recognition module and a processing module.
  • the acquisition module 51 is also configured to acquire audio data played on the terminal device in response to controlling the smart headset to switch to the noise reduction mode;
  • An identification module used to identify the audio data to identify the first vocal data included in the audio data
  • the acquisition module 51 is also used to acquire the first environmental voice data collected by the microphone on the smart headset;
  • a processing module configured to perform noise reduction processing on the first environmental voice data according to the first human voice data.
  • the above processing module is specifically used for:
  • noise reduction processing is performed on the sound data in the first environmental voice data other than the sound data of the first user and the sound data of the second user.
  • At least two voice modes include a communication mode
  • the device further includes: a second determination module and a playback module.
  • the acquisition module 51 is also configured to acquire the second environmental voice data collected by the microphone on the smart earphone in response to controlling the smart earphone to switch to the communication mode;
  • a second determination module configured to determine the voice data of a third user in the second environmental voice data according to the voice data of the second user using the smart headset; wherein the third user and the second user The users are different users;
  • a playback module is used to play the third user's voice data using the smart earphone.
  • At least two voice modes include a communication mode
  • the device further includes:
  • the acquisition module 51 is configured to acquire the first sound direction of the second user using the smart earphone in response to controlling the smart earphone to switch to the communication mode;
  • a collection module configured to control the microphone to collect the sound data of a third user in the environment at a second sound direction other than the first sound direction of the second user; wherein the third user and the second The users are different users;
  • the playback module is also used to play the third user's voice data using the smart earphone.
  • the device further includes:
  • a first shutdown module configured to turn off the noise reduction mode in response to controlling the smart earphone to switch to the communication mode, and prohibit data transmission between the microphone and the terminal device.
  • the device further includes:
  • a second shutdown module configured to turn off the noise reduction mode in response to controlling the smart headset to switch to the communication mode; in response to obtaining a prohibition instruction sent by the terminal device, to prohibit the Data transmission is performed between the microphone and the terminal device.
  • the first determination module 52 is specifically used to:
  • the terminal device In response to the voice segment containing the first target voice segment and not including the second target voice segment and the third target voice segment, it is determined that the terminal device is in the target working scene; wherein the frequency of the first target voice segment The lower limit is greater than the upper frequency limit of the second target speech band, and the upper frequency limit of the first target speech band is less than the lower frequency limit of the third target speech band.
  • the first determination module 52 is specifically used to:
  • the terminal device In response to the source of the audio data being the target source, it is determined that the terminal device is in the target working scene.
  • the audio data played on the terminal device is acquired, and the audio data is identified to determine that the terminal device is in the target working scene.
  • the microphone of the smart headset is in the voice collection state.
  • obtain the mode switching instruction of the smart headset and control the smart headset to switch between at least two voice modes according to the mode switching instruction.
  • this application also proposes an electronic device, including a memory, a processor, and a computer program stored in the memory and executable on the processor.
  • the processor executes the program, the aforementioned method is implemented. methods described in the examples.
  • this application also proposes a non-transitory computer-readable storage medium on which a computer program is stored.
  • the program is executed by a processor, the method as described in the foregoing method embodiments is implemented.
  • this application also proposes a computer program product on which a computer program is stored.
  • the computer program is executed by a processor, the method as described in the foregoing method embodiments is implemented.
  • Figure 6 is a block diagram of an electronic device provided by an embodiment of the present application.
  • the electronic device 800 may be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, or the like.
  • electronic device 800 may include one or more of the following components: processing component 802, memory 804, power component 806, multimedia component 808, audio component 810, input/output (I/O) interface 812, sensor component 814, and communications component 816.
  • processing component 802 memory 804, power component 806, multimedia component 808, audio component 810, input/output (I/O) interface 812, sensor component 814, and communications component 816.
  • memory 804 power component 806, multimedia component 808, audio component 810, input/output (I/O) interface 812, sensor component 814, and communications component 816.
  • I/O input/output
  • Processing component 802 generally controls the overall operations of electronic device 800, such as operations associated with display, phone calls, data communications, camera operations, and recording operations.
  • the processing component 802 may include one or more processors 820 to execute instructions to complete all or part of the steps of the above method.
  • processing component 802 may include one or more modules that facilitate interaction between processing component 802 and other components.
  • processing component 802 may include a multimedia module to facilitate interaction between multimedia component 808 and processing component 802.
  • Memory 804 is configured to store various types of data to support operations at electronic device 800 . Examples of such data include instructions for any application or method operating on the electronic device 800, contact data, phonebook data, messages, pictures, videos, etc.
  • Memory 804 may be implemented by any type of volatile or non-volatile storage device, or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EEPROM), Programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disk.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read-only memory
  • EEPROM erasable programmable read-only memory
  • EPROM Programmable read-only memory
  • PROM programmable read-only memory
  • ROM read-only memory
  • magnetic memory flash memory, magnetic or optical disk.
  • Power component 806 provides power to various components of electronic device 800 .
  • Power components 806 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power to electronic device 800 .
  • Multimedia component 808 includes a screen that provides an output interface between the electronic device 800 and the user.
  • the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user.
  • the touch panel includes one or more touch sensors to sense touches, swipes, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide action.
  • multimedia component 808 includes a front-facing camera and/or a rear-facing camera.
  • the front camera and/or the rear camera may receive external multimedia data.
  • Each front-facing camera and rear-facing camera can be a fixed optical lens system or have a focal length and optical zoom capabilities.
  • Audio component 810 is configured to output and/or input audio signals.
  • audio component 810 includes a microphone (MIC) configured to receive external audio signals when electronic device 800 is in operating modes, such as call mode, recording mode, and voice recognition mode. The received audio signal may be further stored in memory 804 or sent via communication component 816 .
  • audio component 810 also includes a speaker for outputting audio signals.
  • the I/O interface 812 provides an interface between the processing component 802 and a peripheral interface module, which may be a keyboard, a click wheel, a button, etc. These buttons may include, but are not limited to: Home button, Volume buttons, Start button, and Lock button.
  • Sensor component 814 includes one or more sensors for providing various aspects of status assessment for electronic device 800 .
  • the sensor component 814 can detect the open/closed state of the electronic device 800, the relative positioning of the components, such as the display and keypad of the electronic device 800, the sensor component 814 can also detect the electronic device 800 or an electronic device 800.
  • the position of components changes, the presence or absence of user contact with the electronic device 800 , the orientation or acceleration/deceleration of the electronic device 800 and the temperature of the electronic device 800 change.
  • Sensor assembly 814 may include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
  • Sensor assembly 814 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
  • the sensor component 814 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
  • Communication component 816 is configured to facilitate wired or wireless communication between electronic device 800 and other devices.
  • the electronic device 800 can access a wireless network based on a communication standard, such as WiFi, 4G or 5G, or a combination thereof.
  • the communication component 816 receives broadcast signals or broadcast related information from an external broadcast management system via a broadcast channel.
  • the communications component 816 also includes a near field communications (NFC) module to facilitate short-range communications.
  • NFC near field communications
  • the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology and other technologies.
  • RFID radio frequency identification
  • IrDA infrared data association
  • UWB ultra-wideband
  • Bluetooth Bluetooth
  • electronic device 800 may be configured by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable A programmable gate array (FPGA), controller, microcontroller, microprocessor or other electronic component implementation is used to perform the above method.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA field programmable A programmable gate array
  • controller microcontroller, microprocessor or other electronic component implementation is used to perform the above method.
  • a non-transitory computer-readable storage medium including instructions such as a memory 804 including instructions, which can be executed by the processor 820 of the electronic device 800 to complete the above method is also provided.
  • the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, etc.
  • references to the terms “one embodiment,” “some embodiments,” “an example,” “specific examples,” or “some examples” or the like means that specific features are described in connection with the embodiment or example. , structures, materials or features are included in at least one embodiment or example of the present application. In this specification, the schematic expressions of the above terms are not necessarily directed to the same embodiment or example. Furthermore, the specific features, structures, materials or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Furthermore, those skilled in the art may combine and combine different embodiments or examples and features of different embodiments or examples described in this specification unless they are inconsistent with each other.
  • first and second are used for descriptive purposes only and cannot be understood as indicating or implying relative importance or implicitly indicating the quantity of indicated technical features. Therefore, features defined as “first” and “second” may explicitly or implicitly include at least one of these features.
  • “plurality” means at least two, such as two, three, etc., unless otherwise expressly and specifically limited.
  • a "computer-readable medium” may be any device that can contain, store, communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Non-exhaustive list of computer readable media include the following: electrical connections with one or more wires (electronic device), portable computer disk cartridges (magnetic device), random access memory (RAM), Read-only memory (ROM), erasable and programmable read-only memory (EPROM or flash memory), fiber optic devices, and portable compact disc read-only memory (CDROM).
  • the computer-readable medium may even be paper or other suitable medium on which the program may be printed, as the paper or other medium may be optically scanned, for example, and subsequently edited, interpreted, or otherwise suitable as necessary. process to obtain the program electronically and then store it in computer memory.
  • various parts of the present application can be implemented in hardware, software, firmware, or a combination thereof.
  • various steps or methods may be implemented in software or firmware stored in a memory and executed by a suitable instruction execution system.
  • a suitable instruction execution system For example, if it is implemented in hardware, as in another embodiment, it can be implemented by any one of the following technologies known in the art or their combination: discrete logic gate circuits with logic functions for implementing data signals; Logic circuits, application specific integrated circuits with suitable combinational logic gates, programmable gate arrays (PGA), field programmable gate arrays (FPGA), etc.
  • the program can be stored in a computer-readable storage medium.
  • the program can be stored in a computer-readable storage medium.
  • each functional unit in various embodiments of the present application can be integrated into a processing module, or each unit can exist physically alone, or two or more units can be integrated into one module.
  • the above integrated modules can be implemented in the form of hardware or software function modules. If the integrated module is implemented in the form of a software function module and sold or used as an independent product, it can also be stored in a computer-readable storage medium.
  • the storage media mentioned above can be read-only memory, magnetic disks or optical disks, etc.

Landscapes

  • Telephone Function (AREA)

Abstract

A smart earphone control method and apparatus, an electronic device and a storage medium. The method comprises: acquiring audio data played on a terminal device (101); identifying the audio data to determine that the terminal device is in a target working scenario (102), wherein, in the target working scenario, a microphone of a smart earphone is in a voice acquisition state; in the target working scenario, acquiring a mode switching instruction for the smart earphone (103); and, according to the mode switching instruction, controlling the smart earphone to be switched between at least two voice modes (104). In a target working scenario, a smart earphone is controlled to be switched between at least two modes, so that when a user uses the smart earphone, the user does not need to frequently take off or put on the smart earphone, thus improving the convenience of using the smart earphone.

Description

智能耳机的控制方法、装置、电子设备和存储介质Control method, device, electronic device and storage medium for smart headphones 技术领域Technical field
本申请涉及计算机应用技术领域,尤其涉及智能耳机的控制方法、装置、电子设备和存储介质。The present application relates to the field of computer application technology, and in particular to control methods, devices, electronic devices and storage media for smart earphones.
背景技术Background technique
终端设备在使用过程中搭配智能耳机,给人们的日常使用带来了便利,而具有降噪功能的智能耳机,可以隔绝周遭的噪音,目前的使用也越来越普及。Terminal devices are equipped with smart headphones during use, which brings convenience to people's daily use. Smart headphones with noise reduction function can isolate the surrounding noise, and their current use is becoming more and more popular.
而用户在使用具有降噪功能的智能耳机时,通常还需要和周围的人进行沟通,此时,用户就需要频繁的摘下智能耳机以便于和周围的人进行沟通,再戴上耳机,降低了智能耳机使用的便利性。When users use smart earphones with noise reduction function, they usually need to communicate with people around them. At this time, users need to frequently take off the smart earphones to communicate with people around them, and then put on the earphones to reduce the noise level. It improves the convenience of using smart headphones.
发明内容Contents of the invention
本申请提出一种智能耳机的控制方法、装置、电子设备和存储介质,以提高智能耳机使用的便利性。This application proposes a control method, device, electronic device and storage medium for smart earphones to improve the convenience of using smart earphones.
本申请一方面实施例提出了一种智能耳机的控制方法,包括:On the one hand, an embodiment of this application proposes a control method for smart headphones, including:
获取所述终端设备上播放的音频数据;Obtain audio data played on the terminal device;
对所述音频数据进行识别,以确定所述终端设备处于目标工作场景;所述目标工作场景下所述智能耳机的麦克风处于语音采集状态The audio data is identified to determine that the terminal device is in a target working scenario; the microphone of the smart headset is in a voice collection state under the target working scenario.
在所述目标工作场景下,获取所述智能耳机的模式切换指令;Under the target working scenario, obtain the mode switching instruction of the smart headset;
根据所述模式切换指令,控制所述智能耳机在所述至少两个语音模式间切换。According to the mode switching instruction, the smart earphone is controlled to switch between the at least two voice modes.
本申请另一方面实施例提出了一种智能耳机的控制装置,包括:Another embodiment of the present application provides a control device for smart headphones, including:
获取模块,用于获取所述终端设备上播放的音频数据;An acquisition module, used to acquire audio data played on the terminal device;
第一确定模块,用于对所述音频数据进行识别,以确定所述终端设备处于目标工作场景;所述目标工作场景下,所述智能耳机的麦克风处于语音采集状态;The first determination module is used to identify the audio data to determine that the terminal device is in a target working scenario; in the target working scenario, the microphone of the smart headset is in a voice collection state;
所述获取模块,还用于在所述目标工作场景下,获取所述智能耳机的模式切换指令;The acquisition module is also used to acquire the mode switching instruction of the smart headset in the target working scenario;
控制模块,用于根据所述模式切换指令,控制所述智能耳机在所述至少两个语音模式间切换。A control module configured to control the smart earphone to switch between the at least two voice modes according to the mode switching instruction.
本申请另一方面实施例提出了一种电子设备,包括存储器、处理器及存储在存储器上 并可在处理器上运行的计算机程序,所述处理器执行所述程序时,实现如前述一方面所述的方法。Another aspect of the present application provides an electronic device, including a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the program, the aforementioned aspect is implemented. the method described.
本申请另一方面实施例提出了一种非临时性计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现如前述一方面所述的方法。Another embodiment of the present application proposes a non-transitory computer-readable storage medium on which a computer program is stored. When the computer program is executed by a processor, the method as described in the foregoing aspect is implemented.
本申请另一方面实施例提出了一种计算机程序产品,其上存储有计算机程序,所述程序被处理器执行时实现如前述一方面所述的方法。Another aspect of the present application provides a computer program product on which a computer program is stored. When the program is executed by a processor, the method as described in the foregoing aspect is implemented.
本申请提出的智能耳机的控制方法、装置、电子设备和存储介质,获取终端设备上播放的音频数据,对音频数据进行识别,以确定终端设备处于目标工作场景,目标工作场景下智能耳机的麦克风处于语音采集状态,在目标工作场景下,获取智能耳机的模式切换指令,根据模式切换指令,控制智能耳机在至少两个语音模式间切换,通过在目标工作场景下,控制智能耳机在至少两种模式间切换可以帮助用户在使用智能耳机的时候,不需要频繁的摘下或带上智能耳机,提高了智能耳机使用的便利性。The control method, device, electronic equipment and storage medium of the smart headset proposed in this application acquire the audio data played on the terminal device and identify the audio data to determine that the terminal device is in the target working scene and the microphone of the smart headset in the target working scene. In the voice collection state, in the target work scenario, obtain the mode switching instruction of the smart headset, and control the smart headset to switch between at least two voice modes according to the mode switching instruction, by controlling the smart headset to switch between at least two voice modes in the target work scenario. Switching between modes can help users avoid taking off or putting on smart headphones frequently when using smart headphones, which improves the convenience of using smart headphones.
本申请附加的方面和优点将在下面的描述中部分给出,部分将从下面的描述中变得明显,或通过本申请的实践了解到。Additional aspects and advantages of the application will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the application.
附图说明Description of the drawings
本申请上述的和/或附加的方面和优点从下面结合附图对实施例的描述中将变得明显和容易理解,其中:The above and/or additional aspects and advantages of the present application will become apparent and readily understood from the following description of the embodiments in conjunction with the accompanying drawings, in which:
图1为本申请实施例所提供的一种智能耳机的控制方法的流程示意图;Figure 1 is a schematic flowchart of a smart headset control method provided by an embodiment of the present application;
图2为本申请实施例提供的另一种智能耳机的控制方法的流程示意图;Figure 2 is a schematic flow chart of another smart headset control method provided by an embodiment of the present application;
图3为本申请实施例提供的另一种智能耳机的控制方法的流程示意图;Figure 3 is a schematic flow chart of another smart headset control method provided by an embodiment of the present application;
图4为本申请实施例提供的另一种智能耳机的控制方法的流程示意图;Figure 4 is a schematic flow chart of another smart headset control method provided by an embodiment of the present application;
图5为本申请实施例提供的一种智能耳机的控制装置的结构示意图;Figure 5 is a schematic structural diagram of a control device for a smart headset provided by an embodiment of the present application;
图6为本申请实施例提供的一种电子设备的框图。Figure 6 is a block diagram of an electronic device provided by an embodiment of the present application.
具体实施方式Detailed ways
下面详细描述本申请的实施例,所述实施例的示例在附图中示出,其中自始至终相同或类似的标号表示相同或类似的元件或具有相同或类似功能的元件。下面通过参考附图描述的实施例是示例性的,旨在用于解释本申请,而不能理解为对本申请的限制。The embodiments of the present application are described in detail below. Examples of the embodiments are shown in the accompanying drawings, wherein the same or similar reference numerals throughout represent the same or similar elements or elements with the same or similar functions. The embodiments described below with reference to the drawings are exemplary and are intended to explain the present application, but should not be construed as limiting the present application.
下面参考附图描述本申请实施例的智能耳机的控制方法、装置、电子设备和存储介质。The following describes the control method, device, electronic device and storage medium of the smart headset according to the embodiment of the present application with reference to the accompanying drawings.
图1为本申请实施例所提供的一种智能耳机的控制方法的流程示意图。FIG. 1 is a schematic flowchart of a method for controlling a smart headset provided by an embodiment of the present application.
本申请实施例的智能耳机的控制方法的执行主体为智能耳机的控制装置,该装置可设置于电子设备中,电子设备可以为智能耳机,智能耳机为具有降噪功能的降噪耳机。The execution subject of the smart earphone control method in the embodiment of the present application is a control device of the smart earphone. The device can be installed in an electronic device. The electronic device can be a smart earphone. The smart earphone is a noise-cancelling earphone with a noise reduction function.
如图1所示,该方法可以包括以下步骤:As shown in Figure 1, the method may include the following steps:
步骤101,获取终端设备上播放的音频数据。Step 101: Obtain audio data played on the terminal device.
其中,终端设备,可以为智能手机、掌上电脑、智能穿戴设备、电脑等,本实施例中不进行限定。The terminal device may be a smartphone, a handheld computer, a smart wearable device, a computer, etc., which are not limited in this embodiment.
本申请实施例的一个示例中,智能耳机为具有降噪功能的蓝牙耳机,智能耳机和终端设备通过蓝牙进行连接,建立连接后,终端设备将播放的音频数据按照设定的频率发送給智能耳机,例如,实时发送,或者是每200ms发送一次,其中,音频数据可以为人的语音数据、影音娱乐数据等。In an example of the embodiment of this application, the smart earphone is a Bluetooth earphone with a noise reduction function. The smart earphone and the terminal device are connected through Bluetooth. After the connection is established, the terminal device sends the played audio data to the smart earphone according to the set frequency. , for example, sent in real time, or sent every 200ms, where the audio data can be human voice data, audio-visual entertainment data, etc.
步骤102,对音频数据进行识别,以确定终端设备处于目标工作场景。Step 102: Recognize the audio data to determine that the terminal device is in the target working scene.
其中,目标工作场景下,智能耳机的麦克风处于语音采集状态,也就是说当终端设备处于目标工作场景时,智能耳机的麦克风处于开启状态,可用于采集环境的语音数据。Among them, in the target working scenario, the microphone of the smart headset is in the voice collection state, which means that when the terminal device is in the target working scenario, the microphone of the smart headset is turned on and can be used to collect voice data of the environment.
本申请实施例中,通过对音频数据进行识别,以识别出音频数据属于影音娱乐数据,还是属于人的语音数据。其中,影音娱乐数据,例如音乐、电影、短视频等,这类数据均包含背景音乐,若音频数据识别属于语音数据,例如,电话会议场景下的人的语音数据,或即时聊天软件的人的语音数据等,在这种场景下,使用智能耳机的用户的声音数据需要传送給参与电话会议或即时聊天的其它终端设备的用户,从而,智能耳机上的麦克风需要开启以采集使用智能耳机的用户的声音数据,则确定终端设备处于目标工作场景;若音频数据识别为影音娱乐数据,例如,在应用程序上收听的音乐、观看的视频等,在这种场景下,智能耳机用于将终端设备播放的音频数据通过耳机传入人耳,而智能耳机上的麦克风不开启,不需要采集使用智能耳机的用户的声音数据或环境数据。In the embodiment of the present application, the audio data is identified to identify whether the audio data belongs to audio-visual entertainment data or human voice data. Among them, audio-visual entertainment data, such as music, movies, short videos, etc., such data all include background music. If the audio data recognition is voice data, for example, the voice data of people in a conference call scenario, or the voice data of people using instant chat software. Voice data, etc. In this scenario, the voice data of the user using the smart headset needs to be transmitted to users of other terminal devices participating in conference calls or instant chats. Therefore, the microphone on the smart headset needs to be turned on to collect the users using the smart headset. sound data, it is determined that the terminal device is in the target working scene; if the audio data is identified as audio-visual entertainment data, for example, music listened to on the application, video watched, etc., in this scenario, the smart headset is used to use the terminal device The played audio data is transmitted to the human ear through the earphones, and the microphone on the smart earphones is not turned on. There is no need to collect the sound data or environmental data of the user using the smart earphones.
本申请实施例的一种实现方式中,获取设定时长的音频数据,对音频数据进行识别,确定音频数据包括的语音频段,响应于语音频段中包含第一目标语音频段,且不包含第二目标语音频段和第三目标语音频段,确定终端设备处于目标工作场景。其中,第一目标语音频段的频率下限大于第二目标语音频段的频率上限,且第一目标语音频段的频率上限小于第三目标语音频段的频率下限。作为一种示例,语音频段包含低频频段、中频频段和高频频段,则第一目标语音频段为中频频段,第二目标语音频段为低频频段,第三目标语音频段为高频频段,通常获取的用户的声音数据中低频频段和高频频段较少,也就是说用户声音数据的频率主要属于中频频段,而影音娱乐数据中由于包含音乐数据,音乐数据中包含较多的低频频段的数据和高频频段的数据,因此,通过对音频数据包含的频段进行识别, 可确定音频数据是人的语音数据还是影音娱乐数据,在音频数据识别为人的语音数据的情况下,确定终端设备处于目标工作场景。In an implementation manner of the embodiment of the present application, audio data of a set duration is obtained, the audio data is identified, the voice segment included in the audio data is determined, and the response is that the voice segment includes the first target voice segment and does not include the second target voice segment. The target voice band and the third target voice band determine that the terminal device is in the target working scenario. Wherein, the lower frequency limit of the first target speech band is greater than the upper frequency limit of the second target speech band, and the upper frequency limit of the first target speech band is less than the lower frequency limit of the third target speech band. As an example, the speech band includes a low frequency band, a medium frequency band and a high frequency band, then the first target speech band is the medium frequency band, the second target speech band is the low frequency band, and the third target speech band is the high frequency band. Usually The acquired user voice data contains fewer low-frequency and high-frequency bands, which means that the frequency of the user's voice data mainly belongs to the mid-frequency band. Since the audio-visual entertainment data contains music data, the music data contains more low-frequency bands. data and high-frequency band data. Therefore, by identifying the frequency band contained in the audio data, it can be determined whether the audio data is human voice data or audio-visual entertainment data. When the audio data is recognized as human voice data, it is determined that the terminal device is in Target work scenario.
本申请实施例的另一种实现方式中,通过对音频数据进行识别,若识别到音频数据中包括的第一目标音频频段在整个音频频段中占比大于设定阈值,即第二目标音频频段和/或第三目标音频频段在整个音频频段中的占比低于设定阈值,则认为该音频数据为用户的声音数据,即确定终端设备处于目标工作场景。In another implementation manner of the embodiment of the present application, by identifying the audio data, if it is recognized that the proportion of the first target audio frequency band included in the audio data in the entire audio frequency band is greater than the set threshold, that is, the second target audio frequency band and/or the proportion of the third target audio frequency band in the entire audio frequency band is lower than the set threshold, then the audio data is considered to be the user's voice data, that is, it is determined that the terminal device is in the target working scene.
需要说明的是,为了提高识别的准确度,可采集多个设定时长的音频数据进行识别,在多个设定时长的音频数据包含的语音频段均满足上述要求的情况下,确定终端设备处于目标工作场景。It should be noted that in order to improve the accuracy of identification, audio data of multiple set durations can be collected for identification. When the voice segments contained in the audio data of multiple set durations meet the above requirements, it is determined that the terminal device is in Target work scenario.
本申请实施例的又一种实现方式中,对音频数据携带的来源信息进行识别,确定音频数据的来源,响应于音频数据的来源为目标来源,确定终端设备处于目标工作场景,具体来说,可对音频数据携带的来源信息进行识别,其中,来源信息指示了音频数据所属的应用软件,若识别到音频数据所属的应用程序为目标应用程序,则确定终端设备处于目标工作场景。例如,若来源信息指示拼音数据所属的是影音娱乐的应用程序,不是通话类应用程序,例如,电话会议应用程序,则确定终端设备不属于目标工作场景,否则确定终端设备属于目标工作场景。In another implementation manner of the embodiment of the present application, the source information carried by the audio data is identified, the source of the audio data is determined, and in response to the source of the audio data being the target source, it is determined that the terminal device is in the target working scene. Specifically, The source information carried by the audio data can be identified, where the source information indicates the application software to which the audio data belongs. If the application to which the audio data belongs is identified as the target application, it is determined that the terminal device is in the target working scene. For example, if the source information indicates that the pinyin data belongs to an audio-visual entertainment application, not a call application, for example, a conference call application, then it is determined that the terminal device does not belong to the target work scenario, otherwise it is determined that the terminal device belongs to the target work scenario.
步骤103,在目标工作场景下,获取智能耳机的模式切换指令。Step 103: Obtain the mode switching instruction of the smart headset in the target working scenario.
本申请实施例中,在目标工作场景下,智能耳机具有至少两种语音模式,至少两种语音模式可通过模式切换指令切换,目标工作场景,即为多个用户进行交流的场景,例如,电话会议场景,语音聊天场景等。In the embodiment of the present application, the smart headset has at least two voice modes in the target working scenario, which is a scenario where multiple users communicate, for example, a phone call. Meeting scenarios, voice chat scenarios, etc.
本申请实施例的一种实现方式中,在监测到终端设备处于目标工作场景下的情况下,可更改智能耳机的设定按键对应的功能模式,例如,设定按键A在非目标工作场景下,功能模式包括播放和暂停功能,而在目标工作模式下,功能模式切换为包括至少两种语音模式。从而,在目标工作模式下,响应于用户对智能耳机的设定按键的按压操作,可获取智能耳机的模式切换指令,模式切换指令用于将智能耳机切换至模式切换指令指示的语音模式。例如,智能耳机包含两种语音模式,即降噪模式和交流模式,在降噪模式下,外界声音会被降噪处理,即降噪处理后,外界声音小于设定音量;在交流模式下,外界声音不会被降噪处理,而是可以维持原来的声音或者是放大原来的声音,以使得外界声音大于设定音量。若监测到用户按压设定按键,则确定监测到用户按压设定按键前智能耳机所处的语音模式,若智能耳机处于降噪模式,则响应按键的操作,生成交流模式切换指令,同理,若智能耳机处于交流模式,则响应对按键的操作,生成降噪模式切换指令。In an implementation manner of the embodiment of the present application, when it is detected that the terminal device is in a target working scenario, the functional mode corresponding to the setting button of the smart headset can be changed. For example, setting button A in a non-target working scenario , the functional mode includes play and pause functions, and in the target working mode, the functional mode is switched to include at least two voice modes. Therefore, in the target working mode, in response to the user pressing the setting button of the smart earphone, the mode switching instruction of the smart earphone can be obtained, and the mode switching instruction is used to switch the smart earphone to the voice mode indicated by the mode switching instruction. For example, smart headphones include two voice modes, namely noise reduction mode and communication mode. In the noise reduction mode, the external sound will be noise reduction processing, that is, after the noise reduction processing, the external sound is less than the set volume; in the communication mode, The external sound will not be denoised, but the original sound can be maintained or the original sound can be amplified so that the external sound is greater than the set volume. If it is detected that the user presses the setting button, it is determined which voice mode the smart headset was in before the user pressed the setting button. If the smart headset is in the noise reduction mode, it responds to the operation of the button and generates a communication mode switching command. Similarly, If the smart headset is in the communication mode, it will generate a noise reduction mode switching instruction in response to the operation of the button.
本申请实施例的另一种实现方式中,在监测到终端设备处于目标工作场景下的情况下,可通过获取到的语音识别对应的模式切换指令,以更改智能耳机对应的功能模式,具体来说,可对获取到的语音信号进行识别,以识别得到语音信号中包含的关键词,基于关键词确定对应的模式切换指令,其中,模式切换指令用于将智能耳机切换至模式切换指令指示的语音模式。其中,关于智能耳机的语音模式可参照前一种实现方式中的说明,此处不再赘述。In another implementation manner of the embodiment of the present application, when it is detected that the terminal device is in the target working scene, the mode switching instruction corresponding to the obtained voice recognition can be used to change the corresponding functional mode of the smart headset. Specifically, That is, the acquired voice signal can be identified to identify the keywords contained in the voice signal, and the corresponding mode switching instruction is determined based on the keywords, where the mode switching instruction is used to switch the smart headset to the mode indicated by the mode switching instruction. Voice mode. Regarding the voice mode of the smart headset, please refer to the description in the previous implementation, and will not be described again here.
步骤104,根据模式切换指令,控制智能耳机在至少两个语音模式间切换。Step 104: Control the smart headset to switch between at least two voice modes according to the mode switching instruction.
进而,根据模式切换指令,控制智能耳机在至少两个语音模式间切换,实现了在佩戴智能耳机的情况下,即可在至少两个语音模式间切换,提高了语音模式切换的效率,增加了用户使用智能耳机的便利性。Furthermore, according to the mode switching instruction, the smart headset is controlled to switch between at least two voice modes, so that the smart headset can be switched between at least two voice modes while wearing the smart headset, which improves the efficiency of voice mode switching and increases the number of voice modes. User convenience in using smart headphones.
本实施例中以两个语音模式即降噪模式和交流模式为例进行说明,在一种场景下,智能耳机处于降噪模式,也就是说使用智能耳机的用户通过客户端程序进行语音通话,则根据模式切换指令,切换至交流模式,在交流模式下,语音通话的声音会降低至小于设定音量,以使得用户在佩戴智能耳机的情况下,也不会因为终端上设备中播放的音频数据的声音影响外界声音的获取;在另一种场景下,智能耳机处于交流模式,则检测到模式切换指令后,智能耳机会从交流模式切换至降噪模式,在降噪模式下,外界声音会被降噪处理,使得用户听到的外界声音音量小于设定音量。通过在目标工作场景下,控制智能耳机在两种模式的切换可以帮助用户在使用智能耳机的过程中,不需要频繁的摘下或带上智能耳机即可顺畅的沟通和开会,提高了智能耳机使用的便利性。This embodiment takes two voice modes, namely the noise reduction mode and the communication mode, as an example to illustrate. In one scenario, the smart headset is in the noise reduction mode, that is to say, the user using the smart headset makes a voice call through the client program. Then according to the mode switching instruction, switch to the communication mode. In the communication mode, the sound of the voice call will be reduced to less than the set volume, so that when the user wears the smart headset, he will not be affected by the audio played by the device on the terminal. The sound of data affects the acquisition of external sounds; in another scenario, the smart earphones are in communication mode, and after detecting the mode switching command, the smart earphones will switch from the communication mode to the noise reduction mode. In the noise reduction mode, the external sounds It will be processed by noise reduction, so that the volume of external sounds heard by the user is lower than the set volume. By controlling the switching of smart earphones between the two modes in the target work scenario, users can communicate and hold meetings smoothly without having to frequently take off or put on the smart earphones during the use of smart earphones, which improves the efficiency of smart earphones. Ease of use.
本申请实施例的智能耳机的控制方法中,获取终端设备上播放的音频数据,对音频数据进行识别,以确定终端设备处于目标工作场景,目标工作场景下智能耳机的麦克风处于语音采集状态,在目标工作场景下,获取智能耳机的模式切换指令,根据模式切换指令,控制智能耳机在至少两个语音模式间切换,通过在目标工作场景下,控制智能耳机在至少两种模式间切换可以帮助用户在使用智能耳机的时候,不需要频繁的摘下或带上智能耳机,提高了智能耳机使用的便利性。In the control method of the smart headset in the embodiment of the present application, the audio data played on the terminal device is obtained, and the audio data is identified to determine that the terminal device is in the target working scene. In the target working scene, the microphone of the smart headset is in the voice collection state. In the target work scenario, obtain the mode switching instruction of the smart headset, and control the smart headset to switch between at least two voice modes according to the mode switching instruction. By controlling the smart headset to switch between at least two modes in the target work scenario, it can help the user. When using smart headphones, you do not need to frequently take off or put on the smart headphones, which improves the convenience of using smart headphones.
基于上一实施例,图2为本申请实施例提供的另一种智能耳机的控制方法的流程示意图,本实施例中,以语音模式包含降噪模式和交流模式为例进行说明,具体说明了智能耳机切换至降噪模式后,智能耳机对获取到的环境语音数据进行降噪处理,以避免人声数据对目标模式的影响,如图2所示,该方法包含以下步骤:Based on the previous embodiment, FIG. 2 is a schematic flow chart of another smart earphone control method provided by an embodiment of the present application. In this embodiment, the voice mode including the noise reduction mode and the communication mode is used as an example for explanation. Specifically, After the smart headset switches to the noise reduction mode, the smart headset performs noise reduction processing on the acquired environmental voice data to avoid the impact of human voice data on the target mode. As shown in Figure 2, the method includes the following steps:
步骤201,获取终端设备上播放的音频数据。Step 201: Obtain audio data played on the terminal device.
步骤202,对音频数据进行识别,以确定终端设备处于目标工作场景。Step 202: Recognize the audio data to determine that the terminal device is in the target working scene.
其中,目标工作场景下智能耳机的麦克风处于语音采集状态。Among them, the microphone of the smart headset is in the voice collection state in the target work scenario.
步骤203,在目标工作场景下,获取智能耳机的模式切换指令。Step 203: Obtain the mode switching instruction of the smart headset in the target working scenario.
步骤204,根据模式切换指令,控制智能耳机在至少两个语音模式间切换。Step 204: Control the smart headset to switch between at least two voice modes according to the mode switching instruction.
其中,步骤201至步骤204,可参照前述实施例中的解释说明,原理相同,此处不再赘述。For steps 201 to 204, reference may be made to the explanations in the foregoing embodiments. The principles are the same and will not be described again here.
步骤205,响应于控制智能耳机切换至降噪模式,获取终端设备上播放的音频数据。Step 205: In response to controlling the smart headset to switch to the noise reduction mode, obtain the audio data played on the terminal device.
需要说明的是,步骤205至步骤208,在步骤204之后执行仅为一种示例,也就是说,步骤205至步骤208的执行时序不进行限定。It should be noted that the execution of steps 205 to 208 after step 204 is only an example, that is to say, the execution timing of steps 205 to 208 is not limited.
本申请实施例中,响应于根据模式切换指令控制智能耳机切换至降噪模式,获取终端设备上播放的音频数据,音频数据是其它终端设备采集的语音数据,其它终端设备的用户和该终端设备均处于目标工作场景下,该语音数据中包含人的声音数据,为了便于区别,称为第一用户的声音数据。In the embodiment of the present application, in response to controlling the smart headset to switch to the noise reduction mode according to the mode switching instruction, audio data played on the terminal device is obtained. The audio data is voice data collected by other terminal devices. Users of other terminal devices and the terminal device They are all in the target work scenario. The voice data contains human voice data. For the convenience of distinction, it is called the first user's voice data.
例如,目标工作场景为电话会议场景,在电话会议场景下,有3个用户参加电话会议,分别为用户1、用户2和用户3,其中,用户1为终端设备对应的用户,用户2和用户3则对应其它终端设备,则获取到的终端设备上播放的音频数据则包括两个第一用户的声音数据,即用户2和用户3的声音数据。For example, the target work scenario is a conference call scenario. In the conference call scenario, there are three users participating in the conference call, namely User 1, User 2 and User 3. Among them, User 1 is the user corresponding to the terminal device, User 2 and User 3 are participating in the conference call. 3 corresponds to other terminal devices, and the obtained audio data played on the terminal device includes the sound data of the two first users, that is, the sound data of user 2 and user 3.
步骤206,对音频数据进行识别,以识别得到音频数据包括的第一用户的声音数据。Step 206: Recognize the audio data to identify the first user's voice data included in the audio data.
本申请实施例的一种实现方式中,对音频数据进行识别,以识别到音频数据中不同人声对应的声音频段,由于不同人的声音频率不同,因此,可识别出不同人的声音对应的不同声音频段或者说是不同的声音的频谱特征,即可识别得到音频数据包括的第一用户的声音数据,其中,识别得到的第一用户的声音数据可以为一个或多个,与目标工作场景下,参与的人数有关,参与的人数多,识别到的第一用户的声音数据则多。In an implementation manner of the embodiment of the present application, the audio data is identified to identify the sound segments corresponding to different voices in the audio data. Since the voices of different people have different frequencies, therefore, the segments corresponding to the voices of different people can be identified. Different sound frequency bands or different sound spectrum characteristics can be used to identify the first user's sound data included in the audio data. The identified first user's sound data can be one or more, which are related to the target work scene. Next, it is related to the number of participants. The more participants, the more voice data of the first user will be recognized.
本申请实施例的另一种实现方式中,可基于训练得到的语音识别模型,对音频数据进行音色等特征识别,以识别得到音频数据中包含的第一用户的声音数据。In another implementation manner of the embodiment of the present application, feature recognition such as timbre and color can be performed on the audio data based on the trained speech recognition model to identify the first user's voice data contained in the audio data.
例如,在一种电话会议场景下,有3个用户参加电话会议,则获取到的音频数据进行识别得到的音频数据包括两个第一用户的声音数据,可称为第一用户A的声音数据和第一用户B的声音数据。For example, in a conference call scenario where three users participate in the conference call, the audio data obtained after identification includes the voice data of the two first users, which can be called the voice data of the first user A. and the voice data of the first user B.
步骤207,获取智能耳机上的麦克风采集的第一环境语音数据。Step 207: Obtain the first environmental voice data collected by the microphone on the smart headset.
本申请实施例中,智能耳机上的麦克风处于启动状态,可实时采集环境语音数据,为了便于和其它模式下采集的环境语音数据进行区分,称为第一环境语音数据,并发送給智能耳机。In the embodiment of the present application, the microphone on the smart headset is in the activated state and can collect environmental voice data in real time. In order to facilitate differentiation from the environmental voice data collected in other modes, it is called the first environmental voice data and is sent to the smart headset.
步骤208,根据第一用户的声音数据,对第一环境语音数据进行降噪处理。Step 208: Perform noise reduction processing on the first environmental voice data according to the first user's voice data.
本申请实施例中,可对第一环境语音数据进行识别,以识别得到第一环境语音数据中包含的人的声音数据,其中,对第一环境语音数据的识别,以得到人的声音数据的方法可参照前述步骤中的解释说明,原理相同,此处不再赘述。In the embodiment of the present application, the first environmental voice data can be recognized to obtain the human voice data contained in the first environmental voice data, wherein the first environmental voice data can be recognized to obtain the human voice data. For the method, please refer to the explanations in the previous steps. The principles are the same and will not be repeated here.
作为一种实现方式,根据第一用户的声音数据的声音频段,将第一环境语音数据中不属于第一用户的声音数据的声音频段的声音频段的音量降低或者是直接删除,以对第一环境语音数据进行降噪处理,以降低环境语音数据对终端设备上播放的语音数据的影响。As an implementation manner, according to the voice segments of the first user's voice data, the volume of the voice segments in the first environmental voice data that do not belong to the first user's voice data is reduced or directly deleted, so as to improve the first user's voice data. The ambient voice data undergoes noise reduction processing to reduce the impact of the ambient voice data on the voice data played on the terminal device.
作为另一种实现方式,获取使用智能耳机的第二用户的声音数据,根据第一用户的声音数据和第二用户的声音数据,对第一环境语音数据中第一用户的声音数据和第二用户的声音数据以外的声音数据进行降噪处理,其中,降噪处理的降噪级别,可以根据需求进行设置,作为一种实现方式,可将降噪设置为最高,以实现将获取到的第一环境语音数据中第一用户的声音数据和第二用户的声音数据以外的声音数据滤除,以最大限度的提高目标工作场景下通话或交流的效果,实现了对第一环境语音数据中不属于第一用户和第二用户的其它的声音数据进行降噪处理,使得在目标工作场景下,智能耳机在使用降噪模式时可以滤除环境中的声音数据,使得目标工作模式下第二用户听到的终端设备播放的音频数据更加纯净,提高了目标工作场景下的语音效果。As another implementation manner, the voice data of the second user using the smart headset is obtained, and based on the voice data of the first user and the voice data of the second user, the voice data of the first user and the voice data of the second user in the first environmental voice data are compared. Sound data other than the user's sound data is subjected to noise reduction processing. The noise reduction level of the noise reduction processing can be set according to needs. As an implementation method, the noise reduction can be set to the highest to achieve the acquisition of the third sound. The voice data other than the first user's voice data and the second user's voice data are filtered out in the first ambient voice data to maximize the effect of calls or communication in the target work scenario, and achieve the goal of filtering out the first ambient voice data that is not included in the first user's voice data. Other sound data belonging to the first user and the second user are subjected to noise reduction processing, so that in the target working scenario, the smart earphones can filter out the sound data in the environment when using the noise reduction mode, so that the second user in the target working mode can The audio data played by the terminal device is purer, which improves the speech effect in the target work scenario.
本申请实施例的智能耳机的控制方法中,响应于控制智能耳机切换至降噪模式,获取终端设备上播放的音频数据,对音频数据进行识别,以识别得到音频数据包括的第一用户的声音数据,获取智能耳机上的麦克风采集的环境语音数据,根据第一用户的声音数据,对环境语音数据进行降噪处理,以降低环境语音数据对终端设备上播放的语音数据的影响,提高了智能耳机使用的便利性。In the control method of the smart headset in the embodiment of the present application, in response to controlling the smart headset to switch to the noise reduction mode, the audio data played on the terminal device is obtained, and the audio data is identified to identify the first user's voice included in the audio data. data, obtain the ambient voice data collected by the microphone on the smart headset, and perform noise reduction processing on the ambient voice data based on the first user's voice data to reduce the impact of the ambient voice data on the voice data played on the terminal device and improve the intelligence Ease of use of headphones.
基于上一实施例,图3为本申请实施例提供的另一种智能耳机的控制方法的流程示意图,本实施例中,以语音模式包含降噪模式和交流模式为例进行说明,具体说明了智能耳机切换至交流模式后,智能耳机如何获取环境语音数据中使用智能耳机的第二用户外的第三用户的声音数据,实现在不摘取智能耳机的情况下进行交流。如图3所示,该方法包含以下步骤:Based on the previous embodiment, FIG. 3 is a schematic flow chart of another smart earphone control method provided by an embodiment of the present application. In this embodiment, the voice mode including the noise reduction mode and the communication mode is used as an example for explanation. Specifically, After the smart headset switches to communication mode, how does the smart headset obtain the sound data of a third user outside the second user using the smart headset in the ambient voice data, so as to achieve communication without taking off the smart headset. As shown in Figure 3, the method consists of the following steps:
步骤301,获取终端设备上播放的音频数据。Step 301: Obtain audio data played on the terminal device.
步骤302,对音频数据进行识别,以确定终端设备处于目标工作场景。Step 302: Recognize the audio data to determine that the terminal device is in the target working scene.
其中,目标工作场景下智能耳机的麦克风处于语音采集状态。Among them, the microphone of the smart headset is in the voice collection state in the target work scenario.
步骤303,在目标工作场景下,获取智能耳机的模式切换指令。Step 303: Obtain the mode switching instruction of the smart headset in the target working scenario.
步骤304,根据模式切换指令,控制智能耳机在至少两个语音模式间切换。Step 304: Control the smart headset to switch between at least two voice modes according to the mode switching instruction.
其中,步骤301至步骤304,可参照前述实施例中的解释说明,原理相同,此处不再赘述。For steps 301 to 304, reference may be made to the explanations in the foregoing embodiments. The principles are the same and will not be described again here.
步骤305,响应于根据模式切换指令控制智能耳机切换至交流模式,获取使用智能耳机的第二用户的第一声音方位。Step 305: In response to controlling the smart earphone to switch to the communication mode according to the mode switching instruction, obtain the first sound direction of the second user using the smart earphone.
需要说明的是,步骤305至步骤307,在步骤304之后执行仅为一种示例,对于步骤305至步骤307还可以在步骤304之前执行,也就是说步骤305至步骤307的执行时序,本实施例中不进行限定。It should be noted that the execution of steps 305 to 307 after step 304 is only an example. Steps 305 to 307 can also be executed before step 304. That is to say, the execution timing of steps 305 to 307 in this implementation There are no restrictions in the example.
本申请实施例中,智能耳机在交流模式下时,使用智能耳机的第二用户会和周围的用户进行沟通交流。作为一种实现方式,在切换至交流模式的情况下,可通过智能耳机向使用智能耳机的第二用户发出语音提示,提示第二用户说出设定语句,从而确定采集第二用户说设定语句时对应的第一声音方位。作为另一种实现方式,在切换至交流模式的情况下,可获取智能耳机存储的使用智能耳机的第二用户的第一声音方位,由于第二用户在使用智能耳机时智能耳机的佩戴方式通常是固定的,从而智能耳机采集第二用户的声音数据的第一声音方位也是固定的,因此可从智能耳机的存储单元中获取第二用户的第一声音方位。In the embodiment of the present application, when the smart earphone is in the communication mode, the second user using the smart earphone will communicate with surrounding users. As an implementation method, when switching to the communication mode, a voice prompt can be sent to the second user using the smart earphone through the smart earphone, prompting the second user to speak the setting sentence, thereby determining to collect the setting of the second user's voice The first sound position corresponding to the sentence. As another implementation, when switching to the communication mode, the first sound direction of the second user using the smart earphone stored by the smart earphone can be obtained, because the second user usually wears the smart earphone when using the smart earphone. is fixed, so the first sound direction of the second user's sound data collected by the smart earphone is also fixed, so the first sound direction of the second user can be obtained from the storage unit of the smart earphone.
步骤306,控制麦克风在第二用户的第一声音方位以外的第二声音方位上采集环境中第三用户的声音数据。Step 306: Control the microphone to collect the sound data of the third user in the environment at a second sound direction other than the first sound direction of the second user.
其中,第三用户可以为环境中发出声音的用户,可以为和第二用户交流的用户,或者是未和第二用户交流,但是在说话的其它用户。其中,第二声音方位即为环境中第三用户发出声音的方位。The third user may be a user who makes a sound in the environment, may be a user communicating with the second user, or may be another user who is not communicating with the second user but is talking. The second sound direction is the direction from which the third user makes the sound in the environment.
本申请实施例的一种实现方式中,麦克风为麦克风阵列,控制麦克风阵列中用于采集第二用户的第一声音方位的第一麦克风停止采集第二用户的声音数据,控制麦克风阵列中第一麦克风以外的第二麦克风在第二用户的第一声音方位以外的第二声音方位上采集环境中第三用户的声音数据,使得麦克风采集的声音数据中不包含使用第二用户的声音数据,提高了在交流模式下,声音采集的准确性。In one implementation manner of the embodiment of the present application, the microphone is a microphone array, and the first microphone in the microphone array used to collect the first sound direction of the second user is controlled to stop collecting the second user's sound data, and the first microphone in the microphone array is controlled to stop collecting the second user's sound data. A second microphone other than the microphone collects the sound data of the third user in the environment at a second sound direction other than the first sound direction of the second user, so that the sound data collected by the microphone does not include the sound data of the second user, improving Improves the accuracy of sound collection in communication mode.
可选的,若环境中第一声音方位以外的声音方位存在多个,可将声音强度最大的方位作为第二声音方位。Optionally, if there are multiple sound directions other than the first sound direction in the environment, the direction with the strongest sound intensity can be used as the second sound direction.
步骤307,采用智能耳机播放第三用户的声音数据。Step 307: Use smart headphones to play the third user's voice data.
本申请实施例中,将麦克风采集的第三用户的声音数据发送給智能耳机,智能耳机对第三用户的声音数据放大后,播放第三用户的声音数据,提高了第三用户的声音数据的音量大小,提高了第二用户获取交流用户的声音数据的可靠性,使得第二用户不需要摘下耳机也可以清楚的听清楚交流的第三用户的声音,提高了交流的效果。In the embodiment of the present application, the voice data of the third user collected by the microphone is sent to the smart earphone. After the smart earphone amplifies the voice data of the third user, it plays the voice data of the third user, which improves the accuracy of the voice data of the third user. The volume improves the reliability of the second user's acquisition of the communicating user's voice data, allowing the second user to clearly hear the communicating third user's voice without taking off the earphones, thereby improving the communication effect.
进一步,响应于控制智能耳机切换至交流模式,关闭降噪模式,并且禁止麦克风和终端设备间进行数据传输,而禁止麦克风和终端设备间进行数据传输可以由智能耳机执行或者由终端设备执行,下面分别说明。Further, in response to controlling the smart headset to switch to the communication mode, turn off the noise reduction mode, and prohibit data transmission between the microphone and the terminal device, and prohibiting data transmission between the microphone and the terminal device may be performed by the smart headset or performed by the terminal device, as follows Explain separately.
本申请实施例的一种实现方式中,响应于控制智能耳机切换至交流模式,关闭降噪模式,以避免对麦克风采集的环境语音数据进行降噪处理。In one implementation manner of the embodiment of the present application, in response to controlling the smart headset to switch to the communication mode, the noise reduction mode is turned off to avoid noise reduction processing on the environmental voice data collected by the microphone.
同时,为了避免对智能耳机所处的目标工作场景的干扰,在使用耳机的第二用户与周围的第三用户交流的过程中,禁止麦克风和终端设备间进行数据传输,也就是说禁止麦克风将采集的第三用户的声音数据发送至终端设备,从而避免终端设备将第三用户的声音数据在目标工作场景下播放,目标工作场景例如为电话会议场景,也就是避免参加电话会议的其它客户听到第二用户和第三用户的交流声音。At the same time, in order to avoid interference with the target working scene of the smart headset, when the second user using the headset communicates with the surrounding third user, data transmission between the microphone and the terminal device is prohibited, which means that the microphone is prohibited from transmitting data to the surrounding third user. The collected voice data of the third user is sent to the terminal device, thereby preventing the terminal device from playing the voice data of the third user in the target work scenario. The target work scenario is, for example, a conference call scenario, that is, to prevent other customers participating in the conference call from listening. to the second user and the third user's communication voice.
本申请实施例的另一种实现方式中,在根据模式切换指令将智能耳机切换至交流模式的情况下,可提示用户在终端设备的交互界面中触发禁止指令,以使得终端设备将禁止指令通过蓝牙发送至智能耳机,智能耳机根据获取到的终端设备发送的禁止指令,禁止麦克风和终端设备间进行数据传输,具体可以是禁止传输第三用户的声音数据,实现了避免将第二用户与周围的第三用户交流的语音内容传输至终端设备,从而避免终端设备将交流的语音内容在目标工作场景下播放,目标工作场景例如为电话会议场景,也就是避免参加电话会议的其它客户听到第二用户和第三用户的交流的语音内容。In another implementation manner of the embodiment of the present application, when the smart headset is switched to the communication mode according to the mode switching instruction, the user can be prompted to trigger the prohibition instruction in the interactive interface of the terminal device, so that the terminal device passes the prohibition instruction through Bluetooth is sent to the smart headset. The smart headset prohibits data transmission between the microphone and the terminal device based on the obtained prohibition instruction sent by the terminal device. Specifically, it can prohibit the transmission of the third user's voice data, thereby avoiding the second user from being connected to the surroundings. The voice content communicated by the third user is transmitted to the terminal device, thereby preventing the terminal device from playing the communicated voice content in the target work scenario. The target work scenario is, for example, a conference call scenario, that is, to prevent other customers participating in the conference call from hearing the third user. The voice content of the exchange between the second user and the third user.
本申请实施例中,响应于控制智能耳机切换至交流模式,获取使用智能耳机的第二用户的声音方位,控制麦克风在第二用户的声音方位以外的声音方位上采集第三用户的声音数据,采用智能耳机播放第三用户的声音数据,通过不采集使用智能耳机的第二用户的声音数据,而是采集和第二用户交流的第三用户的声音数据,以使得用户不摘下智能耳机也可以顺畅的和周围用户进行交流,提高了在佩戴智能耳机的情况下交流的可靠性。In the embodiment of the present application, in response to controlling the smart headset to switch to the communication mode, obtaining the sound direction of the second user using the smart headset, and controlling the microphone to collect the third user's sound data at a sound direction other than the second user's sound direction, Using smart earphones to play the voice data of a third user does not collect the voice data of the second user using the smart earphones, but collects the voice data of the third user communicating with the second user, so that the user does not take off the smart earphones. It can communicate with surrounding users smoothly, improving the reliability of communication while wearing smart headphones.
基于上一实施例,图4为本申请实施例提供的另一种智能耳机的控制方法的流程示意图,具体说明了智能耳机切换至交流模式后,智能耳机如何获取环境语音数据中使用智能耳机的第二用户外的第三用户的声音数据,实现在不摘取智能耳机的情况下进行交流。如图4所示,该方法包含以下步骤:Based on the previous embodiment, FIG. 4 is a schematic flow chart of another smart headset control method provided by an embodiment of the present application. It specifically illustrates how the smart headset obtains environmental voice data using the smart headset after the smart headset switches to the communication mode. The voice data of the third user outside the second user enables communication without taking off the smart earphones. As shown in Figure 4, the method consists of the following steps:
步骤401,获取终端设备上播放的音频数据。Step 401: Obtain audio data played on the terminal device.
步骤402,对音频数据进行识别,以确定终端设备处于目标工作场景。Step 402: Recognize the audio data to determine that the terminal device is in the target working scene.
其中,目标工作场景下智能耳机的麦克风处于语音采集状态。Among them, the microphone of the smart headset is in the voice collection state in the target work scenario.
步骤403,在目标工作场景下,获取智能耳机的模式切换指令。Step 403: Obtain the mode switching instruction of the smart headset in the target working scenario.
步骤404,根据模式切换指令,控制智能耳机在至少两个语音模式间切换。Step 404: Control the smart headset to switch between at least two voice modes according to the mode switching instruction.
其中,步骤401至步骤404,可参照前述实施例中的解释说明,原理相同,此处不再赘述。For steps 401 to 404, reference may be made to the explanations in the foregoing embodiments. The principles are the same and will not be described again here.
步骤405,响应于控制智能耳机切换至交流模式,获取智能耳机上的麦克风采集的第二环境语音数据。Step 405: In response to controlling the smart headset to switch to the communication mode, obtain the second environmental voice data collected by the microphone on the smart headset.
需要说明的是,步骤405至步骤407,在步骤404之后执行仅为一种示例,对于步骤405至407步骤的执行时序,本实施例中不进行限定。It should be noted that the execution of steps 405 to 407 after step 404 is only an example, and the execution timing of steps 405 to 407 is not limited in this embodiment.
其中,智能耳机上的麦克风采集的第二环境语音数据,和前述实施例中的第一环境语音数据的内容可以相同也可以不同,仅为区分不同模式下采集的的环境语音数据,可参照前述实施例中的第一环境语音数据的解释说明,此处不在赘述。Among them, the content of the second environmental voice data collected by the microphone on the smart headset may be the same as or different from the first environmental voice data in the previous embodiment. It is only to distinguish the environmental voice data collected in different modes. Please refer to the aforementioned The explanation of the first environmental voice data in the embodiment will not be described again here.
步骤406,根据使用智能耳机的第二用户的声音数据,确定第二环境语音数据中的第三用户的声音数据。Step 406: Determine the voice data of the third user in the second environmental voice data based on the voice data of the second user using the smart headset.
其中,第三用户和第二用户为不同的用户,在电话会议场景下,第三用户是和第二用户交流的用户。Among them, the third user and the second user are different users. In the conference call scenario, the third user is the user communicating with the second user.
步骤407,采用智能耳机播放第三用户的声音数据。Step 407: Use smart headphones to play the third user's voice data.
本申请实施例中,对第二用户的声音数据进行识别,以识别得到第二用户的声音数据的声音频段或音色特征信息,根据第二用户的声音数据的声音频段或音色特征信息,识别第二环境语音数据中包含的第二用户的声音数据,从而第二环境语音数据中第二用户的声音数据以外的声音数据即为第三用户的声音数据,进而,将第三用户的声音数据通过智能耳机播放,以提高第三用户的声音大小,可选定的,可将第三用户的声音数据中不属于人声的噪音数据滤除并放大后通过智能耳机播放,确保佩戴智能耳机的第二用户可以清晰的获取到第三用户的声音。In the embodiment of the present application, the second user's voice data is identified to obtain the voice segment or timbre feature information of the second user's voice data, and the second user's voice data is identified based on the voice segment or timbre feature information of the second user's voice data. The second user's voice data included in the second environmental voice data, so the voice data other than the second user's voice data in the second environmental voice data is the third user's voice data, and further, the third user's voice data is passed through Smart earphones play to increase the volume of the third user's voice. Optionally, the noise data that does not belong to the human voice in the third user's voice data can be filtered out and amplified and then played through the smart earphones to ensure that the third user wearing the smart earphones The second user can clearly obtain the voice of the third user.
需要说明的是,第三用户的声音数据,可以为一个用户的声音数据,也可以为多个用户的声音数据。It should be noted that the voice data of the third user may be the voice data of one user or the voice data of multiple users.
其中,对第二用户的声音数据进行识别的方法可参照前前述实施例中的解释说明,原理相同,此处不再赘述。The method of identifying the second user's voice data may refer to the explanations in the foregoing embodiments. The principles are the same and will not be described again here.
进一步,响应于控制智能耳机切换至交流模式,关闭降噪模式,并且禁止麦克风和终端设备间进行数据传输,而禁止麦克风和终端设备间进行数据传输可以由智能耳机执行或者由终端设备执行,下面分别说明。Further, in response to controlling the smart headset to switch to the communication mode, turn off the noise reduction mode, and prohibit data transmission between the microphone and the terminal device, and prohibiting data transmission between the microphone and the terminal device may be performed by the smart headset or performed by the terminal device, as follows Explain separately.
本申请实施例的一种实现方式中,响应于控制智能耳机切换至交流模式,关闭降噪模式,以避免对麦克风采集的环境语音数据进行降噪处理。In one implementation manner of the embodiment of the present application, in response to controlling the smart headset to switch to the communication mode, the noise reduction mode is turned off to avoid noise reduction processing on the environmental voice data collected by the microphone.
同时,为了避免对智能耳机所处的目标工作场景的干扰,在使用耳机的第二用户与周 围的第三用户交流的过程中,禁止麦克风和终端设备间进行数据传输,也就是说禁止麦克风将采集的第二环境语音数据发送至终端设备,从而避免终端设备将包含第三用户的声音数据的第二环境语音数据在目标工作场景下播放,目标工作场景例如为电话会议场景,也就是避免参加电话会议的其它客户听到第二用户和第三用户的交流声音。At the same time, in order to avoid interference with the target working scene of the smart headset, when the second user using the headset communicates with the surrounding third user, data transmission between the microphone and the terminal device is prohibited, which means that the microphone is prohibited from transmitting data to the surrounding third user. The collected second environment voice data is sent to the terminal device, thereby preventing the terminal device from playing the second environment voice data containing the third user's voice data in the target work scenario. The target work scenario is, for example, a conference call scene, that is, to avoid participating in Other users of the conference call hear the exchanges between the second user and the third user.
本申请实施例的另一种实现方式中,在根据模式切换指令将智能耳机切换至交流模式的情况下,可提示用户在终端设备的交互界面中触发禁止指令,以使得终端设备将禁止指令通过蓝牙发送至智能耳机,智能耳机根据获取到的终端设备发送的禁止指令,禁止麦克风和终端设备间进行数据传输,具体可以是禁止传输麦克风采集的第二环境语音数据,实现了避免将第二用户与周围的第三用户交流的语音内容传输至终端设备,从而避免终端设备将交流的语音内容在目标工作场景下播放,目标工作场景例如为电话会议场景,也就是避免参加电话会议的其它客户听到第二用户和第三用户的交流的语音内容。In another implementation manner of the embodiment of the present application, when the smart headset is switched to the communication mode according to the mode switching instruction, the user can be prompted to trigger the prohibition instruction in the interactive interface of the terminal device, so that the terminal device passes the prohibition instruction through Bluetooth is sent to the smart headset. The smart headset prohibits data transmission between the microphone and the terminal device according to the prohibition instruction sent by the terminal device. Specifically, it can prohibit the transmission of the second environment voice data collected by the microphone, thereby avoiding the second user. The voice content communicated with the surrounding third users is transmitted to the terminal device, thereby preventing the terminal device from playing the communicated voice content in the target work scenario. The target work scenario is, for example, a conference call scenario, that is, to prevent other customers participating in the conference call from listening. Speech content of the exchange to the second user and the third user.
本申请实施例的智能耳机的控制方法中,响应于控制智能耳机切换至交流模式,获取智能耳机上的麦克风采集的第二环境语音数据,根据使用智能耳机的第二用户的声音数据,确定第二环境语音数据中的第三用户的声音数据,采用智能耳机播放第三用户的声音数据,通过确定环境语音数据中的第三用户的声音数据并播放第三用户的声音数据,以使得第二用户不摘下智能耳机也可以顺畅的和周围的第三用户进行交流,提高了用户在佩戴智能耳机的情况下交流的可靠性,从而不需要在智能耳机的模式切换时,频繁的摘下或戴上智能耳机,增加了智能耳机使用的便利性。In the control method of the smart headset according to the embodiment of the present application, in response to controlling the smart headset to switch to the communication mode, the second environmental voice data collected by the microphone on the smart headset is obtained, and the second environmental voice data collected by the microphone on the smart headset is determined based on the voice data of the second user using the smart headset. The third user's voice data in the second ambient voice data is used to play the third user's voice data using smart headphones. By determining the third user's voice data in the ambient voice data and playing the third user's voice data, the second Users can communicate smoothly with third-party users around them without taking off their smart headphones, which improves the reliability of communication while wearing smart headphones and eliminates the need to frequently take off or remove the smart headphones when switching modes. Wearing smart headphones increases the convenience of using smart headphones.
为了实现上述实施例,本申请实施例还提出一种智能耳机的控制装置。In order to implement the above embodiments, embodiments of the present application also provide a control device for smart earphones.
图5为本申请实施例提供的一种智能耳机的控制装置的结构示意图。FIG. 5 is a schematic structural diagram of a control device for a smart headset provided by an embodiment of the present application.
如图5所示,该装置可以包括:As shown in Figure 5, the device may include:
获取模块51,用于获取所述终端设备上播放的音频数据。The acquisition module 51 is used to acquire the audio data played on the terminal device.
第一确定模块52,用于对所述音频数据进行识别,以确定所述终端设备处于目标工作场景;所述目标工作场景下,所述智能耳机的麦克风处于语音采集状态。The first determination module 52 is used to identify the audio data to determine that the terminal device is in a target working scenario; in the target working scenario, the microphone of the smart headset is in a voice collection state.
所述获取模块51,还用于在所述目标工作场景下,获取所述智能耳机的模式切换指令。The obtaining module 51 is also used to obtain the mode switching instruction of the smart headset in the target working scenario.
控制模块53,用于根据所述模式切换指令,控制所述智能耳机在所述至少两个语音模式间切换。The control module 53 is configured to control the smart earphone to switch between the at least two voice modes according to the mode switching instruction.
进一步,在本申请实施例的一种实现方式中,至少两个语音模式包括降噪模式,该装置,还包括:识别模块和处理模块。Further, in an implementation manner of the embodiment of the present application, at least two voice modes include a noise reduction mode, and the device further includes: a recognition module and a processing module.
所述获取模块51,还用于响应于控制所述智能耳机切换至所述降噪模式,获取所述终端设备上播放的音频数据;The acquisition module 51 is also configured to acquire audio data played on the terminal device in response to controlling the smart headset to switch to the noise reduction mode;
识别模块,用于对所述音频数据进行识别,以识别得到所述音频数据包括的第一人声数据;An identification module, used to identify the audio data to identify the first vocal data included in the audio data;
所述获取模块51,还用于获取所述智能耳机上的麦克风采集的第一环境语音数据;The acquisition module 51 is also used to acquire the first environmental voice data collected by the microphone on the smart headset;
处理模块,用于根据所述第一人声数据,对所述第一环境语音数据进行降噪处理。A processing module configured to perform noise reduction processing on the first environmental voice data according to the first human voice data.
在本申请实施例的一种实现方式中,上述处理模块,具体用于:In an implementation manner of the embodiment of the present application, the above processing module is specifically used for:
获取使用所述智能耳机的第一目标用户的第二声音数据;Obtain second sound data of the first target user using the smart headset;
根据所述第一声音数据和所述第二声音数据,对所述第一环境语音数据中所述第一用户的声音数据和所述第二用户的声音数据以外的声音数据进行降噪处理。According to the first sound data and the second sound data, noise reduction processing is performed on the sound data in the first environmental voice data other than the sound data of the first user and the sound data of the second user.
在本申请实施例的一种实现方式中,至少两个语音模式包括交流模式,所述装置,还包括:第二确定模块和播放模块。In an implementation manner of the embodiment of the present application, at least two voice modes include a communication mode, and the device further includes: a second determination module and a playback module.
所述获取模块51,还用于响应于控制所述智能耳机切换至所述交流模式,获取所述智能耳机上的麦克风采集的第二环境语音数据;The acquisition module 51 is also configured to acquire the second environmental voice data collected by the microphone on the smart earphone in response to controlling the smart earphone to switch to the communication mode;
第二确定模块,用于根据使用所述智能耳机的第二用户的声音数据,确定所述第二环境语音数据中的第三用户的声音数据;其中,所述第三用户和所述第二用户为不同的用户;A second determination module, configured to determine the voice data of a third user in the second environmental voice data according to the voice data of the second user using the smart headset; wherein the third user and the second user The users are different users;
播放模块,用于采用所述智能耳机播放所述第三用户的声音数据。A playback module is used to play the third user's voice data using the smart earphone.
在本申请实施例的一种实现方式中,至少两个语音模式包括交流模式,所述装置,还包括:In an implementation manner of the embodiment of the present application, at least two voice modes include a communication mode, and the device further includes:
所述获取模块51,用于响应于控制所述智能耳机切换至所述交流模式,获取使用所述智能耳机的第二用户的第一声音方位;The acquisition module 51 is configured to acquire the first sound direction of the second user using the smart earphone in response to controlling the smart earphone to switch to the communication mode;
采集模块,用于控制所述麦克风在所述第二用户的第一声音方位以外的第二声音方位上采集环境中的第三用户的声音数据;其中,所述第三用户和所述第二用户为不同的用户;A collection module configured to control the microphone to collect the sound data of a third user in the environment at a second sound direction other than the first sound direction of the second user; wherein the third user and the second The users are different users;
所述播放模块,还用于采用所述智能耳机播放所述第三用户的声音数据。The playback module is also used to play the third user's voice data using the smart earphone.
在本申请实施例的一种实现方式中,所述装置,还包括:In an implementation manner of the embodiment of the present application, the device further includes:
第一关闭模块,用于响应于控制所述智能耳机切换至所述交流模式,关闭所述降噪模式,且禁止所述麦克风和所述终端设备间进行数据传输。A first shutdown module, configured to turn off the noise reduction mode in response to controlling the smart earphone to switch to the communication mode, and prohibit data transmission between the microphone and the terminal device.
在本申请实施例的一种实现方式中,所述装置,还包括:In an implementation manner of the embodiment of the present application, the device further includes:
第二关闭模块,用于响应于控制所述智能耳机切换至所述交流模式,关闭所述降噪模式;响应于获取到所述终端设备发送的禁止指令,根据所述禁止指令,禁止所述麦克风和所述终端设备间进行数据传输。A second shutdown module, configured to turn off the noise reduction mode in response to controlling the smart headset to switch to the communication mode; in response to obtaining a prohibition instruction sent by the terminal device, to prohibit the Data transmission is performed between the microphone and the terminal device.
在本申请实施例的一种实现方式中,第一确定模块52,具体用于:In an implementation manner of the embodiment of the present application, the first determination module 52 is specifically used to:
对所述音频数据进行识别,确定所述音频数据包括的语音频段;Identify the audio data and determine the voice segments included in the audio data;
响应于所述语音频段中包含第一目标语音频段,且不包含第二目标语音频段和第三目标语音频段,确定所述终端设备处于目标工作场景;其中,所述第一目标语音频段的频率下限大于所述第二目标语音频段的频率上限,且所述第一目标语音频段的频率上限小于所述第三目标语音频段的频率下限。In response to the voice segment containing the first target voice segment and not including the second target voice segment and the third target voice segment, it is determined that the terminal device is in the target working scene; wherein the frequency of the first target voice segment The lower limit is greater than the upper frequency limit of the second target speech band, and the upper frequency limit of the first target speech band is less than the lower frequency limit of the third target speech band.
在本申请实施例的一种实现方式中,第一确定模块52,具体用于:In an implementation manner of the embodiment of the present application, the first determination module 52 is specifically used to:
对所述音频数据携带的来源信息进行识别,确定所述音频数据的来源;Identify the source information carried by the audio data and determine the source of the audio data;
响应于所述音频数据的来源为目标来源,确定所述终端设备处于目标工作场景。In response to the source of the audio data being the target source, it is determined that the terminal device is in the target working scene.
需要说明的是,前述对方法实施例的解释说明也适用于该实施例的装置,此处不再赘述。It should be noted that the foregoing explanation of the method embodiment also applies to the device of this embodiment, and will not be described again here.
本申请实施例的智能耳机的控制装置中,获取终端设备上播放的音频数据,对音频数据进行识别,以确定终端设备处于目标工作场景,目标工作场景下智能耳机的麦克风处于语音采集状态,在目标工作场景下,获取智能耳机的模式切换指令,根据模式切换指令,控制智能耳机在至少两个语音模式间切换,通过在目标工作场景下,控制智能耳机在至少两种模式间切换可以帮助用户在使用智能耳机的时候,不需要频繁的摘下或带上智能耳机,提高了智能耳机使用的便利性。In the control device of the smart headset in the embodiment of the present application, the audio data played on the terminal device is acquired, and the audio data is identified to determine that the terminal device is in the target working scene. In the target working scene, the microphone of the smart headset is in the voice collection state. In the target work scenario, obtain the mode switching instruction of the smart headset, and control the smart headset to switch between at least two voice modes according to the mode switching instruction. By controlling the smart headset to switch between at least two modes in the target work scenario, it can help the user. When using smart headphones, you do not need to frequently take off or put on the smart headphones, which improves the convenience of using smart headphones.
为了实现上述实施例,本申请还提出一种电子设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述程序时,实现如前述方法实施例所述的方法。In order to implement the above embodiments, this application also proposes an electronic device, including a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the program, the aforementioned method is implemented. methods described in the examples.
为了实现上述实施例,本申请还提出一种非临时性计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时,实现如前述方法实施例所述的方法。In order to implement the above embodiments, this application also proposes a non-transitory computer-readable storage medium on which a computer program is stored. When the program is executed by a processor, the method as described in the foregoing method embodiments is implemented.
为了实现上述实施例,本申请还提出一种计算机程序产品,其上存储有计算机程序,所述计算机程序被处理器执行时实现如前述方法实施例所述的方法。In order to implement the above embodiments, this application also proposes a computer program product on which a computer program is stored. When the computer program is executed by a processor, the method as described in the foregoing method embodiments is implemented.
图6为本申请实施例提供的一种电子设备的框图。例如,电子设备800可以是移动电话,计算机,数字广播终端,消息收发设备,游戏控制台,平板设备,医疗设备,健身设备,个人数字助理等。Figure 6 is a block diagram of an electronic device provided by an embodiment of the present application. For example, the electronic device 800 may be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, or the like.
参照图6,电子设备800可以包括以下一个或多个组件:处理组件802,存储器804,电力组件806,多媒体组件808,音频组件810,输入/输出(I/O)接口812,传感器组件814,以及通信组件816。6, electronic device 800 may include one or more of the following components: processing component 802, memory 804, power component 806, multimedia component 808, audio component 810, input/output (I/O) interface 812, sensor component 814, and communications component 816.
处理组件802通常控制电子设备800的整体操作,诸如与显示,电话呼叫,数据通信,相机操作和记录操作相关联的操作。处理组件802可以包括一个或多个处理器820来执行指令,以完成上述的方法的全部或部分步骤。此外,处理组件802可以包括一个或多个模 块,便于处理组件802和其他组件之间的交互。例如,处理组件802可以包括多媒体模块,以方便多媒体组件808和处理组件802之间的交互。 Processing component 802 generally controls the overall operations of electronic device 800, such as operations associated with display, phone calls, data communications, camera operations, and recording operations. The processing component 802 may include one or more processors 820 to execute instructions to complete all or part of the steps of the above method. Additionally, processing component 802 may include one or more modules that facilitate interaction between processing component 802 and other components. For example, processing component 802 may include a multimedia module to facilitate interaction between multimedia component 808 and processing component 802.
存储器804被配置为存储各种类型的数据以支持在电子设备800的操作。这些数据的示例包括用于在电子设备800上操作的任何应用程序或方法的指令,联系人数据,电话簿数据,消息,图片,视频等。存储器804可以由任何类型的易失性或非易失性存储设备或者它们的组合实现,如静态随机存取存储器(SRAM),电可擦除可编程只读存储器(EEPROM),可擦除可编程只读存储器(EPROM),可编程只读存储器(PROM),只读存储器(ROM),磁存储器,快闪存储器,磁盘或光盘。 Memory 804 is configured to store various types of data to support operations at electronic device 800 . Examples of such data include instructions for any application or method operating on the electronic device 800, contact data, phonebook data, messages, pictures, videos, etc. Memory 804 may be implemented by any type of volatile or non-volatile storage device, or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EEPROM), Programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disk.
电力组件806为电子设备800的各种组件提供电力。电力组件806可以包括电源管理系统,一个或多个电源,及其他与为电子设备800生成、管理和分配电力相关联的组件。 Power component 806 provides power to various components of electronic device 800 . Power components 806 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power to electronic device 800 .
多媒体组件808包括在所述电子设备800和用户之间的提供一个输出接口的屏幕。在一些实施例中,屏幕可以包括液晶显示器(LCD)和触摸面板(TP)。如果屏幕包括触摸面板,屏幕可以被实现为触摸屏,以接收来自用户的输入信号。触摸面板包括一个或多个触摸传感器以感测触摸、滑动和触摸面板上的手势。所述触摸传感器可以不仅感测触摸或滑动动作的边界,而且还检测与所述触摸或滑动操作相关的持续时间和压力。在一些实施例中,多媒体组件808包括一个前置摄像头和/或后置摄像头。当电子设备800处于操作模式,如拍摄模式或视频模式时,前置摄像头和/或后置摄像头可以接收外部的多媒体数据。每个前置摄像头和后置摄像头可以是一个固定的光学透镜系统或具有焦距和光学变焦能力。 Multimedia component 808 includes a screen that provides an output interface between the electronic device 800 and the user. In some embodiments, the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user. The touch panel includes one or more touch sensors to sense touches, swipes, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide action. In some embodiments, multimedia component 808 includes a front-facing camera and/or a rear-facing camera. When the electronic device 800 is in an operating mode, such as a shooting mode or a video mode, the front camera and/or the rear camera may receive external multimedia data. Each front-facing camera and rear-facing camera can be a fixed optical lens system or have a focal length and optical zoom capabilities.
音频组件810被配置为输出和/或输入音频信号。例如,音频组件810包括一个麦克风(MIC),当电子设备800处于操作模式,如呼叫模式、记录模式和语音识别模式时,麦克风被配置为接收外部音频信号。所接收的音频信号可以被进一步存储在存储器804或经由通信组件816发送。在一些实施例中,音频组件810还包括一个扬声器,用于输出音频信号。 Audio component 810 is configured to output and/or input audio signals. For example, audio component 810 includes a microphone (MIC) configured to receive external audio signals when electronic device 800 is in operating modes, such as call mode, recording mode, and voice recognition mode. The received audio signal may be further stored in memory 804 or sent via communication component 816 . In some embodiments, audio component 810 also includes a speaker for outputting audio signals.
I/O接口812为处理组件802和外围接口模块之间提供接口,上述外围接口模块可以是键盘,点击轮,按钮等。这些按钮可包括但不限于:主页按钮、音量按钮、启动按钮和锁定按钮。The I/O interface 812 provides an interface between the processing component 802 and a peripheral interface module, which may be a keyboard, a click wheel, a button, etc. These buttons may include, but are not limited to: Home button, Volume buttons, Start button, and Lock button.
传感器组件814包括一个或多个传感器,用于为电子设备800提供各个方面的状态评估。例如,传感器组件814可以检测到电子设备800的打开/关闭状态,组件的相对定位,例如所述组件为电子设备800的显示器和小键盘,传感器组件814还可以检测电子设备800或电子设备800一个组件的位置改变,用户与电子设备800接触的存在或不存在,电子设 备800方位或加速/减速和电子设备800的温度变化。传感器组件814可以包括接近传感器,被配置用来在没有任何的物理接触时检测附近物体的存在。传感器组件814还可以包括光传感器,如CMOS或CCD图像传感器,用于在成像应用中使用。在一些实施例中,该传感器组件814还可以包括加速度传感器,陀螺仪传感器,磁传感器,压力传感器或温度传感器。 Sensor component 814 includes one or more sensors for providing various aspects of status assessment for electronic device 800 . For example, the sensor component 814 can detect the open/closed state of the electronic device 800, the relative positioning of the components, such as the display and keypad of the electronic device 800, the sensor component 814 can also detect the electronic device 800 or an electronic device 800. The position of components changes, the presence or absence of user contact with the electronic device 800 , the orientation or acceleration/deceleration of the electronic device 800 and the temperature of the electronic device 800 change. Sensor assembly 814 may include a proximity sensor configured to detect the presence of nearby objects without any physical contact. Sensor assembly 814 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor component 814 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
通信组件816被配置为便于电子设备800和其他设备之间有线或无线方式的通信。电子设备800可以接入基于通信标准的无线网络,如WiFi,4G或5G,或它们的组合。在一个示例性实施例中,通信组件816经由广播信道接收来自外部广播管理系统的广播信号或广播相关信息。在一个示例性实施例中,所述通信组件816还包括近场通信(NFC)模块,以促进短程通信。例如,在NFC模块可基于射频识别(RFID)技术,红外数据协会(IrDA)技术,超宽带(UWB)技术,蓝牙(BT)技术和其他技术来实现。 Communication component 816 is configured to facilitate wired or wireless communication between electronic device 800 and other devices. The electronic device 800 can access a wireless network based on a communication standard, such as WiFi, 4G or 5G, or a combination thereof. In one exemplary embodiment, the communication component 816 receives broadcast signals or broadcast related information from an external broadcast management system via a broadcast channel. In one exemplary embodiment, the communications component 816 also includes a near field communications (NFC) module to facilitate short-range communications. For example, the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology and other technologies.
在示例性实施例中,电子设备800可以被一个或多个应用专用集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理设备(DSPD)、可编程逻辑器件(PLD)、现场可编程门阵列(FPGA)、控制器、微控制器、微处理器或其他电子元件实现,用于执行上述方法。In an exemplary embodiment, electronic device 800 may be configured by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable A programmable gate array (FPGA), controller, microcontroller, microprocessor or other electronic component implementation is used to perform the above method.
在示例性实施例中,还提供了一种包括指令的非临时性计算机可读存储介质,例如包括指令的存储器804,上述指令可由电子设备800的处理器820执行以完成上述方法。例如,所述非临时性计算机可读存储介质可以是ROM、随机存取存储器(RAM)、CD-ROM、磁带、软盘和光数据存储设备等。In an exemplary embodiment, a non-transitory computer-readable storage medium including instructions, such as a memory 804 including instructions, which can be executed by the processor 820 of the electronic device 800 to complete the above method is also provided. For example, the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, etc.
在本说明书的描述中,参考术语“一个实施例”、“一些实施例”、“示例”、“具体示例”、或“一些示例”等的描述意指结合该实施例或示例描述的具体特征、结构、材料或者特点包含于本申请的至少一个实施例或示例中。在本说明书中,对上述术语的示意性表述不必须针对的是相同的实施例或示例。而且,描述的具体特征、结构、材料或者特点可以在任一个或多个实施例或示例中以合适的方式结合。此外,在不相互矛盾的情况下,本领域的技术人员可以将本说明书中描述的不同实施例或示例以及不同实施例或示例的特征进行结合和组合。In the description of this specification, reference to the terms "one embodiment," "some embodiments," "an example," "specific examples," or "some examples" or the like means that specific features are described in connection with the embodiment or example. , structures, materials or features are included in at least one embodiment or example of the present application. In this specification, the schematic expressions of the above terms are not necessarily directed to the same embodiment or example. Furthermore, the specific features, structures, materials or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Furthermore, those skilled in the art may combine and combine different embodiments or examples and features of different embodiments or examples described in this specification unless they are inconsistent with each other.
此外,术语“第一”、“第二”仅用于描述目的,而不能理解为指示或暗示相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括至少一个该特征。在本申请的描述中,“多个”的含义是至少两个,例如两个,三个等,除非另有明确具体的限定。In addition, the terms “first” and “second” are used for descriptive purposes only and cannot be understood as indicating or implying relative importance or implicitly indicating the quantity of indicated technical features. Therefore, features defined as "first" and "second" may explicitly or implicitly include at least one of these features. In the description of this application, "plurality" means at least two, such as two, three, etc., unless otherwise expressly and specifically limited.
流程图中或在此以其他方式描述的任何过程或方法描述可以被理解为,表示包括一个 或更多个用于实现定制逻辑功能或过程的步骤的可执行指令的代码的模块、片段或部分,并且本申请的优选实施方式的范围包括另外的实现,其中可以不按所示出或讨论的顺序,包括根据所涉及的功能按基本同时的方式或按相反的顺序,来执行功能,这应被本申请的实施例所属技术领域的技术人员所理解。Any process or method descriptions in flowcharts or otherwise described herein may be understood to represent modules, segments, or portions of code that include one or more executable instructions for implementing customized logical functions or steps of the process. , and the scope of the preferred embodiments of the present application includes additional implementations in which functions may be performed out of the order shown or discussed, including in a substantially simultaneous manner or in the reverse order, depending on the functionality involved, which shall It should be understood by those skilled in the technical field to which the embodiments of this application belong.
在流程图中表示或在此以其他方式描述的逻辑和/或步骤,例如,可以被认为是用于实现逻辑功能的可执行指令的定序列表,可以具体实现在任何计算机可读介质中,以供指令执行系统、装置或设备(如基于计算机的系统、包括处理器的系统或其他可以从指令执行系统、装置或设备取指令并执行指令的系统)使用,或结合这些指令执行系统、装置或设备而使用。就本说明书而言,"计算机可读介质"可以是任何可以包含、存储、通信、传播或传输程序以供指令执行系统、装置或设备或结合这些指令执行系统、装置或设备而使用的装置。计算机可读介质的更具体的示例(非穷尽性列表)包括以下:具有一个或多个布线的电连接部(电子装置),便携式计算机盘盒(磁装置),随机存取存储器(RAM),只读存储器(ROM),可擦除可编辑只读存储器(EPROM或闪速存储器),光纤装置,以及便携式光盘只读存储器(CDROM)。另外,计算机可读介质甚至可以是可在其上打印所述程序的纸或其他合适的介质,因为可以例如通过对纸或其他介质进行光学扫描,接着进行编辑、解译或必要时以其他合适方式进行处理来以电子方式获得所述程序,然后将其存储在计算机存储器中。The logic and/or steps represented in the flowcharts or otherwise described herein, for example, may be considered a sequenced list of executable instructions for implementing the logical functions, and may be embodied in any computer-readable medium, For use by, or in combination with, instruction execution systems, devices or devices (such as computer-based systems, systems including processors or other systems that can fetch instructions from and execute instructions from the instruction execution system, device or device) or equipment. For the purposes of this specification, a "computer-readable medium" may be any device that can contain, store, communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. More specific examples (non-exhaustive list) of computer readable media include the following: electrical connections with one or more wires (electronic device), portable computer disk cartridges (magnetic device), random access memory (RAM), Read-only memory (ROM), erasable and programmable read-only memory (EPROM or flash memory), fiber optic devices, and portable compact disc read-only memory (CDROM). Additionally, the computer-readable medium may even be paper or other suitable medium on which the program may be printed, as the paper or other medium may be optically scanned, for example, and subsequently edited, interpreted, or otherwise suitable as necessary. process to obtain the program electronically and then store it in computer memory.
应当理解,本申请的各部分可以用硬件、软件、固件或它们的组合来实现。在上述实施方式中,多个步骤或方法可以用存储在存储器中且由合适的指令执行系统执行的软件或固件来实现。如,如果用硬件来实现和在另一实施方式中一样,可用本领域公知的下列技术中的任一项或他们的组合来实现:具有用于对数据信号实现逻辑功能的逻辑门电路的离散逻辑电路,具有合适的组合逻辑门电路的专用集成电路,可编程门阵列(PGA),现场可编程门阵列(FPGA)等。It should be understood that various parts of the present application can be implemented in hardware, software, firmware, or a combination thereof. In the above embodiments, various steps or methods may be implemented in software or firmware stored in a memory and executed by a suitable instruction execution system. For example, if it is implemented in hardware, as in another embodiment, it can be implemented by any one of the following technologies known in the art or their combination: discrete logic gate circuits with logic functions for implementing data signals; Logic circuits, application specific integrated circuits with suitable combinational logic gates, programmable gate arrays (PGA), field programmable gate arrays (FPGA), etc.
本技术领域的普通技术人员可以理解实现上述实施例方法携带的全部或部分步骤是可以通过程序来指令相关的硬件完成,所述的程序可以存储于一种计算机可读存储介质中,该程序在执行时,包括方法实施例的步骤之一或其组合。Those of ordinary skill in the art can understand that all or part of the steps involved in implementing the methods of the above embodiments can be completed by instructing relevant hardware through a program. The program can be stored in a computer-readable storage medium. The program can be stored in a computer-readable storage medium. When executed, one of the steps of the method embodiment or a combination thereof is included.
此外,在本申请各个实施例中的各功能单元可以集成在一个处理模块中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个模块中。上述集成的模块既可以采用硬件的形式实现,也可以采用软件功能模块的形式实现。所述集成的模块如果以软件功能模块的形式实现并作为独立的产品销售或使用时,也可以存储在一个计算机可读取存储介质中。In addition, each functional unit in various embodiments of the present application can be integrated into a processing module, or each unit can exist physically alone, or two or more units can be integrated into one module. The above integrated modules can be implemented in the form of hardware or software function modules. If the integrated module is implemented in the form of a software function module and sold or used as an independent product, it can also be stored in a computer-readable storage medium.
上述提到的存储介质可以是只读存储器,磁盘或光盘等。尽管上面已经示出和描述了本申请的实施例,可以理解的是,上述实施例是示例性的,不能理解为对本申请的限制,本领域的普通技术人员在本申请的范围内可以对上述实施例进行变化、修改、替换和变型。The storage media mentioned above can be read-only memory, magnetic disks or optical disks, etc. Although the embodiments of the present application have been shown and described above, it can be understood that the above-mentioned embodiments are illustrative and cannot be understood as limitations of the present application. Those of ordinary skill in the art can make modifications to the above-mentioned embodiments within the scope of the present application. The embodiments are subject to changes, modifications, substitutions and variations.

Claims (20)

  1. 一种智能耳机的控制方法,其特征在于,包括:A control method for smart headphones, which is characterized by including:
    获取所述终端设备上播放的音频数据;Obtain audio data played on the terminal device;
    对所述音频数据进行识别,以确定所述终端设备处于目标工作场景;所述目标工作场景下所述智能耳机的麦克风处于语音采集状态The audio data is identified to determine that the terminal device is in a target working scenario; the microphone of the smart headset is in a voice collection state under the target working scenario.
    在所述目标工作场景下,获取所述智能耳机的模式切换指令;Under the target working scenario, obtain the mode switching instruction of the smart headset;
    根据所述模式切换指令,控制所述智能耳机在所述至少两个语音模式间切换。According to the mode switching instruction, the smart earphone is controlled to switch between the at least two voice modes.
  2. 如权利要求1所述的方法,其特征在于,所述至少两个语音模式包括降噪模式,所述方法,还包括:The method of claim 1, wherein the at least two voice modes include a noise reduction mode, and the method further includes:
    响应于控制所述智能耳机切换至所述降噪模式,获取所述终端设备上播放的音频数据;In response to controlling the smart headset to switch to the noise reduction mode, obtain audio data played on the terminal device;
    对所述音频数据进行识别,以识别得到所述音频数据包括的第一用户的声音数据;Identify the audio data to identify the first user's voice data included in the audio data;
    获取所述智能耳机上的麦克风采集的第一环境语音数据;Obtain the first environmental voice data collected by the microphone on the smart headset;
    根据所述第一用户的声音数据,对所述第一环境语音数据进行降噪处理。Perform noise reduction processing on the first environmental voice data according to the first user's voice data.
  3. 如权利要求2所述的方法,其特征在于,所述根据所述第一用户的声音数据,对所述第一环境语音数据进行降噪处理,包括:The method of claim 2, wherein performing noise reduction processing on the first environmental voice data according to the first user's voice data includes:
    获取使用所述智能耳机的第二用户的声音数据;Obtain the voice data of the second user using the smart headset;
    根据所述第一用户的声音数据和所述第二用户的声音数据,对所述第一环境语音数据中所述第一用户的声音数据和所述第二用户的声音数据以外的声音数据进行降噪处理。According to the voice data of the first user and the voice data of the second user, the voice data in the first environmental voice data other than the voice data of the first user and the voice data of the second user are processed. Noise reduction processing.
  4. 如权利要求1所述的方法,其特征在于,所述至少两个语音模式包括交流模式,所述方法,还包括:The method of claim 1, wherein the at least two voice modes include a communication mode, and the method further includes:
    响应于控制所述智能耳机切换至所述交流模式,获取所述智能耳机上的麦克风采集的第二环境语音数据;In response to controlling the smart headset to switch to the communication mode, obtain the second environmental voice data collected by the microphone on the smart headset;
    根据使用所述智能耳机的第二用户的声音数据,确定所述第二环境语音数据中的第三用户的声音数据;其中,所述第三用户和所述第二用户为不同的用户;Determine the voice data of the third user in the second environmental voice data according to the voice data of the second user using the smart headset; wherein the third user and the second user are different users;
    采用所述智能耳机播放所述第三用户的声音数据。The smart earphone is used to play the third user's voice data.
  5. 如权利要求1所述的方法,其特征在于,所述至少两个语音模式包括交流模式,所 述方法,还包括:The method of claim 1, wherein the at least two voice modes include communication modes, and the method further includes:
    响应于控制所述智能耳机切换至所述交流模式,获取使用所述智能耳机的第二用户的第一声音方位;In response to controlling the smart earphone to switch to the communication mode, obtain the first sound direction of the second user using the smart earphone;
    控制所述麦克风在所述第一声音方位以外的第二声音方位上采集环境中的第三用户的声音数据;其中,所述第三用户和所述第二用户为不同的用户;采用所述智能耳机播放所述第三用户的声音数据。Controlling the microphone to collect sound data of a third user in the environment at a second sound direction other than the first sound direction; wherein the third user and the second user are different users; using the The smart earphone plays the voice data of the third user.
  6. 如权利要求4或5所述的方法,其特征在于,所述方法,还包括:The method according to claim 4 or 5, characterized in that the method further includes:
    响应于控制所述智能耳机切换至所述交流模式,关闭所述降噪模式,且禁止所述麦克风和所述终端设备间进行数据传输。In response to controlling the smart earphone to switch to the communication mode, the noise reduction mode is turned off, and data transmission between the microphone and the terminal device is prohibited.
  7. 如权利要求4或5所述的方法,其特征在于,所述方法,还包括:The method according to claim 4 or 5, characterized in that the method further includes:
    响应于控制所述智能耳机切换至所述交流模式,关闭所述降噪模式;In response to controlling the smart headset to switch to the communication mode, turning off the noise reduction mode;
    响应于获取到所述终端设备发送的禁止指令,根据所述禁止指令,禁止所述麦克风和所述终端设备间进行数据传输。In response to obtaining the prohibition instruction sent by the terminal device, data transmission between the microphone and the terminal device is prohibited according to the prohibition instruction.
  8. 如权利要求1所述的方法,其特征在于,所述对所述音频数据进行识别,以确定所述终端设备处于目标工作场景,包括:The method of claim 1, wherein identifying the audio data to determine that the terminal device is in a target working scene includes:
    对所述音频数据进行识别,确定所述音频数据包括的语音频段;Identify the audio data and determine the voice segments included in the audio data;
    响应于所述语音频段中包含第一目标语音频段,且不包含第二目标语音频段和第三目标语音频段,确定所述终端设备处于目标工作场景;其中,所述第一目标语音频段的频率下限大于所述第二目标语音频段的频率上限,且所述第一目标语音频段的频率上限小于所述第三目标语音频段的频率下限。In response to the voice segment containing the first target voice segment and not including the second target voice segment and the third target voice segment, it is determined that the terminal device is in the target working scene; wherein the frequency of the first target voice segment The lower limit is greater than the upper frequency limit of the second target speech band, and the upper frequency limit of the first target speech band is less than the lower frequency limit of the third target speech band.
  9. 如权利要求1所述的方法,其特征在于,所述对所述音频数据进行识别,以确定所述终端设备处于目标工作场景,包括:The method of claim 1, wherein identifying the audio data to determine that the terminal device is in a target working scene includes:
    对所述音频数据携带的来源信息进行识别,确定所述音频数据的来源;Identify the source information carried by the audio data and determine the source of the audio data;
    响应于所述音频数据的来源为目标来源,确定所述终端设备处于目标工作场景。In response to the source of the audio data being the target source, it is determined that the terminal device is in the target working scene.
  10. 一种智能耳机的控制装置,其特征在于,包括:A control device for smart headphones, which is characterized by including:
    获取模块,用于获取所述终端设备上播放的音频数据;An acquisition module, used to acquire audio data played on the terminal device;
    第一确定模块,用于对所述音频数据进行识别,以确定所述终端设备处于目标工作场 景;所述目标工作场景下,所述智能耳机的麦克风处于语音采集状态;The first determination module is used to identify the audio data to determine that the terminal device is in a target working scenario; in the target working scenario, the microphone of the smart headset is in a voice collection state;
    所述获取模块,还用于在所述目标工作场景下,获取所述智能耳机的模式切换指令;The acquisition module is also used to acquire the mode switching instruction of the smart headset in the target working scenario;
    控制模块,用于根据所述模式切换指令,控制所述智能耳机在所述至少两个语音模式间切换。A control module configured to control the smart earphone to switch between the at least two voice modes according to the mode switching instruction.
  11. 如权利要求10所述的装置,其特征在于,所述至少两个语音模式包括降噪模式,所述装置,还包括:The device of claim 10, wherein the at least two voice modes include a noise reduction mode, and the device further includes:
    所述获取模块,还用于响应于控制所述智能耳机切换至所述降噪模式,获取所述终端设备上播放的音频数据;The acquisition module is also configured to acquire audio data played on the terminal device in response to controlling the smart headset to switch to the noise reduction mode;
    识别模块,用于对所述音频数据进行识别,以识别得到所述音频数据包括的第一人声数据;An identification module, used to identify the audio data to identify the first vocal data included in the audio data;
    所述获取模块,还用于获取所述智能耳机上的麦克风采集的第一环境语音数据;The acquisition module is also used to acquire the first environmental voice data collected by the microphone on the smart headset;
    处理模块,用于根据所述第一人声数据,对所述第一环境语音数据进行降噪处理。A processing module configured to perform noise reduction processing on the first environmental voice data according to the first human voice data.
  12. 如权利要求11所述的装置,其特征在于,所述处理模块,具体用于:The device according to claim 11, characterized in that the processing module is specifically used for:
    获取使用所述智能耳机的第一目标用户的第二声音数据;Obtain second sound data of the first target user using the smart headset;
    根据所述第一声音数据和所述第二声音数据,对所述第一环境语音数据中所述第一用户的声音数据和所述第二用户的声音数据以外的声音数据进行降噪处理。According to the first sound data and the second sound data, noise reduction processing is performed on the sound data in the first environmental voice data other than the sound data of the first user and the sound data of the second user.
  13. 如权利要求10所述的装置,其特征在于,所述至少两个语音模式包括交流模式,所述装置,还包括:The device of claim 10, wherein the at least two voice modes include a communication mode, and the device further includes:
    所述获取模块,还用于响应于控制所述智能耳机切换至所述交流模式,获取所述智能耳机上的麦克风采集的第二环境语音数据;The acquisition module is also configured to acquire the second environmental voice data collected by the microphone on the smart earphone in response to controlling the smart earphone to switch to the communication mode;
    第二确定模块,用于根据使用所述智能耳机的第二用户的声音数据,确定所述第二环境语音数据中的第三用户的声音数据;其中,所述第三用户和所述第二用户为不同的用户;A second determination module, configured to determine the voice data of a third user in the second environmental voice data according to the voice data of the second user using the smart headset; wherein the third user and the second user The users are different users;
    播放模块,用于采用所述智能耳机播放所述第三用户的声音数据。A playback module is used to play the third user's voice data using the smart earphone.
  14. 如权利要求10所述的装置,其特征在于,所述至少两个语音模式包括交流模式,所述装置,还包括:The device of claim 10, wherein the at least two voice modes include a communication mode, and the device further includes:
    所述获取模块,还用于响应于控制所述智能耳机切换至所述交流模式,获取使用所述智能耳机的第二用户的第一声音方位;The acquisition module is also configured to acquire the first sound direction of the second user using the smart earphone in response to controlling the smart earphone to switch to the communication mode;
    采集模块,用于控制所述麦克风在所述第二用户的第一声音方位以外的第二声音方位上采集环境中的第三用户的声音数据;其中,所述第三用户和所述第二用户为不同的用户;A collection module configured to control the microphone to collect the sound data of a third user in the environment at a second sound direction other than the first sound direction of the second user; wherein the third user and the second The users are different users;
    所述播放模块,用于采用所述智能耳机播放所述第三用户的声音数据。The playback module is used to play the third user's voice data using the smart earphone.
  15. 如权利要求13或14所述的装置,其特征在于,所述装置,还包括:The device according to claim 13 or 14, characterized in that the device further includes:
    第一关闭模块,用于响应于控制所述智能耳机切换至所述交流模式,关闭所述降噪模式,且禁止所述麦克风和所述终端设备间进行数据传输。A first shutdown module, configured to turn off the noise reduction mode in response to controlling the smart earphone to switch to the communication mode, and prohibit data transmission between the microphone and the terminal device.
  16. 如权利要求13或14所述的装置,其特征在于,所述装置,还包括:The device according to claim 13 or 14, characterized in that the device further includes:
    第二关闭模块,用于响应于控制所述智能耳机切换至所述交流模式,关闭所述降噪模式;响应于获取到所述终端设备发送的禁止指令,根据所述禁止指令,禁止所述麦克风和所述终端设备间进行数据传输。A second shutdown module, configured to turn off the noise reduction mode in response to controlling the smart headset to switch to the communication mode; in response to obtaining a prohibition instruction sent by the terminal device, to prohibit the Data transmission is performed between the microphone and the terminal device.
  17. 如权利要求10所述的装置,其特征在于,所述第一确定模块,具体用于:The device according to claim 10, characterized in that the first determination module is specifically used to:
    对所述音频数据进行识别,确定所述音频数据包括的语音频段;Identify the audio data and determine the voice segments included in the audio data;
    响应于所述语音频段中包含第一目标语音频段,且不包含第二目标语音频段和第三目标语音频段,确定所述终端设备处于目标工作场景;其中,所述第一目标语音频段的频率下限大于所述第二目标语音频段的频率上限,且所述第一目标语音频段的频率上限小于所述第三目标语音频段的频率下限。In response to the voice segment containing the first target voice segment and not including the second target voice segment and the third target voice segment, it is determined that the terminal device is in the target working scene; wherein the frequency of the first target voice segment The lower limit is greater than the upper frequency limit of the second target speech band, and the upper frequency limit of the first target speech band is less than the lower frequency limit of the third target speech band.
  18. 如权利要求10所述的装置,其特征在于,所述第一确定模块,具体用于:The device according to claim 10, characterized in that the first determination module is specifically used to:
    对所述音频数据携带的来源信息进行识别,确定所述音频数据的来源;Identify the source information carried by the audio data and determine the source of the audio data;
    响应于所述音频数据的来源为目标来源,确定所述终端设备处于目标工作场景。In response to the source of the audio data being the target source, it is determined that the terminal device is in the target working scene.
  19. 一种电子设备,其特征在于,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述程序时,实现如权利要求1-9中任一所述的方法。An electronic device, characterized in that it includes a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the program, any one of claims 1-9 is implemented. the method described.
  20. 一种非临时性计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1-9中任一所述的方法。A non-transitory computer-readable storage medium on which a computer program is stored, characterized in that when the computer program is executed by a processor, the method according to any one of claims 1-9 is implemented.
PCT/CN2022/099966 2022-06-20 2022-06-20 Smart earphone control method and apparatus, electronic device and storage medium WO2023245390A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2022/099966 WO2023245390A1 (en) 2022-06-20 2022-06-20 Smart earphone control method and apparatus, electronic device and storage medium
CN202280004138.1A CN117751585A (en) 2022-06-20 2022-06-20 Control method and device of intelligent earphone, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/099966 WO2023245390A1 (en) 2022-06-20 2022-06-20 Smart earphone control method and apparatus, electronic device and storage medium

Publications (1)

Publication Number Publication Date
WO2023245390A1 true WO2023245390A1 (en) 2023-12-28

Family

ID=89378989

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/099966 WO2023245390A1 (en) 2022-06-20 2022-06-20 Smart earphone control method and apparatus, electronic device and storage medium

Country Status (2)

Country Link
CN (1) CN117751585A (en)
WO (1) WO2023245390A1 (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170318374A1 (en) * 2016-05-02 2017-11-02 Microsoft Technology Licensing, Llc Headset, an apparatus and a method with automatic selective voice pass-through
CN108156550A (en) * 2017-12-27 2018-06-12 上海传英信息技术有限公司 The playing method and device of headphone
CN111464905A (en) * 2020-04-09 2020-07-28 电子科技大学 Hearing enhancement method and system based on intelligent wearable device and wearable device
CN113099338A (en) * 2021-03-08 2021-07-09 头领科技(昆山)有限公司 Intelligent control's audio chip and wireless earphone of making an uproar that falls
CN113873378A (en) * 2020-06-30 2021-12-31 华为技术有限公司 Earphone noise processing method and device and earphone
CN113873379A (en) * 2020-06-30 2021-12-31 华为技术有限公司 Mode control method and device and terminal equipment

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170318374A1 (en) * 2016-05-02 2017-11-02 Microsoft Technology Licensing, Llc Headset, an apparatus and a method with automatic selective voice pass-through
CN108156550A (en) * 2017-12-27 2018-06-12 上海传英信息技术有限公司 The playing method and device of headphone
CN111464905A (en) * 2020-04-09 2020-07-28 电子科技大学 Hearing enhancement method and system based on intelligent wearable device and wearable device
CN113873378A (en) * 2020-06-30 2021-12-31 华为技术有限公司 Earphone noise processing method and device and earphone
CN113873379A (en) * 2020-06-30 2021-12-31 华为技术有限公司 Mode control method and device and terminal equipment
CN113099338A (en) * 2021-03-08 2021-07-09 头领科技(昆山)有限公司 Intelligent control's audio chip and wireless earphone of making an uproar that falls

Also Published As

Publication number Publication date
CN117751585A (en) 2024-03-22

Similar Documents

Publication Publication Date Title
EP3163748B1 (en) Method, device and terminal for adjusting volume
US20230315380A1 (en) Devices with enhanced audio
CN105451111B (en) Earphone control method for playing back, device and terminal
CN106454644B (en) Audio playing method and device
CN110166890B (en) Audio playing and collecting method and device and storage medium
KR102538775B1 (en) Method and apparatus for playing audio, electronic device, and storage medium
CN107493500A (en) Multimedia resource player method and device
CN104636110B (en) Control the method and device of volume
CN104991754A (en) Recording method and apparatus
CN109087650B (en) Voice wake-up method and device
JP2017528929A (en) Event notification method and apparatus
CN106888327B (en) Voice playing method and device
CN108549529A (en) Control the method and device of audio output
JP2017508413A (en) Voice call prompt method, apparatus, program, and recording medium
CN108845787A (en) Method, apparatus, terminal and the storage medium that audio is adjusted
CN111009239A (en) Echo cancellation method, echo cancellation device and electronic equipment
CN106375846B (en) The processing method and processing device of live audio
WO2018058331A1 (en) Method and apparatus for controlling volume
CN111988704B (en) Sound signal processing method, device and storage medium
WO2023245390A1 (en) Smart earphone control method and apparatus, electronic device and storage medium
US11388281B2 (en) Adaptive method and apparatus for intelligent terminal, and terminal
CN106936972B (en) Method, device and system for controlling audio channel
CN108491180B (en) Audio playing method and device
CN112118502B (en) Earphone control method, device, equipment and storage medium
CN114567840B (en) Audio output method and device, mobile terminal and storage medium

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 202280004138.1

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22947171

Country of ref document: EP

Kind code of ref document: A1