WO2022002110A1 - 一种模式控制方法、装置及终端设备 - Google Patents

一种模式控制方法、装置及终端设备 Download PDF

Info

Publication number
WO2022002110A1
WO2022002110A1 PCT/CN2021/103435 CN2021103435W WO2022002110A1 WO 2022002110 A1 WO2022002110 A1 WO 2022002110A1 CN 2021103435 W CN2021103435 W CN 2021103435W WO 2022002110 A1 WO2022002110 A1 WO 2022002110A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
processing
function
signal
anc
Prior art date
Application number
PCT/CN2021/103435
Other languages
English (en)
French (fr)
Inventor
陈伟宾
王提政
李玉龙
范泛
仇存收
熊伟
曹天祥
桂振侠
陈志鹏
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to EP21834164.2A priority Critical patent/EP4171060A4/en
Priority to BR112022026923A priority patent/BR112022026923A2/pt
Priority to KR1020237002989A priority patent/KR20230027296A/ko
Publication of WO2022002110A1 publication Critical patent/WO2022002110A1/zh
Priority to US18/148,080 priority patent/US20230164475A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/10Earpieces; Attachments therefor ; Earphones; Monophonic headphones
    • H04R1/1041Mechanical or electronic switches, or control elements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/10Earpieces; Attachments therefor ; Earphones; Monophonic headphones
    • H04R1/1016Earpieces of the intra-aural type
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10KSOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
    • G10K11/00Methods or devices for transmitting, conducting or directing sound in general; Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
    • G10K11/16Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
    • G10K11/175Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound
    • G10K11/178Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase
    • G10K11/1781Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase characterised by the analysis of input or output signals, e.g. frequency range, modes, transfer functions
    • G10K11/17821Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase characterised by the analysis of input or output signals, e.g. frequency range, modes, transfer functions characterised by the analysis of the input signals only
    • G10K11/17823Reference signals, e.g. ambient acoustic environment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/08Mouthpieces; Microphones; Attachments therefor
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/10Earpieces; Attachments therefor ; Earphones; Monophonic headphones
    • H04R1/1083Reduction of ambient noise
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/43Electronic input selection or mixing based on input signal analysis, e.g. mixing or selection between microphone and telecoil or between microphones with different directivity characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/55Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired
    • H04R25/558Remote control, e.g. of amplification, frequency
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • H04R3/005Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R5/00Stereophonic arrangements
    • H04R5/04Circuit arrangements, e.g. for selective connection of amplifier inputs/outputs to loudspeakers, for loudspeaker detection, or for adaptation of settings to personal preferences or hearing impairments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2201/00Details of transducers, loudspeakers or microphones covered by H04R1/00 but not provided for in any of its subgroups
    • H04R2201/10Details of earpieces, attachments therefor, earphones or monophonic headphones covered by H04R1/10 but not provided for in any of its subgroups
    • H04R2201/109Arrangements to adapt hands free headphones for use on both ears
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2225/00Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
    • H04R2225/41Detection or adaptation of hearing aid parameters or programs to listening situation, e.g. pub, forest
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2225/00Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
    • H04R2225/43Signal processing in hearing aids to enhance the speech intelligibility
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2410/00Microphones
    • H04R2410/05Noise reduction with a separate noise microphone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2420/00Details of connection covered by H04R, not provided for in its groups
    • H04R2420/01Input selection or mixing for amplifiers or loudspeakers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2430/00Signal processing covered by H04R, not provided for in its groups
    • H04R2430/01Aspects of volume control, not necessarily automatic, in sound systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2430/00Signal processing covered by H04R, not provided for in its groups
    • H04R2430/03Synergistic effects of band splitting and sub-band processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2460/00Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
    • H04R2460/01Hearing devices using active noise cancellation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2460/00Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
    • H04R2460/03Aspects of the reduction of energy consumption in hearing devices
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2460/00Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
    • H04R2460/05Electronic compensation of the occlusion effect
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2460/00Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
    • H04R2460/13Hearing devices using bone conduction transducers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/50Customised settings for obtaining desired overall acoustical characteristics
    • H04R25/505Customised settings for obtaining desired overall acoustical characteristics using digital signal processing

Definitions

  • the embodiments of the present application relate to the technical field of audio processing, and in particular, to a mode control method, apparatus, and terminal device.
  • headset users In recent years, there have been more and more headset users, and users' functional requirements for headsets have become more and more differentiated. For example, some users do not want to hear external noise when wearing headphones, and can use the active noise control (ANC) function to eliminate the noise in the ears. Some users want to hear the sound outside the earphone, and need to use the ambient sound through (hear through, HT) function to realize the same feeling of the outside sound as without the earphone. Some users may have hearing impairments, and the external signals that the user wants can be delivered to the user through the augmentation hearing (AH) function, and the unwanted signals can be filtered out at the same time.
  • ANC active noise control
  • HT ambient sound through
  • HT ambient sound through
  • Some users may have hearing impairments, and the external signals that the user wants can be delivered to the user through the augmentation hearing (AH) function, and the unwanted signals can be filtered out at the same time.
  • AH augmentation hearing
  • the headset cannot achieve the effect that the user wants to achieve according to the user's needs.
  • Embodiments of the present application provide an earphone noise processing method, device, and earphone, so as to achieve the effect that the user wants to achieve according to the user's needs.
  • an embodiment of the present application provides a noise processing method for an earphone, the earphone has at least two functions of an active noise reduction ANC function, an ambient sound transparent transmission HT function, or an auditory enhancement AH function, and the earphone includes a first microphone and a second microphone, the first microphone is used to collect the first signal, the first signal is used to characterize the sound of the current external environment, the second microphone is used to collect the second signal, and the second signal is used to characterize the wearing The sound of the environment inside the ear canal of the user of the headset.
  • the earphone can be a left earphone or a right earphone.
  • the processing modes used by the left earphone and the right earphone can be the same or different.
  • the earphone receives the first audio signal from the terminal device; obtains a target mode; wherein, the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the earphone to implement the target processing function, so
  • the target processing function is one of the active noise reduction ANC function, the ambient sound transparent transmission HT function or the auditory enhancement AH function; according to the target mode, the first audio signal, the first signal, and the second signal A second audio signal is obtained.
  • the target mode is determined according to the scene type of the external environment of the field, and the user's listening effect can be optimized in real time.
  • the earphone further includes a speaker, and the speaker is used to play the second audio signal.
  • the target processing function is an ANC function
  • the second audio signal played by the speaker can reduce the sound of the user's current environment and the sound inside the user's ear canal. Perception of ambient sound; or, the target processing function is an HT function, and the second audio signal played by the speaker can enhance the user's perception of the sound of the environment where the user is currently located; or, the target processing function
  • the function is the AH function, and the second audio signal played by the speaker can enhance the user's perception of the event sound; the event sound satisfies a preset frequency spectrum.
  • the audio signal played by the speaker of the left earphone can attenuate the sound of the user's left earphone to the current environment of the user (that is, the sound of the current external environment) and the sound inside the ear canal of the user's left earphone. Perception of ambient sound.
  • the audio signal played by the speaker of the right earphone can weaken the perception of the user's right earphone to the sound of the user's current environment (that is, the sound of the current external environment) and the ambient sound inside the ear canal of the user's right earphone.
  • the feeling of the left ear follows the processing mode adopted by the left earphone
  • the feeling of the right ear follows the processing mode adopted by the right earphone.
  • the target processing function is the ANC function
  • the second audio signal is obtained based on the first audio signal, the third signal and the fourth signal
  • the third signal is the first signal.
  • the fourth signal is an inverted signal of the second signal; or, when the target processing function is the HT function, the second audio signal is based on the first audio signal, the first signal and the second audio signal or, when the target processing function is the AH function, the second audio signal is obtained based on the first audio signal, the fifth signal and the fourth signal, and the fifth signal is one of the first audio signals.
  • the event signal is used to represent a specific sound in the current external environment, and the event signal satisfies a preset frequency spectrum.
  • the above design provides a simple and effective way of obtaining the signal output by the speaker under different processing modes.
  • the acquiring target mode includes:
  • a first control instruction from the terminal device is received, where the first control instruction carries the target mode, and the target mode is determined by the terminal device according to the scene type of the current external environment.
  • the terminal device determines the target mode according to the scene type of the external environment and indicates to the earphone, which can optimize the user's listening effect in real time.
  • a second control instruction from the terminal device, where the second control instruction carries a target processing strength, and the target processing strength is used to instruct the headset to perform the target processing function processing strength; obtaining a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal, including: according to the target mode, the target processing strength, the The first audio signal, the first signal, and the second signal result in a second audio signal.
  • the terminal device indicates the processing intensity in the corresponding processing mode of the earphone. Adjusting the processing intensity on the basis of the processing mode further improves the user's hearing experience.
  • the target event corresponding to the event sound in the current external environment is determined according to the first signal, and the target processing intensity in the target mode is determined according to the target event; wherein, the target processing intensity It is used to indicate the processing strength when the earphone realizes the target processing function; obtaining a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal, including: A second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal. where different processing intensities correspond to different events.
  • the processing intensity can be in one-to-one correspondence with events, or one processing intensity corresponds to multiple events. For example, two events can use the same processing intensity, but different processing intensity cannot be used for the same event.
  • the earphone determines the processing intensity according to the event sound in the external environment, realizes different auditory experience in different external environments, can reduce the sense of bottom noise and enhance the noise reduction.
  • the headset further includes a bone conduction sensor, and the bone conduction sensor is used to collect bone conduction signals generated by the vibration of the user's vocal cords;
  • the first scene at the location includes: identifying the first scene where the user is currently located according to the first signal and the bone conduction signal.
  • the target event is a howling event, a wind noise event, a sudden event or a human voice event.
  • the acquiring the target mode includes: identifying, according to the first signal, the scene type of the current external environment as the target scene type (target scene for short, or target type for short), and according to the target scene
  • the target mode adopted by the headset is determined, where the target mode is a processing mode corresponding to the target scene.
  • the different processing modes correspond to different scene types.
  • the processing modes may be in one-to-one correspondence with scene types, or one processing mode may correspond to multiple scene types, for example, two scene types may adopt the same processing mode.
  • the earphone determines the processing mode adopted by the earphone according to the recognized scene type, reduces the time delay, and optimizes the user's hearing experience in real time.
  • the target scene is a walking scene, a running scene, a quiet scene, a multi-person talking scene, a cafe scene, a subway scene, a train scene, a waiting room scene, a dialogue scene, an office scene, an outdoor scene, One of the driving scene, windy scene, airplane scene, siren scene, whistle scene, and crying scene.
  • the audio signal includes: obtaining a second audio signal according to the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
  • the earphone determines the processing mode to be adopted, and instructs the terminal device to adjust the processing intensity and reduce the occupation of processing resources of the earphone.
  • the target processing function is an ANC function
  • the greater the target processing intensity is, the more the user perceives the sound of the current environment of the user and the environment inside the user's ear canal The weaker the sound; or, when the target processing function is the HT function, the greater the target processing intensity, the greater the intensity of the sound of the environment where the user is currently perceived by the user; or, when the target When the processing function is the AH function, the greater the target processing intensity, the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
  • the target mode instructs the headset to implement an ANC function, and obtaining a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal, including:
  • the second audio signal is obtained by performing mixing processing on the fourth audio signal and the first audio signal.
  • FF filtering and FB serial processing are used to implement ANC processing to obtain better noise-reduced signals and enhance the noise-reduction effect.
  • the filter coefficient used in the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the ANC function; or, the third The filter coefficient used in the filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the ANC function.
  • the target mode instructs the earphone to implement the HT function
  • the second audio signal is obtained according to the target mode, the first audio signal, the first signal, and the second signal, including:
  • the first signal processing including a second filtering process (such as HT filtering);
  • a third filtering process (such as FB filtering) on the second filtered signal to obtain a third filtered signal
  • the second audio signal is obtained by performing mixing processing on the third filtered signal and the fifth audio signal.
  • filtering and supplementary processing may be performed on the fifth audio signal to reduce hearing loss.
  • the HT filter undergoes down-mixing processing and filter compensation processing to further reduce hearing loss.
  • performing first signal processing on the first environmental signal to obtain a processed environmental signal includes: performing a second filtering process on the first signal to obtain a second filtered signal;
  • the second signal processing includes de-blocking effect processing.
  • the second signal processing further includes at least one of the following: noise reduction processing, wind noise reduction processing, gain adjustment processing or frequency response adjustment processing.
  • the noise floor and abnormal sounds are reduced, and the user's hearing experience is improved.
  • the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the HT function; or,
  • the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the HT function.
  • the target mode instructs the headset to implement the AH function
  • the second audio signal is obtained according to the target mode, the first audio signal, the first signal, and the second signal, including:
  • Second filtering processing (such as HT filtering) on the first signal to obtain a second filtering signal
  • enhancement processing on the second filtering signal to obtain a filtering enhanced signal
  • Performing a first filtering process (such as FF filtering) on the first signal to obtain a first filtered signal
  • the second audio signal is obtained by mixing the fifth filtered signal, the sixth audio signal and the first filtered signal.
  • the transparently transmitted signal is processed by transparent transmission filtering and enhancement processing, and the transparently transmitted signal is clearer.
  • filtering and compensating the sixth audio signal can avoid losses caused by FB filtering and ensure transparency to the greatest extent. Signal transmission without distortion.
  • performing enhancement processing on the second filtered signal to obtain a filtered enhanced signal including:
  • noise reduction processing includes artificial intelligence AI noise reduction processing and/or wind noise reduction processing
  • the filtered enhanced signal is obtained by performing gain amplification processing and frequency response adjustment on the signal obtained after the noise reduction processing.
  • the transparently transmitted signal is enhanced. Improve the user's sense of hearing for desired external sounds.
  • the headset includes a bone conduction sensor, the bone conduction sensor is used to collect the bone conduction signal of the headset user, and the signal obtained after noise reduction processing is subjected to gain amplification processing, including: : performing harmonic expansion on the bone conduction signal to obtain a harmonically expanded signal; using the first gain coefficient to amplify the signal obtained by the noise reduction process; using the fourth filter coefficient to filter out the signal obtained by the amplification process The harmonically extended signal included in ; wherein the fourth filter coefficient is determined based on the first gain coefficient.
  • an amplification method which only amplifies the specific sound other than the voice of the wearing user, and improves the effect of the specific sound in the transparently transmitted ambient sound.
  • the first gain coefficient is a gain coefficient associated with the target processing intensity in the target mode.
  • performing enhancement processing on the second filtered signal to obtain a filtered enhanced signal including:
  • the audio event signal in the deblocked signal is subjected to gain amplification processing and frequency response adjustment to obtain a filtered enhanced signal.
  • the headset further includes a bone conduction sensor, the bone conduction sensor is used to collect the bone conduction signal of the headset user, and perform gain amplification on the audio event signal in the de-occluded signal processing, including: performing harmonic expansion on the bone conduction signal to obtain a harmonically extended signal; amplifying an audio event signal in the de-occluded signal using a second gain coefficient to obtain an amplified signal; using a second filter coefficient Filtering out the harmonically extended signal included in the amplified signal; wherein the second filter coefficient is determined based on the second gain coefficient.
  • the bone conduction sensor is used to collect the bone conduction signal of the headset user, and perform gain amplification on the audio event signal in the de-occluded signal processing, including: performing harmonic expansion on the bone conduction signal to obtain a harmonically extended signal; amplifying an audio event signal in the de-occluded signal using a second gain coefficient to obtain an amplified signal; using a second filter coefficient Filtering out the harmonically extended signal included in the amplified signal; where
  • the second gain coefficient is a gain coefficient associated with the target processing intensity for the first filtering processing when the first noise processing is performed.
  • the second gain coefficient is a gain coefficient associated with the first scene identifier for the first filtering process when the first noise processing is performed.
  • the filter coefficient used by the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the AH function; or,
  • the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the AH function; or,
  • the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the AH function.
  • the earphone further includes a bone conduction sensor, and the bone conduction sensor is used to collect the bone conduction signal of the user of the earphone; performing de-occlusion effect processing on the second filtered signal, including:
  • the low-frequency component in the signal obtains a third filtered signal, and the high-frequency component in the third filtered signal from which the low-frequency component is removed is amplified.
  • an embodiment of the present invention provides a mode control method, the method is applied to a terminal device, and the method includes: when a scene type of the current external environment is identified as a target scene, determining a target mode according to the target scene; Wherein, the target mode is one of the processing modes supported by the headset, and the processing modes supported by the headset include at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode, or the hearing enhancement AH mode; The earphone sends the target mode, where the target mode is used to instruct the earphone to implement the processing function corresponding to the target mode.
  • Different processing modes correspond to different scene types, and processing modes may be in one-to-one correspondence with scene types, or one processing mode may correspond to multiple scene types, for example, two scene types may adopt the same processing mode.
  • the terminal device controls the processing mode of the headset in real time according to the scene recognition, so as to optimize the user's hearing experience in real time.
  • the method when determining the target mode corresponding to the target scene in the processing mode of the headset, the method further includes: displaying result prompt information, where the result prompt information is used to prompt the user that the headset implements the desired Describe the processing function corresponding to the target mode.
  • the above design enables the user to determine the current processing mode of the headset in real time.
  • the method before sending the first control signaling to the headset, the method further includes: displaying selection prompt information, where the selection prompt information is used to prompt the user whether to adjust the processing mode of the headset to any The target mode is detected; an operation that the user selects to adjust the processing mode of the headset to the target mode is detected.
  • the user can determine whether to adjust the processing mode of the earphone according to the needs, so as to improve the user experience.
  • a first control and a second control are displayed, wherein different positions of the second control on the first control are used to indicate different processing intensities in the target mode;
  • the method further includes: in response to the user touching the second control to move to a first position on the first control, the second control is on the first control
  • the first position of the device indicates the target processing strength in the target mode; the target processing strength is sent to the headset, where the target processing strength is used to indicate the processing when the headset implements the processing function corresponding to the target mode strength.
  • the user can select the processing intensity of the earphone according to the needs, so as to meet the different needs of the user.
  • the shape of the first control is a ring
  • the processing intensity in the target mode is Change from small to large
  • the shape of the first control is a bar
  • the target processing function when the target processing function is an ANC function, the greater the target processing intensity is, the more the user perceives the sound of the current environment of the user and the environment inside the user's ear canal The weaker the sound.
  • the target processing function when the target processing function is the HT function, the greater the target processing intensity, the greater the intensity of the sound of the environment where the user is currently located; or, when the target processing function is AH
  • the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
  • the left earphone and the right earphone may adopt the same processing mode and processing intensity, and the perception of the user's left ear and right ear may be the same.
  • the left earphone and the right earphone may also adopt different processing modes or different processing intensities, so that the perception of the left ear and the right ear is different.
  • an embodiment of the present invention provides a mode control method, wherein the method is applied to a terminal device, and the method includes: acquiring a target mode; the target mode is one of the processing modes supported by the headset, The processing modes supported by the headset include at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode or the auditory enhancement AH mode; the target processing intensity in the target mode is determined according to the scene type of the current external environment; different The scene types corresponding to different processing intensities in the target mode; the target processing intensities are sent to the earphone, where the target processing intensities are used to indicate the processing intensities when the earphones implement the processing functions corresponding to the target modes.
  • the acquiring the target mode includes: receiving the target mode sent by the headset; or, displaying a selection control, where the selection control includes a processing mode supported by the headset, and it is detected that the user has passed the A selection control selects the operation of the target mode among the processing modes of the headset.
  • the selection control includes the processing mode supported by the headset, or is interpreted as an option for the selection control to provide the processing mode supported by the headset, or the selection control displays the processing mode supported by the headset, and the user can perform in the processing mode supported by the headset. choose. .
  • the method before determining the target processing intensity in the target mode according to the scene type of the current external environment, the method further includes: when the target mode sent by the headset is received, displaying a selection Prompt information, the selection prompt information is used to indicate whether the user adjusts the processing mode of the headset to the target mode; an operation of the user selecting to adjust the processing mode of the headset to the target mode is detected.
  • the target processing function is an ANC function
  • the target processing function is the HT function
  • the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
  • the target processing function is the AH function
  • the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
  • an embodiment of the present application provides a mode control method, the method is applied to a terminal device, the method includes: displaying a first interface; the first interface includes a first selection control, and the first selection control includes The processing mode supported by the first target earphone and the processing intensity corresponding to the processing mode supported by the first target earphone; the processing mode of the first target earphone includes active noise reduction ANC mode, ambient sound transparent transmission HT mode or hearing enhancement AH mode at least two kinds of; in response to a first operation performed by the user on the first interface; the first operation is that the user selects the first target mode from the processing modes supported by the first target headset through the first selection control; and The processing intensity in the first target mode is selected to be generated by the first target processing intensity; the first target mode and the first target processing intensity are sent to the first target earphone, and the first target mode is used to indicate The first target earphone implements the processing function corresponding to the first target mode, and the first target processing intensity is used to indicate the processing when the first target ear
  • the first selection control includes the processing mode supported by the first target headset and the processing intensity corresponding to the processing mode supported by the first target headset, which can be interpreted as the first selection control provides the user with multiple processing modes (all are the first target headset. supported) options and adjustments to the processing strength for each processing mode.
  • the user can freely switch the processing mode and intensity corresponding to the effect that the earphone needs to achieve through the UI interface, so as to meet the different needs of the user.
  • the method before displaying the first interface, the method further includes: displaying selection prompt information, where the selection prompt information is used for the user to select whether to adjust the processing mode of the first target headset; detecting The user selects an operation to adjust the processing mode of the first target headset.
  • the method before displaying the first interface, the method further includes: identifying the scene type of the current external environment as a target scene, and the target scene adaptation needs to adjust the processing mode of the first target headset scene type.
  • a method is provided to actively pop up the first interface in a specific scenario, so as to reduce the manual operation process of the user.
  • the method before displaying the first interface, the method further includes: recognizing that the terminal device triggers the first target headset to play audio. Recognizing that the terminal device triggers the first target earphone to play audio may be interpreted as recognizing that the terminal device starts sending an audio signal to the first target earphone.
  • the method before displaying the first interface, the method further includes: detecting that the terminal device establishes a connection with the first target headset.
  • the method before displaying the first interface, further includes: in the case of detecting that the terminal device establishes a connection with the first target headset, detecting that the user executes the operation on the main interface the second operation; wherein, the main interface includes an icon of a first application, the second operation is generated by the user touching the icon of the first application, and the first interface is the first application display interface.
  • the first selection control includes a first control and a second control
  • the second control indicates two different processing modes of the first target earphone at any two different positions of the first control
  • the second control indicates different processing intensities of the first target earphone in the same processing mode at any two different positions of the first control
  • the first operation is that the user moves the second control
  • the control is generated at a first position on the first control in an area corresponding to the first target mode, where the first position corresponds to the first target processing intensity in the first target mode.
  • the shape of the first control is a ring or a bar.
  • the shape of the first control is a ring
  • the ring includes at least two arc segments
  • the second control is located in different arc segments to indicate different processing modes of the first target headset
  • the second The different positions of the controls in the same arc segment indicate different processing intensities of the same processing mode of the first target earphone.
  • the shape of the first control is a bar
  • the bar includes at least two bar segments
  • the second controls are located in different bar segments to indicate different processing modes of the first target earphone, so Different positions of the second control in the same bar segment indicate different processing intensities of the same processing mode of the first target earphone.
  • the method further includes:
  • the first interface further includes a second selection control
  • the second selection control includes a processing mode supported by the second target headset and a processing mode supported by the second target headset
  • the processing modes supported by the first target earphone include at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode, or the hearing enhancement AH mode
  • the third operation is that the user
  • the second selection control selects the second target mode in the processing mode of the second target earphone and selects the processing intensity in the second target mode to be generated by the second target processing intensity; when the first target earphone is the left earphone, the The second target earphone is a right earphone, or the first target earphone is a right earphone, and the second target earphone is a left earphone;
  • the second target processing intensity is used to indicate the processing intensity when the second target earphone implements the processing function corresponding to the second target mode.
  • the user can operate the processing mode and processing intensity of the left earphone and the right earphone separately, so as to meet the user's differentiated requirements for the hearing sense of the left ear and the right ear.
  • an embodiment of the present application further provides a headset control method, the method is applied to a terminal device, and the method includes: establishing a communication connection between the terminal device and the headset; displaying a first interface; the first interface is used to set the headset function; the first interface includes an event sound enhancement function option; the event sound is a sound that meets preset event conditions in the external environment; when the event sound enhancement function option is enabled, control the headset's sound Both the ANC function and the HT function are turned on.
  • an embodiment of the present application further provides an earphone control device, the device is applied to a terminal device, and the device includes:
  • a display module for displaying a first interface; the first interface is used to set the function of the headset; the first interface includes function options for event sound enhancement; the event sound is in the external environment that meets preset event conditions sound;
  • the processing module is configured to control both the ANC function and the HT function of the earphone to be in an enabled state when the option of the event sound enhancement function is enabled.
  • the first interface includes an option for controlling the HT function of the earphone; when the option of the HT function is enabled, the HT of the earphone is activated function; and an option of enhancing the sound of the event is added in the first interface.
  • the target strength of the HT function can also be acquired; the HT function of the earphone is controlled according to the target strength of the HT function. This step can be performed in cooperation with the display module and the processing module.
  • controlling both the ANC function and the HT function of the earphone to be in an on state includes: keeping the HT function in an on state; and activating the ANC function of the earphone. This step may be performed by a processing module.
  • the first interface includes an option for controlling the ANC function of the headset; when the option of the ANC function is enabled, the headset is activated ANC function. Further optionally, a strength option of the ANC function is added in the first interface. This step may be performed by a processing module.
  • the first interface further includes an option for disabling the ANC function, and/or an option for disabling the HT function.
  • the strength options of the ANC function include at least a first steady-state ANC strength option, a second steady-state ANC strength option, and an adaptive ANC strength option; wherein, the The first steady-state ANC strength option and the second steady-state ANC strength option correspond to the first scenario and the second scenario, respectively, and correspond to different ANC functional strengths; the ANC functional strength corresponding to the adaptive ANC strength option It is related to the scene type of the current environment where the terminal device or the headset is located; different scene types of the current environment correspond to different ANC strengths.
  • the scene type of the current environment where the terminal or the headset is located is acquired; according to the scene The type matches the target strength of the ANC function; the ANC function of the headset is controlled according to the target strength.
  • the different scene types of the current environment include a first scene and a second scene.
  • the event sound includes a human voice or other sounds conforming to preset spectral characteristics.
  • the option enablement of the event sound enhancement function, the option enablement of the HT function, or the option enablement of the ANC function includes: responding to the user's request for The click of the corresponding function option, the adaptive switching of the corresponding function, or the shortcut of the corresponding function is triggered.
  • an embodiment of the present application further provides a noise reduction method, the method is applied to an earphone, the earphone supports at least an active noise reduction ANC function, and can also support an ambient sound transparent transmission HT function, and the earphone includes a first a microphone, a second microphone and a speaker; the method includes: collecting a first signal through the first microphone, where the first signal is used to represent the sound of the current external environment; collecting a second signal through the second microphone, the The second signal is used to characterize the sound of the internal environment of the ear canal of the user wearing the headset; receive an instruction for event sound enhancement; the event sound is the sound that meets the preset event conditions in the external environment; control the ANC function to be turned on, even if the Or turn on the ANC function, at least use the ANC function to perform target processing on the first signal and the second signal to obtain a target signal; the signal-to-noise ratio of the event sound in the target signal is greater than that in the first signal.
  • an embodiment of the present application further provides a noise reduction device.
  • the device is applied to an earphone.
  • the earphone supports at least an active noise reduction ANC function, and can also support an ambient sound transparent transmission HT function.
  • the earphone includes a first microphone, a second microphone and a speaker; the device includes: a collection module configured to collect a first signal through the first microphone, where the first signal is used to represent the sound of the current external environment; further configured to use the second microphone Collecting a second signal, the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the headset; the receiving module is used to receive an instruction of event sound enhancement; the event sound is in the external environment and conforms to a preset event conditional sound; a processing module for enabling the ANC function, and at least using the ANC function to perform target processing on the first signal and the second signal to obtain a target signal; the information of the event sound in the target signal
  • the noise ratio is greater than the signal-to-noise ratio
  • control ANC function and the HT function are both turned on; the HT function is used to transparently transmit the first signal to obtain a restored signal; for events in the restored signal The sound signal is enhanced, and the non-event sound signal in the restored signal is weakened to obtain an event sound enhancement signal; the first signal, the second signal and the event sound enhancement signal are processed by using the ANC function The target signal is obtained.
  • the target signal is obtained by processing the first signal, the second signal and the event sound enhancement signal by using an ANC function.
  • the earphone supports at least the active noise reduction ANC function, the ambient sound transparent transmission HT function and the AH function, and the earphone includes an HT filter bank, a feedback filter bank and Feedforward filter bank; obtain the working mode of the headset; when the working mode is the ANC function, call the feedback filter bank and the feedforward filter bank to realize the ANC function; when the working mode is the HT function When the HT filter bank and the feedback filter bank are called to realize the HT function; when the working mode is the AH function, the HT filter bank, the feedforward filter bank and the all The feedback filter bank is used to realize the AH function.
  • an embodiment of the present application further provides a noise reduction method, the method is applied to an earphone, and the earphone supports at least an ANC function; the earphone includes a first microphone and a third microphone; the first microphone is more focused In order to collect the sound of the current external environment, the third microphone focuses more on sound pickup; when the headset enables the ANC function, the first microphone is used to collect the first signal for the current environment; The microphone collects a second signal for the current environment; judges the noise level of the current scene according to the first signal and the second signal; wherein, different noise levels correspond to different ANC intensities; according to the current noise level Controls the ANC function.
  • an embodiment of the present application further provides a noise reduction device, the device is applied to an earphone, and the earphone supports at least an ANC function; the earphone includes a first microphone and a third microphone; the first microphone is more focused In order to collect the sound of the current external environment, the third microphone focuses more on sound pickup; the collection module is used to collect the first microphone for the current environment through the first microphone when the ANC function of the headset is in an on state. signal; collecting a second signal for the current environment through the third microphone; an identification module for judging the noise level of the current scene according to the first signal and the second signal; wherein different noise levels Corresponding to different ANC strengths; the processing module is used to control the ANC function according to the current noise level.
  • the correlation feature of the first signal and the second signal for voice activity detection; perform noise tracking on non-voice signals; if the noise energy is less than the first signal threshold, the current scene is determined to be a quiet scene; or, if the spectral center of the noise is in a low frequency band and the noise energy is greater than the second threshold, the current scene is determined to be a heavy noise scene; or, if the current scene is both When it does not belong to the quiet scene or the heavy noise scene, the current scene is determined to be an ordinary scene; wherein the second threshold is greater than the first threshold.
  • the ANC strengths corresponding to the quiet scene, the normal scene, and the heavy noise scene sequentially increase.
  • the ninth aspect or the tenth aspect in a possible design, if it is detected that the current scene is a new noise level and lasts for a preset period of time; obtain the ANC intensity corresponding to the new noise level; The ANC strength corresponding to the noise level controls the ANC function.
  • an embodiment of the present application further provides a method for controlling an earphone.
  • the method is applied to a terminal device, and the method includes: establishing a communication connection between the terminal device and the earphone; and the earphone supports at least an active noise reduction (ANC) function.
  • ANC active noise reduction
  • the first interface is used to set the function of the headset; the first interface includes an option for controlling the ANC function of the headset; when the option of the ANC function is enabled, activate the ANC function of the headset; adding a strength option of the ANC function in the first interface; performing ANC noise reduction according to the enabling result of the strength option of the ANC function; wherein, the strength option of the ANC function at least includes A first steady-state ANC strength option, a second steady-state ANC strength option, and an adaptive ANC strength option; wherein the first steady-state ANC strength option and the second steady-state ANC strength option correspond to the first scenario and The second scenario corresponds to different and stable ANC functional strengths respectively; the ANC functional strength corresponding to the adaptive ANC strength option is related to the scene type of the current environment where the terminal device or the headset is located; the current environment Different scene types correspond to different ANC strengths.
  • the different scene types of the current environment include a first scene and a second scene.
  • an embodiment of the present application further provides a headset control device, the terminal device establishes a communication connection with the headset; the headset supports at least an active noise reduction ANC function; the device includes: a display module for displaying the first an interface; the first interface is used to set the function of the headset; the first interface includes an option for controlling the ANC function of the headset; a processing module is used to activate the ANC function when the option of the ANC function is enabled The ANC function of the headset; the display module is further configured to add a strength option of the ANC function in the first interface after the option of the ANC function is enabled; the processing module is also configured to ANC noise reduction is performed on the result of enabling the strength option of the ANC function; wherein, the strength option of the ANC function includes at least a first steady-state ANC strength option, a second steady-state ANC strength option, and an adaptive ANC strength option; wherein, The first steady-state ANC strength option and the second steady-state ANC strength option correspond to the
  • the first steady state ANC strength option when the first steady state ANC strength option is enabled, the first ANC function strength corresponding to the first steady state ANC strength option is obtained; The first ANC function strength controls the ANC function; or, when the second steady state ANC strength option is enabled, obtain the second ANC function strength corresponding to the second steady state ANC strength option; according to the second ANC function Intensity controls the ANC function.
  • the adaptive ANC strength option when the adaptive ANC strength option is enabled, the scene type of the current environment where the terminal device or the headset is located is obtained; the ANC strength is determined according to the scene type of the current environment; the ANC function is controlled according to the determined ANC strength .
  • an embodiment of the present application further provides a method for controlling an earphone.
  • the method is applied to a terminal device, and the method includes: establishing a communication connection between the terminal device and an earphone; and the earphone supports at least ambient sound transparent transmission HT function, display the first interface; the first interface is used to set the function of the headset; the first interface includes options for controlling the HT function of the headset; when the option of the HT function is enabled, activate all The HT function of the headset; the option of the event sound enhancement is added in the first interface; the event sound is the sound that meets the preset event conditions in the external environment; when the event sound enhancement option is enabled When , the earphone is controlled to improve the signal-to-noise ratio of the event sound in the signal collected by the earphone; wherein, the higher the signal-to-noise ratio of the event sound, the higher the energy ratio of the event sound in the signal big.
  • an embodiment of the present application further provides an earphone control device, the device is applied to a terminal device, and the terminal device establishes a communication connection with the earphone; the earphone supports at least the HT function of transparent transmission of ambient sound, and the device It includes: a display module for displaying a first interface; the first interface is used to set the function of the headset; the first interface includes options for controlling the HT function of the headset; When the option of the HT function is enabled, the HT function of the headset is activated; the display module is further configured to add the option of the event sound enhancement in the first interface after the option of the HT function is enabled; The event sound is a sound that meets preset event conditions in the external environment; the processing module is further configured to control the earphone for the event in the signal collected by the earphone when the option of the event sound enhancement function is enabled The signal-to-noise ratio of the sound is improved; wherein, the higher the signal-to-noise ratio of the event sound, the greater the
  • the event sound includes human voice or other sounds that conform to preset spectral characteristics.
  • the first interface includes an option for controlling the ANC function of the headset, an option for turning off the ANC function, and/or an option for turning off the HT function.
  • the first intensity of the ANC function is obtained, and the ANC function of the earphone is controlled according to the first intensity; or, the second intensity of the HT function is obtained.
  • an embodiment of the present application further provides a noise processing device, the device is applied to an earphone, and the earphone has at least two functions of an active noise reduction ANC function, an ambient sound transparent transmission HT function, or an auditory enhancement AH function
  • the headset includes a first microphone and a second microphone; the first microphone is used to collect a first signal, and the first signal is used to characterize the sound of the current external environment; the second microphone is used to collect the second signal , the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the earphone.
  • the noise processing apparatus includes corresponding functional modules, which are respectively used to implement the steps in the method of the first aspect above. For details, please refer to the detailed description in the method example, which will not be repeated here.
  • the functions can be implemented by hardware, or by executing corresponding software by hardware.
  • the hardware or software includes one or more modules corresponding to the above functions.
  • the noise processing device includes:
  • a communication module for receiving the first audio signal from the terminal device
  • an acquisition module configured to acquire a target mode; wherein, the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the headset to implement a target processing function, and the target processing function is active noise reduction One of ANC function, ambient sound transparent transmission HT function or auditory enhancement AH function;
  • a first processing module configured to obtain a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal.
  • an embodiment of the present application provides a target earphone, including a left earphone and a right earphone, where the left earphone is used to implement any of the above possible design methods for earphones, or the right earphone is used to implement any of the above possible design approaches regarding headphones.
  • the left and right earphones use different processing modes.
  • an embodiment of the present application provides a target earphone, where the target earphone includes a left earphone and a right earphone.
  • the left earphone includes or the right earphone includes a first microphone, a second microphone, a processor, a memory, and a speaker;
  • the first microphone is used to collect a first signal, and the first signal is used to represent the sound of the current external environment ;
  • the second microphone is used to collect a second signal, and the second signal is used to characterize the sound of the internal environment of the ear canal of the user wearing the headset;
  • the memory is used to store programs or instructions;
  • the processor used to call the program or instruction, so that the electronic device executes any possible earphone-related method to obtain the second audio signal or the target signal;
  • the speaker is used to play the second audio signal or the target signal Signal.
  • an embodiment of the present application provides a mode control apparatus, where the apparatus is applied to a terminal device.
  • the apparatus includes corresponding functional modules, which are respectively used to implement any of the above possible terminal-related steps.
  • the functions can be implemented by hardware, or by executing corresponding software by hardware.
  • the hardware or software includes one or more modules corresponding to the above functions.
  • an embodiment of the present application provides a terminal device, including a memory, a processor, and a display; the display is used for displaying an interface; the memory is used for storing programs or instructions; the processor is used for The program or instruction is invoked to cause the terminal device to execute steps in any possible terminal-related method.
  • the present application provides a computer-readable storage medium, where computer programs or instructions are stored in the computer-readable storage medium, and when the computer program or instructions are executed by the headset, the headset can execute any of the above possible related Headphone design method.
  • the present application provides a computer-readable storage medium.
  • the computer-readable storage medium stores a computer program or instruction.
  • the headset can perform any of the above possible functions.
  • the present application provides a computer program product, the computer program product includes a computer program or an instruction, when the computer program or instruction is executed by the headset, the method in any of the above possible implementations of the headset is implemented .
  • the present application provides a computer program product, the computer program product includes a computer program or instruction, when the computer program or instruction is executed by the headset, the method in any of the possible headset implementations described above is implemented.
  • FIG. 1 is a schematic diagram of a hardware structure of a terminal device 100 in an embodiment of the present application
  • FIG. 2 is a schematic diagram of a software structure of a terminal device 100 in an embodiment of the present application
  • FIG. 3 is a schematic structural diagram of an earphone 200 in an embodiment of the present application.
  • FIG. 4 is a schematic diagram of the AHA pathway in the embodiment of the present application.
  • 5A is a flowchart of ANC processing in an embodiment of the present application.
  • 5B is a schematic diagram of an ANC processing flow in an embodiment of the present application.
  • 6A is a flowchart of HT processing in an embodiment of the present application.
  • 6B is a schematic diagram of a HT processing flow in an embodiment of the present application.
  • 6C is a schematic diagram of another HT processing flow in an embodiment of the present application.
  • FIG. 7 is a schematic diagram of a process flow of de-occlusion effect processing in an embodiment of the present application.
  • FIG. 8A is a flowchart of AH processing in an embodiment of the present application.
  • FIG. 8B is a schematic diagram of an AH processing flow in an embodiment of the present application.
  • 8C is a schematic diagram of another AH processing flow in an embodiment of the present application.
  • FIG. 9 is a schematic diagram of a noise reduction processing flow diagram in an embodiment of the present application.
  • FIG. 10 is a schematic diagram of a process flow of gain amplification in an embodiment of the present application.
  • FIG. 11 is a schematic diagram of another gain amplification processing flow in an embodiment of the present application.
  • 12A is a schematic diagram of a main interface of a terminal device in an embodiment of the present application.
  • 12B is a schematic diagram of a control interface of an earphone application in an embodiment of the present application.
  • 12C is a schematic diagram of the control of a terminal device controlling a headset in an ANC mode in an embodiment of the present application
  • 12D is a schematic diagram of the control of a terminal device controlling an earphone in an HT mode in an embodiment of the present application
  • 12E is a schematic diagram of the control of a terminal device controlling an earphone in an AH mode in an embodiment of the present application
  • 12F is a schematic diagram of a selection control in an embodiment of the present application.
  • 12G is a schematic diagram of another selection control in an embodiment of the present application.
  • 12H is a schematic diagram of triggering a control interface of an earphone in an embodiment of the present application.
  • FIG. 13 is a schematic diagram of another selection control in an embodiment of the application.
  • FIG. 14A is a schematic diagram of control for enabling a smart scene detection function according to an embodiment of the present application.
  • FIG. 14B is a schematic diagram of another kind of smart scene detection function enabling control in an embodiment of the present application.
  • 14C is a schematic diagram of an earphone control interface in an embodiment of the present application.
  • 15 is a schematic diagram of event detection in an embodiment of the present application.
  • 16 is a schematic diagram of an interaction processing mode and processing intensity between a terminal device and a headset in an embodiment of the application;
  • FIG. 17A is a schematic diagram of displaying a scene detection result in an embodiment of the present application.
  • FIG. 17B is a schematic diagram of another scene detection result display in an embodiment of the present application.
  • FIG. 19 is a schematic structural diagram of a noise processing apparatus 1900 in an embodiment of the present application.
  • FIG. 20 is a schematic structural diagram of a mode control apparatus 2000 in an embodiment of the present application.
  • FIG. 21 is a schematic structural diagram of a mode control device 2100 in an embodiment of the present application.
  • FIG. 22 is a schematic structural diagram of a mode control apparatus 2200 in an embodiment of the present application.
  • FIG. 23 is a schematic structural diagram of a terminal device 2300 in an embodiment of the present application.
  • FIG. 24 is a schematic diagram of an earphone control interface in a terminal device in an embodiment of the application.
  • FIG. 25 is a schematic diagram of a headset control interface in a terminal device in an embodiment of the application.
  • FIG. 26 is a schematic diagram of an earphone control interface in a terminal device according to an embodiment of the application.
  • FIG. 27 is a schematic diagram of a headset control interface in a terminal device according to an embodiment of the application.
  • FIG. 28 is a schematic diagram of an earphone control interface in a terminal device according to an embodiment of the present application.
  • the application (application, app) involved in the embodiments of the present application is a software program capable of implementing one or more specific functions.
  • multiple applications can be installed in a terminal device.
  • the applications mentioned below may be system applications that have been installed when the terminal device is shipped from the factory, or may be third-party applications downloaded from the network or obtained from other terminal devices by the user during the use of the terminal device.
  • the human auditory system has a masking effect, that is, strong-frequency sounds will hinder human perception of the nearby weak-frequency sounds, and the basilar membrane of the cochlea has a frequency selection and tuning effect on external sound signals. Therefore, the concept of critical frequency band is introduced, Perceptually measure sound frequency. It is generally believed that there are 24 critical frequency bands within the hearing threshold of 22 Hz to 22 kHz, which can cause vibrations at different positions on the basilar membrane. Each critical band is called a bark subband.
  • VAD Voice endpoint detection
  • At least one (item) refers to one (item) or more (item), and “multiple (item)” refers to two (item) or more than two (item).
  • “And/or”, which describes the association relationship of the associated objects means that there can be three kinds of relationships, for example, A and/or B, it can mean that A exists alone, A and B exist at the same time, and B exists alone, where A, B can be singular or plural.
  • the character “/” generally indicates that the associated objects are an "or” relationship.
  • At least one item(s) below” or similar expressions thereof refer to any combination of these items, including any combination of single item(s) or plural items(s).
  • At least one (a) of a, b, or c may represent: a, b, c, a-b, a-c, b-c, or a-b-c, where a, b, and c may be single or multiple.
  • the symbol "(a, b)" represents an open interval, and the range is greater than a and less than b; "[a, b]” represents a closed interval, and the range is greater than or equal to a and less than or equal to b; "(a ,b]” represents a half-open and half-closed interval with a range greater than a and less than or equal to b; "(a,b]” represents a half-open and half-closed interval with a range greater than a and less than or equal to b.
  • first and second mentioned in the embodiments of this application are used to distinguish multiple objects, and are not used to limit the size, content, order, timing, priority or importance of multiple objects. degree, etc.
  • first microphone and the second microphone are only used to distinguish different microphones, and do not indicate the difference in size, priority, or importance of the two microphones.
  • An embodiment of the present application provides a system, where the system includes a terminal device 100 and an earphone 200 .
  • the terminal device 100 is connected 200 to the earphone, and the connection may be a wireless connection or a wired connection.
  • the terminal device may be connected to the headset through Bluetooth technology, wireless fidelity (Wi-Fi) technology, infrared IR technology, and ultra-wideband technology.
  • the terminal device 100 is a device having a display interface function.
  • the terminal device 100 may be, for example, a product with a display interface such as a mobile phone, a monitor, a tablet computer, and a vehicle-mounted device, as well as a smart display wearable product such as a smart watch and a smart bracelet.
  • a product with a display interface such as a mobile phone, a monitor, a tablet computer, and a vehicle-mounted device, as well as a smart display wearable product such as a smart watch and a smart bracelet.
  • a smart display wearable product such as a smart watch and a smart bracelet.
  • the specific form of the above mobile terminal is not particularly limited in this embodiment of the present application.
  • the earphone 200 includes two sounding units that can be hung around the ears.
  • the one that fits the left ear can be called the left earphone, and the one that fits the right ear can be called the right earphone.
  • the earphone 200 in this embodiment of the present application may be a headphone, an ear-hook earphone, a neck-hung earphone, or an earbud earphone, or the like.
  • In-ear headphones also include in-ear headphones (or referred to as ear canal headphones) or semi-in-ear headphones.
  • the earphone 200 has at least two of the ANC function, the HT function, or the AH function.
  • AHA for the convenience of description, ANC, HT, and AH are collectively referred to as AHA in this embodiment of the present application, and of course, other names may also be ordered, which is not limited in this application.
  • the left earphone and the right earphone have similar structures. Either the left earphone or the right earphone can adopt the earphone structure described below.
  • the earphone structure (left earphone or right earphone) includes a rubber sleeve that can be inserted into the ear canal, an ear bag close to the ear, and an earphone rod suspended on the ear bag. The rubber sleeve guides the sound to the ear canal.
  • the ear bag includes devices such as batteries, speakers, and sensors. Microphones and physical buttons can be arranged on the headphone rod.
  • the earphone rod can be in the shape of a cylinder, a rectangular parallelepiped, an ellipsoid, or the like.
  • the microphone arranged inside the ear can be called the error microphone, and the microphone arranged outside the earphone is called the reference microphone.
  • the error microphone is used to pick up the sound of the external environment. Referring to the microphone, when the user wears the earphone, the sound of the internal environment of the ear canal of the user wearing the earphone is collected.
  • the two microphones can be either analog microphones or digital microphones. After the user wears the earphone, the positional relationship between the two microphones and the speaker is as follows: the error microphone is inside the ear, close to the earphone rubber sleeve.
  • the loudspeaker is located between the error microphone and the reference microphone.
  • the reference microphone is close to the outer structure of the ear and can be arranged on the upper part of the earphone stem.
  • the tube of the error microphone can either face the loudspeaker or the inside of the ear canal.
  • the terminal device 100 is configured to send downlink audio signals and/or control signaling to the headset 200 .
  • the control signaling is used to control the processing mode adopted by the headset 200 .
  • the processing mode adopted by the headset 200 may include at least two of an empty mode indicating no processing, an ANC mode indicating implementing an ANC function, an HT mode indicating implementing an HT function, or an AH mode indicating implementing an AH function.
  • the earphone user's perception of the sound of the current external environment and the ambient sound inside the ear canal of the user wearing the earphone can be weakened.
  • the headset adopts the HT mode the user's perception of the sound of the current external environment can be enhanced.
  • the headset adopts the AH mode the user's perception of the event sound included in the sound of the current external environment can be enhanced.
  • the event sound refers to a preset sound in an external environment, or the event sound satisfies a preset frequency spectrum.
  • the event sound includes a station announcement sound or a whistle sound in a railway station; then the event sound satisfies the frequency spectrum of the station announcement sound or the frequency spectrum of the whistle sound in the railway station.
  • the event sound may include a notification sound in an airplane terminal, a broadcast sound on an airplane, etc.; another example is the sound of a hotel calling number, and the like. It should be understood that both the terminal and the headset can recognize the event sound.
  • the earphone 200 includes a left earphone and a right earphone, and the left earphone and the right earphone may adopt the same processing mode or different processing modes.
  • the left earphone and the right earphone use the same processing mode, the user's left ear wearing the left earphone and the right ear wearing the right earphone may have the same auditory perception.
  • the left earphone and the right earphone adopt different processing modes, the user's left ear wearing the left earphone and the right ear wearing the right earphone have different auditory perceptions.
  • the left earphone using ANC and the right earphone using AH as an example, when the left earphone adopts the ANC mode, the sound of the earphone user's left ear to the current external environment and the environmental sound inside the ear canal of the user wearing the earphone can be reduced. perception.
  • the right earphone adopts the AH mode the perception of the user's right ear to the event sound included in the sound of the current external environment can be enhanced.
  • any one of the following possible ways can be used to achieve it;
  • the terminal device 100 provides a control interface for the user to select the processing mode of the headset 200 according to requirements. For example, under the operation of the user, the terminal device 100 sends control signaling to the headset 200 , and the control signaling is used to indicate the processing mode adopted by the headset 200 .
  • the processing modes adopted by the left earphone and the right earphone in the earphone 200 may be the same or different.
  • a selection control on the control surface is used to select the same processing mode for the left and right earphones.
  • the control interface may include two selection controls, wherein one selection control is used to select a processing mode for the left earphone, and the other selection control is used to select a processing mode for the right earphone. The control interface and selection controls will be described in detail later, and the description will not be repeated here.
  • the terminal device identifies the scene type of the user's current external environment.
  • the processing modes adopted by the headset 200 are different, that is, the processing functions implemented by the headset are different.
  • the earphone 200 recognizes the user's operation, and determines that the earphone 200 selected by the user adopts the ANC mode, the HT mode or the AH mode.
  • the user's operation may be an operation of the user tapping the earphone, or buttons are provided on the earphone, and different buttons indicate different processing modes.
  • the earphone recognizes the scene type of the external environment of the earphone, and the processing mode adopted by the earphone is different in different scenes.
  • FIG. 1 shows a schematic diagram of an optional hardware structure of the terminal device 100 .
  • the terminal device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2 , mobile communication module 150, wireless communication module 160, audio module 170, speaker 170A, receiver 170B, microphone 170C, headphone jack 170D, sensor module 180, buttons 190, motor 191, indicator 192, camera 193, display screen 194, and Subscriber identification module (subscriber identification module, SIM) card interface 195 and so on.
  • SIM Subscriber identification module
  • the sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, and ambient light. Sensor 180L, bone conduction sensor 180M, etc.
  • the structures illustrated in the embodiments of the present invention do not constitute a specific limitation on the terminal device 100 .
  • the terminal device 100 may include more or less components than shown, or some components are combined, or some components are separated, or different components are arranged.
  • the illustrated components may be implemented in hardware, software, or a combination of software and hardware.
  • the processor 110 may include one or more processing units, for example, the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processor (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), controller, video codec, digital signal processor (digital signal processor, DSP), baseband processor, and/or neural-network processing unit (neural-network processing unit, NPU), etc. Wherein, different processing units may be independent devices, or may be integrated in one or more processors.
  • application processor application processor, AP
  • modem processor graphics processor
  • ISP image signal processor
  • controller video codec
  • digital signal processor digital signal processor
  • baseband processor baseband processor
  • neural-network processing unit neural-network processing unit
  • the controller can generate an operation control signal according to the instruction operation code and timing signal, and complete the control of fetching and executing instructions.
  • a memory may also be provided in the processor 110 for storing instructions and data.
  • the memory in processor 110 is cache memory. This memory may hold instructions or data that have just been used or recycled by the processor 110 . If the processor 110 needs to use the instruction or data again, it can be called directly from the memory. Repeated accesses are avoided and the latency of the processor 110 is reduced, thereby increasing the efficiency of the system.
  • the processor 110 may include one or more interfaces.
  • the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, a universal asynchronous transceiver (universal asynchronous transmitter) receiver/transmitter, UART) interface, mobile industry processor interface (MIPI), general-purpose input/output (GPIO) interface, subscriber identity module (SIM) interface, and / or universal serial bus (universal serial bus, USB) interface, etc.
  • I2C integrated circuit
  • I2S integrated circuit built-in audio
  • PCM pulse code modulation
  • PCM pulse code modulation
  • UART universal asynchronous transceiver
  • MIPI mobile industry processor interface
  • GPIO general-purpose input/output
  • SIM subscriber identity module
  • USB universal serial bus
  • the I2C interface is a bidirectional synchronous serial bus that includes a serial data line (SDA) and a serial clock line (SCL).
  • the processor 110 may contain multiple sets of I2C buses.
  • the processor 110 can be respectively coupled to the touch sensor 180K, the charger, the flash, the camera 193 and the like through different I2C bus interfaces.
  • the processor 110 may couple the touch sensor 180K through the I2C interface, so that the processor 110 and the touch sensor 180K communicate with each other through the I2C bus interface, so as to realize the touch function of the terminal device 100 .
  • the I2S interface can be used for audio communication.
  • the processor 110 may contain multiple sets of I2S buses.
  • the processor 110 may be coupled with the audio module 170 through an I2S bus to implement communication between the processor 110 and the audio module 170 .
  • the audio module 170 can transmit an audio signal to the wireless communication module 160 through the I2S interface, so as to realize the function of answering a call through the headset 200 (such as a Bluetooth headset).
  • the PCM interface can also be used for audio communications, sampling, quantizing and encoding analog signals.
  • the audio module 170 and the wireless communication module 160 may be coupled through a PCM bus interface.
  • the audio module 170 can also transmit audio signals to the wireless communication module 160 through the PCM interface, so as to realize the function of answering a call through the Bluetooth headset 200 .
  • Both the I2S interface and the PCM interface can be used for audio communication.
  • the UART interface is a universal serial data bus used for asynchronous communication.
  • the bus may be a bidirectional communication bus. It converts the data to be transmitted between serial communication and parallel communication.
  • a UART interface is typically used to connect the processor 110 with the wireless communication module 160 .
  • the processor 110 communicates with the Bluetooth module in the wireless communication module 160 through the UART interface to implement the Bluetooth function.
  • the audio module 170 can transmit audio signals to the wireless communication module 160 through the UART interface, so as to realize the function of playing music through the Bluetooth headset 200 .
  • the MIPI interface can be used to connect the processor 110 with peripheral devices such as the display screen 194 and the camera 193 .
  • MIPI interfaces include camera serial interface (CSI), display serial interface (DSI), etc.
  • the processor 110 communicates with the camera 193 through the CSI interface, so as to realize the shooting function of the terminal device 100 .
  • the processor 110 communicates with the display screen 194 through the DSI interface to implement the display function of the terminal device 100 .
  • the GPIO interface can be configured by software.
  • the GPIO interface can be configured as a control signal or as a data signal.
  • the GPIO interface may be used to connect the processor 110 with the camera 193, the display screen 194, the wireless communication module 160, the audio module 170, the sensor module 180, and the like.
  • the GPIO interface can also be configured as I2C interface, I2S interface, UART interface, MIPI interface, etc.
  • the USB interface 130 is an interface that conforms to the USB standard specification, and may specifically be a Mini USB interface, a Micro USB interface, a USB Type C interface, and the like.
  • the USB interface 130 can be used to connect a charger to charge the terminal device 100, and can also be used to transmit data between the terminal device 100 and peripheral devices. It can also be used to connect the headset 200 to play audio through the headset 200 .
  • This interface can also be used to connect other terminal devices, such as AR devices.
  • the interface connection relationship between the modules illustrated in the embodiment of the present invention is only a schematic illustration, and does not constitute a structural limitation of the terminal device 100 .
  • the terminal device 100 may also adopt different interface connection manners in the foregoing embodiments, or a combination of multiple interface connection manners.
  • the charging management module 140 is used to receive charging input from the charger.
  • the charger may be a wireless charger or a wired charger.
  • the charging management module 140 may receive charging input from the wired charger through the USB interface 130 .
  • the charging management module 140 may receive wireless charging input through the wireless charging coil of the terminal device 100 . While the charging management module 140 charges the battery 142 , it can also supply power to the terminal device through the power management module 141 .
  • the power management module 141 is used for connecting the battery 142 , the charging management module 140 and the processor 110 .
  • the power management module 141 receives input from the battery 142 and/or the charging management module 140, and supplies power to the processor 110, the internal memory 121, the display screen 194, the camera 193, and the wireless communication module 160.
  • the power management module 141 can also be used to monitor parameters such as battery capacity, battery cycle times, battery health status (leakage, impedance).
  • the power management module 141 may also be provided in the processor 110 .
  • the power management module 141 and the charging management module 140 may also be provided in the same device.
  • the wireless communication function of the terminal device 100 may be implemented by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, the modulation and demodulation processor, the baseband processor, and the like.
  • Antenna 1 and Antenna 2 are used to transmit and receive electromagnetic wave signals.
  • Each antenna in terminal device 100 may be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve antenna utilization.
  • the antenna 1 can be multiplexed as a diversity antenna of the wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
  • the mobile communication module 150 may provide a wireless communication solution including 2G/3G/4G/5G etc. applied on the terminal device 100 .
  • the mobile communication module 150 may include at least one filter, switch, power amplifier, low noise amplifier (LNA), and the like.
  • the mobile communication module 150 can receive electromagnetic waves from the antenna 1, filter and amplify the received electromagnetic waves, and transmit them to the modulation and demodulation processor for demodulation.
  • the mobile communication module 150 can also amplify the signal modulated by the modulation and demodulation processor, and then turn it into an electromagnetic wave for radiation through the antenna 1 .
  • at least part of the functional modules of the mobile communication module 150 may be provided in the processor 110 .
  • at least part of the functional modules of the mobile communication module 150 may be provided in the same device as at least part of the modules of the processor 110 .
  • the modem processor may include a modulator and a demodulator.
  • the modulator is used to modulate the low frequency baseband signal to be sent into a medium and high frequency signal.
  • the demodulator is used to demodulate the received electromagnetic wave signal into a low frequency baseband signal. Then the demodulator transmits the demodulated low-frequency baseband signal to the baseband processor for processing.
  • the low frequency baseband signal is processed by the baseband processor and passed to the application processor.
  • the application processor outputs sound signals through audio devices (not limited to the speaker 170A, the receiver 170B, etc.), or displays images or videos through the display screen 194 .
  • the modem processor may be a stand-alone device.
  • the modem processor may be independent of the processor 110, and may be provided in the same device as the mobile communication module 150 or other functional modules.
  • the wireless communication module 160 can provide applications on the terminal device 100 including wireless local area networks (WLAN) (such as wireless fidelity (Wi-Fi) networks), bluetooth (BT), global navigation satellites Wireless communication solutions such as global navigation satellite system (GNSS), frequency modulation (FM), near field communication (NFC), and infrared technology (IR).
  • WLAN wireless local area networks
  • BT Bluetooth
  • GNSS global navigation satellite system
  • FM frequency modulation
  • NFC near field communication
  • IR infrared technology
  • the wireless communication module 160 may be one or more devices integrating at least one communication processing module.
  • the wireless communication module 160 receives electromagnetic waves via the antenna 2 , frequency modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 110 .
  • the wireless communication module 160 can also receive the signal to be sent from the processor 110 , perform frequency modulation on it, amplify it, and convert it into electromagnetic waves for radiation through the antenna 2 .
  • the wireless communication module 160 includes a Bluetooth module, and the terminal device 100 establishes a wireless connection with the headset 200 through Bluetooth.
  • the wireless communication module 160 includes an infrared module, and the terminal device 100 can establish a wireless connection with the headset 200 through the infrared module.
  • the antenna 1 of the terminal device 100 is coupled with the mobile communication module 150, and the antenna 2 is coupled with the wireless communication module 160, so that the terminal device 100 can communicate with the network and other devices through wireless communication technology.
  • the wireless communication technology may include global system for mobile communications (GSM), general packet radio service (GPRS), code division multiple access (CDMA), broadband Code Division Multiple Access (WCDMA), Time Division Code Division Multiple Access (TD-SCDMA), Long Term Evolution (LTE), BT, GNSS, WLAN, NFC , FM, and/or IR technology, etc.
  • the GNSS may include a global positioning system (global positioning system, GPS), a global navigation satellite system (GLONASS), a Beidou navigation satellite system (BDS), a quasi-zenith satellite system (quasi -zenith satellite system, QZSS) and/or satellite based augmentation systems (SBAS).
  • GPS global positioning system
  • GLONASS global navigation satellite system
  • BDS Beidou navigation satellite system
  • QZSS quasi-zenith satellite system
  • SBAS satellite based augmentation systems
  • the terminal device 100 implements a display function through a GPU, a display screen 194, an application processor, and the like.
  • the GPU is a microprocessor for image processing, and is connected to the display screen 194 and the application processor.
  • the GPU is used to perform mathematical and geometric calculations for graphics rendering.
  • Processor 110 may include one or more GPUs that execute program instructions to generate or alter display information.
  • Display screen 194 is used to display images, videos, and the like.
  • Display screen 194 includes a display panel.
  • the display panel can be a liquid crystal display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode or an active-matrix organic light-emitting diode (active-matrix organic light).
  • LED diode AMOLED
  • flexible light-emitting diode flexible light-emitting diode (flex light-emitting diode, FLED), Miniled, MicroLed, Micro-oLed, quantum dot light-emitting diode (quantum dot light emitting diodes, QLED) and so on.
  • the terminal device 100 may include 1 or N1 display screens 194 , where N1 is a positive integer greater than 1.
  • the terminal device 100 can realize the shooting function through the ISP, the camera 193, the video codec, the GPU, the display screen 194 and the application processor.
  • the ISP is used to process the data fed back by the camera 193 .
  • the shutter is opened, the light is transmitted to the camera photosensitive element through the lens, the light signal is converted into an electrical signal, and the camera photosensitive element transmits the electrical signal to the ISP for processing, and converts it into an image visible to the naked eye.
  • ISP can also perform algorithm optimization on image noise, brightness, and skin tone.
  • ISP can also optimize the exposure, color temperature and other parameters of the shooting scene.
  • the ISP may be provided in the camera 193 .
  • Camera 193 is used to capture still images or video.
  • the object is projected through the lens to generate an optical image onto the photosensitive element.
  • the photosensitive element may be a charge coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor.
  • CMOS complementary metal-oxide-semiconductor
  • the photosensitive element converts the optical signal into an electrical signal, and then transmits the electrical signal to the ISP to convert it into a digital image signal.
  • the ISP outputs the digital image signal to the DSP for processing.
  • DSP converts digital image signals into standard RGB, YUV and other formats of image signals.
  • the processor 110 may trigger the startup of the camera 193 according to a program or an instruction in the internal memory 121, so that the camera 193 captures at least one image and performs corresponding processing on the at least one image according to the program or instruction.
  • the terminal device 100 may include 1 or N2 cameras 193 , where N2 is a positive integer greater than 1.
  • a digital signal processor is used to process digital signals, in addition to processing digital image signals, it can also process other digital signals. For example, when the terminal device 100 selects a frequency point, the digital signal processor is used to perform Fourier transform on the frequency point energy, and the like.
  • Video codecs are used to compress or decompress digital video.
  • the terminal device 100 may support one or more video codecs.
  • the terminal device 100 can play or record videos in various encoding formats, for example, moving picture experts group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4 and so on.
  • MPEG moving picture experts group
  • the NPU is a neural-network (NN) computing processor.
  • NN neural-network
  • Applications such as intelligent cognition of the terminal device 100 can be implemented through the NPU, such as image recognition, face recognition, speech recognition, text understanding, and the like.
  • the external memory interface 120 can be used to connect an external memory card, such as a Micro SD card, to expand the storage capacity of the terminal device 100 .
  • the external memory card communicates with the processor 110 through the external memory interface 120 to realize the data storage function. For example to save files like music, video etc in external memory card.
  • Internal memory 121 may be used to store computer executable program code, which includes instructions.
  • the internal memory 121 may include a storage program area and a storage data area.
  • the program storage area may store an operating system, an application program (such as a camera application) required for at least one function, and the like.
  • the storage data area may store data created during the use of the terminal device 100 (such as images captured by a camera, etc.) and the like.
  • the internal memory 121 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, universal flash storage (UFS), and the like.
  • the processor 110 executes various functional applications and data processing of the terminal device 100 by executing instructions stored in the internal memory 121 and/or instructions stored in a memory provided in the processor.
  • the internal memory 121 may also store the downlink audio signal provided by the embodiment of the present application.
  • the internal memory 121 may also store codes for implementing the functions of controlling the headset 200 .
  • the code stored in the internal memory 121 for implementing the function of controlling the earphone 200 is executed by the processor 110, the earphone 200 is controlled to realize the corresponding function, such as the ANC function, the HT function or the AH function.
  • the code for implementing the function of controlling the headset 200 provided by the embodiment of the present application may also be stored in an external memory.
  • the processor 110 may run the corresponding data stored in the external memory to realize the function of controlling the earphone 200 through the external memory interface 120, so as to control the earphone 200 to realize the corresponding function.
  • the terminal device 100 may implement audio functions through an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, an application processor, and the like. Such as music playback, recording, etc.
  • the audio module 170 is used for converting digital audio information into analog audio signal output, and also for converting analog audio input into digital audio signal. Audio module 170 may also be used to encode and decode audio signals. In some embodiments, the audio module 170 may be provided in the processor 110 , or some functional modules of the audio module 170 may be provided in the processor 110 .
  • Speaker 170A also referred to as a "speaker" is used to convert audio electrical signals into sound signals.
  • the terminal device 100 can listen to music through the speaker 170A, or listen to a hands-free call.
  • the receiver 170B also referred to as "earpiece" is used to convert audio electrical signals into sound signals.
  • the terminal device 100 answers a call or a voice message, the voice can be answered by placing the receiver 170B close to the human ear.
  • the microphone 170C also called “microphone” or “microphone” is used to convert sound signals into electrical signals. When making a call or sending a voice message, the user can make a sound near the microphone 170C through the human mouth, and input the sound signal into the microphone 170C.
  • the terminal device 100 may be provided with at least one microphone 170C. In other embodiments, the terminal device 100 may be provided with two microphones 170C, which may implement a noise reduction function in addition to collecting sound signals. In other embodiments, the terminal device 100 may further be provided with three, four or more microphones 170C to collect sound signals, reduce noise, identify sound sources, and implement directional recording functions.
  • the earphone jack 170D is used to connect wired earphones.
  • the terminal device 100 is connected to the earphone through the earphone interface 170D.
  • the earphone interface 170D can be a USB interface 130, or a 3.5mm open mobile terminal platform (open mobile terminal platform, OMTP) standard interface, a cellular telecommunications industry association of the USA (CTIA) standard interface.
  • OMTP open mobile terminal platform
  • CTIA cellular telecommunications industry association of the USA
  • the pressure sensor 180A is used to sense pressure signals, and can convert the pressure signals into electrical signals.
  • the pressure sensor 180A may be provided on the display screen 194 .
  • the capacitive pressure sensor may be comprised of at least two parallel plates of conductive material. When a force is applied to the pressure sensor 180A, the capacitance between the electrodes changes.
  • the terminal device 100 determines the intensity of the pressure according to the change in capacitance. When a touch operation acts on the display screen 194, the terminal device 100 detects the intensity of the touch operation according to the pressure sensor 180A.
  • the terminal device 100 may also calculate the touched position according to the detection signal of the pressure sensor 180A.
  • touch operations acting on the same touch position but with different touch operation intensities may correspond to different operation instructions. For example, when a touch operation whose intensity is less than the first pressure threshold acts on the short message application icon, the instruction for viewing the short message is executed. When a touch operation with a touch operation intensity greater than or equal to the first pressure threshold acts on the short message application icon, the instruction to create a new short message is executed.
  • the gyro sensor 180B may be used to determine the motion attitude of the terminal device 100 .
  • the angular velocity of the end device 100 about three axes ie, the x, y and z axes
  • the gyro sensor 180B can be used for image stabilization.
  • the gyro sensor 180B detects the shaking angle of the terminal device 100, calculates the distance to be compensated by the lens module according to the angle, and allows the lens to offset the shaking of the terminal device 100 through reverse motion to achieve anti-shake.
  • the gyro sensor 180B can also be used for navigation and somatosensory game scenarios.
  • the air pressure sensor 180C is used to measure air pressure.
  • the terminal device 100 calculates the altitude through the air pressure value measured by the air pressure sensor 180C to assist in positioning and navigation.
  • the magnetic sensor 180D includes a Hall sensor.
  • the terminal device 100 can detect the opening and closing of the flip holster using the magnetic sensor 180D.
  • the terminal device 100 can detect the opening and closing of the flip according to the magnetic sensor 180D. Further, according to the detected opening and closing state of the leather case or the opening and closing state of the flip cover, characteristics such as automatic unlocking of the flip cover are set.
  • the acceleration sensor 180E can detect the magnitude of the acceleration of the terminal device 100 in various directions (generally three axes).
  • the magnitude and direction of gravity can be detected when the terminal device 100 is stationary. It can also be used to identify the posture of terminal devices, and can be used in applications such as horizontal and vertical screen switching, pedometers, etc.
  • the terminal device 100 can measure the distance through infrared or laser. In some embodiments, when shooting a scene, the terminal device 100 can use the distance sensor 180F to measure the distance to achieve fast focusing.
  • Proximity light sensor 180G may include, for example, light emitting diodes (LEDs) and light detectors, such as photodiodes.
  • the light emitting diodes may be infrared light emitting diodes.
  • the terminal device 100 emits infrared light to the outside through the light emitting diode.
  • the terminal device 100 detects infrared reflected light from nearby objects using a photodiode. When sufficient reflected light is detected, it can be determined that there is an object near the terminal device 100 . When insufficient reflected light is detected, the terminal device 100 may determine that there is no object near the terminal device 100 .
  • the terminal device 100 can use the proximity light sensor 180G to detect that the user holds the terminal device 100 close to the ear to talk, so as to automatically turn off the screen to save power.
  • Proximity light sensor 180G can also be used in holster mode, pocket mode automatically unlocks and locks the screen.
  • the ambient light sensor 180L is used to sense ambient light brightness.
  • the terminal device 100 may determine the exposure time of the image according to the ambient light brightness sensed by the ambient light sensor 180L.
  • the terminal device 100 can adaptively adjust the brightness of the display screen 194 according to the perceived ambient light brightness.
  • the ambient light sensor 180L can also be used to automatically adjust the white balance when taking pictures.
  • the ambient light sensor 180L can also cooperate with the proximity light sensor 180G to detect whether the terminal device 100 is in a pocket, so as to prevent accidental touch.
  • the fingerprint sensor 180H is used to collect fingerprints.
  • the terminal device 100 can use the collected fingerprint characteristics to realize fingerprint unlocking, accessing application locks, taking photos with fingerprints, answering incoming calls with fingerprints, and the like.
  • the temperature sensor 180J is used to detect the temperature.
  • the terminal device 100 uses the temperature detected by the temperature sensor 180J to execute the temperature processing strategy. For example, when the temperature reported by the temperature sensor 180J exceeds a threshold value, the terminal device 100 reduces the performance of the processor located near the temperature sensor 180J, so as to reduce power consumption and implement thermal protection.
  • the terminal device 100 when the temperature is lower than another threshold, the terminal device 100 heats the battery 142 to avoid abnormal shutdown of the terminal device 100 caused by the low temperature.
  • the terminal device 100 boosts the output voltage of the battery 142 to avoid abnormal shutdown caused by low temperature.
  • Touch sensor 180K also called “touch device”.
  • the touch sensor 180K may be disposed on the display screen 194 , and the touch sensor 180K and the display screen 194 form a touch screen, also called a “touch screen”.
  • the touch sensor 180K is used to detect a touch operation on or near it.
  • the touch sensor can pass the detected touch operation to the application processor to determine the type of touch event.
  • Visual output related to touch operations may be provided through display screen 194 .
  • the touch sensor 180K may also be disposed on the surface of the terminal device 100 , which is different from the position where the display screen 194 is located.
  • the bone conduction sensor 180M can acquire vibration signals.
  • the bone conduction sensor 180M can acquire the vibration signal of the vibrating bone mass of the human voice.
  • the bone conduction sensor 180M can also contact the pulse of the human body and receive the blood pressure beating signal.
  • the bone conduction sensor 180M can also be disposed in the earphone, combined with the bone conduction earphone.
  • the audio module 170 can analyze the voice signal based on the vibration signal of the vocal vibration bone block obtained by the bone conduction sensor 180M, so as to realize the voice function.
  • the application processor can analyze the heart rate information based on the blood pressure beat signal obtained by the bone conduction sensor 180M, and realize the function of heart rate detection.
  • the keys 190 include a power-on key, a volume key, and the like. Keys 190 may be mechanical keys. It can also be a touch key.
  • the terminal device 100 may receive key input and generate key signal input related to user settings and function control of the terminal device 100 .
  • Motor 191 can generate vibrating cues.
  • the motor 191 can be used for vibrating alerts for incoming calls, and can also be used for touch vibration feedback.
  • touch operations acting on different applications can correspond to different vibration feedback effects.
  • the motor 191 can also correspond to different vibration feedback effects for touch operations on different areas of the display screen 194 .
  • Different application scenarios for example: time reminder, receiving information, alarm clock, games, etc.
  • the touch vibration feedback effect can also support customization.
  • the indicator 192 can be an indicator light, which can be used to indicate the charging state, the change of the power, and can also be used to indicate a message, a missed call, a notification, and the like.
  • the SIM card interface 195 is used to connect a SIM card.
  • the SIM card can be contacted and separated from the terminal device 100 by inserting into the SIM card interface 195 or pulling out from the SIM card interface 195 .
  • the terminal device 100 may support 1 or N3 SIM card interfaces, where N3 is a positive integer greater than 1.
  • the SIM card interface 195 can support Nano SIM card, Micro SIM card, SIM card and so on. Multiple cards can be inserted into the same SIM card interface 195 at the same time. The types of the plurality of cards may be the same or different.
  • the SIM card interface 195 can also be compatible with different types of SIM cards.
  • the SIM card interface 195 is also compatible with external memory cards.
  • the terminal device 100 interacts with the network through the SIM card to realize functions such as calls and data communication.
  • the terminal device 100 adopts an eSIM, that is, an embedded SIM card.
  • the eSIM card can be embedded in the terminal device 100 and cannot be separated from the terminal device 100 .
  • the software system of the terminal device 100 may adopt a layered architecture, an event-driven architecture, a microkernel architecture, a microservice architecture, or a cloud architecture.
  • the embodiments of the present invention take an Android system with a layered architecture as an example to exemplarily describe the software structure of the terminal device 100 .
  • FIG. 2 is a block diagram of a software structure of a terminal device 100 according to an embodiment of the present invention.
  • the layered architecture divides the software into several layers, and each layer has a clear role and division of labor. Layers communicate with each other through software interfaces.
  • the Android system is divided into four layers, which are, from top to bottom, an application layer, an application framework layer, an Android runtime (Android runtime) and a system library, and a kernel layer.
  • the application layer can include a series of application packages.
  • the application package can include applications such as camera, gallery, calendar, call, map, navigation, WLAN, Bluetooth, music, video, short message and so on.
  • the application framework layer provides an application programming interface (application programming interface, API) and a programming framework for applications in the application layer.
  • the application framework layer includes some predefined functions.
  • the application framework layer may include window managers, content providers, view systems, telephony managers, resource managers, notification managers, and the like.
  • a window manager is used to manage window programs.
  • the window manager can get the size of the display screen, determine whether there is a status bar, lock the screen, take screenshots, etc.
  • Content providers are used to store and retrieve data and make these data accessible to applications.
  • the data may include video, images, audio, calls made and received, browsing history and bookmarks, phone book, etc.
  • the view system includes visual controls, such as controls for displaying text, controls for displaying pictures, and so on. View systems can be used to build applications.
  • a display interface can consist of one or more views.
  • the display interface including the short message notification icon may include a view for displaying text and a view for displaying pictures.
  • the telephony manager is used to provide the communication function of the terminal device 100 .
  • the management of call status including connecting, hanging up, etc.).
  • the resource manager provides various resources for the application, such as localization strings, icons, pictures, layout files, video files and so on.
  • the notification manager enables applications to display notification information in the status bar, which can be used to convey notification-type messages, and can disappear automatically after a brief pause without user interaction. For example, the notification manager is used to notify download completion, message reminders, etc.
  • the notification manager can also display notifications in the status bar at the top of the system in the form of graphs or scroll bar text, such as notifications of applications running in the background, and notifications on the screen in the form of dialog windows. For example, text information is prompted in the status bar, a prompt sound is issued, the terminal device vibrates, and the indicator light flashes.
  • Android Runtime includes core libraries and a virtual machine. Android runtime is responsible for scheduling and management of the Android system.
  • the core library consists of two parts: one is the function functions that the java language needs to call, and the other is the core library of Android.
  • the application layer and the application framework layer run in virtual machines.
  • the virtual machine executes the java files of the application layer and the application framework layer as binary files.
  • the virtual machine is used to perform functions such as object lifecycle management, stack management, thread management, safety and exception management, and garbage collection.
  • a system library can include multiple functional modules. For example: surface manager (surface manager), media library (Media Libraries), 3D graphics processing library (eg: OpenGL ES), 2D graphics engine (eg: SGL), etc.
  • surface manager surface manager
  • media library Media Libraries
  • 3D graphics processing library eg: OpenGL ES
  • 2D graphics engine eg: SGL
  • the Surface Manager is used to manage the display subsystem and provides a fusion of 2D and 3D layers for multiple applications.
  • the media library supports playback and recording of a variety of commonly used audio and video formats, as well as still image files.
  • the media library can support a variety of audio and video encoding formats, such as: MPEG4, H.264, MP3, AAC, AMR, JPG, PNG, etc.
  • the 3D graphics processing library is used to implement 3D graphics drawing, image rendering, compositing, and layer processing.
  • 2D graphics engine is a drawing engine for 2D drawing.
  • the kernel layer is the layer between hardware and software.
  • the kernel layer includes at least display drivers, camera drivers, audio drivers, headphone drivers, and sensor drivers.
  • the workflow of the software and hardware of the terminal device 100 is exemplarily described below in conjunction with the audio capture and playback scene.
  • a corresponding hardware interrupt is sent to the kernel layer.
  • the kernel layer processes touch operations into raw input events (including touch coordinates, timestamps of touch operations, etc.). Raw input events are stored at the kernel layer.
  • the application framework layer obtains the original input event from the kernel layer, and identifies the control corresponding to the input event. Take the touch operation as a touch click operation, and the control corresponding to the click operation is an audio application icon control as an example, the audio application invokes the interface of the application framework layer to start the headphone control application, and then starts the headphone driver by calling the kernel layer, The audio signal is sent to the earphone, and the audio signal is played through the earphone 200 .
  • FIG. 3 shows a schematic diagram of an optional hardware structure of the earphone 200 .
  • the earphone 200 includes a left earphone and a right earphone.
  • the left earphone and the right earphone have similar structures.
  • the structures of the earphones include a first microphone 301 , a second microphone 302 and a third microphone 303 .
  • a processor 304 and a speaker 305 may also be included in the headset. It should be understood that the earphone described later may be interpreted as the left earphone, and may also be interpreted as the right earphone.
  • the first microphone 301 is used to collect the sound of the current external environment, and the first microphone 301 may also be referred to as a reference microphone.
  • the first microphone 301 is located outside the earphone, or the second microphone 301 is located outside the ear.
  • the second microphone 302 collects the sound of the internal environment of the user's ear canal.
  • the second microphone 302 may also be referred to as an error microphone.
  • the second microphone 302 is located inside the earphone and close to the ear canal.
  • the third microphone 303 is used to collect call signals.
  • the third microphone 303 may be located outside the earphone. When the user wears the earphone, the third microphone 303 is closer to the user's mouth than the first microphone 301 .
  • the first microphone 301 is used to collect the sound of the current external environment, which can be interpreted as the sound of the external environment where the headset user wears the headset, for example, on a train, the sound of the external environment is the surrounding environment of the user wearing the headset the sound of.
  • the first microphone 301 on the left earphone adopts the sound of the external environment of the left earphone.
  • the first microphone 301 on the right earphone collects the sound of the external environment of the right earphone.
  • the signal collected by the first microphone 301 (reference microphone) is called the first signal
  • the signal collected by the second microphone 302 (error microphone) is called the second signal.
  • the microphone involved in the embodiments of the present application may be an analog microphone or a digital microphone.
  • the microphone is an analog microphone, before filtering the signal collected by the microphone, the analog signal can be converted into a digital signal.
  • the first microphone and the second microphone are both digital microphones as an example for description, and the first signal and the second signal are both digital signals.
  • the processor 304 is used to process downlink audio signals and/or signals collected by the microphones (including the first microphone 301 , the second microphone 302 and the third microphone 303 ), such as performing ANC processing, HT processing, or AH processing.
  • the processor 304 may include a main control unit and a noise reduction processing unit.
  • the main control unit is used for the user to generate control commands for the operation of the earphone or to receive control commands from the terminal device.
  • the noise reduction processing unit is configured to perform ANC processing, HT processing or AH processing on downlink audio signals and signals collected by the microphones (including the first microphone 301 , the second microphone 302 and the third microphone 303 ) according to the control command.
  • the left and right earphones may also include memory for storing programs or instructions executed by the processor 304 .
  • the processor 304 performs ANC processing, HT processing, or AH processing according to programs or instructions stored in the memory.
  • the memory may include random access memory (RAM), flash memory, read-only memory (ROM), programmable read-only memory (PROM), erasable programmable read-only memory (erasable).
  • RAM random access memory
  • ROM read-only memory
  • PROM programmable read-only memory
  • PROM erasable programmable read-only memory
  • the main control unit for example, can be composed of an ARM processing chip, a central processing unit (CPU), a system on chip (SoC), a digital signal processing (DSP), or a micro controller (micro controller). unit, MCU) in one or more implementations.
  • the noise reduction processing unit may include, for example, a codec (coder-decoder, CODEC) chip or a high-fidelity (high-fidelity, HiFi) chip, or the like.
  • the codec is hardened with a filter, an equalizer (EQ), a dynamic range controller (DRC), a limiter (limiter), and a gain adjuster ( Gain), mixer (mixer), etc., are mainly used for filtering, mixing, gain adjustment and other processing of signals.
  • the noise reduction processing unit may further include a DSP, and the DSP may be used for processing such as scene detection, speech enhancement, and occlusion elimination.
  • the headset may further include a wireless communication unit for establishing a communication connection with the terminal device 200 through the wireless communication module 160 in the terminal device 100 .
  • the wireless communication unit can provide applications on the headset including wireless local area networks (WLAN) (such as wireless fidelity (Wi-Fi) networks), Bluetooth (bluetooth, BT), short-range wireless communication technology ( near field communication, NFC), infrared technology (infrared, IR) and other wireless communication solutions.
  • the wireless communication unit may be one or more devices integrating at least one communication processing module.
  • the wireless communication module 160 may be Bluetooth, the wireless communication unit is also Bluetooth, and the headset 200 and the terminal device 100 are connected through Bluetooth.
  • the ANC function when the ANC function is turned on, it means that the signal path of the ANC function is in an active state, and correspondingly, each functional module in the ANC pathway is also in an active state; the turning on of the HT function means that the signal path of the HT function is in an active state. In the activated state, correspondingly, each functional module in the HT pathway is also in the activated state.
  • both the ANC function and the HT function of the earphone are in an on state, it indicates that the signal pathway of the ANC function is in an active state, and the signal pathway of the HT function is also in an active state. That is, it can indicate the working state of the headset, and is not limited to specific operations at a certain moment or changes in functional controls.
  • the active noise reduction processing in the active noise reduction output path may include, but is not limited to: performing noise suppression using the inverted signal of the first signal collected by the reference microphone and the inverted signal of the second signal collected by the error microphone.
  • the active noise reduction output path includes an inverted signal of the first signal and an inverted signal of the second signal. It should be noted that the phase difference between the first signal and the inverted signal of the first signal is 180°.
  • the speaker outputs a signal obtained by superimposing the inverted signal of the first signal and the inverted signal of the second signal, so that the sound of the current external environment played by the speaker and the sound of the external environment actually heard by the ear are cancelled to achieve active denoising. Effect. Therefore, when the headset adopts the ANC mode, the headset user's perception of the sound of the current environment and the ambient sound inside the user's ear canal can be weakened.
  • filter compensation can be performed on the downlink audio signal.
  • the influence of the downlink audio signal can be removed when the inverted signal of the ambient sound is obtained.
  • the first filtering process and the third filtering process may be used.
  • the first filtering process may be a feed forward (FF) filtering process, which may be implemented by a feed forward filter.
  • the third filtering process may be a feedback (feedback, FB) filtering process, which may be implemented by a feedback filter.
  • FF filtering and FB filtering use a parallel processing architecture to enhance the noise reduction effect. The processing flow of the ANC will be described in detail later, and the description will not be repeated here.
  • the ambient sound transparent transmission processing in the outgoing path of ambient sound transmission may include, but is not limited to, including: performing a third filtering process on the first signal collected by the error microphone to achieve part of the active noise reduction function, and performing a second filtering process on the signal collected by the reference microphone as well as HT enhanced processing.
  • the second filtering process may be a transparent transmission (HT) filtering process, which may be implemented by a transparent transmission filter.
  • the audio signal played by the speaker is obtained according to the first signal and the second signal, so that after the audio signal played by the speaker, the user can hear the sound in the external environment through the earphone, which is compared with the sound of the external environment heard when the HT processing is not performed.
  • the sound intensity is better and the effect is better. Therefore, when the headset adopts the HT mode, the user's perception of the intensity of the sound in the environment where the user is currently located can be enhanced.
  • the processing flow of the HT will be described in detail later, and the description will not be repeated here.
  • the transparent transmission processing of ambient sound in the auditory enhancement output path may include, but is not limited to, using the signal collected by the error microphone to realize part of the active noise reduction function, performing first filtering processing on the signal collected by the reference microphone, and hearing enhancement processing to make the environment where the user is located.
  • the sound of the event in the sound is enhanced, and a second filtering process is performed on the signal collected by the reference microphone.
  • the output signal of the speaker is obtained according to the mixed signal of the event signal in the first signal and the inverted signal of the second signal. It should be noted that the phase difference between the second signal and the inverted signal of the second signal is 180°.
  • the speaker outputs the inverted signal of the second signal, the inverted signal of the first signal and the signal after the event signal in the first signal is superimposed, so that the signal output by the speaker and the sound in the environment actually heard by the ear are canceled to achieve the initiative.
  • the processing flow of the AH will be described in detail later, and the description will not be repeated here.
  • the downlink audio signal, the first signal and the second signal may be a frame of signals or signals of a period of time.
  • the downlink audio signal, the first signal and the second signal are all one-frame signals
  • the downlink audio signal, the first signal and the second signal belong to three signal streams respectively
  • the signal frame of the downlink audio signal and the signal of the first signal The frame and the signal frame of the second signal are in the same time period or overlap in time, etc.
  • functional processing such as ANC, HT, or AH
  • the functional processing is continuously performed for the signal flow where the downlink audio signal is located, the signal flow where the first signal is located, and the signal flow of the second signal.
  • the downlink audio signal sent by the terminal device 100 to the headset 200 may also be referred to as the first audio signal in the subsequent description.
  • the first audio signal may be a call signal, a music signal, or the like.
  • the signal collected by the reference microphone is called the first signal
  • the signal collected by the error microphone is called the second signal. Headphones are in ANC mode.
  • the downlink audio signals sent by the terminal device 100 to the left earphone and the right earphone in the earphone 200 may be the same signal or different signals.
  • the terminal device adopts stereo effect, and the terminal device 100 sends different downlink audio signals to the earphone 200 to realize the stereo effect.
  • the terminal device can also send the same downlink audio signal to the left earphone and the right earphone, and the left earphone and the right earphone adopt stereo processing, so as to achieve a stereo effect.
  • the left earphone or the right earphone may perform the process of FIG. 5A or FIG. 5B according to the user's control.
  • S501 Perform a first filtering process on a first signal collected by a reference microphone to obtain a first filtered signal.
  • the first filtered signal is simply referred to as signal A1 in FIG. 5B .
  • the first filtered signal is simply referred to as signal A2 in Figure 5B.
  • S503 Perform sound mixing processing on the first filtered signal and the first filtered signal to obtain a third audio signal.
  • the third audio signal is simply referred to as the signal A3, that is, the signal A3 is obtained by mixing the signal A1 and the signal A2.
  • the fourth audio signal is simply referred to as signal A4 in FIG. 5B .
  • S505. Perform sound mixing processing on the fourth audio signal and the first audio signal to obtain the second audio signal.
  • the speaker is responsible for playing the second audio signal.
  • the second audio signal is simply referred to as A5.
  • the signal output by the speaker is the fourth audio frequency that has not been mixed. Signal.
  • S502 and S505 need not be performed.
  • the first filtering process is an FF filtering process implemented by an FF filter as an example
  • the third filtering process is an FB filtering process implemented by an FB filter as an example.
  • the reference microphone in the earphone 200 picks up the first signal, which is input to the FF filter to obtain the signal A1 after FF filtering processing.
  • the error microphone picks up the second signal and inputs it to the subtractor.
  • the downlink audio signal is also input to the subtractor after filtering and compensating.
  • the subtracter removes the downlink audio signal after filtering and compensation included in the second signal, so as to eliminate the effect of the downlink audio signal. Affect gets signal A2.
  • the signal A3 is obtained after mixing the signal A1 and the signal A2 by the sound mixer, and the signal A4 is input into the FB filter, and the signal A4 is obtained after the FB filtering process. After the signal A4 is mixed with the downlink audio signal, the signal A5 is obtained, and the signal A5 is input to the speaker for playback.
  • the quality of the ANC effect may be determined by the processing intensity of the ANC processing.
  • the processing strength of the ANC processing depends on the FF filter coefficients used for FF filtering and/or the FB filter coefficients used for FB filtering.
  • the default FF filter coefficients in the ANC mode may be used.
  • the FF filter coefficient used when the ANC mode was selected last time can be used.
  • the earphone determines the FF filter coefficient used in the ANC mode according to the recognized scene.
  • the user indicates to the headset the FF filter coefficient used in the ANC mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the ANC mode as the target processing intensity through the UI control provided by the terminal device, and different processing intensities correspond to different FF filter coefficients.
  • the default FB filter coefficient in the ANC mode can be used.
  • the FB filter coefficient used when the ANC mode was selected last time can be used.
  • the earphone determines the adopted FB filter coefficient according to the recognized scene.
  • the user indicates to the headset the FF filter coefficient used in the ANC mode through the UI control provided by the terminal device, for example, the user selects the processing intensity in the ANC mode through the UI control provided by the terminal device as the target processing intensity, Different processing strengths correspond to different FB filter coefficients.
  • the acquisition of the FF filter coefficient and the FB filter coefficient in the ANC mode may be any combination of the above-mentioned methods.
  • the FF filter adopts the default filter coefficient in the ANC mode, and the FB filter coefficient is determined by the headset according to the recognized scene.
  • the FB filter adopts the default filter coefficient in the ANC mode, and the FF filter coefficient is determined by the user through the UI control provided by the terminal device.
  • the FB filter adopts the default filter coefficient in the ANC mode, and the FF filter coefficient is selected and indicated to the headset by the user through the UI control provided by the terminal device.
  • the determination of the processing intensity in the ANC mode will be described in detail later by using a specific example, and the description will not be repeated here.
  • FIG. 6A , FIG. 6B and FIG. 6C it is a schematic diagram of a process flow of transparent transmission of ambient sound.
  • the downlink audio signal sent by the terminal device 100 to the earphone 200 is described below by taking the first audio signal as an example.
  • the first audio signal may be a call signal, a music signal, or the like.
  • the signal collected by the reference microphone is called the first signal
  • the signal collected by the error microphone is called the second signal.
  • the left earphone or the right earphone in the earphone 200 may perform the process of FIG. 6A or FIG. 6B or FIG. 6C according to the user's control.
  • S601 Perform first signal processing on the first signal collected by the reference microphone to obtain a first processed signal.
  • the first processed signal is referred to as signal B1 in Figures 6B and 6C.
  • the first signal processing includes HT filtering.
  • S602. Perform sound mixing processing on the first processed signal and the first audio signal to obtain a fifth audio signal.
  • the fifth audio signal is referred to as signal B2 in Figures 6B and 6C.
  • the signal B2 is obtained by performing mixing processing on the signal B1 and the downlink audio signal (ie, the first audio signal).
  • the second filtered signal is referred to as signal B3 in Figures 6B and 6C. That is, the signal B3 is obtained by filtering out the signal B2 included in the second ambient signal.
  • the third filtered signal is referred to as signal B4 in Figures 6B and 6C. That is, the signal B4 is obtained by FB filtering the signal B3.
  • the first processed signal when the first processed signal is obtained by performing the first signal processing on the first signal collected by the reference microphone, it may be implemented in the following manner:
  • the HT filtering is performed on the first signal to obtain a second filtered signal.
  • the second filtered signal is referred to as signal B5.
  • the second signal processing is performed on the second filtered signal to obtain the second processed signal; the second signal processing may also be referred to as low-latency algorithm processing.
  • the low-latency algorithm processing includes one or more of de-blocking effect processing, noise reduction processing, wind noise reduction processing, gain adjustment processing or frequency response adjustment processing.
  • the HT filtering process may be implemented by a noise reduction processing unit, as shown in FIG. 6B .
  • the noise reduction processing unit of the earphone includes a CODEC.
  • the CODEC includes an HT filter, an FB filter, a subtractor, a first mixer, a second mixer, and a filter compensation unit.
  • FIG. 6B takes as an example that the noise reduction processing unit further includes a DSP.
  • the DSP can be used to perform low-latency algorithmic processing.
  • the reference microphone in the earphone 200 picks up the first signal, which is input to the HT filter to obtain the signal B5 after HT filter processing.
  • the signal B5 is input to the DSP, and the DSP performs low-latency algorithm processing on the signal B5 to obtain the signal B1.
  • the signal B1 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal B1 to obtain the signal B2.
  • After the signal B2 is processed by the filter compensation unit, it is input to the subtractor.
  • the subtractor is used to filter the signal B2 that has undergone filtering and compensation processing included in the second ambient signal picked up by the error microphone to obtain a signal B3.
  • the signal B3 is input to the FB filter, and the FB filter performs the FB filtering process on the signal B3 to obtain the signal B4.
  • Signal B4 is input to the second mixer.
  • the input of the second mixer also includes signal B2.
  • the second sound mixer performs sound mixing processing on the signal B2 and the signal B4 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
  • the HT filtering process can be implemented by DSP, as shown in FIG. 6C .
  • the DSP can be used to perform HT filtering processing as well as low-latency algorithm processing.
  • the noise reduction processing unit of the earphone includes an FB filter, a subtractor, a first sound mixer, a second sound mixer and a filter compensation unit.
  • the reference microphone in the earphone picks up the first signal and inputs it to the DSP.
  • the DSP performs HT filtering processing on the first signal, and performs low-latency algorithm processing to obtain a signal B1.
  • the signal B1 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal B1 to obtain the signal B2.
  • the signal B2 After the signal B2 is processed by the filter compensation unit, it is input to the subtractor.
  • the subtractor is used to filter the signal B2 included in the second signal picked up by the error microphone to obtain the signal B3.
  • the signal B3 is input to the FB filter, and the FB filter performs the FB filtering process on the signal B3 to obtain the signal B4.
  • Signal B4 is input to the second mixer.
  • the input of the second mixer also includes signal B2.
  • the second sound mixer performs sound mixing processing on the signal B2 and the signal B4 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
  • the low-latency algorithm processing includes de-blocking effect processing.
  • the principle of the occlusion effect is explained first.
  • Path 1 from the bone conduction to the periosteum, which is only a low-frequency signal.
  • Path 2 It propagates from the outside air to the periosteum and is perceived, and the signal contains low frequency signals and medium and high frequency signals.
  • the low-frequency signal and the medium-high frequency signal are superimposed, the low-frequency signal is too sufficient, and the low-frequency signal cannot be emitted when the earphone is worn, resulting in low-frequency chaos inside the ear, resulting in an occlusion effect.
  • the signal B5 obtained after the HT filtering process is processed by the de-blocking effect, and the specific method may be as follows.
  • Mode 1 is shown in FIG. 7 .
  • a first voice harmonic signal matching the bone conduction signal in a voice harmonic set where the voice harmonic set includes a plurality of voice harmonic signals.
  • the plurality of speech harmonic signals included in the speech harmonic set correspond to different frequencies. Specifically, the frequency of the bone conduction signal can be determined, and the first voice harmonic signal is determined from the set of voice harmonics according to the frequency of the bone conduction signal.
  • the speech harmonic signal may also be referred to as the speech harmonic component.
  • S702 remove the first voice harmonic signal in the signal B5 obtained after the HT filtering process.
  • the signal C1 is obtained by removing the first voice harmonic signal in the signal B5 obtained after the HT filtering process. Since the voice of a person collected by a general bone conduction sensor is generally a low-frequency harmonic component, S702 is to remove the low-frequency harmonic component in the signal B5, thereby obtaining a signal C1 that does not include the low-frequency harmonic component.
  • S703 Amplify the high frequency components in the signal B5 from which the first voice harmonic signal is removed. That is, the high frequency components of the signal C1 are amplified.
  • the first voice harmonic signal matching the bone conduction signal can be determined, that is, the bone conduction sensor can detect the bone conduction signal, that is, the earphone wearer is currently speaking, such as speaking or singing.
  • the signal obtained by boosting the high-frequency components on the basis of the signal C1 includes only mid- and high-frequency components, so that the signal heard by the earphone wearer has relatively no occlusion effect.
  • the set of speech harmonics may be pre-stored in the headset.
  • the set of speech harmonics may be obtained in an offline manner or in an online manner.
  • the bone conduction signals of multiple people can be collected through the bone conduction sensor, and the following processing is performed on the bone conduction signals of each person in the bone conduction signals of the multiple people:
  • the first bone conduction signal is transformed into a frequency domain signal by FFT; the fundamental frequency signal in the frequency domain signal is determined by means of a pilot frequency to find the fundamental frequency; the harmonic component of the bone conduction signal is determined according to the fundamental frequency signal, Thereby, the mapping relationship between the frequency of the bone conduction signal and the harmonic component is obtained to obtain a set of speech harmonics.
  • the voice harmonic set may include mapping relationships between different frequencies and harmonic components.
  • the second bone conduction signal collected within a set period of time can be obtained through the bone conduction sensor in the headset.
  • the set period of time there may be multiple people using the headset, or only one person may use the headset, that is, the user;
  • the fundamental frequency signal in the frequency domain signal is determined by means of a pilot frequency to find the fundamental frequency. If multiple people use the headset within the set time period, it may be possible to determine multiple fundamental frequency signals corresponding to multiple different time periods within the set time period.
  • Multiple harmonic components of the bone conduction signal may be determined according to multiple fundamental frequency signals, so as to obtain a mapping relationship between frequencies and harmonic components to obtain a voice harmonic set.
  • the voice harmonic set may include mapping relationships between different frequencies and different harmonic components.
  • adaptive filtering can be performed on the signal B5 obtained after the HT filtering process to remove the low frequency component in the signal B5 to obtain the signal C1, that is, the signal of the earphone wearer's voice in the signal B5 is removed.
  • the high frequency components in the third filtered signal from which the low frequency components are removed are amplified, that is, the high frequency components in the signal C1 are amplified.
  • the signal obtained by boosting the high-frequency components on the basis of the signal C1 includes only mid- and high-frequency components, so that the signal heard by the earphone wearer has relatively no occlusion effect.
  • the quality of the HT effect can be determined by the treatment intensity of the HT treatment.
  • the processing strength of the HT processing depends on the HT filter coefficients used for the HT filtering and/or the FB filter coefficients used for the FB filtering.
  • the default HT filter coefficients in the HT mode may be used.
  • the HT filter coefficient used when the HT mode was selected last time can be used.
  • the earphone determines the HT filter coefficient used in the HT mode according to the recognized scene.
  • the user indicates to the headset the HT filter coefficient used in the HT mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the HT mode as the target processing intensity through the UI control provided by the terminal device, and different processing intensities correspond to different HT filter coefficients.
  • the FB filter coefficient in one way, the default FB filter coefficient in the HT mode can be used.
  • the FB filter coefficient used when the HT mode was selected last time can be used.
  • the earphone determines the adopted FB filter coefficient according to the recognized scene.
  • the user indicates to the headset the HT filter coefficient used in the HT mode through the UI control provided by the terminal device, for example, the user selects the processing intensity in the HT mode through the UI control provided by the terminal device as the target processing intensity, Different processing strengths correspond to different FB filter coefficients.
  • the acquisition of the HT filter coefficients and the FB filter coefficients in the HT mode may be any combination of the methods provided above.
  • the downlink audio signal sent by the terminal device 100 to the earphone 110 is referred to as the first audio signal as an example in the subsequent description.
  • the first audio signal may be a call signal, a music signal, a prompt tone, or the like.
  • the signal collected by the reference microphone is called the first signal
  • the signal collected by the error microphone is called the second signal.
  • the left earphone or the right earphone in the earphone 200 may perform the process of FIG. 8A or FIG. 8B or FIG. 8C according to the user's control.
  • S801 Perform HT filtering on the first signal collected by the reference microphone to obtain a second filtered signal (signal C1).
  • the second filtered signal is referred to as signal C1.
  • the first filtered signal is simply referred to as signal C3.
  • Step S804 Perform sound mixing processing on the filtered enhanced signal and the first audio signal to obtain a sixth audio signal.
  • the sixth audio signal is simply referred to as signal C4.
  • Step S804 is to mix the signal C2 and the downlink audio signal to obtain the signal C4.
  • Step S805 filtering the sixth audio signal included in the second signal to obtain a fourth filtered signal.
  • the fourth filtered signal is simply referred to as signal C5.
  • Step S805 is to filter out the signal C4 included in the second environment signal to obtain the signal C5.
  • step S805 when step S805 is performed, the signal C4 may be filtered and compensated first to obtain a compensated signal, and then the compensated signal included in the second signal may be filtered to obtain C5.
  • Step S806 is to perform FB filtering on the signal C5 to obtain the signal C6.
  • Step S807 performing mixing processing on the fifth filtered signal, the sixth audio signal and the first filtered signal to obtain the second audio signal.
  • Step S806 is to mix the signal C6, the signal C4 and the signal C3 to obtain the second audio signal.
  • Mode 1 is shown in FIG. 9 .
  • the manner of performing de-blocking effect processing on the signal C1 may be the same as that for performing the de-blocking effect processing on the signal B5. For details, refer to manners 1 and 2 in scenario 2, which will not be repeated here.
  • noise reduction processing is performed on the signal processed by the de-occlusion effect, and the noise reduction processing includes artificial intelligence AI noise reduction processing and/or wind noise reduction processing; in FIG. 9, the noise reduction processing includes AI noise reduction processing and wind noise reduction processing. Take processing as an example.
  • a gain amplification process is performed on the signal obtained after the wind noise processing, and a feasible way is to directly amplify the signal obtained after the wind noise processing.
  • the method of directly amplifying the signal not only amplifies the external signal, but also amplifies the voice of the wearer.
  • the embodiment of the present application provides a gain amplification processing method, which only amplifies the external signal, and does not amplify the voice signal speaking to the wearer. For example, as shown in FIG. 10 , when the signal obtained after the noise reduction processing is subjected to gain amplification processing, it can be implemented in the following manner.
  • the voice signal of the wearer's speech is conducted to the periosteum through the bone, and the voice signal is concentrated in the low frequency, which is recorded as the bone conduction signal D1.
  • the bone conduction signal D1 is acquired by the bone conduction sensor.
  • the bone conduction signal D1 is harmonically extended to obtain a harmonically extended signal, for example, the harmonically extended signal is called D2.
  • the harmonic extension may be a harmonic enhancement method or a method in which the harmonics of the bone conduction signal D1 are directly upwardly extended.
  • the signal obtained by the noise reduction process is referred to as signal D3.
  • the signal D4 is obtained by amplifying the signal D3 by using the first gain coefficient.
  • the amplification processing here can directly amplify the signal.
  • the first filter coefficient is determined based on the first gain coefficient.
  • the first gain coefficient gain to adjust the strength of the adaptive filtering can also be called the first filter coefficient, that is, the number of dB used for signal amplification by the first gain coefficient is the same as the number of dB filtered out by the adaptive filter, so that the wearer's voice The signal can remain balanced, neither amplified nor reduced.
  • the second method is shown in FIG. 11 .
  • the manner of performing de-blocking effect processing on the signal C1 may be the same as that for performing the de-blocking effect processing on the signal B5. For details, refer to manners 1 and 2 in scenario 2, which will not be repeated here.
  • S1102 Perform audio event detection on the deblocked signal to obtain an audio event signal (which may be referred to as an event signal for short) in the deblocked signal. Audio event signals, such as station announcements, whistles, etc.
  • S1103 Perform gain amplification on the audio event signal in the deblocked signal.
  • the audio event signal in the de-blocked signal such as station announcement sound, whistle sound, etc.
  • the method of performing gain amplification processing on the audio event signal in the de-blocked signal may be the same as when performing gain amplification processing on the signal obtained after the noise reduction processing, and will not be repeated here.
  • the noise reduction processing unit includes a CODEC and a DSP.
  • the CODEC of the earphone includes an HT filter, a FB filter, a FF filter, a subtractor, a first mixer, a second mixer, and a filter compensation unit.
  • the HT filtering process is implemented by the CODEC.
  • DSP can be used to perform enhancement processing.
  • the reference microphone in the earphone 110 picks up the first signal, which is input to the HT filter to obtain the signal C1 after HT filter processing.
  • the signal C1 is input to the DSP, and the DSP performs enhancement processing on the signal C1 to obtain the signal C2.
  • the signal C2 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal C2 to obtain the signal C4.
  • the signal C4 is processed by the filter compensation unit, it is input to the subtractor.
  • the subtractor is used for filtering the filtered and compensated signal C4 included in the second ambient signal picked up by the error microphone to obtain the signal C5.
  • the signal C5 is input to the FB filter, and the FB filter performs the FB filtering process on the signal C5 to obtain the signal C6.
  • Signal C6 is input to the second mixer.
  • the input of the second mixer also includes signal C4 and signal C3.
  • the second audio mixer performs audio mixing processing on the signal C3, the signal C4 and the signal C6 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
  • the noise reduction processing unit includes a CODEC and a DSP as an example.
  • the DSP can be used to perform HT filtering processing as well as enhancement processing.
  • the CODEC of the earphone includes a FB filter, an FF filter, a subtractor, a first mixer, a second mixer, and a filter compensation unit.
  • the reference microphone in the earphone 110 picks up the first signal and inputs it to the DSP.
  • the DSP performs HT filtering processing on the first signal to obtain a signal C1.
  • the DSP performs enhancement processing on the signal C1 to obtain a signal C2.
  • the signal C2 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal C2 to obtain the signal C4.
  • the signal C4 is processed by the filter compensation unit, it is input to the subtractor.
  • the subtractor is used for filtering the filtered and compensated signal C4 included in the second ambient signal picked up by the error microphone to obtain the signal C5.
  • the signal C5 is input to the FB filter, and the FB filter performs the FB filtering process on the signal C5 to obtain the signal C6.
  • Signal C6 is input to the second mixer.
  • the input of the second mixer also includes signal C4 and signal C3.
  • the second audio mixer performs audio mixing processing on the signal C3, the signal C4 and the signal C6 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
  • the quality of the AH effect can be determined by the treatment intensity of the AH treatment.
  • the processing strength of the AH processing depends on at least one of the HT filter coefficient, the FB filter coefficient, or the FF filter coefficient.
  • the default FF filter coefficient in the AH mode can be used.
  • the FF filter coefficient used when the AH mode was selected last time can be used.
  • the earphone determines the FF filter coefficient used in the AH mode according to the recognized scene.
  • the user indicates to the headset the FF filter coefficient used in the AH mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the AH mode as the target processing intensity through a UI control provided by the terminal device, and different processing intensities correspond to different FF filter coefficients.
  • the default HT filter coefficients in the AH mode may be used.
  • the HT filter coefficient used when the AH mode was selected last time can be used.
  • the earphone determines the HT filter coefficient used in the AH mode according to the recognized scene.
  • the user indicates to the headset the HT filter coefficients used in the AH mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the AH mode as the target processing intensity through the UI control provided by the terminal device, and different processing intensities correspond to different HT filter coefficients.
  • the FB filter coefficient in one way, the default FB filter coefficient in the AH mode can be used. In another way, the FB filter coefficient used when the AH mode was selected last time can be used.
  • the earphone determines the adopted FB filter coefficient according to the recognized scene.
  • the user indicates to the headset the HT filter coefficient used in the AH mode through the UI control provided by the terminal device, for example, the user selects the processing intensity in the AH mode through the UI control provided by the terminal device as the target processing intensity, Different processing strengths correspond to different FB filter coefficients.
  • the acquisition of the HT filter coefficient, the FB filter coefficient or the FF filter coefficient in the AH mode may be any combination of the above-mentioned methods.
  • the determination of the processing mode adopted by the earphone 200 can be selected by the user through the UI control on the terminal device 100 and indicated to the earphone, or the terminal device can determine and indicate to the earphone according to the adaptively recognized scene, Alternatively, it can also be determined by the headset according to adaptive scene recognition.
  • Example 1 a single control controls the left and right earphones.
  • the terminal device 100 provides a control interface for the user to select the processing mode of the earphone 200 (including the left earphone and the right earphone) according to requirements: empty mode, ANC mode, HT mode or AH mode. No processing is done in empty mode.
  • the processing modes of the earphones for the user to select on the control interface are all processing modes supported by the earphones.
  • the left earphone and the right earphone have the same processing function, or support the same processing mode.
  • the left earphone supports AHA
  • the right earphone also supports AHA.
  • the terminal device is installed with an earphone application adapted to the earphone 200, and during the adaptation process, the processing function of the earphone can be known.
  • the terminal device can determine the processing function possessed by the headset according to the function parameters.
  • the control interface includes user interface (user interface, UI) controls.
  • the UI control is used for the user to select the processing mode of the headset 200 , and for the convenience of distinction, the UI control used for the user to select the processing mode of the headset is called a selection control.
  • the processing mode includes at least two of ANC mode, HT mode or AH mode.
  • the terminal device 100 sends control signaling 1 to the left earphone and the right earphone respectively in response to the user selecting the target mode from the processing modes supported by the earphone through the selection control. Control signaling 1 carries the target mode.
  • the selection controls can also be used to select the processing intensity in target mode.
  • the shape of the selection control can be a circle, a bar or other shapes.
  • the selection controls may include a first control and a second control. Any two different positions of the second control on the first control correspond to different processing modes of the headset or two different positions of the second control on the first control correspond to different processing intensities in the same processing mode of the headset.
  • the user selects different processing modes and controls the processing intensity by moving the position of the second control representing the user's selection on the first control on the display screen.
  • control of the processing modes of the left earphone and the right earphone is implemented through an earphone application (appliaction, APP).
  • the terminal device 100 includes a headset control application for controlling the headset, which is referred to as a headset application.
  • a headset application For example, refer to the main interface of the terminal device shown in FIG. 12A .
  • the terminal device can respond to the user's operation of clicking the icon 001, start the headset application, and display the control of the headset application on the display screen interface, or it is interpreted that when the headset application is started, the control interface of the headset application pops up.
  • the name of the application can be called Audio Assistant, and this control function can also be integrated in the settings options in the terminal system.
  • the first control in the annular selection control in FIG. 12B includes three circular arc segments, corresponding to the ANC mode, the HT mode, and the AH mode, respectively. If the second control is located in the arc segment of the ANC mode, it is determined to be the ANC mode, and different positions of the second control located in the arc segment of the ANC mode correspond to different processing intensities in the ANC mode.
  • the HT mode is determined, and different positions of the second control located in the arc segment of the HT mode correspond to different processing intensities in the HT mode.
  • the second control is located in the arc segment of the AH mode, then it is determined to be the AH mode, and the different positions of the second control located in the arc segment of the AH mode correspond to different processing intensities in the AH mode.
  • the strongest noise reduction position of ANC is connected to the position of the weakest HT transparent transmission strength, and the auditory effect can be smoothly transitioned; can also transition smoothly.
  • the highlighted black dots on the ring represent the second control for the user to select the processing intensity.
  • the user can select different processing modes and control the processing intensity by moving the position of the black dots on the circumference.
  • the terminal device 100 (such as a processor) responds to operation 1 performed by the user on the control interface, for example, operation 1 is generated when the user moves the position of the second control on the display screen representing the second control selected by the user.
  • the terminal device 100 sends a control instruction 1 to the left earphone and the right earphone respectively, and the control instruction 1 is used to indicate the target mode and the target processing intensity.
  • the target mode is the ANC mode.
  • control instruction 1 may include an ANC identifier and a parameter value indicating a target processing intensity when performing ANC processing.
  • the FB filter coefficients and/or the FF filter coefficients corresponding to different processing intensities are different.
  • the control command 1 includes radians.
  • the corresponding processing mode can be determined according to the range in which the radian is located. Different radian values correspond to the processing strength in this processing mode. Referring to Fig. 12B, the processing mode corresponding to (0, 180] is the ANC mode, the processing mode corresponding to (180, 270] is the HT mode, and the processing mode corresponding to (270, 360] is the AH mode.
  • the left earphone and the right earphone can be Including the mapping relationship between the radian range and the processing mode, and the mapping relationship between the radian value and the filter coefficient. Taking the ANC mode as an example, the FB filter coefficient and the FF filter coefficient corresponding to different radian values are different.
  • the user can touch the black dot in the disc to rotate clockwise from 0 degrees to 360 degrees.
  • the corresponding FF filter coefficient and FB filter coefficient at 0 degrees make the ANC noise reduction effect the strongest, that is, the user
  • the perceived sound of the user's current environment and the ambient sound inside the user's ear canal are weaker.
  • the FF filter coefficient and the FB filter coefficient change, so that the active noise reduction effect gradually weakens, and the active noise reduction effect is the weakest when it reaches 180 degrees, similar to the headphones without any noise reduction processing.
  • From 180 degrees to 270 degrees is to control the transparent transmission of the ambient sound.
  • the corresponding HT filter coefficients and FB filter coefficients at 180 degrees make The transparent transmission effect of ambient sound is the weakest, that is, the lower the sound intensity of the user's current environment perceived by the user, is similar to the use of empty mode after wearing headphones.
  • the HT filter coefficient and the FB filter coefficient change, which makes the transparent transmission effect of ambient sound stronger and stronger.
  • 270 degrees to 360 degrees is used to control the hearing enhancement, that is, the user touches the black dot in the disc
  • the corresponding FF filter coefficient, HT filter coefficient and FB filter coefficient at 180 degrees make the hearing enhancement effect the weakest, and the user perceives The sound of the event included in the sound of the environment where the user is currently located is weaker.
  • the FF filter coefficient, HT filter coefficient and FB filter coefficient change, so that the auditory enhancement effect becomes stronger and stronger, that is, the event signal that the user wants to hear becomes stronger and stronger to assist hearing.
  • the terminal device 100 sends a control command 1 to the left earphone and the right earphone respectively through Bluetooth in response to the user's operation 1 .
  • the control command 1 may include the ANC identification and the parameter value of the target processing strength.
  • the operations performed after the left earphone and the right earphone receive the control command 1 are similar, and the processing of the left earphone is taken as an example in the subsequent description.
  • the main control unit of the left earphone acquires the FF filter coefficient and the FB filter coefficient processed by the ANC from the coefficient library according to the ANC identifier and the target processing strength.
  • the coefficient library includes the mapping relationship shown in Table 1.
  • Table 1 is only an example, and constitutes a specific definition of the mapping relationship.
  • the parameter value of the target processing intensity is intensity 1
  • the main control unit of the left earphone obtains the FF filter coefficient corresponding to intensity 1 according to Table 1 as the coefficient FF1
  • the FB filter coefficient is the coefficient FB1
  • the main control unit controls the FF filter to use the coefficient FF1
  • a signal A1 is obtained by performing FF filtering processing on the first signal collected by the reference microphone.
  • the main control unit controls the FB filter to use the coefficient FB1 to perform FB filtering processing on the signal A3 to obtain the second audio signal.
  • the main control unit writes the coefficient FF1 and the coefficient FB1 into the AHA core, so that the AHA core executes the steps of S501-S504 to obtain the second audio signal.
  • the terminal device 100 sends a control instruction 1 to the left earphone and the right earphone respectively through Bluetooth in response to the user's operation 1 .
  • the control instruction 1 may include an HT identifier and a target processing intensity, where the target processing intensity is used to indicate the processing intensity when the HT processing is performed.
  • the operations performed after the left earphone and the right earphone receive the control command 1 are similar, and the processing of the left earphone is taken as an example in the subsequent description.
  • the main control unit of the left earphone acquires the HT filter coefficient and/or the FB filter coefficient of the HT processing from the coefficient library according to the HT identifier and the target processing strength.
  • the value of the target processing intensity is intensity 5
  • the main control unit of the left earphone obtains the HT filter coefficient corresponding to intensity 5 according to Table 1 as the coefficient HT1
  • the FB filter coefficient is the coefficient FB5
  • the main control unit controls the HT filter.
  • the controller uses the coefficient HT1 to perform HT filtering processing on the first signal collected by the reference microphone.
  • the main control unit controls the FB filter to use the coefficient FB5 to perform FB filtering processing on the signal B3.
  • the main control unit writes the coefficient HT1 and the coefficient FB5 into the AHA core, so that the AHA core executes the steps of S601-S605 to obtain the second audio signal.
  • the terminal device 100 sends a control instruction 1 to the left earphone and the right earphone respectively through Bluetooth in response to the user's operation 1 .
  • the control instruction 1 may include the AH identifier and the parameter value of the target processing strength.
  • the operations performed after the left earphone and the right earphone receive the control command 1 are similar, and the processing of the left earphone is taken as an example in the subsequent description.
  • the main control unit of the left earphone obtains the HT filter coefficient, FF filter coefficient and FB filter coefficient of the AH process from the coefficient library according to the HT identifier and the target processing strength.
  • the value of the target processing strength is indication 7
  • the main control unit of the left earphone obtains the HT filter coefficient corresponding to indication 7 according to Table 1 as the coefficient HT3
  • the FB filter coefficient is the coefficient FB7
  • the FF filter coefficient is the coefficient FF5.
  • the main control unit controls the HT filter to perform HT filter processing on the first signal collected by the reference microphone using the coefficient HT3.
  • the main control unit controls the FB filter to use the coefficient FB7 to perform FB filtering processing on the signal C5.
  • the main control unit controls the FF filter to use the coefficient FF5 to perform FF filtering processing on the first signal.
  • the main control unit writes the coefficient HT3, the coefficient FB7 and the coefficient FF5 into the AHA core, so that the AHA core executes the steps of S801-S807 to obtain the second audio signal.
  • the selection control may take the form of a bar.
  • the selection controls include a first control and a second control.
  • the bar of the first control may be divided into a plurality of bar segments, which are divided according to the number of processing modes supported by the headset.
  • the second control indicates different processing modes in different bar segments of the first control.
  • the second control indicates different processing intensities of the same processing mode at different positions of the same bar segment of the first control.
  • the bar of the first control includes three bar segments.
  • the user can touch the black bar to slide left and right.
  • the black bar is at the K1 position
  • the corresponding FF filter coefficient and FB filter coefficient make the ANC noise reduction effect the strongest.
  • the FF filter coefficient and FB filter coefficient are the strongest.
  • the filter coefficient changes, so that the active noise reduction effect gradually weakens.
  • the K2 position is reached, the active noise reduction effect is the weakest, similar to the headphones without any noise reduction processing.
  • the area between K2 and K3 is the part that controls the transparent transmission of ambient sound.
  • Position K3 to position K4 is used to control auditory enhancement, that is, the user moves from position K3 to position K4 by touching the black bar, and when the black bar moves to position K3, the corresponding FF filter coefficient, HT filter coefficient and FB filter The coefficient makes the hearing enhancement effect the weakest. Moving from position K3 to position K4, the FF filter coefficient, HT filter coefficient and FB filter coefficient change, making the hearing enhancement effect stronger and stronger, that is, the voice signal that the user wants to hear is getting stronger and stronger. , to assist hearing.
  • the selection controls in (a) include buttons corresponding to different processing modes, including an ANC button, an HT button, and an AH button.
  • the terminal device 100 displays the display interface shown in FIG. 12G(b) in response to the user's operation of clicking the ANC button.
  • the display interface of (b) includes controls 002 for selecting treatment intensity. Users can touch the black bar and slide up and down to determine the processing intensity of ANC, that is, to select the corresponding FF filter coefficient and FB filter coefficient. Black bars slide in the area of L1-L2. When the black bar is located at the L1 position, the corresponding FF filter coefficient and FB filter coefficient make the ANC noise reduction effect the strongest. After sliding down, the FF filter coefficient and the FB filter coefficient change, so that the active noise reduction effect gradually weakens. When it reaches the L2 position Active noise reduction has the weakest effect, similar to wearing headphones without any noise reduction.
  • the headset APP when the headset 200 establishes a connection with the terminal device, the headset APP can be triggered to start, and a control interface including selection controls is displayed, such as the control interface illustrated in FIG. 12A , FIG. 12B , FIG. 12F or FIG. 12G .
  • the interface displayed by the terminal device is interface 1
  • the terminal device recognizes that the headset 200 is connected to the terminal device, and can jump from interface 1 to the control interface.
  • the headset APP can be triggered to start, that is, a control interface including selection controls is displayed, such as FIG. 12A , FIG. 12B , FIG. 12C or Figure 12D illustrates the display interface.
  • a control interface including selection controls is displayed, such as FIG. 12A , FIG. 12B , FIG. 12C or Figure 12D illustrates the display interface.
  • the terminal device triggers the headset to play audio it may be that after the terminal device establishes a connection with the headset, a song is played, and a control interface including selection controls may be displayed.
  • a control interface including selection controls can be displayed.
  • the identified scene type of the current external environment is the target scene, and the target scene adaptation needs to adjust all settings.
  • prompt information can be displayed, and the prompt information is used to prompt the user whether to adjust the processing mode of the headset.
  • a control interface including selection controls may be displayed, such as the control interface illustrated in Figures 12A, 12B, 12C, or 12D.
  • FIG. 12E takes the example of displaying the control interface of FIG. 12A as an example.
  • the terminal device recognizes that the current scene the user is in is a noisy scene.
  • the user may need to start the processing mode, so as to display selection prompt information (such as in the form of a prompt box) to prompt the user whether to adjust the processing mode of the headset .
  • selection prompt information such as in the form of a prompt box
  • the terminal device recognizes that the scene type of the external environment is a noisy scene.
  • the user may need to start the processing mode, so as to display a prompt box to prompt the user whether to adjust the processing mode of the headset.
  • the scene types that trigger the display of the prompt box may include: noisy scene, terminal building scene, train station scene, bus station scene, road scene, and the like.
  • the strength of the identified signal reaches a set threshold, it is considered a noisy scene.
  • the specific sound of an aircraft announcing the station is recognized, and the terminal building scene is determined to be recognized.
  • the sound of the train time notification is recognized, and the train station scene is determined to be recognized; for another example, the bus ticket broadcast is recognized, and the bus station scene is determined to be recognized.
  • the ticking of a signal light or the whistle of a car is recognized, and the road scene is determined to be recognized.
  • a control interface including selection controls is displayed according to the identified scene where the user is currently located.
  • Example 2 dual controls to control left and right earphones
  • the terminal device 100 provides a control interface for the user to select the processing modes of the left earphone and the right earphone respectively according to requirements, and the processing modes of the left earphone and the right earphone may be different.
  • the left earphone selects ANC mode
  • the right earphone selects HT mode.
  • the control interface includes selection controls for the left earphone and selection controls for the right earphone.
  • the selection control of the left earphone is called the first selection control
  • the selection control of the right earphone is called the second selection control.
  • the first selection control is used for the user to select the processing mode of the left earphone
  • the second selection control is used for the user to select the processing mode of the right earphone.
  • the shape of the first selection control and the second selection control may be a circular ring, a bar or other shapes.
  • the shapes of the first selection control and the second selection control may be the same or different.
  • the user realizes the selection of different processing modes and the control of processing intensity by moving the position of the control representing the user's selection on the display screen.
  • the shapes of the controls adopted by the left earphone and the right earphone may refer to the description in Example 1, and details are not repeated here.
  • both the first selection control and the second selection control include a first control and a second control.
  • Two different positions of the second control on the first control correspond to different processing modes, or, two different positions of the second control on the first control correspond to different processing intensities in the same processing mode. See the control interface shown in Figure 13 as an example.
  • the user can select the different processing modes and control the processing intensity implemented by the left earphone through the position of the second control (black dot) in the first selection control of the left earphone moving on the circumference of the first control.
  • the user can select the different processing modes implemented by the right earphone and control the processing intensity by moving the position of the second control on the first control in the second selection control of the right earphone.
  • the user can select different processing modes for the left and right earphones, or the same processing intensity in the same processing mode, or different processing intensities in the same processing mode, so as to match the ear differences or match the needs of different applications. need.
  • Example 2 the manner of triggering the display of the control interface including the first selection control and the second selection control refers to the description of Example 1, and details are not repeated here.
  • Example 3 terminal device smart scene detection.
  • the terminal device identifies the scene the user is currently in, and in different scenes, the headset adopts different processing modes.
  • the terminal device recognizes that the scene type indication of the current external environment is the first scene, it determines the target mode corresponding to the first scene in the processing mode of the earphone, and sends control signaling 2 to the left earphone and the right earphone respectively.
  • Control signaling 2 is used to indicate the target mode. Different target modes correspond to different scene types.
  • the terminal device determines which function to instruct the earphone to perform according to the identified scene. Correspond the functions of AHA to the scene type, and select the most suitable function for the scene type, so that users can automatically feel the most needed effect.
  • the scene types may include walking scene, running scene, quiet scene, multi-person talking scene, cafe scene, subway scene, train scene, car scene, waiting room scene, conversation scene, office scene, outdoor scene, driving Scene, windy scene, airplane scene, siren scene, whistle scene, cry scene, etc.
  • AI models can be built offline and stored on terminal devices. For example, a microphone on a terminal device records a large amount of noise and sensor data and/or video processing unit (vpu) data in different scenarios, and manually labels the data for its corresponding scenarios.
  • vpu noise and sensor data and/or video processing unit
  • initialize and build an AI model which can be one of convolutional neural networks (CNN)/deep neural networks (DNN)/long short-term memory (LSTM) , can also be a mixture of different models.
  • CNN convolutional neural networks
  • DNN deep neural networks
  • LSTM long short-term memory
  • use the labeled data for model training to obtain the corresponding AI model.
  • the processing modes suitable for different scene types are exemplified.
  • the processing mode corresponding to the scene type is indicated in brackets.
  • the noise of the plane is relatively loud when flying, so it is suitable to use the ANC mode.
  • the HT mode For another example, in walking scenes, running scenes and quiet scenes, it is suitable to use the HT mode, and you can hear the sound of sudden events. For another example, in the cafe scene, the user needs to be quiet, and the ANC mode can be used. For another example, in a light music scene, the HT mode can be used. For another example, the alarm sound scene (AH), the whistle sound scene (AH), and the crying sound scene (AH) are used to hear the preset sound, and the AH mode is suitable.
  • the AH mode is suitable.
  • the terminal device 100 when the terminal device 100 recognizes that the scene type of the current external environment is an airplane scene, it can send control signaling 2 to the headset, and the control signaling 2 is used to indicate that the headset needs to implement the ANC function, that is, to indicate Headphones are in ANC mode. After the left earphone and the right earphone receive the control signaling 2, the processes of S501-S504 are performed respectively.
  • the terminal device 100 when the terminal device 100 recognizes that the scene type of the current external environment is a walking scene, it can send control signaling 2 to the headset, and the control signaling 2 is used to indicate that the headset needs to implement the HT function, that is, the headset Use HT mode. After the left earphone and the right earphone receive the control signaling 2, the processes of S601-S605 are performed respectively.
  • the terminal device 100 when the terminal device 100 recognizes that the scene type of the current external environment is a train station scene, it can send control signaling 2 to the headset, and the control signaling 2 is used to indicate that the headset needs to implement the AH function , that is, the headset adopts AH mode. After the left earphone and the right earphone receive the control signaling 2, the processes of S801-S807 are performed respectively.
  • the terminal device After the headset is connected to the terminal device, the terminal device starts scene detection. After the detection of the terminal device is completed, the detection result can also be displayed to the user, so that the user can know the processing mode adopted by the headset. For example, the detection result is displayed to the user in the form of a prompt box.
  • the detection result may include the detected scene, and may also include a processing mode corresponding to the detected scene. For example, when the terminal device recognizes that the scene is the first scene, it determines the target mode corresponding to the first scene in the processing mode of the headset, and can display the detection result, that is, the first scene and the target mode to the user. Then send control signaling 2 to the left earphone and the right earphone respectively. Control signaling 2 is used to indicate the target mode.
  • a function switch for enabling smart scene detection is configured on the terminal device, and the terminal device triggers scene detection when the terminal device responds to the user initiating the function of smart scene detection.
  • the scene in which it is identified is the first scene
  • the target mode corresponding to the first scene in the processing mode of the earphone is determined, and then control signaling 2 is sent to the left earphone and the right earphone respectively. Control signaling 2 is used to indicate the target mode.
  • the detection result can also be displayed to the user, so that the user can know the processing mode adopted by the headset.
  • the detection result may include the detected scene, and may also include a processing mode corresponding to the detected scene.
  • the terminal device determines the target mode corresponding to the first scene in the processing mode of the headset, and can display the detection result, that is, the first scene and the target mode to the user. Then send control signaling 2 to the left earphone and the right earphone respectively. Control signaling 2 is used to indicate the target mode.
  • the left earphone and the right earphone send control signaling 2 again.
  • the function switch for enabling smart scene detection configured by the terminal device may be configured on the control interface of the headset application, or may also be configured in the system setting menu bar of the terminal device.
  • the terminal device can control the processing mode adopted by the headset by identifying the scene, and the terminal device can also identify the user operation on the selection control on the control interface to control the processing mode adopted by the headset. processing mode.
  • the terminal device can determine whether the smart scene detection function is enabled according to requirements. When the smart scene detection function is not enabled, the processing mode adopted by the headset can be manually selected by using Example 1.
  • the terminal device 100 identifies the current user scene at.
  • the user enables the smart scene detection function, he can update to another interface based on the interface for manually selecting the processing mode, or display the detection results based on the interface for manually selecting the processing function.
  • the processing function selected by the user on the terminal device is the HT function; after the smart scene detection function is enabled, the terminal device recognizes that the current scene the user is in is an airplane scene, and the ANC function is suitable.
  • the user starts the headset application, and the control interface of the headset application is displayed on the display screen, taking a circular ring as an example.
  • the processing function selected by the user is the HT function, as shown in (a) of FIG. 14A .
  • the control interface includes an option control for whether to enable the smart scene detection function.
  • the user triggers the smart scene detection function, executes the scene detection to obtain the detection result, and changes the position of the control representing the processing function selected by the user to the area belonging to the ANC function.
  • the position of the black dot on the disk may be the default value under the ANC function or the position where the processing intensity selected by the last user selection of the ANC function is located. For example, as shown in (b) of FIG. 14A . In (b) of FIG. 14A , the scene where an airplane is detected is taken as an example.
  • the terminal device 100 sends control signaling 2 to the left earphone and the right earphone respectively, and the control signaling 2 is used to indicate the ANC function.
  • the user starts the headset application, and the control interface of the headset application is displayed on the display screen, taking a circular ring as an example.
  • the processing function selected by the user is the HT function, as shown in (a) of FIG. 14B .
  • the control interface includes an option control for whether to enable the smart scene detection function. After the user triggers the option control for enabling the smart scene detection function, the user triggers the smart scene detection function, executes the scene detection to obtain the detection result, and displays the detection result on the detection result interface.
  • the detection interface interface may further include a scene that the terminal device can recognize, and a processing function corresponding to the scene. For example, as shown in (b) of FIG. 14B , the detection result is an airplane scene, and the corresponding processing function is the ANC function.
  • the terminal device 100 sends control signaling 2 to the left earphone and the right earphone respectively, and the control signaling 2 is used to indicate the ANC function.
  • the target processing intensity in the target mode can be determined by any one of the following methods.
  • Method 1 The headset adopts the default target processing intensity in target mode.
  • the terminal device After the terminal device sends the control signaling 2 to the left earphone and the right earphone respectively, taking the left earphone as an example, after the left earphone receives the control signaling 2, it determines that the processing mode used is the target mode, and the control signaling 2 does not indicate the target processing intensity. , the headset determines the default target processing strength. Taking the target mode as the ANC mode as an example, after receiving the control signaling 2, the left earphone determines to use the ANC mode, and obtains the default FF filter coefficient and FB filter coefficient in the ANC mode from the left earphone.
  • Method 2 Use the processing intensity used when the target mode was used last time as the target processing intensity.
  • the target processing strength is determined by the terminal device, and is indicated to the left earphone and the right earphone through control signaling. After the terminal device performs scene detection and determines the target mode according to the detected scene, obtains the processing intensity used when the target mode was used last time as the target processing intensity, and sends control signaling 2 to the left earphone and the right earphone respectively. Let 2 indicate the target mode and target processing strength.
  • the headset itself determines the processing intensity in the target mode.
  • the terminal device After the terminal device performs scene detection, after determining the target mode according to the detected scene, it sends control signaling 2 to the left earphone and the right earphone respectively, and the control signaling 2 indicates the target mode.
  • the left earphone and the right earphone receive the control signaling 2 , determine the adopted processing mode as the target mode, and obtain the saved processing intensity of the last time the target mode was adopted as the target processing intensity.
  • the target mode as ANC as an example, obtain the saved FF filter coefficients and FB filter coefficients when the ANC mode was last adopted, and perform ANC processing.
  • Manner 3 The terminal device determines the target processing intensity according to the recognized scene.
  • the function switch for enabling smart scene detection is not configured on the terminal device, after the terminal device recognizes the scene, it can determine the target processing intensity according to the recognized scene.
  • the processing modes determined in different scenarios are the same, and the processing intensities corresponding to different scenarios are different.
  • the walking scene, the running scene, and the quiet scene are all suitable for using the HT mode.
  • the walking scene, the running scene, and the quiet scene have different processing intensities.
  • ANC mode is suitable for multi-person talking scenes, cafe scenes, train scenes, airplane scenes, windy scenes, and office scenes.
  • the processing intensity corresponding to the scene, the windy scene, and the office scene is different.
  • the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene are all suitable for using the AH mode. When the AH mode is used, the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene correspond respectively different processing intensity.
  • the terminal device sends control signaling 2 to the left earphone and right earphone according to the stored correspondence between the scene type, the target mode and the processing intensity.
  • the control signaling 2 indicates the target mode and the target processing intensity in the target mode. Therefore, after receiving the control signaling 2, the earphone determines to adopt the target mode according to the control signaling 2, and determines the filter coefficient corresponding to the target processing intensity.
  • the target mode as AH as an example
  • the FF filter coefficient, the FB filter coefficient and the HT filter coefficient are determined according to the target processing intensity, and S801-S807 are performed according to the FF filter coefficient, the FB filter coefficient and the HT filter coefficient.
  • the user indicates to the headset the processing intensity used in the target mode through a UI control provided by the terminal device.
  • the detection result is displayed on the display interface of the terminal device, and the detection result includes the detected scene and the target mode corresponding to the detected scene.
  • the display interface may include a control for selecting the treatment intensity.
  • the control for selecting the treatment intensity is called an intensity control.
  • Intensity controls can include Control 1 and Control 2. Different positions of control 1 are used to indicate different processing intensity in target mode. Intensity controls can be circles, bars, or other shapes.
  • the detected scene is a terminal building scene. Taking the control 1 in the intensity control as a circle and the control 2 as a circular black dot as an example, in response to the user touching the control 2 and moving to the position of the control 1 1.
  • Position 1 represents the target processing intensity in the user-selected target mode.
  • Send control instruction 2 to the left earphone and right earphone, and control instruction 2 indicates the target mode and the target processing intensity corresponding to position 1.
  • the target mode and target processing intensity can be sent to the left and right earphones through different control commands.
  • the terminal device After determining the target mode according to the detected scene, the terminal device sends control signaling indicating the target mode to the left earphone and the right earphone.
  • the left earphone and the right earphone adopt the default processing intensity in the target mode, that is, adopt the default filter coefficient in the target mode to realize the target processing corresponding to the target mode.
  • the control signaling indicating the target processing intensity is sent to the left earphone and the right earphone, and further, the left earphone and the right earphone use the filter coefficient corresponding to the target processing intensity to achieve the target.
  • the target processing corresponding to the mode.
  • the user triggers the smart scene detection function, executes the scene detection to obtain the detection result, and sets the position of the control representing the processing function selected by the user by the position of the control. Changed to the area belonging to the ANC function.
  • the position of the black dot on the disk may be the default value under the ANC function or the position where the processing intensity selected by the last user selection of the ANC function is located.
  • the user selects the processing intensity in ANC mode by moving the position of the black dot. And send control signaling 2 to the left earphone and right earphone, and the control signaling 2 indicates the ANC mode and the target processing strength.
  • Example 4 scene detection of headphones, different scenes correspond to different processing functions.
  • the headset has a scene detection function.
  • the headset recognizes the scene the user is currently in.
  • the processing functions implemented by the headset are different.
  • the left earphone may have the scene detection function, or the right earphone may have the scene detection function, or both the left earphone and the right earphone may have the scene detection function.
  • one of the left earphone and the right earphone is used to perform scene detection.
  • the left earphone performs scene detection, and sends the detection result to the right earphone, so that both the left earphone and the right earphone are based on the detection result of the left earphone. Executes processing for realizing the processing function corresponding to the detection result.
  • both the left earphone and the right earphone execute processing for realizing the processing function corresponding to the detection result according to the detection result of the right earphone.
  • both the left earphone and the right earphone perform scene detection, the left earphone performs the processing for realizing the processing function corresponding to the detection result according to the detection result of the left earphone, and the right earphone performs the processing for realizing the detection according to the detection result of the right earphone The result corresponds to the processing of the processing function.
  • the enabling of the scene detection function of the headset may be controlled by the user through the headset or through the terminal device.
  • the headset is provided with a button for activating the scene detection function.
  • the user can turn on or turn off the scene detection function of the headset by touching the button.
  • the headset identifies the scene where the current user is located (or the scene where the headset is currently located), the corresponding relationship between the scene and the processing mode determines the processing mode corresponding to the identified scene, and realizes the processing corresponding to the processing mode.
  • the user can turn on or turn off the scene detection function of the earphone by tapping on the earphone, such as three consecutive taps.
  • the headset responds to the operation of the user tapping the headset three times to enable the scene detection function of the headset.
  • the scene detection function of the earphone is enabled, the earphone turns off the scene detection function of the earphone in response to the operation of the user tapping the earphone three times.
  • the headset identifies the scene where the current user is located (or the scene where the headset is currently located), the corresponding relationship between the scene and the processing mode determines the processing mode corresponding to the identified scene, and realizes the processing corresponding to the processing mode.
  • the activation of the scene detection function of the left earphone or the right earphone is controlled by the terminal device 100 .
  • the headphone control interface includes a switch button for the headphone scene detection function.
  • the terminal device can determine whether to enable the scene detection function of the headset according to user requirements. When the scene detection function of the headset is not enabled, Example 1 can be used to manually select the processing function to be implemented by the headset.
  • the headset identifies the scene type of the current external environment, determines the processing mode corresponding to the identified scene type according to the corresponding relationship between the scene type and the processing mode, and implements the processing function corresponding to the processing mode.
  • the terminal device 100 sends control signaling 3 to the headset 200, where the control signaling 3 is used to instruct the headset to start the scene detection function.
  • the headset 200 starts to perform scene detection according to the control signaling 3 .
  • the headset 200 determines the processing function to be executed according to the detected scene type of the current external environment, such as an ANC function, and the headset 200 executes the ANC processing and executes S501-S504.
  • the headset after the headset is connected to the terminal device, the headset starts scene detection; or when the headset receives the downlink audio signal sent by the terminal device, the scene detection starts.
  • the detection result may also be sent to the terminal device.
  • the detection result may be included in the indication information and sent to the terminal device.
  • the detection result may include the detected scene and the processing mode corresponding to the scene.
  • the terminal device receives the detection result, it displays the detection result to the user, so that the user knows the processing mode adopted by the headset.
  • the detection result is displayed to the user in the form of a prompt box.
  • only the detected scene may be included in the detection result.
  • the terminal device determines the processing mode corresponding to the scene detected by the headset, and displays the scene detected by the headset and the processing mode corresponding to the scene to the user. For example, when the earphone recognizes that the scene is the first scene, it determines the target mode corresponding to the first scene in the processing mode of the earphone, and can display the detection result, that is, the first scene and the target mode to the user.
  • the processing function of the processing mode corresponding to the scene is not executed immediately, the detection result is sent to the terminal device, and the terminal device displays the detection result to the user.
  • the terminal device responds to the user's operation of determining the processing mode, it sends a confirmation instruction to the earphone, and when the earphone receives the confirmation instruction, it adopts the processing mode corresponding to the scene detected by the earphone to execute the processing function.
  • the types of scenes that can be recognized by the headset may include walking scenes, running scenes, quiet scenes, multi-person talking scenes, cafe scenes, subway scenes, train scenes, car scenes, waiting room scenes, conversation scenes, office scenes, outdoor scenes.
  • the processing modes suitable for different scene types are exemplified.
  • the processing mode corresponding to the scene type is indicated in brackets.
  • the noise of the plane is relatively loud when flying, so it is suitable to use the ANC mode.
  • the HT mode For another example, in walking scenes, running scenes and quiet scenes, it is suitable to use the HT mode, and you can hear the sound of sudden events. For another example, in the cafe scene, the user needs to be quiet, and the ANC mode can be used. For another example, in a light music scene, the HT mode can be used. For another example, the alarm sound scene (AH), the whistle sound scene (AH), and the crying sound scene (AH) are used to hear the preset sound, and the AH mode is suitable.
  • the AH mode is suitable.
  • the processing of S601-S605 is performed respectively for the left earphone and the right earphone.
  • the scene type is the train station scene
  • the left earphone and the right earphone perform the processing of S801-S807, respectively.
  • the target processing intensity in the target mode may be determined in any of the following manners.
  • Method 1 The headset adopts the default target processing intensity in target mode.
  • the earphone determines the adopted processing mode as the target mode according to the detected scene, and the left earphone and the right earphone determine to adopt the default target processing intensity.
  • the target mode as the ANC mode as an example
  • the left earphone and the right earphone obtain the default FF filter coefficient and FB filter coefficient in the ANC mode.
  • Method 2 Use the processing intensity used when the target mode was used last time as the target processing intensity.
  • the earphone determines the processing intensity in the target mode by itself. After the headset performs scene detection, after determining the target mode according to the detected scene, the saved processing intensity of the last time the target mode was used is obtained as the target processing intensity. Taking the target mode as ANC as an example, obtain the saved FF filter coefficients and FB filter coefficients when the ANC mode was last adopted, and perform ANC processing.
  • the target processing strength is determined by the terminal device, and is indicated to the left earphone and the right earphone through control signaling.
  • the headset After the headset performs scene detection, it sends the detection result to the terminal device, so that the terminal device obtains the processing intensity used when the target mode was used last time as the target processing intensity, and sends control signaling 4 to the left headset and the right headset respectively. Let 4 indicate the target processing strength.
  • the headset determines the target processing intensity according to the recognized scene.
  • the target processing intensity can be determined according to the recognized scene.
  • the processing modes determined in different scenarios are the same, and the processing intensities corresponding to different scenarios are different.
  • the walking scene, the running scene, and the quiet scene are all suitable for using the HT mode.
  • the walking scene, the running scene, and the quiet scene have different processing intensities.
  • ANC mode is suitable for multi-person talking scenes, cafe scenes, train scenes, airplane scenes, windy scenes, and office scenes.
  • the processing intensity corresponding to the scene, the windy scene, and the office scene is different.
  • the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene are all suitable for using the AH mode. When the AH mode is used, the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene correspond respectively different processing intensity.
  • the left earphone and the right earphone determine the target mode corresponding to the detected scene and the target processing intensity in the target mode according to the stored correspondence between the scene type, the target mode, and the processing intensity. Thereby, the left earphone and the right earphone obtain the filter coefficient corresponding to the target processing intensity.
  • the target mode as AH as an example
  • the FF filter coefficient, the FB filter coefficient and the HT filter coefficient are determined according to the target processing intensity
  • S801-S807 are performed according to the FF filter coefficient, the FB filter coefficient and the HT filter coefficient.
  • the headset may also perform emergency event detection to determine a target event (or referred to as a target event scenario).
  • a target event scenario For example, emergent events include one or more of wind noise events, whistling events, emergencies, human voice events, or no emergencies, and different events have different processing intensities.
  • the headset performs scene detection, as well as event detection.
  • target mode different events correspond to different filter coefficients. Taking ANC as an example, different events correspond to different FF filter coefficients and/or different FB filter coefficients.
  • the left earphone can obtain the FF filter coefficient or FB corresponding to the detected event from the coefficient library according to the detection result.
  • the coefficient library stores processing modes, events, and the mapping relationship between FF filter coefficients and FB filter coefficients.
  • the processing effect of ANC mainly depends on FB filtering and/or FF filtering.
  • the filter coefficient of the FF filter is controlled according to the detected scene, and the FB filter coefficient adopts a fixed value.
  • the filter coefficient of the FB filter is controlled according to the detected scene, and the FF filter coefficient adopts a fixed value.
  • the FF filter coefficient and the FB filter coefficient are controlled according to the detected scene.
  • the event includes a howling event, a wind noise event, a sudden event, a human voice event, or no sudden event.
  • the earphone 200 detects the event sound in the external environment, and can determine the target event corresponding to the event sound in the external environment according to the signal collected by the reference microphone.
  • the signal collected by the reference microphone includes a signal of a preset frequency spectrum, and then an event corresponding to the signal of the preset frequency spectrum is determined.
  • the signal collected by the reference microphone includes a wind sound signal, that is, the collected signal includes a signal matching the frequency spectrum of the wind sound, and the event corresponding to the detected event sound in the external environment is determined to be a wind noise event.
  • a spectrum matching method or a deep neural network (deep neural networks, DNN) matching method may be used.
  • the headset 200 can determine the event in the current environment of the user according to the signal collected by the reference microphone in the following manner, as shown in FIG. 15 .
  • Bone conduction sensors are also included in the headset 200 .
  • the bone conduction sensor is used to collect the bone conduction signal of the earphone user.
  • the bone conduction sensor obtains the bone conduction signal by collecting the bone conduction signal, that is, collecting the periosteal vibration signal generated by the user's speaking.
  • the activation of the scene detection function of the left earphone or the right earphone may be controlled by the terminal device 100, or may be controlled by a user's operation on the earphone, such as an operation of tapping the left earphone or the right earphone.
  • a bone conduction sensor is included on the earphone, and the user's upper and lower teeth touch to generate a tooth touching sound, so that the bone conduction sensor activates the scene detection function by detecting the audio signal generated by the user's upper and lower teeth touching.
  • the third signal collected by the reference microphone is a signal collected by the reference microphone after the headset starts the scene detection function.
  • the energy of the bone conduction signal collected by the bone conduction sensor is small.
  • the signal AA1 is the third signal.
  • the earphone 200 may first determine the energy of the bone conduction signal, and if the energy of the bone conduction signal is less than the set threshold, the filtering operation is not performed, that is, S1501 is not performed, when it is determined that the energy of the bone conduction signal is greater than or equal to the set threshold When the threshold value is reached, execute S1501.
  • the earphone 200 performs spectrum analysis on the signal AA1 to obtain the energy characteristic of the signal AA1. For example, the earphone 200 performs spectral analysis on the signal to obtain the energy of the entire frame of the signal AA1 and the energy of each bark subband of the signal AA1, thereby forming the energy characteristic of the signal AA1 represented by a vector.
  • S1503 Determine a first energy feature of the energy features included in the energy feature set that matches the energy feature of the filtered signal, where different energy features included in the energy feature set correspond to different event identifiers.
  • S1504 Determine that the event identified by the event identifier corresponding to the first energy feature is an event in the current environment of the user, that is, a detection result of event detection.
  • the energy feature set may be generated by: performing wind noise detection, burst noise detection, howling detection, and human voice detection on the signals collected by the first microphone, the second microphone, and the third microphone to obtain wind noise signal, burst noise signal and howling signal. Then, spectrum analysis is performed on the wind noise signal, the burst noise signal, the howling signal and the human voice signal, respectively, to obtain the sub-band energy characteristics of the wind noise signal, the sub-band energy characteristics of the burst noise signal, and the sub-band of the howling signal. Energy signature and sub-band energy signature of the vocal signal.
  • the sub-band energy features of the wind noise signal, the sub-band energy features of the burst noise signal, the sub-band energy features of the howling signal, and the sub-band energy features of the human voice signal constitute the energy feature set. It should be understood that in a quiet scene, the sub-band energy of the noise is relatively weak.
  • a spectrum matching method can be used, and a deep neural network (DNN) can also be used.
  • DNN deep neural network
  • matching degree between the energy feature of the filtered signal and each energy feature included in the energy feature set can be determined through the DNN, and the event identified by the event identifier corresponding to the first energy feature with the highest matching degree is Test results.
  • the main control unit in the headset 200 may implement the determination of an event in the current environment of the user according to the signal collected by the reference microphone.
  • the main control unit includes a DSP for executing S1501-S1504.
  • the user indicates to the headset the processing intensity used in the target mode through a UI control provided by the terminal device.
  • the detection result is sent to the terminal device, and the terminal device displays the detection result to the user.
  • the detection result is displayed on the display interface of the terminal device, and the detection result includes the scene detected by the headset and the target mode corresponding to the detected scene.
  • the display interface further includes a control for selecting the treatment intensity.
  • the control for selecting the treatment intensity is called an intensity control.
  • Intensity controls can include Control 1 and Control 2. Different positions of control 1 are used to indicate different processing intensity in target mode. Intensity controls can be circles, bars, or other shapes. Referring to FIG.
  • position 2 represents the target processing intensity in the user-selected target mode.
  • the target mode is HT as an example.
  • the terminal device 100 sends control signaling 3 to the headset 200 in response to the user initiating the function of scene detection in the headset, where the control signaling 3 is used to instruct the headset to activate the scene detection function.
  • the headset 200 starts and executes scene detection according to the control signaling 3 to obtain the detection result.
  • the headset 200 may send the detection result to the terminal device 100, so that the terminal device 100 displays it to the user, and displays the detected scene corresponding to the processing mode that the headset needs to adopt to the user.
  • the user may update the interface to manually select the processing mode to another interface, or display the detection results based on the interface for manually selecting the processing function.
  • the processing function selected by the user on the terminal device is the HT function; after the headset scene detection function is enabled, the headset 200 recognizes that the scene the user is currently in is an airplane scene, and the ANC function is suitable for detecting The result, the aircraft scene, the ANC function, is sent to the end device.
  • the user starts the headset application, and the control interface of the headset application is displayed on the display screen, taking a circular ring as an example.
  • the processing function selected by the user is the HT function, as shown in (a) of FIG. 17A .
  • the control interface includes an option control for whether to enable the headset scene detection function.
  • the terminal device triggers the headset scene detection function, and sends control signaling 3 to the headset 200, where the control signaling 3 is used to instruct the headset to activate the scene detection function.
  • the headset 200 starts and executes scene detection according to the control signaling 3 to obtain the detection result.
  • the headset 200 sends the detection result to the terminal device 100, and after receiving the detection result, the terminal device 100 changes the position of the control representing the processing function selected by the user to an area belonging to the ANC function.
  • the user selects the processing intensity in ANC mode by moving the position of the black dot on the disc. For example, as shown in (b) of FIG. 17A . In (b) of FIG. 17A , a scene where an airplane is detected is taken as an example.
  • the user starts the earphone application, and the control interface of the earphone application is displayed on the display screen, taking a circular ring as an example.
  • the processing function selected by the user is the HT function, as shown in (a) of FIG. 17B .
  • the control interface includes an option control for whether to enable the headset scene detection function.
  • the terminal device triggers the headset scene detection function, and sends control signaling 3 to the headset 200, where the control signaling 3 is used to instruct the headset to activate the scene detection function.
  • the headset 200 starts and executes scene detection according to the control signaling 3 to obtain the detection result.
  • the headset 200 sends the detection result to the terminal device 100, and after receiving the detection result, the terminal device 100 displays the detection result on the detection result interface.
  • the detection interface interface may also include a scene that the headset can recognize, and a processing mode corresponding to the scene.
  • the user selects the processing intensity in ANC mode by moving the position of the black dot on the disc. For example, as shown in (b) of FIG. 17B , the detection result is an airplane scene, and the corresponding processing mode is the ANC mode.
  • an artificial intelligence (artificial intelligence, AI) model can be used to perform detection and classification.
  • the AI model can be configured in the headset. Another way can determine the scene type according to the signal collected by the reference microphone. For example, the headset 200 can determine the scene where the user is currently located according to the signal collected by the reference microphone in the following manner, as shown in FIG. 18 .
  • S1801 Perform spectrum analysis on the first signal collected by the reference microphone, divide the first signal into multiple subbands, and calculate the energy of each subband. For example, the first information collected by the reference microphone is divided into subbands in the frequency domain according to the bark subband division method, and the energy of each subband is calculated.
  • S1802 Determine the VAD to obtain the noise segment in the first signal, and obtain the smoothed energy of each subband in the noise segment.
  • the VAD is determined as follows:
  • S1803 Determine the scene type according to the smoothed energy of each subband in the noise segment.
  • the processing is performed for the determined noise segment to determine the scene type:
  • Example 5 After determining the processing mode, the headset performs event detection in the processing mode. In the processing mode, the filter coefficients corresponding to different events (ie, the processing intensity in the processing mode) are different.
  • the earphone recognizes the user's operation, and determines that the earphone 200 selected by the user needs to implement ANC processing, HT processing or AH processing.
  • the processing mode adopted by the earphone 200 is the ANC mode function.
  • the user's operation may be an operation of the user tapping the earphone, and the processing mode is determined to be the ANC mode, the HT mode, or the AH mode through different operations.
  • buttons are provided on the earphone, and different buttons indicate different processing modes. The user selects the processing mode employed by the headset by pressing a button.
  • the left earphone and the right earphone perform ANC processing, and specifically perform S501-S504.
  • the processing mode to be implemented by the earphone is selected and controlled by the terminal device 100 .
  • Either the left earphone or the right earphone can have an event detection function.
  • one of the left earphone and the right earphone is used to perform event detection.
  • the left earphone performs event detection and sends the detection result to the right earphone, or the right earphone performs event detection and sends the detection result. to the left earphone.
  • different events correspond to different FF filter coefficients and FB filter coefficients.
  • the left earphone or the right earphone performs event detection, taking the left earphone as an example, the left earphone can obtain the FF filter coefficient or FB filter coefficient corresponding to the detected event in the ANC mode from the coefficient library according to the detection result.
  • the coefficient library For example, refer to the content included in the coefficient library shown in Table 2, and take an event including a howling event, a wind noise event, a sudden event or a human voice event as an example.
  • the headset includes hardware structures and/or software modules corresponding to executing each function.
  • modules and method steps of each example described in conjunction with the embodiments disclosed in the present application can be implemented in the form of hardware or a combination of hardware and computer software. Whether a function is performed by hardware or computer software-driven hardware depends on the specific application scenarios and design constraints of the technical solution.
  • an embodiment of the present application further provides a noise processing apparatus 1900 .
  • the noise processing device 1900 is applied to an earphone.
  • the headset has at least two functions in the ANC function, the HT function or the AH function; the headset includes a first microphone and a second microphone; the first microphone is used to collect a first signal, and the first signal is used to represent the current the sound of the external environment; the second microphone is used to collect a second signal, and the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the earphone.
  • the noise processing apparatus 1900 can be used to implement the function of the earphone in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
  • the apparatus may include a communication module 1901 , an acquisition module 1902 , and a first processing module 1903 .
  • a communication module 1901 configured to receive a first audio signal from the terminal device
  • the acquiring module 1902 is configured to acquire a target mode; wherein, the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the earphone to implement a target processing function, and the target processing function is an active reduction.
  • the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the earphone to implement a target processing function, and the target processing function is an active reduction.
  • the first processing module 1903 is configured to obtain a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal.
  • the apparatus further includes: a playing module, configured to play the second audio signal.
  • a playing module configured to play the second audio signal.
  • the playback module it is not shown in FIG. 19 .
  • the target processing function is an ANC function
  • the second audio signal played by the playback module can attenuate the user's sound to the user's current environment and the user's ear canal Perception of internal ambient sounds; or,
  • the target processing function is the HT function, and the second audio signal played by the playback module can enhance the user's perception of the sound of the environment where the user is currently located; or,
  • the target processing function is the AH function, and the second audio signal played by the playing module can enhance the user's perception of the event sound; the event sound satisfies a preset frequency spectrum.
  • the target processing function is an ANC function
  • the second audio signal is obtained based on the first audio signal, the third signal and the fourth signal, and the third signal is the first signal
  • the inverted signal of the fourth signal is the inverted signal of the second signal
  • the target processing function is the HT function
  • the second audio signal is obtained based on the first audio signal, the first signal and the second signal; or,
  • the target processing function is the AH function
  • the second audio signal is obtained based on the first audio signal, the fifth signal and the fourth signal
  • the fifth signal is an event signal in the first signal
  • the The event signal satisfies the preset spectrum.
  • the communication module 1901 is further configured to receive a first control instruction from the terminal device, where the first control instruction carries the target mode, and the target mode is the terminal The device is determined according to the scene type of the current external environment;
  • the target mode is sent to the obtaining module 1902 .
  • the communication module 1901 is further configured to:
  • the second control instruction carries a target processing strength, and the target processing strength is used to indicate the processing strength when the headset implements the target processing function;
  • the first processing module 1903 is specifically used for:
  • a second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
  • it also includes:
  • the second processing module 1904 is configured to determine the target event corresponding to the event sound in the current external environment according to the first signal, and determine the target processing intensity in the target mode according to the target event; wherein, the target processing intensity used to indicate the processing strength when the headset implements the target processing function;
  • the first processing module 1903 is specifically used for:
  • a second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
  • the headset further includes a bone conduction sensor, and the bone conduction sensor is used to collect bone conduction signals generated by the vibration of the user's vocal cords;
  • the first processing module 1901 is specifically configured to determine the target event corresponding to the event sound in the current external environment according to the first signal and the bone conduction signal.
  • the target event includes one of a howling event, a wind noise event, a sudden event or a human voice event.
  • it also includes:
  • the third processing module 1905 is configured to identify the scene type of the current external environment as a target scene according to the first signal, and determine the target mode adopted by the headset according to the target scene, where the target mode corresponds to the target scene. The processing mode of the target scene.
  • the target scene includes a walking scene, a running scene, a quiet scene, a multi-person talking scene, a cafe scene, a subway scene, a train scene, a waiting room scene, a dialogue scene, an office scene, and an outdoor scene , driving scene, windy scene, airplane scene, siren scene, whistle scene, cry scene.
  • the communication module 1901 is further configured to send indication information to the terminal device, where the indication information carries the target mode;
  • the first processing module 1903 is specifically used for:
  • a second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
  • the target processing function is an ANC function
  • the greater the target processing intensity is the more the user perceives the sound of the current environment of the user and the sound inside the user's ear canal.
  • the target processing function is the HT function
  • the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
  • the target processing function is the AH function
  • the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
  • the earphone is a left earphone, or the earphone is a right earphone.
  • the target mode instructs the headset to implement the ANC function
  • the first processing module 1903 is specifically configured to:
  • the second audio signal is obtained by performing mixing processing on the fourth audio signal and the first audio signal.
  • the filter coefficient used in the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the ANC function; or,
  • the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the ANC function.
  • the target mode indicates that the headset implements the HT function
  • the first processing module 1903 is specifically configured to:
  • the second audio signal is obtained by performing mixing processing on the third filtered signal and the fifth audio signal.
  • the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the HT function; or,
  • the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the HT function.
  • the target mode indicates that the headset implements the AH function
  • the first processing module 1903 is specifically configured to:
  • the second audio signal is obtained by mixing the fifth filtered signal, the sixth audio signal and the first filtered signal.
  • the filter coefficient used in the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the AH function; or,
  • the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the AH function; or,
  • the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the AH function.
  • the terminal device includes corresponding hardware structures and/or software modules for executing each function.
  • modules and method steps of each example described in conjunction with the embodiments disclosed in the present application can be implemented in the form of hardware or a combination of hardware and computer software. Whether a function is performed by hardware or computer software-driven hardware depends on the specific application scenarios and design constraints of the technical solution.
  • an embodiment of the present application further provides a mode control apparatus 2000 .
  • the mode control apparatus 2000 is applied to the terminal device 100, and can be used to implement the functions of the terminal device in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
  • the mode control apparatus 2000 includes a first detection module 2001 and a transmission module 2002 , and may further include a display module 2003 and a second detection module 2004 .
  • a first detection module 2001 configured to identify a target mode according to the target scene when the scene type of the current external environment is a target scene;
  • the target mode is one of the processing modes supported by the headset, different processing modes correspond to different scene types, and the processing modes supported by the headset include active noise reduction ANC mode, ambient sound transparent transmission HT mode or hearing enhancement At least two of the AH modes.
  • the sending module 2002 is configured to send the target mode to the earphone, where the target mode is used to instruct the earphone to implement the processing function corresponding to the target mode.
  • it also includes:
  • the display module 2003 is configured to display result prompt information when the target mode is determined according to the target scene, and the result prompt information is used to prompt the user that the headset implements the processing function corresponding to the target mode.
  • the display module 2003 is configured to display selection prompt information before sending the first control signaling to the headset, where the selection prompt information is used to prompt the user whether to adjust the processing mode of the headset is the target mode;
  • the second detection module 2004 is configured to detect an operation in which the user selects the processing mode of the earphone as the target mode.
  • the display module 2003 is further configured to display a first control and a second control, wherein different positions of the second control on the first control are used to indicate the target Different processing intensity in mode;
  • the second detection module 2004 is further configured to detect the operation of the user touching the second control to move to the first position on the first control before the sending module 2002 sends the first control signaling to the headset , the first position of the second control on the first control indicates the target processing intensity in the target mode;
  • the sending module 2002 is further configured to send the target processing intensity to the earphone, where the target processing intensity is used to indicate the processing intensity when the earphone implements the processing function corresponding to the target mode.
  • the shape of the first control is a ring
  • the processing in the target mode The intensity varies from small to large; or,
  • the user touches the second control to move on the first control in a counterclockwise direction, and the processing intensity in the target mode changes from small to large;
  • the shape of the first control is a bar
  • the processing in the target mode The intensity varies from small to large; or
  • the target processing function is an ANC function
  • the greater the target processing intensity is the more the user perceives the sound of the current environment where the user is located and the sound inside the user's ear canal.
  • the target processing function is the HT function
  • the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
  • the target processing function is the AH function
  • the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
  • an embodiment of the present application further provides a mode control apparatus 2100 .
  • the mode control apparatus 2100 is applied to the terminal device 100, and can be used to implement the functions of the terminal device in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
  • the mode control device 2100 includes a processing module 2101 , a sending module 2102 , a receiving module 2103 , a display module 2104 and a detection module 2105 .
  • the processing module 2101 is used to obtain a target mode;
  • the target mode is one of the processing modes supported by the headset, and the processing modes supported by the headset include active noise reduction ANC mode, ambient sound transparent transmission HT mode or auditory enhancement AH mode at least two of them;
  • the processing module 2101 is further configured to determine the target processing intensity in the target mode according to the scene type of the current external environment; different scene types correspond to different processing intensities in the target mode;
  • the sending module 2102 is configured to send the target processing strength to the earphone, where the target processing strength is used to indicate the processing strength when the earphone implements the processing function corresponding to the target mode.
  • it also includes:
  • a receiving module 2103 configured to receive the target mode sent by the headset.
  • it also includes:
  • the display module 2104 is configured to display a selection control, where the selection control includes a processing mode supported by the headset, and detects an operation of the user selecting the target mode in the processing mode of the headset through the selection control.
  • the display module 2104 is further used for:
  • the processing module 2101 determines the target processing intensity in the target mode according to the scene type of the current external environment
  • the receiving module 2103 receives the target mode sent by the headset
  • a selection prompt is displayed,
  • the selection prompt information is used to indicate whether the user adjusts the processing mode of the headset to the target mode;
  • the detection module 2105 is configured to detect the operation of the user selecting to adjust the processing mode of the earphone to the target mode.
  • the target processing function is an ANC function
  • the greater the target processing intensity is the more the user perceives the sound of the current environment of the user and the sound inside the user's ear canal.
  • the target processing function is the HT function
  • the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
  • the target processing function is the AH function
  • the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
  • an embodiment of the present application further provides a mode control apparatus 2200 .
  • the mode control apparatus 2200 is applied to the terminal device 100, and can be used to implement the functions of the terminal device in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
  • the mode control device 2100 includes a display module 2201 , a detection module 2202 , a transmission module 2203 , a processing module 2204 , and an identification module 2205 .
  • the display module 2201 is used to include a first selection control on the first interface, the first selection control includes the processing mode supported by the first target headset and the processing intensity corresponding to the processing mode supported by the first target headset; the first target The processing mode of the headset includes at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode, or the hearing enhancement AH mode;
  • the detection module 2202 detects the first operation performed by the user on the first interface; the first operation is that the user selects the first target mode and selects the first target mode from the processing modes supported by the first target headset through the first selection control.
  • the processing intensity in a target mode is generated by the first target processing intensity;
  • a sending module 2203 configured to send the first target mode and the first target processing intensity to the first target earphone, where the first target mode is used to instruct the first target earphone to achieve the first target
  • the processing function corresponding to the mode, and the first target processing intensity is used to indicate the processing intensity when the first target earphone implements the processing function corresponding to the first target mode.
  • the display module 2201 is further configured to display selection prompt information before displaying the first interface, where the selection prompt information is used for the user to choose whether to adjust the first target headset processing mode;
  • the detection module 2202 is further configured to detect the user's operation of selecting and adjusting the processing mode of the first target earphone.
  • it also includes:
  • the identification module 2205 is configured to identify the scene type of the current external environment as the target scene before the display module 2201 displays the first interface, and the target scene is adapted to the scene type that needs to adjust the processing mode of the first target headset.
  • it also includes:
  • the identification module 2205 is configured to identify that the terminal device triggers the first target headset to play audio before the display module 2201 displays the first interface.
  • the detection module 2202 is further configured to detect that the terminal device establishes a connection with the first target headset before the display module displays the first interface.
  • the detection module 2202 before the display module 2201 displays the first interface, detects that the terminal device establishes a connection with the first target headset, and detects the The second operation performed by the user on the main interface.
  • the main interface includes an icon of a first application, the second operation is generated by the user touching the icon of the first application, and the first interface is a display interface of the first application.
  • the first selection control includes a first control and a second control
  • the second control indicates two different processing modes of the first target earphone at any two different positions of the first control , or, the second control indicates different processing intensities in the same processing mode of the first target earphone at any two different positions of the first control;
  • the first operation is generated by the user moving a first position of the second control in an area corresponding to the first target mode on the first control, where the first position corresponds to the first The first target processing intensity in target mode.
  • the shape of the first control is a circular ring, the circular ring includes at least two arc segments, and the second controls are located in different arc segments to indicate that the first target headset is different processing mode, the second controls are located at different positions of the same arc segment to indicate different processing intensities of the same processing mode of the first target earphone; or,
  • the shape of the first control is a bar, the bar includes at least two bar segments, the second controls are located in different bar segments to indicate different processing modes of the first target earphone, and the second control is located in different bar segments. Different positions of the controls on the same bar segment indicate different processing intensities of the same processing mode of the first target earphone.
  • the detection module 2202 is further configured to detect a third operation performed by the user on the first interface, the first interface further includes a second selection control, and the second selection control includes The processing mode supported by the second target earphone and the processing intensity corresponding to the processing mode supported by the second target earphone, the processing mode supported by the first target earphone includes active noise reduction ANC mode, ambient sound transparent transmission HT mode or auditory enhancement AH mode
  • the third operation is that the user selects the second target mode in the processing mode of the second target headset through the second selection control and selects the processing intensity in the second target mode as the second target
  • the processing intensity is generated; when the first target earphone is the left earphone, the second target earphone is the right earphone, or, the first target earphone is the right earphone, and the second target earphone is the left earphone;
  • the sending module 2203 is further configured to send the second target mode and the second target processing intensity to the second target headset, where the second target mode is used to instruct the second target headset to implement the The processing function corresponding to the second target mode, and the second target processing intensity is used to indicate the processing intensity when the second target earphone implements the processing function corresponding to the second target mode.
  • the terminal device includes a processor 2301 , a memory 2302 , a communication interface 2303 and a display 2304 .
  • the memory 2302 is used to store instructions or programs executed by the processor 2301, or input data required by the processor 2301 to execute the instructions or programs, or to store data generated after the processor 2301 executes the instructions or programs.
  • the processor 2301 is configured to execute the instructions or programs stored in the memory 2302 to perform the functions performed by the terminal device in the above method.
  • the processor 2301 is configured to execute the functions of the first detection module 2001 , the sending module 2002 , the display module 2003 , and the second detection module 2004 .
  • the processor 2301 is configured to execute the functions of the first detection module 2001 and the second detection module 2004 .
  • the function of the sending module 2002 is realized by the communication interface 2303
  • the function of the display module 2003 can be realized by the display 2304 .
  • the processing module 2101 , the sending module 2102 , the receiving module 2103 , the display module 2104 and the detection module 2105 may be implemented by the processor 2301 .
  • the processor 2301 can be used to execute the functions of the processing module 2101 and the detection module 2105
  • the functions of the sending module 2102 and the receiving module 2103 can be implemented by the communication interface 2303
  • the functions of the display module 2104 can be implemented by the display 2304 .
  • the display module 2201 , the detection module 2202 , the sending module 2203 , the processing module 2204 , and the identification module 2205 may be implemented by the processor 2301 .
  • the functions of the processing module 2204 , the detection module 2202 and the identification module 2205 can all be implemented by the processor 2301 .
  • the function of the sending module 2203 can be implemented by the communication interface 2303
  • the function of the display module 2201 can be implemented by the display 2304 .
  • the processor in the embodiment of the present application may be a central processing unit (central processing unit, CPU), and may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSP), application-specific integrated circuits (application specific integrated circuit, ASIC), field programmable gate array (field programmable gate array, FPGA) or other programmable logic devices, transistor logic devices, hardware components or any combination thereof.
  • a general-purpose processor may be a microprocessor or any conventional processor.
  • this application will describe some optional earphone usage examples and signal processing examples with reference to some specific application scenarios.
  • Users can use it in conjunction with the terminal and the headset.
  • the headset can support at least one of the ANC function, the HT function, and the AH function, and of course it can also include an empty mode; among them, the ultimate goal of the ANC function is to eliminate the noise in actual hearing; the ultimate goal of the HT function is to eliminate the external
  • the effect of the sound entering the human ear makes the effect of the external environment signal heard by the user through the earphone is equivalent to the effect of the user's naked ear hearing the sound of the current external environment; equivalence can refer to the same or approximately the same effect.
  • the terminal device establishes a communication connection with the headset
  • the first interface is used to set the function of the headset; the first interface may include an enabling option of an ANC function, an HT function, an AH function or an empty mode, or a disabling option thereof.
  • a possible example of the first interface is shown in Figure 24.
  • the noise control mode supported by the headset may include options of empty mode (off), ANC function (noise reduction), and HT function (transparent transmission). It should be understood that the first interface may further include more setting contents or options, which are not all shown in the drawings in this application.
  • the transparent transmission strength of the HT function can be further acquired, and the HT function of the earphone can be controlled according to the acquired transparent transmission strength.
  • an option of event sound enhancement is added in the first interface; it should be noted that, optionally, this option may not be available before the HT option is enabled.
  • the event sound is a sound in the external environment that meets the preset event conditions; the event sound may include human voice or other sounds that meet the preset frequency spectrum characteristics, and for details, refer to the relevant descriptions in the foregoing embodiments.
  • the terminal can control the earphone to improve the signal-to-noise ratio of the event sound in the signal collected by the earphone; wherein, the event sound The higher the signal-to-noise ratio of , the greater the proportion of the energy of the event sound in the signal.
  • two methods for improving the signal-to-noise ratio of the event sound in the signal collected by the earphone will be described in detail in the following examples. See S4001-S4005 and S5001-S5005.
  • FIG. 26 A possible implementation form can be seen in Fig. 26.
  • the HT function transparent transmission
  • the user's click instruction can be received or the preset vocal enhancement enabling instruction can be obtained, so that the Vocal Enhancement This option is enabled.
  • the terminal can control both the ANC function and the HT function of the headset to be in an enabled state.
  • the ANC function of the earphone may be activated again when the HT function is kept in an on state.
  • the ANC function and the HT function are jointly enabled to process the sound signal of the external environment collected by the earphone, and the processing includes improving the signal-to-noise ratio of the event sound in the collected signal.
  • the terminal may enable the ANC function of the earphone, and control the earphone to improve the signal-to-noise ratio of the event sound in the signal collected by the earphone in combination with some enhancement and noise reduction algorithms.
  • the intensity of event sound enhancement may be further acquired, and the event sound enhancement function of the headset is controlled according to the acquired event sound enhancement intensity.
  • the stronger the intensity of the event sound enhancement the higher the signal-to-noise ratio of the event sound.
  • a strength option of the ANC function is added in the first interface. It should be noted that, optionally, this option may not appear in the first interface before the ANC option is enabled, or may appear in the first interface simultaneously with the ANC and HT function options.
  • the strength options of the ANC function include at least a first steady-state ANC strength option, a second steady-state ANC strength option, and an adaptive ANC strength option; wherein the first steady-state ANC strength option and the second steady-state ANC strength option respectively correspond to the first steady-state ANC strength option and the second steady-state ANC strength option.
  • the first scene and the second scene correspond to different and stable ANC function strengths respectively; the ANC function strength corresponding to the adaptive ANC strength option is related to the scene type of the current environment in which the terminal device or headset is located; scenes with different current environments Types correspond to different ANC strengths.
  • the different scene types of the current environment may include the first scene and the second scene.
  • the option of the intensity option (noise reduction mode) of the ANC function is added to the interface.
  • the strength option (noise reduction mode) of the ANC function may further include multiple options, as shown in FIG. 28 , including but not limited to noise reduction modes such as light, balanced, deep, and intelligent dynamic noise reduction.
  • the light level is suitable for quieter environments
  • the depth level is suitable for very noisy environments, that is, scenes that do not belong to the light level but do not belong to the depth level can be classified as balanced levels, that is, ordinary scenes.
  • the noise reduction intensity of the ANC control function of the headset corresponds to the light level
  • the noise reduction intensity of the control headset ANC corresponds to the medium level
  • the noise reduction intensity of the control headphone ANC corresponds to the depth level.
  • the noise reduction intensity of the ANC function of the headset corresponds to the medium level.
  • the noise reduction intensity of the ANC function of the headset corresponds to the depth level.
  • the ANC noise reduction depth corresponding to the light level may include 20-28dB
  • the ANC noise reduction depth corresponding to the balanced level may include 30-36dB
  • the ANC noise reduction depth corresponding to the depth level may be greater than 40dB.
  • the environmental scenes corresponding to the light level may include but are not limited to offices, bedrooms, quiet living rooms, etc.; the environmental scenes corresponding to the balanced level may include but are not limited to supermarkets, squares, waiting rooms, roads, cafes, shopping malls, etc.;
  • the environmental scenes corresponding to the depth level may include but are not limited to subways, high-speed trains, taxis, and airplanes.
  • the embodiment of the present application also includes a kind of intelligent dynamic noise reduction, that is, a kind of adaptive environmental noise reduction; that is, the ANC intensity can be determined according to the scene type of the current environment by acquiring the scene type of the environment where the terminal or the headset is located; The ANC function is controlled according to the determined ANC strength.
  • a kind of intelligent dynamic noise reduction that is, a kind of adaptive environmental noise noise reduction
  • the ANC intensity can be determined according to the scene type of the current environment by acquiring the scene type of the environment where the terminal or the headset is located
  • the ANC function is controlled according to the determined ANC strength.
  • the adaptive ambient noise noise reduction may include, but is not limited to, at least one of the above-mentioned light, balanced, and deep levels.
  • adaptive environmental noise reduction can perform corresponding levels of ANC noise reduction in three levels: light, balanced, and deep according to the state of the environment. Detecting whether the current environment is light, balanced, or deep can be performed by earphones or terminals.
  • adaptive ambient noise noise reduction allows users to perform adaptive noise reduction at different levels according to changes in the environment without manual operation, improving user experience.
  • enabling methods for ANC, HT, event sound enhancement, noise reduction mode or closing options include, but are not limited to, receiving user clicks on corresponding function items, terminal adaptive switching, and headset adaptive switching. , or the trigger of the shortcut.
  • a user's selection operation for the option of the event sound enhancement function, the selection operation for the option for the HT function, or the selection operation for the option of the ANC function is received; or; it is recognized that the current environment is related to the event sound Enhance the scene corresponding to the function and activate the option of the event sound enhancement function, recognize that the current environment is a scene corresponding to the HT function and activate the option of the HT function, or recognize that the current environment is corresponding to the ANC function scene and activate the option of the ANC function; or; in response to a user's pressing operation on the headset, the pressing operation is used for at least two of the event sound enhancement function, the HT function or the ANC function switch between functions.
  • a pressure sensor may be included in the headset, and the pressure sensor may predefine some shortcut operations, such as switching between noise reduction modes.
  • the strength of the ANC function mentioned in this application can be understood as the noise reduction strength of ANC, or simply the noise reduction strength; the strength of the HT function can be understood as the transparent transmission strength; the strength of the AH function can be understood as the enhancement strength. Different intensities will affect the relevant filter coefficients. For details, reference may be made to the relevant descriptions of the foregoing embodiments, which will not be repeated here.
  • the earphone includes a first microphone (reference microphone), a second microphone (error microphone) and a speaker; in combination with the possible implementation forms of the earphone in the foregoing embodiment, the earphone can perform the following methods:
  • S4001 collect a first signal through a first microphone, and the first signal is used to characterize the sound of the current external environment; in the prior art, the signal collected by the reference microphone is also called a reference signal
  • S4002 Collect a second signal through the second microphone, where the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the headset.
  • the signal collected by the error microphone is also called an error signal.
  • the sound of the internal environment of the ear canal can be understood as the comprehensive sound of the ambient sound combined with the sound that the earphone may play, the algorithm being used by the earphone (such as noise reduction, transparent transmission, etc.), the human ear environment and other factors after the user puts the earphone on perception.
  • the sound of the internal environment of the ear canal can be understood as, but not limited to, the comprehensive sound representation of the environmental sound collected by the error microphone combined with the human ear environment.
  • the sound of the internal environment of the ear canal can be understood as, but not limited to, the environmental sound collected by the error microphone combined with the sound played by the headset microphone and the comprehensive sound representation of the human ear environment.
  • the sound of the internal environment of the ear canal can be understood as, but not limited to, the ambient sound collected by the error microphone combined with the sound processed by the earphone microphone to play the algorithm-processed sound and the comprehensive sound representation of the human ear environment.
  • the event sound is a sound that meets preset event conditions in the external environment
  • control the ANC function and the HT function to be in an open state at the same time; at least use the HT function and the ANC function to perform target processing on the first signal and the second signal to obtain the target signal; the signal-to-noise ratio of the event sound in the target signal is greater than that in the first signal. The signal-to-noise ratio of the event sound.
  • the HT function is used to transparently transmit the first signal collected by the reference microphone to obtain the restored signal C1; the event sound signal (for example, human voice) in the restored signal C1 is enhanced, and for the restored signal C1 The non-event sound signal is attenuated, and the event sound enhancement signal C2 is obtained;
  • the event sound signal for example, human voice
  • the target signal is obtained by processing the first signal C1, the signal collected by the error microphone, and the event sound enhancement signal C2 by using the ANC function.
  • the related enhancement processing may refer to but not limited to the embodiment of FIG. 9 .
  • the headset can also perform the following methods:
  • the first microphone is used to collect a first signal, and the first signal is used to represent the sound of the current external environment;
  • the second microphone is used to collect a second signal, and the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the headset;
  • the event sound is a sound that meets preset event conditions in the external environment
  • S5004 enable the ANC function; enhance the event sound signal in the first signal, and attenuate the non-event sound signal in the first signal to obtain an event sound enhancement signal; use the ANC function to The signal and the event sound enhancement signal are processed to obtain the target signal; the signal-to-noise ratio of the event sound in the target signal is greater than the signal-to-noise ratio of the event sound in the first signal; wherein, the higher the signal-to-noise ratio of the event sound in the signal, the Indicates that the greater the proportion of the energy of the event sound in the signal.
  • the related enhancement processing may refer to but not limited to the embodiment of FIG. 9 .
  • the earphone supports at least the ANC function, and the earphone includes a first microphone and a third microphone;
  • the first microphone here can be understood as the reference microphone in the foregoing embodiment, which is more focused on collecting the sound of the current external environment , the third microphone is more focused on sound pickup; when the user wears the headset, the third microphone is closer to the user's mouth than the first microphone, so the third microphone can pick up the user's voice more clearly than the first microphone. signal; the headset can also perform the following methods:
  • the headset enables or enables the ANC function
  • S6004 Determine the noise level of the current scene according to the first signal and the second signal; wherein, different noise levels correspond to different ANC intensities.
  • the correlation feature of the first signal and the second signal can be used for voice activity detection; noise tracking is performed on non-voice signals; if the noise energy is less than the first threshold, the current scene is determined as a quiet scene; or, if The spectral gravity center of the noise is located in the low frequency band, and the noise energy is greater than the second threshold, the current scene is determined to be a heavy noise scene; or, if the current scene is neither a quiet scene nor a heavy noise scene, the current scene is determined to be Ordinary scene; wherein, the second threshold (such as but not limited to a certain value in [-80dB, -65dB]) is greater than the first threshold (such as but not limited to a certain value in [-40dB, -30dB]).
  • the ANC noise reduction intensity corresponding to the quiet scene, the normal scene, and the heavy noise scene increases sequentially.
  • the ANC intensity can be preset with various intensity adjustment modes.
  • the filter of the ANC algorithm is adjusted according to the ANC noise reduction intensity corresponding to the noise level of the response, and the ANC function can be controlled.
  • the manner of controlling the ANC function may further include the methods for controlling the strength of the ANC as described in the above S3007 and S3008, which will not be repeated here.
  • Switching is allowed after the threshold is met for 1 second, that is, the newly detected noise level lasts for a preset period of time before switching to the new noise level; 2) The current noise reduction level is maintained for at least 10 seconds before switching is allowed A scenario, that is, the newly switched state is maintained for at least a second preset time; 3) There may be switching back and forth between heavy noise, normal or quiet noise levels, and if the switching is too fast, it will cause discomfort. Therefore, the switching frequency within a period of time can be monitored, and if abnormal, the threshold value of the judgment level can be raised. For example, within the preset time period, if the number of switching exceeds the preset number (for example, 4 times in 2 minutes), the threshold value favoring the normal mode is increased, the frequent switching of the mode is reduced, and the user experience is improved.
  • the preset number for example, 4 times in 2 minutes
  • the embodiments of the present application also provide a headset control device, which is applied to terminal equipment, and the terminal equipment establishes a communication connection with the headset; the headset supports the active noise reduction ANC function and the ambient sound transparent transmission HT function; the device includes:
  • a display module for displaying a first interface; the first interface is used to set the function of the headset; the first interface includes function options for event sound enhancement; the event sound is in the external environment that meets preset event conditions sound.
  • the first interface includes an option for controlling the HT function of the earphone; when the option of the HT function is enabled, the processing module is used for the HT function of the earphone; the display module is further used for: in the first interface Added an option to enhance the sound of said events in .
  • the processing module is configured to control both the ANC function and the HT function of the earphone to be in an enabled state when the option of the event sound enhancement function is enabled.
  • the processing module is further configured to activate the ANC function of the headset.
  • the embodiments of the present application also provide an earphone control device, which is applied to a terminal device, and the terminal device establishes a communication connection with the earphone; the earphone supports at least the active noise reduction ANC function; the device includes:
  • a display module for displaying a first interface; the first interface is used to set the function of the headset; the first interface includes options for controlling the ANC function of the headset;
  • the display module is further configured to add a strength option of the ANC function in the first interface after the option of the ANC function is enabled;
  • the processing module is further configured to perform ANC noise reduction according to the enabling result of the strength option of the ANC function;
  • the strength options of the ANC function include at least a first steady state ANC strength option, a second steady state ANC strength option, and an adaptive ANC strength option; wherein the first steady state ANC strength option and the second steady state ANC strength option
  • the state ANC strength options correspond to the first scene and the second scene respectively, and respectively correspond to different and stable ANC function strengths; the ANC function strength corresponding to the adaptive ANC strength option is the same as that of the terminal device or the headset. It is related to the scene type of the current environment at the location; different scene types of the current environment correspond to different ANC strengths.
  • the processing module is specifically configured to: when the first steady state ANC strength option is enabled, obtain the first ANC function strength corresponding to the first steady state ANC strength option; control the ANC function according to the first ANC function strength Or, when the second steady state ANC strength option is enabled, obtain the second ANC function strength corresponding to the second steady state ANC strength option; control the ANC function according to the second ANC function strength;
  • the adaptive ANC strength option is enabled, the scene type of the current environment where the terminal device or the headset is located is obtained; the ANC strength is determined according to the scene type of the current environment; the ANC function is controlled according to the determined ANC strength.
  • an embodiment of the present application also provides an earphone control device, the device is applied to a terminal device, and the terminal device establishes a communication connection with the earphone; the earphone supports at least an ambient sound transparent transmission HT function, and the device includes: :
  • a display module for displaying a first interface; the first interface is used to set the function of the headset; the first interface includes options for controlling the HT function of the headset;
  • a processing module configured to activate the HT function of the headset when the option of the HT function is enabled
  • the display module is further configured to add an option for enhancing the event sound in the first interface after the option of the HT function is enabled; the event sound is a sound that meets preset event conditions in the external environment;
  • the processing module is further configured to control the earphone to improve the signal-to-noise ratio of the event sound in the signal collected by the earphone when the option of the event sound enhancement function is enabled; A higher signal-to-noise ratio indicates a greater proportion of the energy of the event sound in the signal.
  • the processing module is further specifically configured to: obtain the first strength of the ANC function, and control the ANC function of the earphone according to the first strength; or obtain the second strength of the HT function, and control according to the second strength The HT function of the earphone; or, acquiring a third intensity of event sound enhancement, and controlling the event sound enhancement function of the earphone according to the third intensity.
  • the embodiment of the present application also provides a noise reduction device, which is applied to an earphone, the earphone supports at least the active noise reduction ANC function and the ambient sound transparent transmission HT function, and the earphone includes a first microphone, a third Two microphones and speakers; the device includes:
  • the acquisition module is used to collect a first signal through the first microphone, where the first signal is used to represent the sound of the current external environment; and is also used to collect a second signal through the second microphone, and the second signal is used for sound to characterize the internal environment of the ear canal of the user wearing the headset;
  • a receiving module for receiving an instruction of event sound enhancement;
  • the event sound is a sound that meets preset event conditions in the external environment;
  • a processing module configured to control both the ANC function and the HT function to be on, and at least use the HT function and the ANC function to perform target processing on the first signal and the second signal to obtain a target signal; the target The signal-to-noise ratio of the event sound in the signal is greater than the signal-to-noise ratio of the event sound in the first signal.
  • the processing module is specifically configured to: transparently transmit the first signal by using the HT function to obtain a restored signal; enhance the event sound signal in the restored signal, and attenuate the non-event sound signal in the restored signal, Obtain the event sound enhancement signal; use the ANC function to process the first signal, the second signal and the event sound enhancement signal to obtain the target signal.
  • Playing module playing the target signal through the speaker.
  • the embodiments of the present application also provide a noise reduction device, which is applied to an earphone, the earphone supports at least an active noise reduction ANC function, and the earphone includes a first microphone, a second microphone and a speaker; the The device includes:
  • the acquisition module is used to collect a first signal through the first microphone, where the first signal is used to represent the sound of the current external environment; and is also used to collect a second signal through the second microphone, and the second signal is used for sound to characterize the internal environment of the ear canal of the user wearing the headset;
  • a receiving module for receiving an instruction of event sound enhancement;
  • the event sound is a sound that meets preset event conditions in the external environment;
  • a processing module for enabling the ANC function; enhancing the event sound signal in the first signal, and weakening the non-event sound signal in the first signal to obtain an event sound enhancement signal; using the ANC function
  • the target signal is obtained by processing the first signal, the second signal and the event sound enhancement signal; the signal-to-noise ratio of the event sound in the target signal is greater than that of all the event sounds in the first signal.
  • the signal-to-noise ratio of the event sound wherein, the higher the signal-to-noise ratio of the event sound in the signal, the greater the energy ratio of the event sound in the signal;
  • Playing module playing the target signal through the speaker.
  • the embodiments of the present application also provide a noise reduction device, which is applied to an earphone, the earphone supports at least the active noise reduction ANC function, the ambient sound transparent transmission HT function and the AH function, and the earphone includes an HT filter A filter bank, a feedback filter bank and a feedforward filter bank; the device includes:
  • the acquisition module is used to acquire the working mode of the headset
  • a calling module for: when the working mode is the ANC function, calling the feedback filter bank and the feedforward filter bank to realize the ANC function; when the working mode is the HT function, calling the HT function The filter bank and the feedback filter bank realize the HT function; when the working mode is the AH function, the HT filter bank, the feedforward filter bank and the feedback filter bank are called, Implement the AH function.
  • an embodiment of the present application also provides a noise reduction device, which is applied to an earphone, and the earphone supports at least the ANC function; the earphone includes a first microphone and a third microphone; the first microphone is more focused In order to collect the sound of the current external environment, the third microphone focuses more on sound pickup;
  • a collection module configured to collect a first signal for the current environment through the first microphone when the headset enables the ANC function; collect a second signal for the current environment through the third microphone;
  • the identification module is configured to judge the noise level of the current scene according to the first signal and the second signal; wherein, different noise levels correspond to different ANC strengths.
  • the identification module is specifically configured to: use the correlation feature of the first signal and the second signal for voice activity detection; perform noise tracking on non-speech signals; if the noise energy is less than the first threshold, the current The scene is determined to be a quiet scene; or, if the spectral center of gravity of the noise is located in a low frequency band, and the noise energy is greater than the second threshold, the current scene is determined to be a heavy noise scene; or, if the current scene does not belong to the quiet scene When the scene does not belong to the heavy noise scene, the current scene is determined to be an ordinary scene; wherein, the second threshold is greater than the first threshold.
  • the processing module is used for controlling the ANC function according to the current noise level.
  • the processing module is specifically configured to: if it is detected that the current scene is a new noise level and lasts for a preset period of time; obtain the ANC intensity corresponding to the new noise level; according to the ANC corresponding to the new noise level Intensity controls the ANC function.
  • the method steps in the embodiments of the present application may be implemented in a hardware manner, or may be implemented in a manner in which a processor executes software instructions.
  • the software instructions can be composed of corresponding software modules, and the software modules can be stored in random access memory (RAM), flash memory, read-only memory (ROM), programmable read-only memory (programmable ROM, PROM), erasable programmable read-only memory (erasable PROM, EPROM), electrically erasable programmable read-only memory (electrically ePROM, EEPROM), registers, hard disk, removable hard disk, CD-ROM or any known in the art other forms of storage media.
  • An exemplary storage medium is coupled to the processor, such that the processor can read information from, and write information to, the storage medium.
  • the storage medium can also be an integral part of the processor.
  • the processor and storage medium may reside in an ASIC. Alternatively, the ASIC may be located in the end device. Of course, the processor and the storage medium may also exist in the terminal device as
  • the above-mentioned embodiments it may be implemented in whole or in part by software, hardware, firmware or any combination thereof.
  • software it can be implemented in whole or in part in the form of a computer program product.
  • the computer program product includes one or more computer programs or instructions.
  • the processes or functions described in the embodiments of the present application are executed in whole or in part.
  • the computer may be a general purpose computer, a special purpose computer, a computer network, user equipment, or other programmable apparatus.
  • the computer program or instructions may be stored in a computer-readable storage medium or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer program or instructions may be downloaded from a website site, computer, A server or data center transmits by wire or wireless to another website site, computer, server or data center.
  • the computer-readable storage medium may be any available medium that can be accessed by a computer, or a data storage device such as a server, data center, or the like that integrates one or more available media.
  • the usable medium can be a magnetic medium, such as a floppy disk, a hard disk, and a magnetic tape; it can also be an optical medium, such as a digital video disc (DVD); it can also be a semiconductor medium, such as a solid state drive (solid state drive). , SSD).
  • a magnetic medium such as a floppy disk, a hard disk, and a magnetic tape
  • an optical medium such as a digital video disc (DVD)
  • DVD digital video disc
  • it can also be a semiconductor medium, such as a solid state drive (solid state drive). , SSD).

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Otolaryngology (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Neurosurgery (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Telephone Function (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • User Interface Of Digital Computer (AREA)
  • Soundproofing, Sound Blocking, And Sound Damping (AREA)

Abstract

本申请实施例提供一种模式控制方法、装置及终端设备,以实现根据用户需求实现用户想要达到的效果。终端设备根据外部环境的场景类型来确定耳机采用的处理模式为目标模式,所述目标模式为耳机支持的处理模式中的一个,不同的处理模式对应于不同的场景类型,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种。目标模式下的处理强度可以根据当前外部环境中的事件来实现自动调节。另外,终端设备还可以耳机提供控制界面,为用户提供选择处理模式和处理强度的选择控件,用户可以根据需求对耳机的处理模式和处理强度进行控制。

Description

一种模式控制方法、装置及终端设备 技术领域
本申请实施例涉及音频处理技术领域,尤其涉及一种模式控制方法、装置及终端设备。
背景技术
近年来,耳机用户越来越多,用户对耳机的功能诉求也越来越有差异化。比如有的用户在佩戴耳机时不想听见外界噪声,可以通过主动降噪(active noise control,ANC)功能对耳朵的噪声进行消除。有的用户想听见耳机外的声音,需要通过环境声透传(hear through,HT)功能来实现与不带耳机一样感受外界声音。有的用户可能听觉有障碍,可以通过听觉增强功能(augment hearing,AH)将用户想要的外界信号输送给用户,同时滤除不想要的信号。
但是目前耳机并不能根据用户需求来实现用户想要达到的效果。
发明内容
本申请实施例提供一种耳机噪声处理方法、装置及耳机,以实现根据用户需求实现用户想要达到的效果。
第一方面,本申请实施例提供一种耳机噪声处理方法,所述耳机具备主动降噪ANC功能、环境声透传HT功能或听觉增强AH功能中至少两个功能,所述耳机包括第一麦克风和第二麦克风,第一麦克风用于采集第一信号,所述第一信号用于表征当前外部环境的声音,所述第二麦克风用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音。耳机可以为左耳机或者右耳机。左耳机和右耳机采用的处理模式可以相同或者不同。耳机接收来自所述终端设备的第一音频信号;获取目标模式;其中,所述目标模式是基于当前外部环境的场景类型确定的,所述目标模式用于指示所述耳机实现目标处理功能,所述目标处理功能为主动降噪ANC功能、环境声透传HT功能或者听觉增强AH功能中一个;根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
通过上述方法,通过场外部环境的场景类型来确定目标模式,能够实时优化用户的听感效果。
在一种可能的设计中,所述耳机还包括扬声器,所述扬声器用于播放所述第二音频信号。
在一种可能的设计中,所述目标处理功能为ANC功能,所述扬声器播放的所述第二音频信号,能够减弱用户对所述用户当前所处环境的声音以及所述用户耳道内部的环境声音的感知;或者,所述目标处理功能为HT功能,所述扬声器播放的所述第二音频信号,能够强化用户对所述用户当前所处环境的声音的感知;或者,所述目标处理功能为AH功能,所述扬声器播放的所述第二音频信号,能够强化用户对事件声音的感知;所述事件声音满足预设频谱。
应理解的是,左耳机采用ANC模式时,左耳机的扬声器播放的音频信号,能够减弱用户左耳机对用户当前所处环境的声音(即当前外部环境的声音)以及用户左耳机耳道内部的环境声音的感知。右耳机采用ANC模式时,右耳机的扬声器播放的音频信号,能够减弱用户右 耳机对用户当前所处环境的声音(即当前外部环境的声音)以及用户右耳机耳道内部的环境声音的感知。同理,针对HT和AH模式来说,左耳朵感受跟随左耳机采用的处理模式,右耳朵的感受跟随右耳机采用的处理模式。
在一种可能的设计中,当所述目标处理功能为ANC功能时,第二音频信号是基于第一音频信号、第三信号以及第四信号得到的,所述第三信号为第一信号的反相信号,所述第四信号为所述第二信号的反相信号;或者,当所述目标处理功能为HT功能时,第二音频信号是基于第一音频信号、第一信号以及第二信号得到的;或者,当所述目标处理功能为AH功能时,第二音频信号是基于第一音频信号、第五信号以及第四信号得到的,所述第五信号为所述第一信号中的事件信号,所述事件信号用于表征当前外部环境中的特定声音,所述事件信号满足预设频谱。
上述设计提供在不同的处理模式下,扬声器所输出的信号的获得方式,简单有效。
在一种可能的设计中,所述获取目标模式,包括:
接收来自所述终端设备的第一控制指令,所述第一控制指令携带所述目标模式,所述目标模式是所述终端设备根据当前外部环境的场景类型确定的。
上述设计中,终端设备通过外部环境的场景类型来确定目标模式并向耳机指示,能够实时优化用户的听感效果。
在一种可能的设计中,接收来自所述终端设备的第二控制指令,所述第二控制指令携带目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度;根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号,包括:根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
通过上述设计,由终端设备指示耳机相应的处理模式下的处理强度。在处理模式的基础上调整处理强度进一步提高用户的听觉感受。
在一种可能的设计中,根据所述第一信号确定当前外部环境中事件声音对应的目标事件,根据所述目标事件确定在所述目标模式下的目标处理强度;其中,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度;根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号,包括:根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。其中不同的处理强度对应于不同事件。处理强度可以与事件为一一对应的,或者一个处理强度对应于多个事件,比如某两个事件可采用同一种处理强度,同一事件不可采用不同的处理强度。
通过上述设计,耳机根据外部环境中的事件声音来确定处理强度,实现在不同的外部环境下不同的听觉感受,可以降低底噪感,增强降噪力度。
在一种可能的设计中,所述耳机还包括骨传导传感器,所述骨传导传感器用于采集所述用户声带震动产生的骨传导信号;所述根据所述第一信号识别所述用户当前所处的第一场景,包括:根据所述第一信号和所述骨传导信号识别所述用户当前所处的第一场景。
在一种可能的设计中,所述目标事件为啸叫事件、风噪事件、突发事件或者人声事件。
在一种可能的设计中,所述获取目标模式,包括:根据所述第一信号识别当前外部环境的场景类型为目标场景类型(简称目标场景,或者简称为目标类型),根据所述目标场景确定所述耳机采用的所述目标模式,所述目标模式是对应于所述目标场景的处理模式。其中不同的处理模式对应于不同的场景类型。处理模式可以与场景类型为一一对应的,或者一个处理模式可以对应于多个场景类型,比如某两个场景类型可采用同一种处理模式。
上述设计中,耳机根据识别到的场景类型来确定耳机采用的处理模式,减少时延,并且到达实时优化用户的听觉感受。
在一种可能的设计中,所述目标场景为走路场景、跑步场景、安静场景、多人说话场景、咖啡厅场景、地铁场景、火车场景、候车厅场景、对话场景、办公室场景、户外场景、开车场景、大风场景、飞机场景、警报声场景、鸣笛声场景、哭叫声场景中的一项。
在一种可能的设计中,还包括:
向所述终端设备发送指示信息,所述指示信息携带所述目标模式;接收来自所述终端设备的第三控制信令,所述第三控制信令包括所述目标模式下的目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度;根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号,包括:根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
上述设计中,耳机来确定采用的处理模式,并且指示给终端设备,由终端设备来实现调节处理强度,降低耳机的处理资源的占用。
在一种可能的设计中,当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
在一种可能的设计中,所述目标模式指示所述耳机实现ANC功能,根据所述目标模式、所述第一音频信号、第一信号、以及第二信号得到第二音频信号,包括:
对所述第一信号进行所述第一滤波处理(比如,前馈FF滤波)得到第一滤波信号;
滤除所述第二信号中包括的第一音频信号得到第一经滤除信号;
对所述第一滤波信号、所述经滤除信号进行混音处理得到第三音频信号;
对所述第三音频信号进行所述第三滤波处理(比如,反馈FB滤波)得到第四音频信号;
对所述第四音频信号以及所述第一音频信号进行混音处理得到所述第二音频信号。
上述设计中,采用FF滤波和FB串行处理的方式来实现ANC处理,得到更好降噪后的信号,增强降噪效果。
在一种可能的设计中,所述第一滤波处理采用的滤波系数为所述ANC功能下,针对所述第一滤波处理,所述目标处理强度所关联的滤波系数;或者,所述第三滤波处理采用的滤波系数为所述ANC功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
在上述设计中,在不同的处理强度下采用不同的滤波系数,相较采用固定滤波系数来说,达到更优的ANC效果,提高用户的听觉感受。
在一种可能的设计中,所述目标模式指示所述耳机实现HT功能,根据所述目标模式、第一音频信号、第一信号、以及第二信号得到第二音频信号,包括:
对所述第一信号进行第一信号处理得到第一经处理信号,所述第一信号处理包括第二滤波处理(比如HT滤波);
对所述第一经处理信号和所述第一音频信号进行混音处理得到第五音频信号;
滤波所述第二信号包括的所述第五音频信号得到第二经滤除信号;
对所述第二经滤除信号进行第三滤波处理(比如FB滤波)得到第三经滤除信号;
将所述第三经滤除信号与所述第五音频信号进行混音处理得到所述第二音频信号。
其中,滤波所述第二信号包括的所述第五音频信号之前,还可以对第五音频信号进行滤波补充处理,减少听感损失。上述设计中,HT滤波时经过下行混音处理,并进行滤波补偿处理,进一步减少听感损失。
在一种可能的设计中,对所述第一环境信号进行第一信号处理得到经处理环境信号,包括:对所述第一信号进行第二滤波处理得到第二滤波信号;
对所述第二滤波信号进行第二信号处理得到第二经处理信号;
所述第二信号处理包括去闭塞效应处理。
上述设计,对HT滤波后的信号进行去闭塞效应处理能够使用听到的环境声音更清晰。
在一种可能的设计中,所述第二信号处理还包括如下至少一项:降底噪处理、降风噪处理、增益调节处理或者频响调节处理。
经过上述第二信号处理后,减少底噪、减少异常音,提高用户的听觉感受。
在一种可能的设计中,所述第二滤波处理采用的滤波系数为所述HT功能下,针对所述第二滤波处理,所述目标处理强度所关联的滤波系数;或者,
所述第三滤波处理采用的滤波系数为所述HT功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
在一种可能的设计中,目标模式指示所述耳机实现AH功能,根据所述目标模式、所述第一音频信号、第一信号、以及第二信号得到第二音频信号,包括:
对所述第一信号进行第二滤波处理(比如HT滤波)得到第二滤波信号,并对所述第二滤波信号进行增强处理得到滤波增强信号;
对所述第一信号进行第一滤波处理(比如FF滤波)得到第一滤波信号;
对所述滤波增强信号和所述第一音频信号进行混音处理得到第六音频信号;
滤波所述第二信号包括的所述第六音频信号得到第四经滤除信号;
对所述第四经滤除信号进行第三滤波处理(比如FB滤波)得到第五经滤除信号;
对所述第五经滤除信号、所述第六音频信号以及所述第一滤波信号进行混音处理得到所述第二音频信号。
上述设计中,主动降噪与环境声透出并行实现,透传的信号经过透传滤波处理以及增强处理,透传的信号更清晰。
可选地,滤波所述第二信号包括的所述第六音频信号得到第四经滤除信号之前,对第六音频信号进行滤波补偿处理,可以避免FB滤波带来的损失,最大程度保证透传信号的不失真。
在一种可能的设计中,对所述第二滤波信号进行增强处理得到滤波增强信号,包括:
对所述第二滤波信号进行去闭塞效应处理,并对去闭塞效应处理后的信号进行降噪处理,所述降噪处理包括人工智能AI降噪处理和/或降风噪处理;
对所述降噪处理后得到的信号进行增益放大处理以及频响调节得到所述滤波增强信号。
上述设计中,对透传的信号进行增强处理。提高用户对需要的外界声音的听感。
在一种可能的设计中,所述耳机中包括骨传导传感器,所述骨传导传感器用于采集所述耳机用户的骨传导信号,所述对降噪处理后得到的信号进行增益放大处理,包括:对所述骨传导信号进行谐波扩展得到经谐波扩展信号;使用第一增益系数对所述降噪处理得到的信号进行放大处理;使用第四滤波系数滤除所述放大处理得到的信号中包括的经谐波扩展信号;其中,所述第四滤波系数基于所述第一增益系数确定。
上述设计中,提供一种放大方式,仅放大佩戴用户说话声音以外的特定声音,提高透传 的环境声中的特定声音的效果。
在一种可能的设计中,所述第一增益系数为所述目标模式下,所述目标处理强度所关联的增益系数。
在一种可能的设计中,对所述第二滤波信号进行增强处理得到滤波增强信号,包括:
对所述第二滤波信号进行去闭塞效应处理得到经去闭塞信号;
对所述经去闭塞信号进行音频事件检测,以获取所述经去闭塞信号中的音频事件信号;
对所述经去闭塞信号中的音频事件信号进行增益放大处理以及频响调节得到滤波增强信号。
在一种可能的设计中,所述耳机中还包括骨传导传感器,所述骨传导传感器用于采集所述耳机用户的骨传导信号,对所述经去闭塞信号中的音频事件信号进行增益放大处理,包括:对所述骨传导信号进行谐波扩展得到经谐波扩展信号;使用第二增益系数对所述经去闭塞信号中的音频事件信号进行放大得到经放大信号;使用第二滤波系数滤除所述经放大信号中包括的经谐波扩展信号;其中,所述第二滤波系数基于所述第二增益系数确定。
在一种可能的设计中,所述第二增益系数为执行所述第一噪声处理的情况下,针对所述第一滤波处理,所述目标处理强度所关联的增益系数;或者,
所述第二增益系数为执行所述第一噪声处理的情况下,针对所述第一滤波处理,所述第一场景标识所关联的增益系数。
在一种可能的设计中所述第一滤波处理采用的滤波系数为所述AH功能下,针对所述第一滤波处理,所述目标处理强度所关联的滤波系数;或者,
所述第二滤波处理采用的滤波系数为所述AH功能下,针对所述第二滤波处理,所述目标处理强度所关联的滤波系数;或者,
所述第三滤波处理采用的滤波系数为所述AH功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
在一种可能的设计中,所述耳机还包括骨传导传感器,所述骨传导传感器用于采集所述耳机的用户的骨传导信号;对所述第二滤波信号进行去闭塞效应处理,包括:
在语音谐波集合中确定与所述骨传导信号匹配的第一语音谐波信号,所述语音谐波集合中包括多个语音谐波信号;去除所述第二滤波信号中的所述第一语音谐波信号,将去除所述第一语音谐波信号的第二滤波信号中的高频分量进行放大;或者,对所述第二滤波信号进行自适应滤波处理,以去除所述第二滤波信号中的低频分量得到第三滤波信号,将去除所述低频分量的第三滤波信号中的高频分量进行放大。
第二方面,本发明实施例提供一种模式控制方法,所述方法应用于终端设备,所述方法包括:当识别当前外部环境的场景类型为目标场景时,根据所述目标场景确定目标模式;其中,所述目标模式为耳机支持的处理模式中的一个,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;向所述耳机发送所述目标模式,所述目标模式用于指示所述耳机实现所述目标模式所对应的处理功能。不同的处理模式对应于不同的场景类型,处理模式可以与场景类型为一一对应的,或者一个处理模式可以对应于多个场景类型,比如某两个场景类型可采用同一种处理模式。
上述设计中,终端设备根据场景识别来实时控制耳机的处理模式,从而达到实时优化用户的听觉感受。
在一种可能的设计中,确定耳机的处理模式中对应于所述目标场景的目标模式时,所述方法还包括:显示结果提示信息,所述结果提示信息用于提示用户所述耳机实现所述目标模 式所对应的处理功能。上述设计,使得用户能够实时确定耳机当前的处理模式。
在一种可能的设计中,向所述耳机发送第一控制信令之前,所述方法还包括:显示选择提示信息,所述选择提示信息用于提示用户是否调整所述耳机的处理模式为所述目标模式;检测到用户选择调整所述耳机的处理模式为所述目标模式的操作。
上述设计中,用户可以根据需求确定是否调整耳机的处理模式,提高用户感受。
在一种可能的设计中,显示第一控件和第二控件,其中,所述第二控件在所述第一控件上的不同位置用于指示所述目标模式下的不同处理强度;向所述耳机发送第一控制信令之前,所述方法还包括:响应于用户触控所述第二控件移动到所述第一控件上的第一位置,所述第二控件在所述第一控件上的第一位置指示所述目标模式下的目标处理强度;向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。上述设计中,用户可以根据需要选择耳机的处理强度,满足用户的不同需求。
在一种可能的设计中,所述第一控件的形状为圆环,所述用户触控所述第二控件按顺时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件按逆时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
在一种可能的设计中,所述第一控件的形状为条形,所述用户触控所述第二控件从上向下在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件从下向上在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件从左向右在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件从右向左在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
在一种可能的设计中,当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱。或者,当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
需要说明的是,左耳机与右耳机可以采用相同的处理模式以及处理强度,则用户左耳朵和右耳朵的感知可以相同。左耳机和右耳机还可以采用不同的处理模式或者不同的处理强度,则左耳朵和右耳朵的感知不同。
第三方面,本发明实施例提供一种模式控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:获取目标模式;所述目标模式为耳机支持的处理模式中的一个,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;根据当前外部环境的场景类型确定所述目标模式下的目标处理强度;不同的场景类型对应于目标模式下的不同处理强度;向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
在一种可能的设计中,所述获取目标模式,包括:接收所述耳机发送的所述目标模式;或者,显示选择控件,所述选择控件包括耳机支持的处理模式,检测到用户通过所述选择控件在所述耳机的处理模式中选择所述目标模式的操作。所述选择控件包括耳机支持的处理模式,或者解释为选择控件用于提供耳机支持的处理模式的选项,或者选择控件上显示有耳机支持的处理模式,并且用户可以在耳机支持的处理模式中进行选择。。
在一种可能的设计中,根据当前外部环境的场景类型确定所述目标模式下的目标处理强度之前,所述方法还包括:接收到所述耳机发送的所述目标模式的情况下,显示选择提示信息,所述选择提示信息用于指示所述用户是否调整所述耳机的处理模式为所述目标模式;检测到用户选择调整所述耳机的处理模式为所述目标模式的操作。
在一种可能的设计中当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,
当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,
当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
第四方面,本申请实施例提供一种模式控制方法,所述方法应用于终端设备,所述方法包括:显示第一界面;第一界面上包括第一选择控件,所述第一选择控件包括第一目标耳机支持的处理模式以及第一目标耳机支持的处理模式对应的处理强度;所述第一目标耳机的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;响应于用户针对所述第一界面执行的第一操作;所述第一操作是用户通过所述第一选择控件在第一目标耳机支持的处理模式中选择第一目标模式以及选择第一目标模式下的处理强度为第一目标处理强度产生的;向所述第一目标耳机发送所述第一目标模式以及所述第一目标处理强度,所述第一目标模式用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能,所述第一目标处理强度用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能时的处理强度。
所述第一选择控件包括第一目标耳机支持的处理模式以及第一目标耳机支持的处理模式对应的处理强度,可以解释为第一选择控件为用户提供多个处理模式(均为第一目标耳机所支持的)的选项以及每个处理模式下处理强度的调整项。
上述设计中,用户可以通过UI界面自如的切换需要耳机达到的效果对应的处理模式以及力度,满足用户的不同需求。
在一种可能的设计中,显示第一界面之前,所述方法还包括:显示选择提示信息,所述选择提示信息用于所述用户选择是否调整所述第一目标耳机的处理模式;检测到用户选择调整所述第一目标耳机的处理模式的操作。
上述设计中,用户可以根据需求来决定是否调整当前的处理模式。
在一种可能的设计中,显示所述第一界面之前,所述方法还包括:识别当前外部环境的场景类型为目标场景,所述目标场景适配需要调整所述第一目标耳机的处理模式的场景类型。
上述设计中,提供一种在特定的场景下主动弹出第一界面,减少用户的手动操作过程。
在一种可能的设计中,显示所述第一界面之前,所述方法还包括:识别到所述终端设备触发所述第一目标耳机播放音频。识别到所述终端设备触发所述第一目标耳机播放音频,可以解释为识别终端设备启动向第一目标耳机发送音频信号。
上述设计中,提供一种主动弹出第一界面的方式,减少用户的手动操作过程。
在一种可能的设计中,显示所述第一界面之前,所述方法还包括:检测到所述终端设备与所述第一目标耳机建立连接。
上述设计中,提供一种主动弹出第一界面的方式,减少用户的手动操作过程。
在一种可能的设计中,显示所述第一界面之前,所述方法还包括:检测到所述终端设备与所述第一目标耳机建立连接的情况下,检测到所述用户在主界面执行的第二操作;其中, 所述主界面包括第一应用的图标,所述第二操作是所述用户触控所述第一应用的图标产生的,所述第一界面为所述第一应用的显示界面。
在一种可能的设计中,所述第一选择控件包括第一控件和第二控件,第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的两种不同处理模式,或者,所述第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的的同一种处理模式下的不同处理强度;所述第一操作是所述用户移动所述第二控件位于所述第一控件上所述第一目标模式对应的区域内的第一位置产生的,所述第一位置对应于所述第一目标模式下的第一目标处理强度。
在一种可能的设计中,所述第一控件的形状为圆环或者条形。
比如,所述第一控件的形状为圆环,所述圆环包括至少两个圆弧段,所述第二控件位于不同的圆弧段指示第一目标耳机的不同处理模式,所述第二控件位于同一圆弧段的不同位置指示第一目标耳机的同一处理模式的不同处理强度。
再比如,所述第一控件的形状为条形,所述条形包括至少两个条形段,所述第二控件位于不同的条形段指示所述第一目标耳机的不同处理模式,所述第二控件位于同一条形段的不同位置指示所述第一目标耳机的同一处理模式的不同处理强度。
在一种可能的设计中,所述方法还包括:
响应于用户在第一界面执行的第三操作,所述第一界面上还包括第二选择控件,所述第二选择控件包括第二目标耳机支持的处理模式以及第二目标耳机支持的处理模式对应的处理强度,所述第一目标耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种,所述第三操作是所述用户通过所述第二选择控件在第二目标耳机的处理模式中选择第二目标模式以及选择第二目标模式下的处理强度为第二目标处理强度产生的;所述第一目标耳机为左耳机时,所述第二目标耳机为右耳机,或者,所述第一目标耳机为右耳机,所述第二目标耳机为左耳机;
向所述第二目标耳机发送所述第二目标模式以及所述第二目标处理强度,所述第二目标模式用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能,所述第二目标处理强度用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能时的处理强度。
上述设计中,用户可以分开操作左耳机和右耳机的处理模式以及处理强度,满足用户对左耳朵和右耳朵听感的差异化需求。
第五方面,本申请实施例还提供一种耳机控制方法,方法应用于终端设备,方法包括:所述终端设备与耳机建立通信连接;显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括事件声音增强的功能选项;所述事件声音为外部环境中符合预设事件条件的声音;当所述事件声音增强功能的选项使能时,控制所述耳机的ANC功能和HT功能均处于开启状态。
第六方面,本申请实施例还提供一种耳机控制装置,装置应用于终端设备,装置包括:
显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括事件声音增强的功能选项;所述事件声音为外部环境中符合预设事件条件的声音;
处理模块,用于当所述事件声音增强功能的选项使能时,控制所述耳机的ANC功能和HT功能均处于开启状态。
根据第五方面或第六方面,在一种可能的设计中,第一界面中包括用于控制所述耳机HT功能的选项;当所述HT功能的选项使能时,激活所述耳机的HT功能;并在所述第一界面中新增所述事件声音增强的选项。进一步可选地,还可以获取HT功能的目标强度;根据所述HT功能的目标强度控制所述耳机的HT功能。该步骤可以由显示模块和处理模块来配合 执行。
根据第五方面或第六方面,在一种可能的设计中,控制所述耳机的ANC功能和HT功能均处于开启状态包括:保持所述HT功能处于开启状态;激活所述耳机的ANC功能。该步骤可以由处理模块执行。
根据第五方面或第六方面,在一种可能的设计中,所述第一界面中包括用于控制所述耳机ANC功能的选项;当所述ANC功能的选项使能时,激活所述耳机的ANC功能。进一步可选地,在所述第一界面中新增所述ANC功能的强度选项。该步骤可以由处理模块执行。
根据第五方面或第六方面,在一种可能的设计中,所述第一界面中还包括用于关闭所述ANC功能的选项,和/或关闭所述HT功能的选项。
根据第五方面或第六方面,在一种可能的设计中,ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,所述第一稳态ANC强度选项和所述第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的ANC功能强度;所述自适应ANC强度选项对应的ANC功能强度与所述终端设备或所述耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。
根据第五方面或第六方面,在一种可能的设计中,当所述ANC强度自适应选项被使能时,获取所述终端或所述耳机所处当前环境的场景类型;根据所述场景类型匹配出ANC功能的目标强度;根据所述目标强度控制所述耳机的ANC功能。所述当前环境不同的场景类型包括第一场景和第二场景。
根据第五方面或第六方面,在一种可能的设计中,所述事件声音包括人声或者符合预设频谱特性的其它声音。
根据第五方面或第六方面,在一种可能的设计中,所述事件声音增强功能的选项使能、所述HT功能的选项使能或所述ANC功能的选项使能包括:响应用户针对相应功能选项的点选、相应功能的自适应切换,或者相应功能的快捷方式触发。
第七方面,本申请实施例还提供一种降噪方法,所述方法应用于耳机,所述耳机至少支持主动降噪ANC功能,还可以支持环境声透传HT功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述方法包括:通过所述第一麦克风采集第一信号,所述第一信号用于表征当前外部环境的声音;通过所述第二麦克风采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;控制ANC功能开启,即使能或开启ANC功能,至少利用所述ANC功能对所述第一信号和所述第二信号进行目标处理得到目标信号;所述目标信号中的事件声音的信噪比大于所述第一信号中的事件声音的信噪比;通过所述扬声器播放所述目标信号。
第八方面,本申请实施例还提供一种降噪装置,所述装置应用于耳机,耳机至少支持主动降噪ANC功能,还可以支持环境声透传HT功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述装置包括:采集模块,用于通过所述第一麦克风采集第一信号,所述第一信号用于表征当前外部环境的声音;还用于通过所述第二麦克风采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;接收模块,用于接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;处理模块,用于使能ANC功能,并至少利用所述ANC功能对所述第一信号和所述第二信号进行目标处理得到目标信号;所述目标信号中的事件声音的信噪比大于所述第一信号中的事件声音的信噪比;播放模块,通过 所述扬声器播放所述目标信号。
根据第七方面或第八方面,在一种可能的设计中,控制ANC功能和HT功能均处于开启状态;利用HT功能透传所述第一信号得到还原信号;针对所述还原信号中的事件声音信号进行增强,并对于所述还原信号中的非事件声音信号进行减弱,得到事件声音增强信号;利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号。
根据第七方面或第八方面,在一种可能的设计中,利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号。
根据第七方面或第八方面,在一种可能的设计中,耳机至少支持主动降噪ANC功能、环境声透传HT功能和AH功能,所述耳机包括HT滤波器组、反馈滤波器组和前馈滤波器组;获取耳机的工作模式;当所述工作模式为ANC功能时,调用所述反馈滤波器组和前馈滤波器组,实现所述ANC功能;当所述工作模式为HT功能时,调用所述HT滤波器组和所述反馈滤波器组,实现所述HT功能;当所述工作模式为AH功能时,调用所述HT滤波器组、所述前馈滤波器组和所述反馈滤波器组,实现所述AH功能。
第九方面,本申请实施例还提供一种降噪方法,所述方法应用于耳机,所述耳机至少支持ANC功能;所述耳机包括第一麦克风、第三麦克风;所述第一麦克风更侧重于采集当前外部环境的声音,所述第三麦克风更侧重于拾音;当所述耳机使能所述ANC功能时,通过所述第一麦克风针对当前环境采集第一信号;通过所述第三麦克风针对所述当前环境采集第二信号;根据所述第一信号和所述第二信号判断所述当前场景的噪声等级;其中,不同的噪声等级对应于不同的ANC强度;根据当前的噪声等级控制所述ANC功能。
第十方面,本申请实施例还提供一种降噪装置,所述装置应用于耳机,所述耳机至少支持ANC功能;所述耳机包括第一麦克风、第三麦克风;所述第一麦克风更侧重于采集当前外部环境的声音,所述第三麦克风更侧重于拾音;采集模块,用于当所述耳机的所述ANC功能处于开启状态时,通过所述第一麦克风针对当前环境采集第一信号;通过所述第三麦克风针对所述当前环境采集第二信号;识别模块,用于根据所述第一信号和所述第二信号判断所述当前场景的噪声等级;其中,不同的噪声等级对应于不同的ANC强度;处理模块,用于根据当前的噪声等级控制所述ANC功能。
根据第九方面或第十方面,在一种可能的设计中,利用第一信号和第二信号的相关性特征作语音活动检测;对非语音信号进行噪声跟踪;若所述噪声能量小于第一阈值,则将当前场景确定为安静场景;或者,若所述噪声的频谱重心位于低频段,且噪声能量大于第二阈值,则将当前场景确定为重噪场景;或者,若所述当前场景既不属于所述安静场景也不属于所述重噪场景时,则将当前场景确定为普通场景;其中,所述第二阈值大于所述第一阈值。所述安静场景、所述普通场景、所述重噪场景对应的ANC强度依次变大。
根据第九方面或第十方面,在一种可能的设计中,若检测到当前场景为新的噪声等级且持续预设时长;获取所述新的噪声等级对应的ANC强度;根据所述新的噪声等级对应的ANC强度控制所述ANC功能。
第十一方面,本申请实施例还提供一种耳机控制方法,所述方法应用于终端设备,所述方法包括:所述终端设备与耳机建立通信连接;所述耳机至少支持主动降噪ANC功能;显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机ANC 功能的选项;当所述ANC功能的选项使能时,激活所述耳机的ANC功能;在所述第一界面中新增所述ANC功能的强度选项;根据所述ANC功能的强度选项的使能结果进行ANC降噪;其中,所述ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,所述第一稳态ANC强度选项和所述第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的且稳定的ANC功能强度;所述自适应ANC强度选项对应的ANC功能强度与所述终端设备或所述耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。可选地,所述当前环境不同的场景类型包括第一场景和第二场景。
第十二方面,本申请实施例还提供一种耳机控制装置,所述终端设备与耳机建立通信连接;所述耳机至少支持主动降噪ANC功能;所述装置包括:显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机ANC功能的选项;处理模块,用于当所述ANC功能的选项使能时,激活所述耳机的ANC功能;所述显示模块还用于在所述ANC功能的选项使能之后在所述第一界面中新增所述ANC功能的强度选项;所述处理模块还用于根据所述ANC功能的强度选项的使能结果进行ANC降噪;其中,所述ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,所述第一稳态ANC强度选项和所述第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的且稳定的ANC功能强度;所述自适应ANC强度选项对应的ANC功能强度与所述终端设备或所述耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。
根据第十一方面或第十二方面,在一种可能的设计中,当第一稳态ANC强度选项使能时,获取第一稳态的ANC强度选项对应的第一ANC功能强度;根据所述第一ANC功能强度控制所述ANC功能;或者,当第二稳态ANC强度选项使能时,获取第二稳态的ANC强度选项对应的第二ANC功能强度;根据所述第二ANC功能强度控制所述ANC功能。当所述自适应ANC强度选项使能时,获取所述终端设备或所述耳机所处当前环境的场景类型;根据当前环境的场景类型确定ANC强度;根据确定出来的ANC强度控制所述ANC功能。
第十三方面,本申请实施例还提供一种耳机控制方法,所述方法应用于终端设备,所述方法包括:所述终端设备与耳机建立通信连接;所述耳机至少支持环境声透传HT功能,显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机HT功能的选项;当所述HT功能的选项使能时,激活所述耳机的HT功能;在所述第一界面中新增所述事件声音增强的选项;所述事件声音为外部环境中符合预设事件条件的声音;当所述事件声音增强功能的选项使能时,控制所述耳机针对所述耳机采集的信号中所述事件声音的信噪比进行提升;其中,所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大。
第十四方面,本申请实施例还提供一种耳机控制装置,所述装置应用于终端设备,所述终端设备与耳机建立通信连接;所述耳机至少支持环境声透传HT功能,所述装置包括:显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机HT功能的选项;处理模块,用于当所述HT功能的选项使能时,激活所述耳机的HT功能;显示模块还用于在所述HT功能的选项使能后,在所述第一界面中新增所述事件声音增强的选项;所述事件声音为外部环境中符合预设事件条件的声音;所述处理模块还用于当所述事件声音增强功能的选项使能时,控制所述耳机针对所述耳机采集的信号中所 述事件声音的信噪比进行提升;其中,所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大。
根据第十三方面或第十四方面,在一种可能的设计中,所述事件声音包括人声或者符合预设频谱特性的其它声音。可选地,所述第一界面中包括用于控制所述耳机ANC功能的选项、用于关闭所述ANC功能的选项,和/或关闭所述HT功能的选项。
根据第十三方面或第十四方面,在一种可能的设计中,获取ANC功能的第一强度,并根据所述第一强度控制所述耳机的ANC功能;或者,获取HT功能的第二强度,并根据所述第二强度控制所述耳机的HT功能;或者,获取事件声音增强的第三强度,并根据所述第三强度控制所述耳机的事件声音增强功能。
第十五方面,本申请实施例还提供一种噪声处理装置,所述装置应用于耳机,所述耳机具备主动降噪ANC功能、环境声透传HT功能或听觉增强AH功能中至少两个功能;所述耳机包括第一麦克风、第二麦克风;第一麦克风,用于采集第一信号,所述第一信号用于表征当前外部环境的声音;所述第二麦克风,用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音。
噪声处理装置包括相应的功能模块,分别用于实现以上第一方面的方法中的步骤,具体参见方法示例中的详细描述,此处不做赘述。功能可以通过硬件实现,也可以通过硬件执行相应的软件实现。硬件或软件包括一个或多个与上述功能相对应的模块。例如,噪声处理装置包括:
通信模块,用于接收来自所述终端设备的第一音频信号;
获取模块,用于获取目标模式;其中,所述目标模式是基于当前外部环境的场景类型确定的,所述目标模式用于指示所述耳机实现目标处理功能,所述目标处理功能为主动降噪ANC功能、环境声透传HT功能或者听觉增强AH功能中一个;
第一处理模块,用于根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
第十六方面,本申请实施例提供一种目标耳机,包括左耳机和右耳机,左耳机用于实现上述任一种可能的关于耳机的设计方法,或者,右耳机用于实现上述任一种可能的关于耳机的设计方法。
在一种可能的设计中,左耳机和右耳机采用的处理模式不同。
第十七方面,本申请实施例提供一种目标耳机,所述目标耳机包括左耳机和右耳机。所述左耳机包括或者右耳机包括第一麦克风、第二麦克风、处理器、存储器以及扬声器;所述第一麦克风,用于采集第一信号,所述第一信号用于表征当前外部环境的声音;所述第二麦克风,用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;所述存储器,用于存储程序或指令;所述处理器,用于调用所述程序或指令,以使得所述电子设备执行任意一种可能的耳机相关的方法得到第二音频信号或目标信号;所述扬声器,用于播放所述第二音频信号或目标信号。
第十八方面,本申请实施例提供一种模式控制装置,装置应用于终端设备。装置包括相应的功能模块,分别用于实现以上任意一种可能的终端相关的步骤,具体参见方法示例中的详细描述,此处不做赘述。功能可以通过硬件实现,也可以通过硬件执行相应的软件实现。硬件或软件包括一个或多个与上述功能相对应的模块。
第十九方面,本申请实施例提供一种终端设备,包括存储器、处理器和显示器;所述显 示器,用于显示界面;所述存储器,用于存储程序或指令;所述处理器,用于调用所述程序或指令,以使得所述终端设备执行任意一种可能的终端相关的方法中的步骤。
第二十方面,本申请提供一种计算机可读存储介质,计算机可读存储介质中存储有计算机程序或指令,当计算机程序或指令被耳机执行时,使得耳机执行上述任意一种可能的相关的耳机设计方法。
第二十一方面,本申请提供一种计算机可读存储介质,计算机可读存储介质中存储有计算机程序或指令,当计算机程序或指令被终端设备执行时,使得耳机执行上述任意一种可能的终端的相关设计中的方法。
第二十二方面,本申请提供一种计算机程序产品,该计算机程序产品包括计算机程序或指令,当该计算机程序或指令被耳机执行时,实现上述任意一种可能的耳机的实现方式中的方法。
第二十三方面,本申请提供一种计算机程序产品,该计算机程序产品包括计算机程序或指令,当该计算机程序或指令被耳机执行时,实现上述任意可能的耳机的实现方式中的方法。
应理解,上述任一种可能的设计在不违背自然规律的前提下可以自由组合。
附图说明
图1为本申请实施例中终端设备100的硬件结构示意图;
图2为本申请实施例中终端设备100的软件结构示意图;
图3为本申请实施例中耳机200的结构示意图;
图4为本申请实施例中AHA通路示意图;
图5A为本申请实施例中ANC处理流程图;
图5B为本申请实施例中一种ANC处理流程示意图;
图6A为本申请实施例中HT处理流程图;
图6B为本申请实施例中一种HT处理流程示意图;
图6C为本申请实施例中另一种HT处理流程示意图;
图7为本申请实施例中去闭塞效应处理流程示意图;
图8A为本申请实施例中AH处理流程图;
图8B为本申请实施例中一种AH处理流程示意图;
图8C为本申请实施例中另一种AH处理流程示意图;
图9为本申请实施例中降噪处理流程示意图;
图10为本申请实施例中一种增益放大处理流程示意图;
图11为本申请实施例中另一增益放大处理流程示意图;
图12A为本申请实施例中终端设备的主界面示意图;
图12B为本申请实施例中耳机应用的控制界面示意图;
图12C为本申请实施例中ANC模式下终端设备控制耳机的控制示意图;
图12D为本申请实施例中HT模式下终端设备控制耳机的控制示意图;
图12E为本申请实施例中AH模式下终端设备控制耳机的控制示意图;
图12F为本申请实施例中一种选择控件的示意图;
图12G为本申请实施例中另一种选择控件的示意图;
图12H为本申请实施例中耳机的控制界面触发示意图;
图13为本申请实施例中又一种选择控件的示意图;
图14A为本申请实施例中一种智慧场景检测功能开启控制示意图;
图14B为本申请实施例中另一种智慧场景检测功能开启控制示意图;
图14C为本申请实施例中一种耳机控制界面示意图;
图15为本申请实施例中事件检测示意图;
图16为本申请实施例中终端设备与耳机交互处理模式和处理强度的示意图;
图17A为本申请实施例中一种场景检测结果显示示意图;
图17B为本申请实施例中另一种场景检测结果显示示意图;
图18为本申请实施例中场景检测示意图;
图19为本申请实施例中噪声处理装置1900结构示意图;
图20为本申请实施例中模式控制装置2000结构示意图;
图21为本申请实施例中模式控制装置2100结构示意图;
图22为本申请实施例中模式控制装置2200结构示意图;
图23为本申请实施例中终端设备2300结构示意图;
图24为本申请实施例中终端设备中一种耳机控制界面示意图;
图25为本申请实施例中终端设备中一种耳机控制界面示意图;
图26为本申请实施例中终端设备中一种耳机控制界面示意图;
图27为本申请实施例中终端设备中一种耳机控制界面示意图;
图28为本申请实施例中终端设备中一种耳机控制界面示意图。
具体实施方式
下面将结合附图,对本申请实施例进行详细描述。本申请的实施方式部分使用的术语仅用于对本申请的具体实施例进行解释,而非旨在限定本申请。显然,所描述的实施例仅仅是本申请一部分实施例,并不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。
以下,先对本申请实施例中的部分用语进行解释说明,以便于本领域技术人员理解。
1)本申请实施例涉及的应用(application,app),为能够实现某项或多项特定功能的软件程序。通常,终端设备中可以安装多个应用。比如,相机应用、邮箱应用、耳机控制应用等。下文中提到的应用,可以是终端设备出厂时已安装的系统应用,也可以是用户在使用终端设备的过程中从网络下载或从其他终端设备获取的第三方应用。
2)bark子带。
人类听觉系统具有掩蔽效应,即强频音会阻碍人类对其附近同时发生的弱频音的感知,且耳蜗的基底膜对外来声音信号有频率选择和调谐作用,因此,引入临界频带的概念,从感知上度量声音频率。通常认为,在22HZ~22kHZ的听阈内有24个临界频带,会引起基底膜上不同位置的震动。每个临界频带称为bark子带。
3)语音端点检测(voice activity detection,VAD)。VAD是从带有噪声的语音中准确的定位出语音的开始和结束点,因为一段语音中含有很长的静音部分,也就是把静音和实际语音分离开来,是一种对语音数据的原始处理。
4)本申请实施例中“至少一个(项)”是指一个(项)或者多(项),“多个(项)”是指两个(项)或两个(项)以上。“和/或”,描述关联对象的关联关系,表示可以存在三种关系, 例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B的情况,其中A,B可以是单数或者复数。字符“/”一般表示前后关联对象是一种“或”的关系。“以下至少一项(个)”或其类似表达,是指的这些项中的任意组合,包括单项(个)或复数项(个)的任意组合。例如,a、b或c中的至少一项(个),可以表示:a、b、c、a-b、a-c、b-c或a-b-c,其中a、b、c可以是单个,也可以是多个。本申请中,符号“(a,b)”表示开区间,范围为大于a且小于b;“[a,b]”表示闭区间,范围为大于或等于a且小于或等于b;“(a,b]”表示半开半闭区间,范围为大于a且小于或等于b;“(a,b]”表示半开半闭区间,范围为大于a且小于或等于b。以及,除非有相反的说明,本申请实施例提及“第一”、“第二”等序数词是用于对多个对象进行区分,不用于限定多个对象的大小、内容、顺序、时序、优先级或者重要程度等。例如,第一麦克风和第二麦克风,只是为了区分不同的麦克风,而并不是表示这两个麦克风的大小、优先级或者重要程度等的不同。
本申请实施例提供一种系统,系统包括终端设备100和耳机200。终端设备100与耳机连接200,连接可以是无线连接,也可以为有线连接。对于无线连接,比如可以是终端设备通过蓝牙技术、无线高保真(wireless fidelity,Wi-Fi)技术、红外IR技术、超宽带技术与所述耳机连接。
本申请实施例中,终端设备100是具备显示界面功能的设备。终端设备100例如可以为手机、显示器、平板电脑、车载设备等具有显示界面的产品,以及智能手表、智能手环等智能显示穿戴产品等备。本申请实施例对上述移动终端的具体形式不做特殊限制。
耳机200包括可以挂在耳朵边上的两个发音单元。适配于左耳朵的可以称为左耳机,适配于右耳朵的可以称为右耳机。从佩戴角度来说,本申请实施例中耳机200可以是头戴式耳机、耳挂式耳机、颈挂式耳机或者耳塞式耳机等。耳塞式耳机还包括入耳式耳机(或者称为耳道式耳机)或者半入耳式耳机。耳机200具备ANC功能、HT功能或者AH功能中的至少两种。为了描述方便,本申请实施例中将ANC、HT和AH统称为AHA,当然还可以命令为其它的名字,本申请对此不作限定。
作为一种示例,以入耳式耳机为例。左耳机和右耳机采用的结构类似。左耳机或者右耳机均可以采用如下所描述的耳机结构。耳机结构(左耳机或右耳机)包括可以塞入耳道内的胶套、贴近耳朵的耳包、悬挂在耳包上的耳机杆。胶套将声音导向耳道,耳包内包括电池、扬声器、传感器等器件,耳机杆上可布置麦克风、物理按键等。耳机杆可以是圆柱、长方体、椭圆体等形状。布置在耳朵内部的麦克风可以称为误差麦克风,布置在耳机外部的麦克风称为参考麦克风。误差麦克风用于采集外部环境的声音。参考麦克风,在用户佩戴该耳机时,采集佩戴耳机的用户耳道内部环境的声音。两个麦克风即可以是模拟麦克风、也可以是数字麦克风。在用户佩戴耳机后,两个麦克风与扬声器摆放位置关系是:误差麦克风在耳朵内部,靠近耳机胶套。扬声器位于误差麦克风与参考麦克风之间。参考麦克风靠近耳朵外部结构,可以布置在耳机杆上部。误差麦克风的管道可面向扬声器、也可以面向耳道内部。参考麦克风附近有耳机开孔,用于透传外部环境声音进入参考麦克风。
本申请实施例中,终端设备100,用于向耳机200发送下行音频信号和/或控制信令。比如,控制信令用于对耳机200采用的处理模式进行控制。耳机200采用的处理模式可以包括指示不作任何处理的空模式、指示实现ANC功能的ANC模式、指示实现HT功能的HT模式或者指示实现AH功能的AH模式中的至少两项。
当耳机采用ANC模式时,能够减弱耳机用户对当前外部环境的声音以及佩戴所述耳机的用户耳道内部的环境声音的感知。当耳机采用HT模式时,能够强化用户对当前外部环境的 声音的感知。当耳机采用AH模式时,能够强化用户对当前外部环境的声音包括的事件声音的感知。其中,事件声音是指外部环境中的预设声音,或者事件声音满足预设频谱。比如事件声音包括火车站中的报站声音或者鸣笛声音;则事件声音满足火车站中的报站声音的频谱或者鸣笛声音的频谱。再比如,事件声音可以包括飞机航站楼中的通知声音,飞机上的广播声音等;再比如饭店的叫号的声音等等。应理解,终端和耳机都可以识别事件声音。
应理解的是,耳机200包括左耳机和右耳机,左耳机和右耳机可以采用相同的处理模式或者不同的处理模式。左耳机与右耳机采用相同的处理模式时,用户佩戴左耳机的左耳朵和佩戴右耳机的右耳朵听觉感知可以相同。当左耳机与右耳机采用不同的处理模式时,用户佩戴左耳机的左耳朵和佩戴右耳机的右耳朵听觉感知不同。以左耳机采用ANC和右耳机采用AH为例,当左耳机采用ANC模式时,能够减弱耳机用户的左耳朵对当前外部环境的声音以及佩戴所述耳机的用户左耳朵耳道内部的环境声音的感知。当右耳机采用AH模式时,能够强化用户的右耳朵对当前外部环境的声音包括的事件声音的感知。
在确定耳机的处理模式时,可以采用如下任一种可能的方式来实现;
第一种可能的方式,终端设备100提供控制界面,供用户根据需求选择耳机200的处理模式。比如终端设备100在用户的操作下,向耳机200发送控制信令,控制信令用来指示耳机200采用的处理模式。
需要说明的是,耳机200中的左耳机和右耳机采用的处理模式可以相同也可以不同。比如,控制界面上一个选择控件用于为左耳机和右耳机选择相同的处理模式。再比如,控制界面可以包括两个选择控件,其中一个选择控件用于为左耳机选择处理模式,另一个选择控件用于为右耳机选择处理模式。针对控制界面以及选择控件,后续会详细描述,此处不再重复描述。
第二种可能的方式,终端设备识别用户当前外部环境的场景类型,在不同的场景下,耳机200所采用的处理模式不同,即耳机实现的处理功能不同。
第三种可能的方式,耳机200识别用户的操作,确定用户选择的耳机200采用ANC模式、HT模式或者AH模式。比如,用户的操作可以是用户敲击耳机的操作,或者耳机上设置有按钮,不同的按钮指示不同的处理模式。
第四种可能的方式,耳机识别耳机外部环境的场景类型,不同的场景,耳机所采用的处理模式不同。
针对上述第一种可能的方式至第四种可能的方式,后续会详细描述,此处不再重复描述。
图1示出了终端设备100的一种可选的硬件结构示意图。
终端设备100可以包括处理器110,外部存储器接口120,内部存储器121,通用串行总线(universal serial bus,USB)接口130,充电管理模块140,电源管理模块141,电池142,天线1,天线2,移动通信模块150,无线通信模块160,音频模块170,扬声器170A,受话器170B,麦克风170C,耳机接口170D,传感器模块180,按键190,马达191,指示器192,摄像头193,显示屏194,以及用户标识模块(subscriber identification module,SIM)卡接口195等。其中传感器模块180可以包括压力传感器180A,陀螺仪传感器180B,气压传感器180C,磁传感器180D,加速度传感器180E,距离传感器180F,接近光传感器180G,指纹传感器180H,温度传感器180J,触摸传感器180K,环境光传感器180L,骨传导传感器180M等。
可以理解的是,本发明实施例示意的结构并不构成对终端设备100的具体限定。在本申请另一些实施例中,终端设备100可以包括比图示更多或更少的部件,或者组合某些部件, 或者拆分某些部件,或者不同的部件布置。图示的部件可以以硬件,软件或软件和硬件的组合实现。
处理器110可以包括一个或多个处理单元,例如:处理器110可以包括应用处理器(application processor,AP),调制解调处理器,图形处理器(graphics processing unit,GPU),图像信号处理器(image signal processor,ISP),控制器,视频编解码器,数字信号处理器(digital signal processor,DSP),基带处理器,和/或神经网络处理器(neural-network processing unit,NPU)等。其中,不同的处理单元可以是独立的器件,也可以集成在一个或多个处理器中。
控制器可以根据指令操作码和时序信号,产生操作控制信号,完成取指令和执行指令的控制。
处理器110中还可以设置存储器,用于存储指令和数据。在一些实施例中,处理器110中的存储器为高速缓冲存储器。该存储器可以保存处理器110刚用过或循环使用的指令或数据。如果处理器110需要再次使用该指令或数据,可从所述存储器中直接调用。避免了重复存取,减少了处理器110的等待时间,因而提高了系统的效率。
在一些实施例中,处理器110可以包括一个或多个接口。接口可以包括集成电路(inter-integrated circuit,I2C)接口,集成电路内置音频(inter-integrated circuit sound,I2S)接口,脉冲编码调制(pulse code modulation,PCM)接口,通用异步收发传输器(universal asynchronous receiver/transmitter,UART)接口,移动产业处理器接口(mobile industry processor interface,MIPI),通用输入输出(general-purpose input/output,GPIO)接口,用户标识模块(subscriber identity module,SIM)接口,和/或通用串行总线(universal serial bus,USB)接口等。
I2C接口是一种双向同步串行总线,包括一根串行数据线(serial data line,SDA)和一根串行时钟线(derail clock line,SCL)。在一些实施例中,处理器110可以包含多组I2C总线。处理器110可以通过不同的I2C总线接口分别耦合触摸传感器180K,充电器,闪光灯,摄像头193等。例如:处理器110可以通过I2C接口耦合触摸传感器180K,使处理器110与触摸传感器180K通过I2C总线接口通信,实现终端设备100的触摸功能。
I2S接口可以用于音频通信。在一些实施例中,处理器110可以包含多组I2S总线。处理器110可以通过I2S总线与音频模块170耦合,实现处理器110与音频模块170之间的通信。在一些实施例中,音频模块170可以通过I2S接口向无线通信模块160传递音频信号,实现通过耳机200(比如蓝牙耳机)接听电话的功能。
PCM接口也可以用于音频通信,将模拟信号抽样,量化和编码。在一些实施例中,音频模块170与无线通信模块160可以通过PCM总线接口耦合。在一些实施例中,音频模块170也可以通过PCM接口向无线通信模块160传递音频信号,实现通过蓝牙耳机200接听电话的功能。所述I2S接口和所述PCM接口都可以用于音频通信。
UART接口是一种通用串行数据总线,用于异步通信。该总线可以为双向通信总线。它将要传输的数据在串行通信与并行通信之间转换。在一些实施例中,UART接口通常被用于连接处理器110与无线通信模块160。例如:处理器110通过UART接口与无线通信模块160中的蓝牙模块通信,实现蓝牙功能。在一些实施例中,音频模块170可以通过UART接口向无线通信模块160传递音频信号,实现通过蓝牙耳机200播放音乐的功能。
MIPI接口可以被用于连接处理器110与显示屏194,摄像头193等外围器件。MIPI接口包括摄像头串行接口(camera serial interface,CSI),显示屏串行接口(display serial interface,DSI)等。在一些实施例中,处理器110和摄像头193通过CSI接口通信,实现终端设备100 的拍摄功能。处理器110和显示屏194通过DSI接口通信,实现终端设备100的显示功能。
GPIO接口可以通过软件配置。GPIO接口可以被配置为控制信号,也可被配置为数据信号。在一些实施例中,GPIO接口可以用于连接处理器110与摄像头193,显示屏194,无线通信模块160,音频模块170,传感器模块180等。GPIO接口还可以被配置为I2C接口,I2S接口,UART接口,MIPI接口等。
USB接口130是符合USB标准规范的接口,具体可以是Mini USB接口,Micro USB接口,USB Type C接口等。USB接口130可以用于连接充电器为终端设备100充电,也可以用于终端设备100与外围设备之间传输数据。也可以用于连接耳机200,通过耳机200播放音频。该接口还可以用于连接其他终端设备,例如AR设备等。
可以理解的是,本发明实施例示意的各模块间的接口连接关系,只是示意性说明,并不构成对终端设备100的结构限定。在本申请另一些实施例中,终端设备100也可以采用上述实施例中不同的接口连接方式,或多种接口连接方式的组合。
充电管理模块140用于从充电器接收充电输入。其中,充电器可以是无线充电器,也可以是有线充电器。在一些有线充电的实施例中,充电管理模块140可以通过USB接口130接收有线充电器的充电输入。在一些无线充电的实施例中,充电管理模块140可以通过终端设备100的无线充电线圈接收无线充电输入。充电管理模块140为电池142充电的同时,还可以通过电源管理模块141为终端设备供电。
电源管理模块141用于连接电池142,充电管理模块140与处理器110。电源管理模块141接收电池142和/或充电管理模块140的输入,为处理器110,内部存储器121,显示屏194,摄像头193,和无线通信模块160等供电。电源管理模块141还可以用于监测电池容量,电池循环次数,电池健康状态(漏电,阻抗)等参数。在其他一些实施例中,电源管理模块141也可以设置于处理器110中。在另一些实施例中,电源管理模块141和充电管理模块140也可以设置于同一个器件中。
终端设备100的无线通信功能可以通过天线1,天线2,移动通信模块150,无线通信模块160,调制解调处理器以及基带处理器等实现。
天线1和天线2用于发射和接收电磁波信号。终端设备100中的每个天线可用于覆盖单个或多个通信频带。不同的天线还可以复用,以提高天线的利用率。例如:可以将天线1复用为无线局域网的分集天线。在另外一些实施例中,天线可以和调谐开关结合使用。
移动通信模块150可以提供应用在终端设备100上的包括2G/3G/4G/5G等无线通信的解决方案。移动通信模块150可以包括至少一个滤波器,开关,功率放大器,低噪声放大器(low noise amplifier,LNA)等。移动通信模块150可以由天线1接收电磁波,并对接收的电磁波进行滤波,放大等处理,传送至调制解调处理器进行解调。移动通信模块150还可以对经调制解调处理器调制后的信号放大,经天线1转为电磁波辐射出去。在一些实施例中,移动通信模块150的至少部分功能模块可以被设置于处理器110中。在一些实施例中,移动通信模块150的至少部分功能模块可以与处理器110的至少部分模块被设置在同一个器件中。
调制解调处理器可以包括调制器和解调器。其中,调制器用于将待发送的低频基带信号调制成中高频信号。解调器用于将接收的电磁波信号解调为低频基带信号。随后解调器将解调得到的低频基带信号传送至基带处理器处理。低频基带信号经基带处理器处理后,被传递给应用处理器。应用处理器通过音频设备(不限于扬声器170A,受话器170B等)输出声音信号,或通过显示屏194显示图像或视频。在一些实施例中,调制解调处理器可以是独立的器件。在另一些实施例中,调制解调处理器可以独立于处理器110,与移动通信模块150或 其他功能模块设置在同一个器件中。
无线通信模块160可以提供应用在终端设备100上的包括无线局域网(wireless local area networks,WLAN)(如无线保真(wireless fidelity,Wi-Fi)网络),蓝牙(bluetooth,BT),全球导航卫星系统(global navigation satellite system,GNSS),调频(frequency modulation,FM),近距离无线通信技术(near field communication,NFC),红外技术(infrared,IR)等无线通信的解决方案。无线通信模块160可以是集成至少一个通信处理模块的一个或多个器件。无线通信模块160经由天线2接收电磁波,将电磁波信号调频以及滤波处理,将处理后的信号发送到处理器110。无线通信模块160还可以从处理器110接收待发送的信号,对其进行调频,放大,经天线2转为电磁波辐射出去。比如,无线通信模块160包括蓝牙模块,终端设备100通过蓝牙与耳机200建立无线连接。再比如,无线通信模块160包括红外模块,终端设备100可以通过红外模块与耳机200建立无线连接。
在一些实施例中,终端设备100的天线1和移动通信模块150耦合,天线2和无线通信模块160耦合,使得终端设备100可以通过无线通信技术与网络以及其他设备通信。所述无线通信技术可以包括全球移动通讯系统(global system for mobile communications,GSM),通用分组无线服务(general packet radio service,GPRS),码分多址接入(code division multiple access,CDMA),宽带码分多址(wideband code division multiple access,WCDMA),时分码分多址(time-division code division multiple access,TD-SCDMA),长期演进(long term evolution,LTE),BT,GNSS,WLAN,NFC,FM,和/或IR技术等。所述GNSS可以包括全球卫星定位系统(global positioning system,GPS),全球导航卫星系统(global navigation satellite system,GLONASS),北斗卫星导航系统(beidou navigation satellite system,BDS),准天顶卫星系统(quasi-zenith satellite system,QZSS)和/或星基增强系统(satellite based augmentation systems,SBAS)。
终端设备100通过GPU,显示屏194,以及应用处理器等实现显示功能。GPU为图像处理的微处理器,连接显示屏194和应用处理器。GPU用于执行数学和几何计算,用于图形渲染。处理器110可包括一个或多个GPU,其执行程序指令以生成或改变显示信息。
显示屏194用于显示图像,视频等。显示屏194包括显示面板。显示面板可以采用液晶显示屏(liquid crystal display,LCD),有机发光二极管(organic light-emitting diode,OLED),有源矩阵有机发光二极体或主动矩阵有机发光二极体(active-matrix organic light emitting diode的,AMOLED),柔性发光二极管(flex light-emitting diode,FLED),Miniled,MicroLed,Micro-oLed,量子点发光二极管(quantum dot light emitting diodes,QLED)等。在一些实施例中,终端设备100可以包括1个或N1个显示屏194,N1为大于1的正整数。
终端设备100可以通过ISP,摄像头193,视频编解码器,GPU,显示屏194以及应用处理器等实现拍摄功能。
ISP用于处理摄像头193反馈的数据。例如,拍照时,打开快门,光线通过镜头被传递到摄像头感光元件上,光信号转换为电信号,摄像头感光元件将所述电信号传递给ISP处理,转化为肉眼可见的图像。ISP还可以对图像的噪点,亮度,肤色进行算法优化。ISP还可以对拍摄场景的曝光,色温等参数优化。在一些实施例中,ISP可以设置在摄像头193中。
摄像头193用于捕获静态图像或视频。物体通过镜头生成光学图像投射到感光元件。感光元件可以是电荷耦合器件(charge coupled device,CCD)或互补金属氧化物半导体(complementary metal-oxide-semiconductor,CMOS)光电晶体管。感光元件把光信号转换成电信号,之后将电信号传递给ISP转换成数字图像信号。ISP将数字图像信号输出到DSP加 工处理。DSP将数字图像信号转换成标准的RGB,YUV等格式的图像信号。在一些实施例中,处理器110可以根据内部存储器121中的程序或者指令触发启动摄像头193,从而摄像头193采集至少一个图像,并根据程序或者指令对至少一个图像进行相应的处理。在一些实施例中,终端设备100可以包括1个或N2个摄像头193,N2为大于1的正整数。
数字信号处理器用于处理数字信号,除了可以处理数字图像信号,还可以处理其他数字信号。例如,当终端设备100在频点选择时,数字信号处理器用于对频点能量进行傅里叶变换等。
视频编解码器用于对数字视频压缩或解压缩。终端设备100可以支持一种或多种视频编解码器。这样,终端设备100可以播放或录制多种编码格式的视频,例如:动态图像专家组(moving picture experts group,MPEG)1,MPEG2,MPEG3,MPEG4等。
NPU为神经网络(neural-network,NN)计算处理器,通过借鉴生物神经网络结构,例如借鉴人脑神经元之间传递模式,对输入信息快速处理,还可以不断的自学习。通过NPU可以实现终端设备100的智能认知等应用,例如:图像识别,人脸识别,语音识别,文本理解等。
外部存储器接口120可以用于连接外部存储卡,例如Micro SD卡,实现扩展终端设备100的存储能力。外部存储卡通过外部存储器接口120与处理器110通信,实现数据存储功能。例如将音乐,视频等文件保存在外部存储卡中。
内部存储器121可以用于存储计算机可执行程序代码,所述可执行程序代码包括指令。内部存储器121可以包括存储程序区和存储数据区。其中,存储程序区可存储操作系统,至少一个功能所需的应用程序(比如相机应用)等。存储数据区可存储终端设备100使用过程中所创建的数据(比如摄像头采集的图像等)等。此外,内部存储器121可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件,闪存器件,通用闪存存储器(universal flash storage,UFS)等。处理器110通过运行存储在内部存储器121的指令,和/或存储在设置于处理器中的存储器的指令,执行终端设备100的各种功能应用以及数据处理。内部存储器121还可以存储本申请实施例提供的下行音频信号。内部存储器121中还可以存储用于实现控制耳机200功能的代码。当内部存储器121中存储的用于实现控制耳机200功能的代码被处理器110运行时,控制耳机200实现相应的功能,比如ANC功能、HT功能或者AH功能。当然,本申请实施例提供的用于实现控制耳机200功能的代码还可以存储在外部存储器中。这种情况下,处理器110可以通过外部存储器接口120运行存储在外部存储器中实现控制耳机200功能的相应数据,以控制耳机200实现相应功能。
终端设备100可以通过音频模块170,扬声器170A,受话器170B,麦克风170C,耳机接口170D,以及应用处理器等实现音频功能。例如音乐播放,录音等。
音频模块170用于将数字音频信息转换成模拟音频信号输出,也用于将模拟音频输入转换为数字音频信号。音频模块170还可以用于对音频信号编码和解码。在一些实施例中,音频模块170可以设置于处理器110中,或将音频模块170的部分功能模块设置于处理器110中。
扬声器170A,也称“喇叭”,用于将音频电信号转换为声音信号。终端设备100可以通过扬声器170A收听音乐,或收听免提通话。
受话器170B,也称“听筒”,用于将音频电信号转换成声音信号。当终端设备100接听电话或语音信息时,可以通过将受话器170B靠近人耳接听语音。
麦克风170C,也称“话筒”,“传声器”,用于将声音信号转换为电信号。当拨打电话或 发送语音信息时,用户可以通过人嘴靠近麦克风170C发声,将声音信号输入到麦克风170C。终端设备100可以设置至少一个麦克风170C。在另一些实施例中,终端设备100可以设置两个麦克风170C,除了采集声音信号,还可以实现降噪功能。在另一些实施例中,终端设备100还可以设置三个,四个或更多麦克风170C,实现采集声音信号,降噪,还可以识别声音来源,实现定向录音功能等。
耳机接口170D用于连接有线耳机。当本申请实施例提供的耳机200为有线耳机时,终端设备100通过耳机接口170D与耳机连接。耳机接口170D可以是USB接口130,也可以是3.5mm的开放移动终端设备平台(open mobile terminal platform,OMTP)标准接口,美国蜂窝电信工业协会(cellular telecommunications industry association of the USA,CTIA)标准接口。
压力传感器180A用于感受压力信号,可以将压力信号转换成电信号。在一些实施例中,压力传感器180A可以设置于显示屏194。压力传感器180A的种类很多,如电阻式压力传感器,电感式压力传感器,电容式压力传感器等。电容式压力传感器可以是包括至少两个具有导电材料的平行板。当有力作用于压力传感器180A,电极之间的电容改变。终端设备100根据电容的变化确定压力的强度。当有触摸操作作用于显示屏194,终端设备100根据压力传感器180A检测所述触摸操作强度。终端设备100也可以根据压力传感器180A的检测信号计算触摸的位置。在一些实施例中,作用于相同触摸位置,但不同触摸操作强度的触摸操作,可以对应不同的操作指令。例如:当有触摸操作强度小于第一压力阈值的触摸操作作用于短消息应用图标时,执行查看短消息的指令。当有触摸操作强度大于或等于第一压力阈值的触摸操作作用于短消息应用图标时,执行新建短消息的指令。
陀螺仪传感器180B可以用于确定终端设备100的运动姿态。在一些实施例中,可以通过陀螺仪传感器180B确定终端设备100围绕三个轴(即,x,y和z轴)的角速度。陀螺仪传感器180B可以用于拍摄防抖。示例性的,当按下快门,陀螺仪传感器180B检测终端设备100抖动的角度,根据角度计算出镜头模组需要补偿的距离,让镜头通过反向运动抵消终端设备100的抖动,实现防抖。陀螺仪传感器180B还可以用于导航,体感游戏场景。
气压传感器180C用于测量气压。在一些实施例中,终端设备100通过气压传感器180C测得的气压值计算海拔高度,辅助定位和导航。
磁传感器180D包括霍尔传感器。终端设备100可以利用磁传感器180D检测翻盖皮套的开合。在一些实施例中,当终端设备100是翻盖机时,终端设备100可以根据磁传感器180D检测翻盖的开合。进而根据检测到的皮套的开合状态或翻盖的开合状态,设置翻盖自动解锁等特性。
加速度传感器180E可检测终端设备100在各个方向上(一般为三轴)加速度的大小。当终端设备100静止时可检测出重力的大小及方向。还可以用于识别终端设备姿态,应用于横竖屏切换,计步器等应用。
距离传感器180F,用于测量距离。终端设备100可以通过红外或激光测量距离。在一些实施例中,拍摄场景,终端设备100可以利用距离传感器180F测距以实现快速对焦。
接近光传感器180G可以包括例如发光二极管(LED)和光检测器,例如光电二极管。发光二极管可以是红外发光二极管。终端设备100通过发光二极管向外发射红外光。终端设备100使用光电二极管检测来自附近物体的红外反射光。当检测到充分的反射光时,可以确定终端设备100附近有物体。当检测到不充分的反射光时,终端设备100可以确定终端设备100附近没有物体。终端设备100可以利用接近光传感器180G检测用户手持终端设备100贴 近耳朵通话,以便自动熄灭屏幕达到省电的目的。接近光传感器180G也可用于皮套模式,口袋模式自动解锁与锁屏。
环境光传感器180L用于感知环境光亮度。在一些实施例中,终端设备100可以根据环境光传感器180L感知的环境光亮度来确定图像的曝光时间。在一些实施例中,终端设备100可以根据感知的环境光亮度自适应调节显示屏194亮度。环境光传感器180L也可用于拍照时自动调节白平衡。环境光传感器180L还可以与接近光传感器180G配合,检测终端设备100是否在口袋里,以防误触。
指纹传感器180H用于采集指纹。终端设备100可以利用采集的指纹特性实现指纹解锁,访问应用锁,指纹拍照,指纹接听来电等。
温度传感器180J用于检测温度。在一些实施例中,终端设备100利用温度传感器180J检测的温度,执行温度处理策略。例如,当温度传感器180J上报的温度超过阈值,终端设备100执行降低位于温度传感器180J附近的处理器的性能,以便降低功耗实施热保护。在另一些实施例中,当温度低于另一阈值时,终端设备100对电池142加热,以避免低温导致终端设备100异常关机。在其他一些实施例中,当温度低于又一阈值时,终端设备100对电池142的输出电压执行升压,以避免低温导致的异常关机。
触摸传感器180K,也称“触控器件”。触摸传感器180K可以设置于显示屏194,由触摸传感器180K与显示屏194组成触摸屏,也称“触控屏”。触摸传感器180K用于检测作用于其上或附近的触摸操作。触摸传感器可以将检测到的触摸操作传递给应用处理器,以确定触摸事件类型。可以通过显示屏194提供与触摸操作相关的视觉输出。在另一些实施例中,触摸传感器180K也可以设置于终端设备100的表面,与显示屏194所处的位置不同。
骨传导传感器180M可以获取振动信号。在一些实施例中,骨传导传感器180M可以获取人体声部振动骨块的振动信号。骨传导传感器180M也可以接触人体脉搏,接收血压跳动信号。在一些实施例中,骨传导传感器180M也可以设置于耳机中,结合成骨传导耳机。音频模块170可以基于所述骨传导传感器180M获取的声部振动骨块的振动信号,解析出语音信号,实现语音功能。应用处理器可以基于所述骨传导传感器180M获取的血压跳动信号解析心率信息,实现心率检测功能。
按键190包括开机键,音量键等。按键190可以是机械按键。也可以是触摸式按键。终端设备100可以接收按键输入,产生与终端设备100的用户设置以及功能控制有关的键信号输入。
马达191可以产生振动提示。马达191可以用于来电振动提示,也可以用于触摸振动反馈。例如,作用于不同应用(例如拍照,音频播放等)的触摸操作,可以对应不同的振动反馈效果。作用于显示屏194不同区域的触摸操作,马达191也可对应不同的振动反馈效果。不同的应用场景(例如:时间提醒,接收信息,闹钟,游戏等)也可以对应不同的振动反馈效果。触摸振动反馈效果还可以支持自定义。
指示器192可以是指示灯,可以用于指示充电状态,电量变化,也可以用于指示消息,未接来电,通知等。
SIM卡接口195用于连接SIM卡。SIM卡可以通过插入SIM卡接口195,或从SIM卡接口195拔出,实现和终端设备100的接触和分离。终端设备100可以支持1个或N3个SIM卡接口,N3为大于1的正整数。SIM卡接口195可以支持Nano SIM卡,Micro SIM卡,SIM卡等。同一个SIM卡接口195可以同时插入多张卡。所述多张卡的类型可以相同,也可以不同。SIM卡接口195也可以兼容不同类型的SIM卡。SIM卡接口195也可以兼容外部存储卡。 终端设备100通过SIM卡和网络交互,实现通话以及数据通信等功能。在一些实施例中,终端设备100采用eSIM,即:嵌入式SIM卡。eSIM卡可以嵌在终端设备100中,不能和终端设备100分离。
终端设备100的软件系统可以采用分层架构,事件驱动架构,微核架构,微服务架构,或云架构。本发明实施例以分层架构的Android系统为例,示例性说明终端设备100的软件结构。
图2是本发明实施例的终端设备100的软件结构框图。
分层架构将软件分成若干个层,每一层都有清晰的角色和分工。层与层之间通过软件接口通信。在一些实施例中,将Android系统分为四层,从上至下分别为应用程序层,应用程序框架层,安卓运行时(Android runtime)和系统库,以及内核层。应用程序层可以包括一系列应用程序包。
如图2所示,应用程序包可以包括相机,图库,日历,通话,地图,导航,WLAN,蓝牙,音乐,视频,短信息等应用程序。
应用程序框架层为应用程序层的应用程序提供应用编程接口(application programming interface,API)和编程框架。应用程序框架层包括一些预先定义的函数。
如图2所示,应用程序框架层可以包括窗口管理器,内容提供器,视图系统,电话管理器,资源管理器,通知管理器等。
窗口管理器用于管理窗口程序。窗口管理器可以获取显示屏大小,判断是否有状态栏,锁定屏幕,截取屏幕等。
内容提供器用来存放和获取数据,并使这些数据可以被应用程序访问。所述数据可以包括视频,图像,音频,拨打和接听的电话,浏览历史和书签,电话簿等。
视图系统包括可视控件,例如显示文字的控件,显示图片的控件等。视图系统可用于构建应用程序。显示界面可以由一个或多个视图组成的。例如,包括短信通知图标的显示界面,可以包括显示文字的视图以及显示图片的视图。
电话管理器用于提供终端设备100的通信功能。例如通话状态的管理(包括接通,挂断等)。
资源管理器为应用程序提供各种资源,比如本地化字符串,图标,图片,布局文件,视频文件等等。
通知管理器使应用程序可以在状态栏中显示通知信息,可以用于传达告知类型的消息,可以短暂停留后自动消失,无需用户交互。比如通知管理器被用于告知下载完成,消息提醒等。通知管理器还可以是以图表或者滚动条文本形式出现在系统顶部状态栏的通知,例如后台运行的应用程序的通知,还可以是以对话窗口形式出现在屏幕上的通知。例如在状态栏提示文本信息,发出提示音,终端设备振动,指示灯闪烁等。
Android Runtime包括核心库和虚拟机。Android runtime负责安卓系统的调度和管理。
核心库包含两部分:一部分是java语言需要调用的功能函数,另一部分是安卓的核心库。
应用程序层和应用程序框架层运行在虚拟机中。虚拟机将应用程序层和应用程序框架层的java文件执行为二进制文件。虚拟机用于执行对象生命周期的管理,堆栈管理,线程管理,安全和异常的管理,以及垃圾回收等功能。
系统库可以包括多个功能模块。例如:表面管理器(surface manager),媒体库(Media Libraries),三维图形处理库(例如:OpenGL ES),2D图形引擎(例如:SGL)等。
表面管理器用于对显示子系统进行管理,并且为多个应用程序提供了2D和3D图层的融合。
媒体库支持多种常用的音频,视频格式回放和录制,以及静态图像文件等。媒体库可以支持多种音视频编码格式,例如:MPEG4,H.264,MP3,AAC,AMR,JPG,PNG等。
三维图形处理库用于实现三维图形绘图,图像渲染,合成,和图层处理等。
2D图形引擎是2D绘图的绘图引擎。
内核层是硬件和软件之间的层。内核层至少包含显示驱动,摄像头驱动,音频驱动,耳机驱动、传感器驱动。
下面结合捕获播放音频场景,示例性说明终端设备100软件以及硬件的工作流程。
当触摸传感器180K接收到触摸操作,相应的硬件中断被发给内核层。内核层将触摸操作加工成原始输入事件(包括触摸坐标,触摸操作的时间戳等信息)。原始输入事件被存储在内核层。应用程序框架层从内核层获取原始输入事件,识别该输入事件所对应的控件。以该触摸操作是触摸单击操作,该单击操作所对应的控件为音频应用图标的控件为例,音频应用调用应用框架层的接口,启动耳机控制应用,进而通过调用内核层启动耳机驱动,并向耳机发送音频信号,通过耳机200播放音频信号。
图3示出了耳机200的一种可选的硬件结构示意图。耳机200包括左耳机和右耳机。左耳机和右耳机采用的结构类似。耳机(包括左耳机和右耳机)的结构均包括包括第一麦克风301、第二麦克风302以及第三麦克风303。耳机中还可以包括处理器304和扬声器305。应理解的是,后续所描述的耳机可解释为左耳机,也可以解释为右耳机。
第一麦克风301用于采集当前外部环境的声音,第一麦克风301还可以称为参考麦克风。当用户佩戴上该耳机时,第一麦克风301位于耳机外侧,或者说第二麦克风301位于耳朵外部。当用户佩戴上耳机的情况下,第二麦克风302采集用户耳道内部环境的声音。第二麦克风302也可以称为误差麦克风。当用户佩戴上该耳机时,第二麦克风302位于耳机内部贴近耳道。第三麦克风303用于采集通话信号。第三麦克风303可以位于耳机外侧,当用户佩戴上耳机时,第三麦克风303相对于第一麦克风301来说,更靠近用户的嘴部。
需要说明的是,第一麦克风301用于采集当前外部环境的声音,可以解释为,耳机用户佩戴耳机处在的外部环境的声音,比如在火车上,外部环境的声音为佩戴耳机的用户周围环境的声音。左耳机上的第一麦克风301采用的左耳机的外部环境的声音。右耳机上的第一麦克风301采集右耳机的外部环境的声音。
为了便于区分,将第一麦克风301(参考麦克风)采集的信号称为第一信号,将第二麦克风302(误差麦克风)采集的信号称为第二信号。本申请实施例中涉及的麦克风可以是模拟麦克风,也可以是数字麦克风。麦克风为模拟麦克风时,在对麦克风采集的信号进行滤波处理之前,可以将模拟信号转换为数字信号。本申请实施例中以第一麦克风和第二麦克风均为数字麦克风为例进行说明,第一信号和第二信号均为数字信号。
处理器304用于下行音频信号和/或麦克风(包括第一麦克风301、第二麦克风302以及第三麦克风303)采集的信号进行处理,比如执行ANC处理、HT处理或AH处理等。示例性地,处理器304可以包括主控制单元以及降噪处理单元。主控制单元用于用户针对耳机操作产生控制命令或者接收来自终端设备的控制命令等。降噪处理单元用于根据控制命令对下行音频信号以及麦克风(包括第一麦克风301、第二麦克风302以及第三麦克风303)采集的信号进行ANC处理、HT处理或者AH处理。
左耳机和右耳机还可以包括存储器,存储器用于存储处理器304执行的程序或者指令。 处理器304根据存储器存储的程序或者指令执行ANC处理、HT处理或者AH处理。存储器可以包括随机存取存储器(random access memory,RAM)、闪存、只读存储器(read-only memory,ROM)、可编程只读存储器(programmableROM,PROM)、可擦除可编程只读存储器(erasable PROM,EPROM)、电可擦除可编程只读存储器(electrically ePROM,EEPROM)、寄存器、硬盘、移动硬盘、CD-ROM或者本领域熟知的任何其它形式的存储介质中的一个或者多个。
主控制单元,例如可以由ARM处理芯片、中央处理器(central processing unit,CPU)、片上系统(system on chip,SoC)、数字信号处理器(digital signal processing,DSP)或者微控制单元(micro controller unit,MCU)中的一项或多项实现。降噪处理单元例如可以包括编译解码(coder-decoder,CODEC)芯片或者高保真(high-fidelity,HiFi)芯片等。以降噪处理单元包括CODEC芯片为例,codec中硬化有滤波器、均衡器(Equaliser,,EQ)、动态范围控制器(Dynamic Range Control,DRC)、限幅器(limiter)、增益调节器(gain)、混音器(mixer)等,主要用于对信号做滤波、混音、增益调节等处理。降噪处理单元还可以包括DSP,DSP可以用于场景检测、语音增强、闭塞消除等处理。
耳机中还可以包括无线通信单元,用于通过终端设备100中的无线通信模块160与终端设备200建立通信连接。无线通信单元可以提供应用在耳机上的包括无线局域网(wireless local area networks,WLAN)(如无线保真(wireless fidelity,Wi-Fi)网络),蓝牙(bluetooth,BT),近距离无线通信技术(near field communication,NFC),红外技术(infrared,IR)等无线通信的解决方案。无线通信单元可以是集成至少一个通信处理模块的一个或多个器件。比如,无线通信模块160可以是蓝牙,无线通信单元也为蓝牙,耳机200和终端设备100通过蓝牙连接。
作为一种可能的示例,针对三种不同的工作模式,通过三种不同的信号通路输出,分别为主动降噪输出通路、环境声透传输出通路以及听觉增强输出通路。例如,不同的输出通路所采用的处理方式不同,参见图4所示。
本领域技术人员应当理解,ANC功能开启意味着ANC功能的信号通路是处于激活状态的,相应地,ANC通路中的各个功能模块也是处于激活状态;HT功能开启意味着HT功能的信号通路是处于激活状态的,相应地,HT通路中的各个功能模块也是处于激活状态。同理,当所述耳机的ANC功能和HT功能均处于开启状态时,表明ANC功能的信号通路是处于激活状态的,并且HT功能的信号通路也是处于激活状态的。即可以表明耳机的工作状态,而并不限定于某一时刻的具体操作或者功能控件的变化。
主动降噪输出通路中的主动降噪处理可以包括但不仅限于:利用对参考麦克风采集的第一信号的反相信号和误差麦克风采集的第二信号的反相信号进行噪声抑制。主动降噪输出通路中包括第一信号的反相信号和第二信号的反相信号。需要说明的是,第一信号与第一信号的反相信号的相位差为180°。扬声器输出第一信号的反相信号与第二信号的反相信号叠加后的信号,从而使得扬声器播放的当前外部环境的声音与耳朵实际听见的外部环境的声音进行了抵消而达到主动去噪的效果。因此,耳机采用ANC模式时,能够减弱耳机用户对当前所处环境的声音以及所述用户耳道内部的环境声音的感知。
可选地,当有下行音频信号时,可以对下行音频信号进行滤波补偿。并且可以在获取环境声音的反相信号时可以去除下行音频信号的影响。
在获取第一信号的反相信号和第二信号的反相信号时,可以采用第一滤波处理和第三滤波处理。例如,第一滤波处理可以是前馈(feed forward,FF)滤波处理,可以由前馈滤波器 来实现。第三滤波处理可以是反馈(feed back,FB)滤波处理,可以由反馈滤波器来实现。参见图4所示,FF滤波与FB滤波采用并行处理架构来增强降噪效果。后续会对ANC的处理流程进行详细说明,此处不再重复描述。
环境声透传输出通路中的环境声透传处理可以但不仅限于包括:对误差麦克风采集的第一信号进行第三滤波处理实现部分主动降噪功能,对参考麦克风采集的信号进行第二滤波处理以及HT增强处理。例如第二滤波处理可以是透传(HT)滤波处理,可以由透传滤波器来实现。扬声器播放的音频信号是根据第一信号与第二信号得到的,从而使得扬声器播放的音频信号后,用户能够通过耳机听见外部环境中的声音,相比未执行HT处理时所听见的外部环境的声音强度更好,效果更好。因此,耳机采用HT模式时,能够强化用户对所述用户当前所处环境的声音的强度的感知。后续会对HT的处理流程进行详细说明,此处不再重复描述。
听觉增强输出通路中的环境声透传处理可以但不仅限于包括:利用误差麦克风采集的信号实现部分主动降噪功能、对参考麦克风采集的信号进行第一滤波处理以及听觉增强处理使得用户所处环境的声音中的事件声音得到增强,以及对参考麦克风采集的信号进行第二滤波处理。扬声器的输出信号是根据第一信号中的事件信号以及第二信号的反相信号混音后的信号得到的。需要说明的是,第二信号与第二信号的反相信号的相位差为180°。扬声器输出第二信号的反相信号、第一信号的反相信号与第一信号中的事件信号叠加后的信号,使得扬声器输出的信号与耳朵实际听见的环境中的声音进行了抵消而达到主动去噪的效果,并且输出环境中的事件声音,使得用户能够清楚的听到环境中用户所需的预设信号。因此,耳机采用AH模式时,能够强化耳机用户对当前外部环境的声音包括的事件声音的感知。后续会对AH的处理流程进行详细说明,此处不再重复描述。
应当理解的是,下行音频信号、第一信号以及第二信号可以为一帧信号或者一段时间的信号。比如,下行音频信号、第一信号以及第二信号均为一帧信号时,下行音频信号、第一信号以及第二信号分别属于三路信号流,下行音频信号的信号帧与第一信号的信号帧以及第二信号的信号帧在同一时间段内或者在时间上存在重叠等。本申请实施例中在执行功能处理(比如ANC、HT或者AH)时,是针对下行音频信号所在的信号流以及第一信号所在的信号流以及第二信号的信号流持续执行功能处理。
首先,下面针对主动降噪通路的处理流程进行详细说明。
参见图5A和图5B所示,为主动降噪处理流程示意图。终端设备100向耳机200发送的下行音频信号,后续描述也可称为第一音频信号。第一音频信号可以是通话信号、音乐信号等。以参考麦克风采集的信号称为第一信号,误差麦克风采集的信号称为第二信号为例进行说明。耳机采用ANC模式。
需要说明的是,终端设备100向耳机200中的左耳机和右耳机发送的下行音频信号可以是相同的信号,也可以是不同的信号。比如,终端设备采用立体声效果,终端设备100向耳机200发送的不同的下行音频信号,以实现立体声效果。当然,终端设备还可以向左耳机和右耳机发送相同的下行音频信号,左耳机和右耳机采用的立体声处理,从而达到立体声效果。左耳机或者右耳机可以可以根据用户的控制执行图5A或图5B的处理。
S501,对参考麦克风采集的第一信号进行第一滤波处理得到第一滤波信号。在图5B中将第一滤波信号简称为信号A1。
S502,滤除误差麦克风采集的第二信号中包括的第一音频信号得到第一经滤除信号。在 图5B中将第一经滤除信号简称为信号A2。
S503,对所述第一滤波信号、所述第一经滤除信号进行混音处理得到第三音频信号。在图5B中将第三音频信号简称为信号A3,即对信号A1和信号A2进行混音处理得到信号A3。
S504,对所述第三音频信号(信号A3)进行所述第三滤波处理得到第四音频信号。在图5B中将第四音频信号简称为信号A4。
S505,对第四音频信号和第一音频信号进行混音处理得到所述第二音频信号。扬声器负责播放第二音频信号。图5B中,将第二音频信号简称为A5。
需要说明的是,当无下行音频信号的情况下,也就是终端设备未向耳机发送第一音频信号的情况下,耳机采用ANC模式时,扬声器输出的信号是未经过混音处理的第四音频信号。在该情况下,无需执行S502和S505。
图5B中,以第一滤波处理为FF滤波处理,采用FF滤波器实现为例,第三滤波处理为FB滤波处理,采用FB滤波器实现为例。耳机200中参考麦克风拾取第一信号,输入到FF滤波器经过FF滤波处理得到信号A1。误差麦克风拾取第二信号,输入到减法器,下行音频信号经过滤波补偿后也输入到减法器,减法器将第二信号中包括的经过滤波补偿后的下行音频信号去除,以消除下行音频信号的影响得到信号A2。通过混音器对信号A1和信号A2进行混音处理后得到信号A3,将信号A4输入FB滤波器,经过FB滤波处理得到信号A4。信号A4与下行音频信号混音后得到信号A5,信号A5输入到扬声器进行播放。
在一种可能的实施方式中,ANC效果的好坏可以由ANC处理的处理强度来决定。ANC处理的处理强度取决于FF滤波采用的FF滤波系数和/或FB滤波采用的FB滤波系数。
针对FF滤波系数来说,一种方式中,可以采用ANC模式下默认的FF滤波系数。另一种方式中,可以通过上一次选择ANC模式时所采用的FF滤波系数。又一种方式中,由耳机根据识别到的场景来确定ANC模式下采用的FF滤波系数。再一种方式中,由用户通过终端设备提供的UI控件来向耳机指示ANC模式下采用的FF滤波系数。比如用户通过终端设备提供的UI控件来选择ANC模式下的处理强度为目标处理强度,不同的处理强度对应于不同的FF滤波系数。
针对FB滤波系数来说,一种方式中,可以采用ANC模式下默认的FB滤波系数。另一种方式中,可以通过上一次选择ANC模式时所采用的FB滤波系数。又一种方式中,由耳机根据识别到的场景来确定采用的FB滤波系数。再一种方式中,由用户通过终端设备提供的UI控件来向耳机指示ANC模式下采用的FF滤波系数,比如用户通过终端设备提供的UI控件来选择ANC模式下的处理强度为目标处理强度,不同的处理强度对应于不同的FB滤波系数。
针对ANC模式下FF滤波系数和FB滤波系数的获取可以是上述提供的方式中的任意组合。作为一种示例,FF滤波系采用ANC模式下默认的滤波系数,FB滤波系数由耳机根据识别到的场景来确定采用的FB滤波系数。作为又一种示例,FB滤波系采用ANC模式下默认的滤波系数,FF滤波系数由用户通过终端设备提供的UI控件来确定。作为又一种示例,FB滤波系采用ANC模式下默认的滤波系数,FF滤波系数是由用户通过终端设备提供的UI控件选择向耳机指示的。后续会通过具体例子来对ANC模式下的处理强度的确定进行详细说明,此处不再重复描述。
其次,下面针对环境声透传通路的处理流程进行详细说明。
参见图6A、图6B以及图6C所示,为环境声透传处理流程示意图。终端设备100向耳 机200发送的下行音频信号,后续描述以称为第一音频信号为例。第一音频信号可以是通话信号、音乐信号等。以参考麦克风采集的信号称为第一信号,误差麦克风采集的信号称为第二信号为例进行说明。耳机200中的左耳机或者右耳机可以可以根据用户的控制执行图6A或图6B或图6C的处理。
S601,对参考麦克风采集的第一信号进行第一信号处理得到第一经处理信号。在图6B和图6C中将第一经处理信号称为信号B1。所述第一信号处理包括HT滤波。
S602,对所述第一经处理信号和所述第一音频信号进行混音处理得到第五音频信号。在图6B和图6C中将第五音频信号称为信号B2。
也就是,对信号B1和下行音频信号(即第一音频信号)进行混音处理得到信号B2。
S603,滤除所述第二信号包括的所述第五音频信号得到第二经滤除信号。在图6B和图6C中将第二经滤除信号称为信号B3。也就是,滤除第二环境信号包括的信号B2得到信号B3。
S604,对所述第二经滤除信号进行FB滤波得到第三经滤除信号。在图6B和图6C中将第三经滤除信号称为信号B4。也就是,对信号B3进行FB滤波得到信号B4。
S605,将所述第三经滤除信号与所述第五音频信号进行混音处理得到所述第二音频信号。即将信号B4和信号B2进行混音处理得到第二音频信号。
作为一种示例,在对参考麦克风采集的第一信号进行第一信号处理得到第一经处理信号时,可以通过如下方式实现:
对所述第一信号进行HT滤波处理得到第二滤波信号,在图6B和图6C中,将第二滤波信号称为信号B5。进一步地,对所述第二滤波信号进行第二信号处理得到所述第二经处理信号;第二信号处理也可以称为低时延算法处理。低时延算法处理包括去闭塞效应处理、降底噪处理、降风噪处理、增益调节处理或者频响调节处理中的一项或者多项。
一种可能的方式中,HT滤波处理可以由降噪处理单元来实现,参见图6B所示。以耳机的降噪处理单元中包括CODEC为例。CODEC包括HT滤波器、FB滤波器、减法器、第一混音器、第二混音器以及滤波补偿单元。图6B以降噪处理单元还包括DSP为例。DSP可以用于执行低时延算法处理。耳机200中参考麦克风拾取第一信号,输入到HT滤波器经过HT滤波处理得到信号B5。信号B5输入到DSP,DSP对信号B5进行低时延算法处理得到信号B1。信号B1输入到第一混音器,第一混音器对下行音频信号和信号B1进行混音处理得到信号B2。信号B2经过滤波补偿单元的滤波补偿处理后,输入到减法器。减法器用于滤除误差麦克风拾取的第二环境信号中包括的经过滤波补偿处理的信号B2得到信号B3。信号B3输入到FB滤波器,FB滤波器对信号B3进行FB滤波处理后得到信号B4。信号B4输入到第二混音器。另外第二混音器的输入还包括信号B2。第二混音器对信号B2和信号B4进行混音处理得到第二音频信号,第二音频信号输入到扬声器进行播放。
另一种可能的方式中,HT滤波处理可以由DSP来实现,参见图6C所示。DSP可以用于执行HT滤波处理以及低时延算法处理。耳机的降噪处理单元中包括FB滤波器、减法器、第一混音器、第二混音器以及滤波补偿单元。耳机中参考麦克风拾取第一信号,输入到DSP。DSP对第一信号进行HT滤波处理,以及低时延算法处理得到信号B1。信号B1输入到第一混音器,第一混音器对下行音频信号和信号B1进行混音处理得到信号B2。信号B2经过滤波补偿单元的滤波补偿处理后,输入到减法器。减法器用于滤除误差麦克风拾取的第二信号中包括的信号B2得到信号B3。信号B3输入到FB滤波器,FB滤波器对信号B3进行FB滤波处理后得到信号B4。信号B4输入到第二混音器。另外第二混音器的输入还包括信号B2。 第二混音器对信号B2和信号B4进行混音处理得到第二音频信号,第二音频信号输入到扬声器进行播放。
作为一种示例,低时延算法处理包括去闭塞效应处理。在介绍去闭塞效应处理方法之前,先对闭塞效应产生的原理进行说明。耳机佩戴者说话声音有两个途径被感知:途径1、从骨传导到骨膜处被感知,该信号只有低频信号。途径2、从外部空气传播到骨膜处被感知,该信号包含低频信号和中高频信号。当低频信号和中高频信号叠加后,由于低频信号太足,同时由于耳机佩戴时低频信号无法散发出去,导致耳朵内部低频混沌,从而产生闭塞效应。
HT滤波处理后得到的信号B5经过去闭塞效应处理,具体可以采用如下方式。
方式一,参见图7所示。
S701,在语音谐波集合中确定与所述骨传导信号匹配的第一语音谐波信号,所述语音谐波集合中包括多个语音谐波信号。语音谐波集合中包括的多个语音谐波信号对应于不同的频率。具体的,可以确定骨传导信号的频率,根据骨传导信号的频率从语音谐波集合中确定第一语音谐波信号。语音谐波信号也可以称为语音谐波分量。
S702,去除HT滤波处理后得到的信号B5中的所述第一语音谐波信号。比如去除HT滤波处理后得到的信号B5中的第一语音谐波信号得到信号C1。由于一般骨传导传感器采集到的人说话的声音一般为低频谐波分量,因此,S702为去除信号B5中的低频谐波分量,从而得到不包括低频谐波分量的信号C1。
S703,将去除所述第一语音谐波信号的信号B5中的高频分量进行放大。即将信号C1的高频分量进行放大。
在语音谐波集合中能够确定与骨传导信号匹配的第一语音谐波信号,也就是说,骨传导传感器能够检测到骨传导信号,即耳机佩戴者目前正在发音,比如说话或者唱歌等。在信号C1的基础上提升高频分量得到的信号中仅包括中高频分量,从而耳机佩戴者听见的信号相对无闭塞效应。
语音谐波集合可以是预先存储在耳机中的。作为一种示例,语音谐波集合可以采用离线方式获得或者采用在线方式获得。
采集离线方式获得时,可以通过骨传导传感器采集多个人的骨传导信号,对多个人的骨传导信号中每个人的骨传导信号均执行如下处理:以第一骨传导信号为例,对所述第一骨传导信号进行FFT变化到频域信号;通过导频找基频的方式确定所述频域信号中的基频信号;根据所述基频信号确定所述骨传导信号的谐波分量,从而获得骨传导信号的频率与谐波分量的映射关系得到语音谐波集合。语音谐波集合中可以包括不同频率与谐波分量的映射关系。
采用在线方式获得时,可以通过耳机中的骨传导传感器设定时长内采集的第二骨传导信号,设定时长内可能有多个人使用耳机,也可能仅有一个人使用耳机,即该用户;针对第二骨传导信号时执行如下处理:
对所述第二骨传导信号进行FFT变化到频域信号;通过导频找基频的方式确定所述频域信号中的基频信号。如果设定时长内有多个人使用耳机,可能确定设定时长内多个不同时间段分别对应的多个基频信号。可以根据多个基频信号确定所述骨传导信号的多个谐波分量,从而获得频率与谐波分量的映射关系得到语音谐波集合。语音谐波集合中可以包括不同频率与不同谐波分量的映射关系。
方式二,可以采用对经过HT滤波处理后得到的信号B5进行自适应滤波处理,以去除信号B5中的低频分量得到信号C1,即将信号B5中耳机佩戴者发音的信号抠除。将去除所述 低频分量的第三滤波信号中的高频分量进行放大,即对信号C1中的高频分量进行放大。在信号C1的基础上提升高频分量得到的信号中仅包括中高频分量,从而耳机佩戴者听见的信号相对无闭塞效应。
在一种可能的实施方式中,HT效果的好坏可以由HT处理的处理强度来决定。HT处理的处理强度取决于HT滤波采用的HT滤波系数和/或FB滤波采用的FB滤波系数。
针对HT滤波系数来说,一种方式中,可以采用HT模式下默认的HT滤波系数。另一种方式中,可以通过上一次选择HT模式时所采用的HT滤波系数。又一种方式中,由耳机根据识别到的场景来确定HT模式下采用的HT滤波系数。再一种方式中,由用户通过终端设备提供的UI控件来向耳机指示HT模式下采用的HT滤波系数。比如用户通过终端设备提供的UI控件来选择HT模式下的处理强度为目标处理强度,不同的处理强度对应于不同的HT滤波系数。针对FB滤波系数来说,一种方式中,可以采用HT模式下默认的FB滤波系数。另一种方式中,可以通过上一次选择HT模式时所采用的FB滤波系数。又一种方式中,由耳机根据识别到的场景来确定采用的FB滤波系数。再一种方式中,由用户通过终端设备提供的UI控件来向耳机指示HT模式下采用的HT滤波系数,比如用户通过终端设备提供的UI控件来选择HT模式下的处理强度为目标处理强度,不同的处理强度对应于不同的FB滤波系数。
针对HT模式下HT滤波系数和FB滤波系数的获取可以是上述提供的方式中的任意组合。
再次,下面针对听觉增强通路的处理流程进行详细说明。
参见图8A、图8B以及图8C所示,为听觉增强处理流程示意图。终端设备100向耳机110发送的下行音频信号,后续描述以称为第一音频信号为例。第一音频信号可以是通话信号、音乐信号、提示音等。以参考麦克风采集的信号称为第一信号,误差麦克风采集的信号称为第二信号为例进行说明。耳机200中的左耳机或者右耳机可以可以根据用户的控制执行图8A或图8B或图8C的处理。
S801,对参考麦克风采集的第一信号进行HT滤波得到第二滤波信号(信号C1)。在图8B和图8C中,第二滤波信号称为信号C1。
S802,对所述第二滤波信号(即信号C1)进行增强处理得到滤波增强信号。在图8B和图8C中,将滤波增强信号简称为信号C2。
S803,对所述第一信号进行FF滤波得到第一滤波信号。在图8B和图8C中,将第一滤波信号简称为信号C3。
S804,对所述滤波增强信号和所述第一音频信号进行混音处理得到第六音频信号。在图8B和图8C中,将第六音频信号简称为信号C4。步骤S804也就是将信号C2和下行音频信号进行混音处理得到信号C4。
S805,滤波所述第二信号包括的所述第六音频信号得到第四经滤除信号。在图8B和图8C中,将第四经滤除信号简称为信号C5。步骤S805也就是滤除第二环境信号包括的信号C4得到信号C5。
作为一种示例,在执行步骤S805时,可以先对信号C4进行滤波补偿处理得到经补偿信号,然后滤除第二信号包括的经补偿信号得到C5。
S806,对所述第四经滤除信号进行FB滤波得到第五经滤除信号。
在图8B和图8C中,将第四经滤除信号简称为信号C6。步骤S806也就是对信号C5进行FB滤波得到信号C6。
S807,对所述第五经滤除信号、所述第六音频信号以及所述第一滤波信号进行混音处理 得到所述第二音频信号。步骤S806也就是将信号C6、信号C4以及信号C3进行混音处理得到第二音频信号。
在一种可能的实施方式中,对所述第二滤波信号(即信号C1)进行增强处理得到滤波增强信号(信号C2),可以通过如下方式一或方式二实现。
方式一,参见图9所示。
S901,对所述第二滤波信号(即信号C1)进行去闭塞效应处理。
其中,对信号C1进行去闭塞效应处理的方式,可以采用与对信号B5进行去闭塞效应处理的方式相同,具体参见场景二中的方式一和方式二,此处不再赘述。
然后对去闭塞效应处理后的信号进行降噪处理,所述降噪处理包括人工智能AI降噪处理和/或降风噪处理;图9中以降噪处理包括AI降噪处理和降风噪处理为例。
S902,对去闭塞效应处理后的信号进行AI降噪处理。
S903,对AI降噪处理后的信号进行降风噪处理。
S904,对所述降风噪处理后得到的信号进行增益放大处理。
S905,对增益放大处理后的信号进行频响调节得到所述滤波增强信号。
上述S904中,对将风噪处理后得到的信号进行增益放大处理,一种可行的方式是,直接放大风噪处理后得到的信号。直接放大信号的方式在放大外界信号的同时,也放大了佩戴者说话的声音。本申请实施例提供一种增益放大处理方式,仅放大外界信号,对佩戴者说话的语音信号不放大。比如,参见图10所示,对降噪处理后得到的信号进行增益放大处理时,可以通过如下方式实现。
佩戴者说话的语音信号经过骨传导到骨膜处,语音信号集中在低频,记为骨传导信号D1。该骨传导信号D1通过骨传导传感器采集得到。
1、对骨传导信号D1进行谐波扩展得到经谐波扩展信号,比如将经谐波扩展信号称为D2。示例性地,谐波扩展可以采用谐波增强方法或者采用骨传导信号D1的谐波直接向上扩展的方法。
2,使用第一增益系数(gain)对降噪处理得到的信号进行放大处理。为了描述方便,将降噪处理得到的信号称为信号D3。使用第一增益系数对信号D3进行放大处理得到信号D4。这里的放大处理,可以信号直接放大。
3,使用第一滤波系数滤除所述放大处理得到的信号中包括的经谐波扩展信号得到信号D5。使用第一滤波系数采用自适应滤波的方式滤除信号D4中包括的D2。此时信号D5即是已经滤除了佩戴者说话语音的信号。其中,所述第一滤波系数基于所述第一增益系数确定。使用第一增益系数gain调整自适应滤波的力度,也可以称为第一滤波系数,即采用第一增益系数进行信号放大的dB数与自适应滤波就滤除的dB数相同,使得佩戴者语音信号能够保持平衡,不被放大也不被缩小。
方式二,参见图11所示。
S1101,对所述第二滤波信号(即信号C1)进行去闭塞效应处理得到经去闭塞信号;
其中,对信号C1进行去闭塞效应处理的方式,可以采用与对信号B5进行去闭塞效应处理的方式相同,具体参见场景二中的方式一和方式二,此处不再赘述。
S1102,对所述经去闭塞信号进行音频事件检测,以获取所述经去闭塞信号中的音频事件信号(可以简称为事件信号)。音频事件信号,比如报站音、鸣笛音等。
S1103,对所述经去闭塞信号中的音频事件信号进行增益放大处理。
对所述经去闭塞信号中的音频事件信号,比如报站音、鸣笛音等进行增益放大处理,使 得耳机佩戴者能够清晰的听见报站音或者鸣笛音。
S1104,对增益放大处理后的信号进行频响调节得到所述滤波增强信号。
方式二中,对所述经去闭塞信号中的音频事件信号进行增益放大处理的方式,可以采用与对降噪处理后得到的信号进行增益放大处理时,同样的方式,此处不再赘述。
一种可能的方式中,参见图8B所示,以降噪处理单元包括CODEC和DSP为例。耳机的CODEC中包括HT滤波器、FB滤波器、FF滤波器、减法器、第一混音器、第二混音器以及滤波补偿单元。HT滤波处理由CODEC来实现。DSP可以用于执行增强处理。耳机110中参考麦克风拾取第一信号,输入到HT滤波器经过HT滤波处理得到信号C1。信号C1输入到DSP,DSP对信号C1进行增强处理得到信号C2。信号C2输入到第一混音器,第一混音器对下行音频信号和信号C2进行混音处理得到信号C4。信号C4经过滤波补偿单元的滤波补偿处理后,输入到减法器。减法器用于滤除误差麦克风拾取的第二环境信号中包括的经过滤波补偿后的信号C4得到信号C5。信号C5输入到FB滤波器,FB滤波器对信号C5进行FB滤波处理后得到信号C6。信号C6输入到第二混音器。另外第二混音器的输入还包括信号C4和信号C3。第二混音器对信号C3、信号C4和信号C6进行混音处理得到第二音频信号,第二音频信号输入到扬声器进行播放。
另一种可能的方式中,参见图8C所示,以降噪处理单元包括CODEC和DSP为例。DSP可以用于执行HT滤波处理以及增强处理。耳机的CODEC中包括FB滤波器、FF滤波器、减法器、第一混音器、第二混音器以及滤波补偿单元。耳机110中参考麦克风拾取第一信号,输入到DSP。DSP对第一信号进行HT滤波处理得到信号C1。DSP对信号C1进行增强处理得到信号C2。信号C2输入到第一混音器,第一混音器对下行音频信号和信号C2进行混音处理得到信号C4。信号C4经过滤波补偿单元的滤波补偿处理后,输入到减法器。减法器用于滤除误差麦克风拾取的第二环境信号中包括的经过滤波补偿后的信号C4得到信号C5。信号C5输入到FB滤波器,FB滤波器对信号C5进行FB滤波处理后得到信号C6。信号C6输入到第二混音器。另外第二混音器的输入还包括信号C4和信号C3。第二混音器对信号C3、信号C4和信号C6进行混音处理得到第二音频信号,第二音频信号输入到扬声器进行播放。
在一种可能的实施方式中,AH效果的好坏可以由AH处理的处理强度来决定。AH处理的处理强度取决于HT滤波系数、FB滤波系数或者FF滤波系数中的至少一个。
针对FF滤波系数来说,一种方式中,可以采用AH模式下默认的FF滤波系数。另一种方式中,可以通过上一次选择AH模式时所采用的FF滤波系数。又一种方式中,由耳机根据识别到的场景来确定AH模式下采用的FF滤波系数。再一种方式中,由用户通过终端设备提供的UI控件来向耳机指示AH模式下采用的FF滤波系数。比如用户通过终端设备提供的UI控件来选择AH模式下的处理强度为目标处理强度,不同的处理强度对应于不同的FF滤波系数。针对HT滤波系数来说,一种方式中,可以采用AH模式下默认的HT滤波系数。另一种方式中,可以通过上一次选择AH模式时所采用的HT滤波系数。又一种方式中,由耳机根据识别到的场景来确定AH模式下采用的HT滤波系数。再一种方式中,由用户通过终端设备提供的UI控件来向耳机指示AH模式下采用的HT滤波系数。比如用户通过终端设备提供的UI控件来选择AH模式下的处理强度为目标处理强度,不同的处理强度对应于不同的HT滤波系数。针对FB滤波系数来说,一种方式中,可以采用AH模式下默认的FB滤波系数。另一种方式中,可以通过上一次选择AH模式时所采用的FB滤波系数。又一种方式中,由耳机根据识别到的场景来确定采用的FB滤波系数。再一种方式中,由用户通过终端设备提 供的UI控件来向耳机指示AH模式下采用的HT滤波系数,比如用户通过终端设备提供的UI控件来选择AH模式下的处理强度为目标处理强度,不同的处理强度对应于不同的FB滤波系数。
针对AH模式下HT滤波系数、FB滤波系数或者FF滤波系数获取可以是上述提供的方式中的任意组合。
耳机200(包括左耳机和右耳机)采用的处理模式的确定,可以由用户通过终端设备100上UI控件进行选择并向耳机指示,或者终端设备根据自适应识别的场景来确定并向耳机指示,或者还可以由耳机根据自适应的场景识别来确定。
下面以举例的方式,对耳机的处理模式的确定方式进行说明。
示例1,单控件控制左耳机和右耳机。
终端设备100提供控制界面,供用户根据需求选择耳机200(包括左耳机和右耳机)的处理模式:空模式、ANC模式、HT模式或者AH模式。空模式下不进行任何处理。应理解的是,控制界面上供用户选择的耳机的处理模式均为耳机支持的处理模式。在示例1中,左耳机和右耳机所具备的处理功能相同,或者说支持的处理模式相同。例如,左耳机支持AHA,右耳机也支持AHA。比如,终端设备安装有适配耳机200的耳机应用,在适配过程中,能够获知耳机所具备的处理功能。再比如,耳机200与终端设备建立连接的通信过程中,通过向终端设备传输功能参数,从而终端设备能够根据功能参数确定耳机所具备的处理功能。
以用户选择ANC模式为例。控制界面中包括用户界面(user interface,UI)控件。UI控件用于供用户选择耳机200的处理模式,为了便于区分,将用于供用户选择耳机的处理模式的UI控件称为选择控件。处理模式包括ANC模式、HT模式或者AH模式中的至少两项。终端设备100响应于用户通过选择控件在耳机支持的处理模式中选择目标模式时,向左耳机和右耳机分别发送控制信令1。控制信令1携带目标模式。选择控件还可以用于选择目标模式下的处理强度。选择控件的形状可以是圆环形,条形或者其它形状。选择控件可以包括第一控件和第二控件。第二控件在第一控件的任意两个不同位置对应于耳机的不同处理模式或者第二控件在第一控件的两个不同位置对应于耳机的同一处理模式下的不同处理强度。用户通过移动代表用户选择的第二控件在显示屏的第一控件上位置来实现对不同处理模式的选择以及处理强度的控制。
一种可能的实施方式中,通过耳机应用(appliaction,APP)来实现对左耳机和右耳机的处理模式的控制。
终端设备100中包括用于控制耳机的耳机控制应用,简称耳机应用。比如参见图12A所示的终端设备的主界面。在耳机与终端设备连接后,当用户点击桌面上耳机应用(appliaction,APP)的图标001时,终端设备可以响应于用户点击图标001的操作,启动耳机应用,在显示屏上显示耳机应用的控制界面,或者解释为启动耳机应用时,弹出耳机应用的控制界面。应用的名称可以叫做音频助手,该控制功能也可以集成在终端系统中的设置选项中。
例如以选择控件为圆环形为例,参见图12B所示。图12B中以左耳机和右耳机均支持ANC模式、HT模式和AH模式为例。图12B中圆环形的选择控件中的第一控件包括三部分的圆弧段,分别对应于ANC模式、HT模式和AH模式。第二控件位于ANC模式的圆弧段,则确定为ANC模式,第二控件位于ANC模式的圆弧段的不同位置对应于ANC模式下的不同处理强度。第二控件位于HT模式的圆弧段,则确定为HT模式,第二控件位于HT模式的圆弧段的不同位置对应于HT模式下的不同处理强度。第二控件位于AH模式的圆弧段,则 确定为AH模式,第二控件位于AH模式的圆弧段的不同位置对应于AH模式下的不同处理强度。作为一种可选设置,ANC最强降噪位置与HT透传强度最弱位置相接,听觉效果上可以平滑过渡;HT透传强度最强位置与AH增强强度最弱位置相接,听觉效果上也可以平滑过渡。
圆环(或者圆周)上的突出显示的黑色圆点表示用户选择处理强度的第二控件,用户可以通过在圆周上移动黑色圆点的位置,来实现对不同处理模式的选择以及处理强度的控制。终端设备100(比如处理器)响应于用户在控制界面上执行的操作1,比如操作1是用户移动代表用户选择的第二控件在显示屏的第二控件的位置时产生的。终端设备100向左耳机和右耳机分别发送控制指令1,控制指令1用于指示目标模式以及目标处理强度。图12B中以目标模式为ANC模式。
一种示例中,控制指令1可以包括ANC标识和指示执行ANC处理时的目标处理强度的参数值。在ANC模式下,不同的处理强度(即不同的处理强度的取值)对应的FB滤波系数和/或FF滤波系数不同。
另一种示例中,控制指令1中包括弧度。根据弧度所在的范围可以确定对应的处理模式。不同的弧度值对应该处理模式下的处理强度。参见图12B所示,(0,180]对应的处理模式为ANC模式,(180,270]对应的处理模式为HT模式,(270,360]对应的处理模式为AH模式。左耳机和右耳机中可以包括弧度范围与处理模式的映射关系,以及弧度值与滤波系数的映射关系。以ANC模式为例,不同的弧度值对应的FB滤波系数和FF滤波系数不同。
以图12B为例,用户可以通过触控圆盘中黑色圆点,从0度到360度顺时针旋转,0度时对应的FF滤波系数和FB滤波系数使得ANC降噪效果最强,即用户感知到的用户当前所处环境的声音以及用户耳道内部的环境声音越弱。旋转后FF滤波系数和FB滤波系数发生变化,使得主动降噪效果逐渐减弱,到180度时主动降噪效果最弱,类似于耳机佩戴后不做任何降噪处理。从180度之后到270度是控制环境声透传部分,用户通过触控圆盘中黑色圆点,从180度之后到270度顺时针旋转,180度时对应的HT滤波系数和FB滤波系数使得环境声透传效果最弱,即用户感知到的用户当前所处环境的声音强度越小,类似于耳机佩戴后采用空模式。顺时针旋转后,HT滤波系数和FB滤波系数发生变化,使得环境声透传效果越来越强。270度到360度用于控制听觉增强,即用户通过触控圆盘中黑色圆点,180度时对应的FF滤波系数、HT滤波系数和FB滤波系数使得听觉增强效果最弱,用户感知到的用户当前所处环境声音中包括的事件声音越弱。顺时针旋转后,FF滤波系数、HT滤波系数和FB滤波系数发生变化,使得听觉增强效果越来越强,即用户希望听见的事件信号越来越强,辅助听觉。
以终端设备100与左耳机和右耳机通过蓝牙连接为例。
以选择ANC模式为例,参见图12C所示,终端设备100响应于用户的操作1,通过蓝牙向左耳机和右耳机分别发送控制指令1。以控制指令1可以包括ANC标识和目标处理强度的参数值。左耳机和右耳机接收到控制指令1后执行的操作类似,后续描述时以左耳机的处理为例。左耳机的主控制单元接收到控制指令1后,根据所述ANC标识和目标处理强度从系数库中获取ANC处理的FF滤波系数和FB滤波系数。
比如,系数库中包括表1所示的映射关系。表1仅是一种示例,并构成映射关系的具体限定。例如,目标处理强度的参数值为强度1,左耳机的主控制单元根据表1获取强度1对应的FF滤波系数为系数FF1,FB滤波系数为系数FB1,主控制单元控制FF滤波器采用系数FF1对参考麦克风采集的第一信号进行FF滤波处理得到信号A1。主控制单元控制FB滤波 器采用系数FB1对信号A3进行FB滤波处理得到所述第二音频信号。具体的,主控制单元将系数FF1和系数FB1写入AHA内核中,从而AHA内核执行S501-S504的步骤得到第二音频信号。
表1
Figure PCTCN2021103435-appb-000001
以选择HT模式为例,参见图12D所示,终端设备100响应于用户的操作1,通过蓝牙向左耳机和右耳机分别发送控制指令1。以控制指令1可以包括HT标识和目标处理强度,目标处理强度用于指示执行HT处理时的处理强度。左耳机和右耳机接收到控制指令1后执行的操作类似,后续描述时以左耳机的处理为例。左耳机的主控制单元接收到控制指令1后,根据所述HT标识和目标处理强度从系数库中获取HT处理的HT滤波系数和/或FB滤波系数。
以表1为例,目标处理强度的取值为强度5,左耳机的主控制单元根据表1获取强度5对应的HT滤波系数为系数HT1,FB滤波系数为系数FB5,主控制单元控制HT滤波器采用系数HT1对参考麦克风采集的第一信号进行HT滤波处理。主控制单元控制FB滤波器采用系数FB5对信号B3进行FB滤波处理。具体的,主控制单元将系数HT1和系数FB5写入AHA内核中,从而AHA内核执行S601-S605的步骤得到第二音频信号。
以选择AH模式为例,参见图12E所示,终端设备100响应于用户的操作1,通过蓝牙向左耳机和右耳机分别发送控制指令1。以控制指令1可以包括AH标识和目标处理强度的参数值。左耳机和右耳机接收到控制指令1后执行的操作类似,后续描述时以左耳机的处理为例。左耳机的主控制单元接收到控制指令1后,根据所述HT标识和目标处理强度从系数库中获取AH处理的HT滤波系数、FF滤波系数和FB滤波系数。
以表1为例,目标处理强度的取值为指示7,左耳机的主控制单元根据表1获取指示7对应的HT滤波系数为系数HT3,FB滤波系数为系数FB7,FF滤波系数为系数FF5。主控制单元控制HT滤波器采用系数HT3对参考麦克风采集的第一信号进行HT滤波处理。主控制单元控制FB滤波器采用系数FB7对信号C5进行FB滤波处理。主控制单元控制FF滤波器采用系数FF5对第一信号进行FF滤波处理。具体的,主控制单元将系数HT3、系数FB7以及系数FF5写入AHA内核中,从而AHA内核执行S801-S807的步骤得到第二音频信号。
示例性地,参见图12F所示,选择控件可以采用条形。选择控件包括第一控件和第二控件。第一控件的条形可以分为多个条形段,根据耳机支持的处理模式的数量划分。第二控件 在第一控件的不同的条形段指示不同的处理模式。第二控件在第一控件的同一条形段的不同位置指示同一处理模式的不同处理强度。图12F以左耳机和右耳机均支持AHA为例,第一控件的条形包括三个条形段。
以图12F为例,用户可以通过触控黑色条形左右滑动,黑色条形位于K1位置时对应的FF滤波系数和FB滤波系数使得ANC降噪效果最强,向右滑动后FF滤波系数和FB滤波系数发生变化,使得主动降噪效果逐渐减弱,到K2位置时主动降噪效果最弱,类似于耳机佩戴后不做任何降噪处理。从K2到K3之间的区域是控制环境声透传部分,用户通过触控黑色条形,从位置K2移动到K3,黑色条形移动到位置K2时,对应的HT滤波系数和FB滤波系数使得环境声透传效果最弱,移动到位置K3,HT滤波系数和FB滤波系数发生变化,使得环境声透传效果越来越强。位置K3到位置K4度用于控制听觉增强,即用户通过触控黑色条形,从位置K3移动到位置K4,黑色条形移动到位置K3时,对应的FF滤波系数、HT滤波系数和FB滤波系数使得听觉增强效果最弱,从位置K3向位置K4移动,FF滤波系数、HT滤波系数和FB滤波系数发生变化,使得听觉增强效果越来越强,即用户希望听见的语音信号越来越强,辅助听觉。
示例性地,参见图12G所示,在(a)中选择控件包括不同的处理模式对应的按钮,包括ANC按钮、HT按钮和AH按钮。以ANC模式为例,终端设备100响应于用户点击ANC按钮的操作,显示图12G(b)的显示界面。(b)的显示界面包括用于选择处理强度的控件002。用户可以通过触控黑色条形上下滑动来确定ANC的处理强度,即选择对应的FF滤波系数和FB滤波系数。黑色条形在L1-L2的区域内滑动。黑色条形位于L1位置时对应的FF滤波系数和FB滤波系数使得ANC降噪效果最强,向下滑动后FF滤波系数和FB滤波系数发生变化,使得主动降噪效果逐渐减弱,到L2位置时主动降噪效果最弱,类似于耳机佩戴后不做任何降噪处理。
另一种可能的实施方式中,耳机200与终端设备建立连接时,可以触发启动耳机APP,显示包括选择控件的控制界面,比如图12A、图12B、图12F或者图12G示例的控件界面。
比如终端设备显示的界面为界面1,终端设备识别到耳机200与终端设备建立连接,可以由界面1跳转到控制界面。
又一种可能的实施方式中,耳机与终端设备建立连接后,终端设备触发耳机播放音频时,可以触发启动耳机APP,即显示包括选择控件的控制界面,比如图12A、图12B、图12C或者图12D示例的显示界面。比如,终端设备触发耳机播放音频时,可以是终端设备与耳机建立连接后,播放歌曲,可以显示包括选择控件的控制界面。再比如,终端设备与耳机建立连接后,播放视频,可以显示包括选择控件的控制界面。
再一种可能的实施方式中,耳机与终端设备建立连接后,在终端设备通过耳机播放音频过程中,根据识别到的当前外部环境的场景类型为目标场景,所述目标场景适配需要调整所述第一目标耳机的处理模式的场景类型,可以显示提示信息,提示信息用于提示用户是否调整所述耳机的处理模式,参见图12H所示,以提示信息为提示框的形式为例。响应于用户选择调整耳机的处理模式的操作时,可以显示包括选择控件的控制界面,比如图12A、图12B、图12C或者图12D示例的控制界面。图12E以显示图12A示例的控制界面为例。
例如:终端设备识别用户当前所处的场景为嘈杂场景,在这种场景下,用户可能需要启动处理模式,从而显示选择提示信息(比如采用提示框的形式)来提示用户是否调整耳机的处理模式。例如:终端设备识别外部环境的场景类型为嘈杂场景,在这种场景下,用户可能 需要启动处理模式,从而显示提示框来提示用户是否调整耳机的处理模式。
作为一种举例,触发显示提示框的场景类型可以包括:嘈杂场景、航站楼场景、火车站场景、汽车站场景、马路场景等。
比如,识别到信号的强度达到设定阈值认为是嘈杂场景。又比如,识别到飞机报站的特定声音,确定识别到航站楼场景。再比如,识别到火车时间通知的声音,确定识别到火车站场景;再比如,识别到汽车票务广播,确定识别到汽车站场景。再比如,识别到信号灯的滴答声或者汽车的鸣笛声等,确定识别到马路场景。
再一种可能的场景中,耳机与终端设备建立连接后在终端设备通过耳机播放音频过程中,根据识别到的用户当前所处的场景,显示包括选择控件的控制界面。
示例2,双控件控制左右耳机
终端设备100提供控制界面,供用户根据需求分别选择左耳机和右耳机的处理模式,左耳机和右耳机的处理模式可以不同。比如左耳机选择ANC模式,右耳机选择HT模式。控制界面中包括左耳机的选择控件和右耳机的选择控件。为了便于区分,将左耳机的选择控件称为第一选择控件,右耳机的选择控件称为第二选择控件。第一选择控件用于供用户选择左耳机的处理模式,第二选择控件用于供用户选择右耳机的处理模式。第一选择控件、第二选择控件的形状可以是圆环形,条形或者其它形状。第一选择控件和第二选择控件的形态可以相同,也可以不同。用户通过移动代表用户选择的控件在显示屏上位置来实现对不同处理模式的选择以及处理强度的控制。左耳机和右耳机采用的控件的形状可以参见示例1中描述,此处不再赘述。
以左耳机和右耳机均采用圆环形的选择控件为例进行说明。参见图13所示,第一选择控件和第二选择控件均包括第一控件和第二控件。第二控件在第一控件的不同的两个位置对应于不同的处理模式,或者,第二控件在第一控件的不同的两个位置对应于同一处理模式下的不同处理强度。参见图13所示的控制界面为例。用户可以通过左耳机的第一选择控件中第二控件(黑色圆点)在第一控件的圆周上移动的位置,来实现对左耳机所实现的不同处理模式的选择以及处理强度的控制。用户可以通过移动右耳机的第二选择控件中第二控件在第一控件上的位置,来实现对右耳机所实现的不同处理模式的选择以及处理强度的控制。在示例2中,用户可以针对左右耳机可以选择不同的处理模式,或者相同的处理模式下相同的处理强度,或者相同的处理模式下不同的处理强度,从而可以匹配耳朵差异性或配合不同应用的需要。
在示例2中,触发显示包含第一选择控件和第二选择控件的控制界面的方式参见示例1的描述,此处不再赘述。
示例3,终端设备智慧场景检测。
终端设备识别用户当前所处的场景,在不同的场景下,耳机所采用的处理模式不同。终端设备识别到当前外部环境的场景类型指示为第一场景时,确定耳机的处理模式中对应于第一场景的目标模式,分别向左耳机和右耳机发送控制信令2。控制信令2用于指示目标模式。不同的目标模式对应于不同的场景类型。
本申请实施例中终端设备根据识别出的场景确定指示耳机执行哪种功能。将AHA的功能对应到场景类型上,选取该场景类型最合适的功能,让用户自动感受到最需要的效果。
作为一种示例,场景类型可以包括走路场景、跑步场景、安静场景、多人说话场景、咖 啡厅场景、地铁场景、火车场景、汽车场景、候车厅场景、对话场景、办公室场景、户外场景、开车场景、大风场景、飞机场景、警报声场景、鸣笛声场景、哭叫声场景等。
终端设备在进行智慧场景检测是可以通过人工智能(artificial intelligence,AI)模型方式进行检测分类。AI模型可以采用离线方式构建并存储到终端设备上。比如,终端设备上的麦克风录制大量噪声和不同场景下的传感器(sensor)数据和/或视频处理单元(video processing unit,vpu)数据,将数据人工标注其对应的场景。其次,初始化构建一个AI模型,模型可以是卷积神经网络(convolutional neural networks,CNN)/深度神经网络(deep neural networks,DNN)/长短期记忆网络(long short-term memory,LSTM)的一种、也可以是不同模型的混合使用。再次,利用标注的数据进行模型训练,得到相应的AI模型。使用时将实时采集的外部环境的声音信号输入AI模型计算,得到分类结果。
作为一种示例,例举不同的场景类型所适合的处理模式。走路场景(HT)、跑步场景(HT)、安静场景(HT)、多人说话场景(ANC)、咖啡厅场景(ANC)、地铁场景(AH)、火车场景(ANC)、候车厅场景(AH)、对话场景(AH)、办公室场景(ANC)、户外场景(ANC)、开车场景(ANC)、大风场景(ANC)、飞机场景(ANC)、警报声场景(AH)、鸣笛声场景(AH)、哭叫声场景(AH)等,括号中表示该场景类型对应的处理模式。比如飞机场景,飞机飞行时噪声较大,适合采用ANC模式。再比如,走路场景和跑步场景以及安静场景中,适合采用HT模式,可以听见突发的事件声音。再比如,咖啡厅场景中,用户需求安静,可以采用ANC模式。再比如,轻音乐场景中,可以采用HT模式。再比如,警报声场景(AH)、鸣笛声场景(AH)、哭叫声场景(AH),用于需要听见预设声音,适合采用AH模式。
作为一种举例,飞机场景中,当终端设备100识别到当前外部环境的场景类型为飞机场景时,可以向耳机发送控制信令2,控制信令2用于指示耳机需实现ANC功能,即指示耳机采用ANC模式。左耳机和右耳机接收到控制信令2后,分别执行S501-S504的处理。
作为一种举例,走路场景中,当终端设备100识别到当前外部环境的场景类型为走路场景时,可以向耳机发送控制信令2,控制信令2用于指示耳机需实现HT功能,即耳机采用HT模式。左耳机和右耳机接收到控制信令2后,分别执行S601-S605的处理。
作为又一种举例,警报声场景中,当终端设备100识别到当前外部环境的场景类型为火车站场景时,可以向耳机发送控制信令2,控制信令2用于指示耳机需实现AH功能,即耳机采用AH模式。左耳机和右耳机接收到控制信令2后,分别执行S801-S807的处理。
在一种可能的实施方式中,耳机与终端设备建立连接后,即终端设备启动场景检测。终端设备检测完成后,还可以将检测结果显示给用户,使得用户知晓耳机采用的处理模式。比如通过提示框的形式将检测结果显示给用户。检测结果中可以包括所检测到的场景,还可以包括对应于检测到的场景的处理模式。比如,终端设备识别到场景为第一场景时,确定耳机的处理模式中对应于第一场景的目标模式,可以将检测结果,即第一场景以及目标模式显示给用户。然后分别向左耳机和右耳机发送控制信令2。控制信令2用于指示目标模式。
另一种可能的实施方式中,终端设备上配置有启动智慧场景检测的功能开关,终端设备响应于用户启动智慧场景检测的功能时,终端设备触发场景检测。识别到所处的场景为第一场景时,确定耳机的处理模式中对应于第一场景的目标模式,然后分别向左耳机和右耳机发送控制信令2。控制信令2用于指示目标模式。
终端设备检测完成后,还可以将检测结果显示给用户,使得用户知晓耳机采用的处理模式。检测结果中可以包括所检测到的场景,还可以包括对应于检测到的场景的处理模式。比 如,终端设备识别到场景为第一场景时,确定耳机的处理模式中对应于第一场景的目标模式,可以将检测结果,即第一场景以及目标模式显示给用户。然后分别向左耳机和右耳机发送控制信令2。控制信令2用于指示目标模式。可选地,在将检测结果显示用户后,响应于用户确定目标模式的操作是,再执行左耳机和右耳机发送控制信令2。
比如,终端设备配置的启动智慧场景检测的功能开关,可以配置在耳机应用的控制界面上,还可以配置的终端设备的系统设置菜单栏中。以设置在耳机应用的控制界面上为例,终端设备既可以通过识别场景的方式控制耳机所采用的处理模式,终端设备还可以识别控制界面上的选择控件上的用户操作来控制耳机所采用的处理模式。终端设备可以根据需求确定是否开启了智慧场景检测功能,当没有开启智慧场景检测功能时,可以采用示例1手动选择耳机采用的处理模式,当开启了智慧场景检测功能,终端设备100识别当前用户所处的场景。当用户开启了智慧场景检测功能后可以在手动选择处理模式界面的基础上更新为另一个界面,也可以在手动选择处理功能的界面的基础上进行检测结果显示。
比如,用户开启智慧场景检测功能前,终端设备上用户选择的处理功能为HT功能;当开启了智慧场景检测功能后,终端设备识别用户当前所处的场景为飞机场景,适合用ANC功能。一种示例中,用户启动耳机应用,在显示屏上显示耳机应用的控制界面,以圆环形为例。用户选择的处理功能为HT功能,参见图14A中(a)所示。控制界面中包括是否开启智慧场景检测功能的选项控件。当用户触发开启智慧场景检测功能的选项控件后,用户触发智慧场景检测功能,执行场景检测得到检测结果,通过将代表用户选择的处理功能的控件的位置变更为属于ANC功能的区域。黑色圆点在圆盘上的位置可以是该ANC功能下的默认值或者上一次用户选择ANC功能所选择的处理强度所在的位置。比如图14A中(b)所示。图14A中(b)以检测到飞机场景为例。终端设备100向左耳机和右耳机分别发送控制信令2,控制信令2用于指示ANC功能。另一种示例中,用户启动耳机应用,在显示屏上显示耳机应用的控制界面,以圆环形为例。用户选择的处理功能为HT功能,参见图14B中(a)所示。控制界面中包括是否开启智慧场景检测功能的选项控件。当用户触发开启智慧场景检测功能的选项控件后,用户触发智慧场景检测功能,执行场景检测得到检测结果,在检测结果界面显示检测结果。检测界面界面还可以包括终端设备能够识别的场景,以及场景对应的处理功能。例如,参见图14B中(b)所示,检测结果为飞机场景,对应的处理功能为ANC功能。终端设备100向左耳机和右耳机分别发送控制信令2,控制信令2用于指示ANC功能。
在终端设备智慧场景检测的方式下,目标模式下的目标处理强度可以采用如下方式中的任一种方式来确定。
方式一:耳机采用目标模式下默认的目标处理强度。
终端设备分别向左耳机和右耳机发送控制信令2后,以左耳机为例,左耳机接收到控制信令2后,确定采用的处理模式为目标模式,控制信令2未指示目标处理强度,耳机确定采用默认的目标处理强度。以目标模式为ANC模式为例,左耳机接收到控制信令2后,确定采用ANC模式,从左耳机中获取ANC模式下默认的FF滤波系数和FB滤波系数。
方式二:使用上一次采用目标模式时所采用的处理强度作为目标处理强度。
一种示例中,由终端设备确定目标处理强度,并通过控制信令向左耳机和右耳机指示。终端设备执行场景检测后,根据检测到的场景确定目标模式后,获取上一次采用目标模式时,所采用的处理强度作为目标处理强度,向左耳机和右耳机分别发送控制信令2,控制信令2指示目标模式以及目标处理强度。
另一种示例中,而耳机自己确定目标模式下的处理强度。终端设备执行场景检测后,根 据检测到的场景确定目标模式后,向左耳机和右耳机分别发送控制信令2,控制信令2指示目标模式,左耳机和右耳机接收到控制信令2后,确定采用的处理模式为目标模式,获取保存的上一次采用目标模式时的处理强度作为目标处理强度。以目标模式为ANC为例,获取保存的上一次采用ANC模式时的FF滤波系数和FB滤波系数执行ANC处理。
方式三,终端设备根据识别到的场景确定目标处理强度。
终端设备上未配置有启动智慧场景检测的功能开关的情况下,终端设备识别到场景后,可以根据识别的场景确定目标处理强度。
一种示例中,不同场景下确定的处理模式相同,且不同的场景对应的处理强度不同。比如,走路场景、跑步场景、安静场景均适合采用HT模式,在采用HT模式下,走路场景、跑步场景、安静场景分别对应的处理强度不同。再比如,多人说话场景、咖啡厅场景、火车场景、飞机场景、大风场景和办公室场景等,均适合采用ANC模式,在采用ANC模式下,多人说话场景、咖啡厅场景、火车场景、飞机场景、大风场景和办公室场景分别对应的处理强度不同。再比如,对话场景、警报声场景、鸣笛声场景、哭叫声场景,均适合采用AH模式,在采用AH模式时,对话场景、警报声场景、鸣笛声场景、哭叫声场景分别对应的处理强度不同。
基于此,终端设备根据存储的场景类型与目标模式、处理强度的对应关系,向左耳机和右耳机发送控制信令2,控制信令2指示目标模式以及目标模式下的目标处理强度。从而耳机接收到控制信令2后,根据控制信令2确定采用目标模式,确定目标处理强度对应的滤波系数。以目标模式为AH为例,根据目标处理强度确定FF滤波系数、FB滤波系数以及HT滤波系数,并根据FF滤波系数、FB滤波系数以及HT滤波系数执行S801-S807。
方式四,由用户通过终端设备提供的UI控件来向耳机指示目标模式下采用的处理强度。
一种示例中,终端设备执行场景检测后,在终端设备的显示界面上显示检测结果,检测结果中包括检测到的场景以及检测到的场景对应的目标模式。显示界面上可以包括用于选择处理强度的控件,为了描述方便,将该用于选择处理强度的控件称为强度控件。强度控件可以包括控件1和控件2。控制1的不同位置用于指示目标模式下的不同处理强度。强度控件可以是圆形、条形或者其它形状。参见图14C所示,检测到的场景为航站楼场景,以强度控件中的控件1为圆形,控件2为圆形黑点为例,响应于用户触控控件2移动到控件1的位置1,位置1表征所述用户选择目标模式下的目标处理强度。向左耳机和右耳机发送控制指令2,控制指令2指示目标模式以及位置1对应的目标处理强度。
可选地,目标模式和目标处理强度可以通过不同的控制指令发送给左耳机和右耳机。终端设备根据检测到的场景确定目标模式后,向左耳机和右耳机发送指示目标模式的控制信令。左耳机和右耳机接收到指示目标模式的控制信令后,采用目标模式下默认的处理强度,即采用目标模式下默认的滤波系数实现目标模式对应的目标处理。响应于用户触控控件2移动到控件1的位置1时,向左耳机和右耳机发送指示目标处理强度的控制信令,进一步地,左耳机和右耳机采用目标处理强度对应的滤波系数实现目标模式对应的目标处理。
另一种示例中,接着图14A的示例,用户触发开启智慧场景检测功能的选项控件后,用户触发智慧场景检测功能,执行场景检测得到检测结果,通过将代表用户选择的处理功能的控件的位置变更为属于ANC功能的区域。黑色圆点在圆盘上的位置可以是该ANC功能下的默认值或者上一次用户选择ANC功能所选择的处理强度所在的位置。用户通过移动黑色圆点的位置来选择ANC模式下的处理强度。并向左耳机和右耳机发送控制信令2,控制信令2指示ANC模式以及目标处理强度。
示例4,耳机的场景检测,不同的场景对应不同的处理功能。
耳机具备场景检测功能,耳机识别用户当前所处的场景,在所检测的场景类型不同时,耳机所实现的处理功能不同。耳机中左耳机可以具备场景检测功能、或者右耳机具备场景检测功能,或者左耳机和右耳机均具备场景检测功能。作为一种举例,左耳机和右耳机中其中一个耳机用于执行场景检测,比如由左耳机执行场景检测,并将检测结果发送给右耳机,从而左耳机和右耳机均根据左耳机的检测结果执行用于实现检测结果对应的处理功能的处理。或者由右耳机执行场景检测,并将检测结果发送给左耳机。从而左耳机和右耳机均根据右耳机的检测结果执行用于实现检测结果对应的处理功能的处理。作为另一种举例,左耳机和右耳机均执行场景检测,左耳机根据左耳机的检测结果执行用于实现检测结果对应的处理功能的处理,右耳机根据右耳机的检测结果执行用于实现检测结果对应的处理功能的处理。
一种可能的实施方式中,耳机的场景检测功能的开启,可以由用户通过耳机来控制或者通过终端设备来控制。
一种方式中,耳机设置有用于启动场景检测功能的按钮。用户通过触控按钮来开启或者关闭耳机的场景检测功能。当开启耳机的场景检测功能后,耳机识别当前用户所处的场景(或者说耳机当前所处的场景),场景与处理模式的对应关系确定识别到场景对应的处理模式,实现处理模式对应的处理功能。
另一种方式中,用户通过敲击耳机上操作,来开启或者关闭耳机的场景检测功能,比如连续敲击三下。在耳机的场景检测功能关闭的情况下,耳机响应于用户敲击三下耳机的操作,开启耳机的场景检测功能。在耳机的场景检测功能开启的情况下,耳机响应于用户敲击三下耳机的操作,关闭耳机的场景检测功能。当开启耳机的场景检测功能后,耳机识别当前用户所处的场景(或者说耳机当前所处的场景),场景与处理模式的对应关系确定识别到场景对应的处理模式,实现处理模式对应的处理功能。
又一种方式中,左耳机或者右耳机的场景检测功能的开启,由终端设备100来控制。比如,在耳机控制界面上包括耳机场景检测功能的开关按钮。终端设备可以根据用户需求确定是否开启耳机的场景检测功能,当不开启耳机的场景检测功能时,可以采用示例1手动选择耳机需实现的处理功能。当开启耳机的场景检测功能后,耳机识别当前外部环境的场景类型,根据场景类型与处理模式的对应关系确定识别到场景类型对应的处理模式,实现处理模式对应的处理功能。终端设备100响应于用户启动耳机的场景检测功能时,向耳机200发送控制信令3,控制信令3用于指示耳机启动场景检测功能。耳机200根据控制信令3,启动执行场景检测。耳机200根据检测到的当前外部环境的场景类型,确定所需执行的处理功能,比如为ANC功能,则耳机200执行ANC处理,执行S501-S504。
另一种可能的实施方式中,耳机与终端设备建立连接后,即耳机启动场景检测;或者耳机接收终端设备发送的下行音频信号时,即启动场景检测。
在示例4中,作为一种示例,耳机检测完成后,还可以将检测结果发送给终端设备。比如,检测结果可以包括在指示信息中发送给终端设备。检测结果中可以包括检测到的场景以及场景对应的处理模式。终端设备接收到检测结果时,将检测结果显示给用户,使得用户知晓耳机采用的处理模式。比如通过提示框的形式将检测结果显示给用户。可选地,检测结果中可以仅包括检测到场景。终端设备接收到检测结果后,确定耳机检测到的场景所对应的处理模式,并将耳机检测到的场景以及场景对应的处理模式显示给用户。比如,耳机识别到场景为第一场景时,确定耳机的处理模式中对应于第一场景的目标模式,可以将检测结果,即 第一场景以及目标模式显示给用户。
作为另一种示例,耳机检测完成后,不立即执行场景对应的处理模式的处理功能,将检测结果发送给终端设备,终端设备将检测结果显示给用户。终端设备响应于用户确定处理模式的操作时,向耳机发送确认指令,耳机接收到确认指令时,采用耳机检测到的场景对应的处理模式,执行处理功能。
示例性地,耳机能够识别的场景类型可以包括走路场景、跑步场景、安静场景、多人说话场景、咖啡厅场景、地铁场景、火车场景、汽车场景、候车厅场景、对话场景、办公室场景、户外场景、开车场景、大风场景、飞机场景、警报声场景、鸣笛声场景、哭叫声场景等。
作为一种示例,例举不同的场景类型所适合的处理模式。走路场景(HT)、跑步场景(HT)、安静场景(HT)、多人说话场景(ANC)、咖啡厅场景(ANC)、地铁场景(AH)、火车场景(ANC)、候车厅场景(AH)、对话场景(AH)、办公室场景(ANC)、户外场景(ANC)、开车场景(ANC)、大风场景(ANC)、飞机场景(ANC)、警报声场景(AH)、鸣笛声场景(AH)、哭叫声场景(AH)等,括号中表示该场景类型对应的处理模式。比如飞机场景,飞机飞行时噪声较大,适合采用ANC模式。再比如,走路场景和跑步场景以及安静场景中,适合采用HT模式,可以听见突发的事件声音。再比如,咖啡厅场景中,用户需求安静,可以采用ANC模式。再比如,轻音乐场景中,可以采用HT模式。再比如,警报声场景(AH)、鸣笛声场景(AH)、哭叫声场景(AH),用于需要听见预设声音,适合采用AH模式。
作为一种举例,飞机场景中,当识别到场景类型为飞机场景时,确定采用ANC模式,左耳机和右耳机分别执行S501-S504的处理。
作为一种举例,走路场景中,当识别到场景类型为走路场景时,确定采用HT模式,左耳机和右耳机分别执行S601-S605的处理。
作为又一种举例,火车站场景中,当识别到场景类型为火车站场景时,确定采用AH模式。左耳机和右耳机分别执行S801-S807的处理。
在耳机执行场景检测的方式下,目标模式下的目标处理强度可以采用如下方式中的任一种方式来确定。
方式一:耳机采用目标模式下默认的目标处理强度。
耳机(左耳机或右耳机)根据检测出到的场景确定采用的处理模式为目标模式,左耳机和右耳机确定采用默认的目标处理强度。以目标模式为ANC模式为例,左耳机和右耳机获取ANC模式下默认的FF滤波系数和FB滤波系数。
方式二:使用上一次采用目标模式时所采用的处理强度作为目标处理强度。
一种示例中,而耳机(左耳机或者右耳机)自己确定目标模式下的处理强度。耳机执行场景检测后,根据检测到的场景确定目标模式后,获取保存的上一次采用目标模式时的处理强度作为目标处理强度。以目标模式为ANC为例,获取保存的上一次采用ANC模式时的FF滤波系数和FB滤波系数执行ANC处理。
另一种示例中,由终端设备确定目标处理强度,并通过控制信令向左耳机和右耳机指示。耳机执行场景检测后,将检测结果发送给终端设备,从而终端设备获取上一次采用目标模式时,所采用的处理强度作为目标处理强度,向左耳机和右耳机分别发送控制信令4,控制信令4指示目标处理强度。
方式三,耳机根据识别到的场景确定目标处理强度。
耳机识别到场景后,可以根据识别的场景确定目标处理强度。
一种示例中,一种示例中,不同场景下确定的处理模式相同,且不同的场景对应的处理 强度不同。比如,走路场景、跑步场景、安静场景均适合采用HT模式,在采用HT模式下,走路场景、跑步场景、安静场景分别对应的处理强度不同。再比如,多人说话场景、咖啡厅场景、火车场景、飞机场景、大风场景和办公室场景等,均适合采用ANC模式,在采用ANC模式下,多人说话场景、咖啡厅场景、火车场景、飞机场景、大风场景和办公室场景分别对应的处理强度不同。再比如,对话场景、警报声场景、鸣笛声场景、哭叫声场景,均适合采用AH模式,在采用AH模式时,对话场景、警报声场景、鸣笛声场景、哭叫声场景分别对应的处理强度不同。
基于此,左耳机和右耳机根据存储的场景类型与目标模式、处理强度的对应关系,确定检测到的场景所对应的目标模式以及目标模式下的目标处理强度。从而左耳机和右耳机获取目标处理强度对应的滤波系数。以目标模式为AH为例,根据目标处理强度确定FF滤波系数、FB滤波系数以及HT滤波系数,并根据FF滤波系数、FB滤波系数以及HT滤波系数执行S801-S807。
另一种示例中,不同的场景下,耳机还可以进行突发事件检测确定目标事件(或者称为目标事件场景)。突发事件比如包括风噪事件、啸叫事件、突发事件、人声事件或无突发事件中的一项或多项,不同的事件对应的处理强度不同。耳机进行场景检测,以及进行事件检测。在目标模式下,不同的事件对应不同滤波系数。以ANC为例,不同的事件对应的不同FF滤波系数和/或不同的FB滤波系数。左耳机或者右耳机执行场景和事件检测后,以左耳机和ANC模式为例,左耳机可以根据检测结果从系数库中获取在实现ANC功能的情况下,检测到事件对应的FF滤波系数或者FB滤波系数。系数库中存储有处理模式、事件以及FF滤波系数、FB滤波系数的映射关系。ANC的处理效果好坏主要依靠FB滤波和/或FF滤波。例如,根据检测的场景对FF滤波器的滤波系数进行控制,FB滤波系数采用固定值。再比如,根据检测的场景对FB滤波器的滤波系数进行控制,FF滤波系数采用固定值。又比如,根据检测的场景对FF滤波系数和FB滤波系数进行控制。例如,参见表2所示,以事件包括啸叫事件、风噪事件、突发事件、人声事件或者无突发事件为例。
表2
Figure PCTCN2021103435-appb-000002
Figure PCTCN2021103435-appb-000003
示例性地,耳机200检测外部环境中的事件声音,可以根据参考麦克风采集的信号确定外部环境中的事件声音对应的目标事件。比如参考麦克风采集的信号中包括预设频谱的信号,则确定预设频谱的信号对应的事件。例如风噪事件,参考麦克风采集的信号包括风声信号,即采集的信号中包括与风声的频谱匹配的信号,则确定检测到外部环境中的事件声音对应的事件为风噪事件。确定参考麦克风采集的信号中包括预设频谱的信号时,可以采用频谱匹配方式,还可以采用深度神经网络(deep neural networks,DNN)匹配方式。
比如,耳机200可以通过如下方式实现根据参考麦克风采集的信号确定用户当前环境中的事件,参见图15所示。在耳机200中还包括骨传导传感器。骨传导传感器用于采集耳机用户的骨传导信号。当用户佩戴耳机200且发出声音,比如说话或者唱歌等,骨传导传感器通过采集骨传导信号,即采集用户说话产生的骨膜震动信号得到骨传导信号。
左耳机或者右耳机的场景检测功能的开启,可以由终端设备100来控制,也可以由用户在耳机上的操作来控制,比如敲击左耳机或者右耳机的操作。或者耳机上包括骨传导传感器,用户上下牙齿触碰产生牙齿触碰声音,从而骨传导传感器通过检测用户上下牙齿触碰产生的音频信号来启动场景检测功能。
S1501,滤除参考麦克风采集的第三信号中的骨传导信号得到经滤除信号,简称为信号AA1。
其中,步骤S1501中,参考麦克风采集的第三信号,是在耳机启动场景检测功能后,参考麦克风采集到的信号。
需要理解的是,当用户没有发出声音,比如佩戴耳机时没有说话也没有唱歌等,则骨传导传感器采集的骨传导信号的能量较小,比如骨传导信号的能量小于设定阈值时,可以无需执行S1501,则信号AA1即为第三信号。作为一种示例,耳机200可以先确定骨传导信号的能量,如果骨传导信号的能量小于设定阈值,则不执行滤波操作,即不执行S1501,当确定骨传导信号的能量大于或者等于设定阈值时,则执行S1501。
S1502,对经滤除信号进行频谱分析得到所述经滤除信号的能量特征。
也就是,耳机200对信号AA1进行频谱分析得到信号AA1的能量特征。例如,耳机200对信号进行频谱分析,获得信号AA1的整帧的能量以及信号AA1的每个bark子带的能量,从而形成通过向量表示的信号AA1的能量特征。
S1503,确定能量特征集合包括的能量特征中与所述经滤除信号的能量特征匹配的第一能量特征,其中,能量特征集合中包括的不同的能量特征对应于不同的事件标识。
S1504,确定所述第一能量特征对应的事件标识所标识的事件为所述用户当前环境中的事件,即事件检测的检测结果。
作为一种示例,能量特征集合可以通过如下方式来生成:针对第一麦克风、第二麦克风以及第三麦克风采集的信号执行风噪检测、突发噪声检测、啸叫检测以及人声检测获取风噪信号、突发噪声信号以及啸叫信号。然后对风噪信号、突发噪声信号、啸叫信号以及人声信号分别进行频谱分析,以获取风噪信号的子带能量特征、突发噪声信号的子带能量特征、啸叫信号的子带能量特征以及人声信号的子带能量特征。风噪信号的子带能量特征、突发噪声信号的子带能量特征、啸叫信号的子带能量特征以及人声信号的子带能量特征构成能量特征集合。应理解的是,安静场景下,噪声的子带能量较弱。
可选地,在确定能量特征集合包括的能量特征中与所述经滤除信号的能量特征匹配的第一能量特征时,可以采用频谱匹配方式,还可以采用深度神经网络(deep neural networks,DNN)匹配方式。例如,采集DNN匹配方式时,可以通过DNN来确定经滤除信号的能量特征与能量特征集合包括的各个能量特征的匹配度,匹配度最高的第一能量特征对应的事件标识所标识的事件为检测结果。
在本申请实施例中,可以由耳机200中的主控制单元来实现根据参考麦克风采集的信号确定用户当前环境中的事件。例如主控制单元包括DSP,DSP用于执行S1501-S1504。
方式四,由用户通过终端设备提供的UI控件来向耳机指示目标模式下采用的处理强度。
一种示例中,耳机执行场景检测后,将检测结果发送给终端设备,终端设备将检测结果显示给用户。在终端设备的显示界面上显示检测结果,检测结果中包括耳机检测到的场景以及检测到的场景对应的目标模式。显示界面还包括用于选择处理强度的控件,为了描述方便,将该用于选择处理强度的控件称为强度控件。强度控件可以包括控件1和控件2。控制1的不同位置用于指示目标模式下的不同处理强度。强度控件可以是圆形、条形或者其它形状。参见图16所示,以强度控件为圆形为例,响应于用户触控控件2移动到控件1的位置2,位置2表征所述用户选择目标模式下的目标处理强度。向左耳机和右耳机发送控制指令5,控制指令5指示位置2对应的目标处理强度。图16中以目标模式为HT为例。
作为一种举例,终端设备100响应于用户启动耳机场景检测的功能时,向耳机200发送控制信令3,控制信令3用于指示耳机启动场景检测功能。耳机200根据控制信令3,启动执行场景检测得到检测结果。耳机200可以将检测结果发送给终端设备100,从而终端设备100显示给用户,并将检测到的场景对应需要耳机采用的处理模式显示给用户。
进一步地,当用户开启耳机场景检测功能后可以在手动选择处理模式界面的基础上更新为另一个界面,也可以在手动选择处理功能的界面的基础上进行检测结果显示。
比如,用户开启耳机场景检测功能前,终端设备上用户选择的处理功能为HT功能;当开启耳机场景检测功能后,耳机200识别用户当前所处的场景为飞机场景,适合用ANC功能,将检测结果,即飞机场景,ANC功能,发送给终端设备。一种示例中,用户启动耳机应用,在显示屏上显示耳机应用的控制界面,以圆环形为例。用户选择的处理功能为HT功能,参见图17A中(a)所示。控制界面中包括是否开启耳机场景检测功能的选项控件。当用户触发开启耳机场景检测功能的选项控件后,终端设备触发耳机场景检测功能,向耳机200发送控制信令3,控制信令3用于指示耳机启动场景检测功能。耳机200根据控制信令3,启动执行场景检测得到检测结果。耳机200将检测结果发送给终端设备100,终端设备100接收到检测结果后,通过将代表用户选择的处理功能的控件的位置变更为属于ANC功能的区域。用户通过移动黑色圆点在圆盘上的位置选择ANC模式下的处理强度。比如图17A中(b)所示。图17A中(b)以检测到飞机场景为例。
作为另一种举例中,用户启动耳机应用,在显示屏上显示耳机应用的控制界面,以圆环形为例。用户选择的处理功能为HT功能,参见图17B中(a)所示。控制界面中包括是否开启耳机场景检测功能的选项控件。当用户触发开启耳机场景检测功能的选项控件后,终端设备触发耳机场景检测功能,向耳机200发送控制信令3,控制信令3用于指示耳机启动场景检测功能。耳机200根据控制信令3,启动执行场景检测得到检测结果。耳机200将检测结果发送给终端设备100,终端设备100接收到检测结果后,在检测结果界面显示检测结果。检测界面界面还可以包括耳机能够识别的场景,以及场景对应的处理模式。用户通过移动黑 色圆点在圆盘上的位置选择ANC模式下的处理强度。例如,参见图17B中(b)所示,检测结果为飞机场景,对应的处理模式为ANC模式。
示例性地,耳机200检测场景时,一种方式可以采用人工智能(artificial intelligence,AI)模型方式进行检测分类。AI模型可以配置在耳机中。另一种方式可以根据参考麦克风采集的信号确定场景类型。比如,耳机200可以通过如下方式实现根据参考麦克风采集的信号确定用户当前所处的场景,参见图18所示。
S1801,对参考麦克风采集的第一信号进行频谱分析,将第一信号划分为多个子带,并计算每个子带的能量。比如,将参考麦克风采集的第一信息在频域按照bark子带划分方法进行子带划分,并计算每个子带的能量。
S1802,确定VAD,以获取第一信号中的噪声段,并获取噪声段中每个子带的平滑能量。
作为一种示例,VAD确定方式如下:
计算参考麦克风和通话麦信号的互相关性,得到互相关系数A;
计算参考麦克风的自相关系数B;
当A<alpha(第一阈值),B<beta(第二阈值)时,确定该VAD对应的信号段为噪声段,否则确定该VAD对应的信号段判决为语音段;
S1803,根据噪声段中每个子带的平滑能量确定场景类型。
作为一种示例,以确定安静场景、低频重噪声、人声场景为例。针对确定的噪声段执行如下处理来确定场景类型:
(1),计算噪声段中50~1KHz子带能量平均值C、1K~2KHz子带能量平均值D、2K~3KHz子带能量平均值E。如果C/D/E连续N帧满足均小于阈值gama的条件则判决为安静场景。
(2),a=D/C,如果a小于阈值t、C和D均大于阈值k,且连续M帧均满足a小于阈值t、C和D均大于阈值k,则判决为低频重噪场景;
(3),如果a大于阈值k、持续P帧均不是噪声帧,则判决为人声(或者音乐)场景。
示例5,耳机确定处理模式后,进行处理模式下的事件检测,在该处理模式下,不同的事件对应的滤波系数(即该处理模式下的处理强度)不同。
耳机识别用户的操作,确定用户选择的耳机200需要实现ANC处理、HT处理或者AH处理。以耳机200采用的处理模式为ANC模式功能。一种可能的方式,用户的操作可以是用户敲击耳机的操作,通过不同的操作确定处理模式为ANC模式、HT模式或者AH模式。另一种可能的方式,耳机上设置有按钮,不同的按钮指示不同的处理模式。用户通过按下按钮来选择耳机采用的处理模式。比如,耳机200接收到用户触发的ANC模式的操作指令后,左耳机和右耳机执行ANC处理,具体执行S501-S504。又一种可能的方式,选择耳机需要实现的处理模式,由终端设备100来控制。
左耳机或者右耳机可以具备事件检测功能。作为一种举例,左耳机和右耳机中其中一个耳机用于执行事件检测,比如由左耳机执行事件检测,并将检测结果发送给右耳机,或者由右耳机执行事件检测,并将检测结果发送给左耳机。在ANC模式下,不同的事件对应不同的FF滤波系数和FB滤波系数。左耳机或者右耳机执行事件检测后,以左耳机为例,左耳机可以根据检测结果从系数库中获取在ANC模式下,检测到的事件对应的FF滤波系数或者FB滤波系数。例如,参见表2所示系数库中包括的内容,以事件包括啸叫事件事件、风噪事件、 突发事件或者人声事件为例。
可以理解的是,为了实现上述方法实施例中功能,耳机包括了执行各个功能相应的硬件结构和/或软件模块。本领域技术人员应该很容易意识到,结合本申请中所公开的实施例描述的各示例的模块及方法步骤,本申请能够以硬件或硬件和计算机软件相结合的形式来实现。某个功能究竟以硬件还是计算机软件驱动硬件的方式来执行,取决于技术方案的特定应用场景和设计约束条件。
基于与上述方法同样的发明构思,如图19所示,本申请实施例还提供一种噪声处理装置1900。噪声处理装置1900应用于耳机中。所述耳机具备ANC功能、HT功能或AH功能中至少两个功能;所述耳机包括第一麦克风、第二麦克风;第一麦克风,用于采集第一信号,所述第一信号用于表征当前外部环境的声音;所述第二麦克风,用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音。噪声处理装置1900可以用于实现上述方法实施例中耳机的功能,因此可以实现上述方法实施例所具备的有益效果。装置可以包括通信模块1901、获取模块1902、第一处理模块1903。
通信模块1901,用于接收来自所述终端设备的第一音频信号;
获取模块1902,用于获取目标模式;其中,所述目标模式是基于当前外部环境的场景类型确定的,所述目标模式用于指示所述耳机实现目标处理功能,所述目标处理功能为主动降噪ANC功能、环境声透传HT功能或者听觉增强AH功能中一个;
第一处理模块1903,用于根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
在一种可能的实施方式中,所述装置还包括:播放模块,用于播放所述第二音频信号。针对播放模块,图19中未示出。
在一种可能的实施方式中,所述目标处理功能为ANC功能,所述播放模块播放的所述第二音频信号,能够减弱用户对所述用户当前所处环境的声音以及所述用户耳道内部的环境声音的感知;或者,
所述目标处理功能为HT功能,所述播放模块播放的所述第二音频信号,能够强化用户对所述用户当前所处环境的声音的感知;或者,
所述目标处理功能为AH功能,所述播放模块播放的所述第二音频信号,能够强化用户对事件声音的感知;所述事件声音满足预设频谱。
在一种可能的实施方式中,当所述目标处理功能为ANC功能时,第二音频信号是基于第一音频信号、第三信号以及第四信号得到的,所述第三信号为第一信号的反相信号,所述第四信号为所述第二信号的反相信号;或者,
当所述目标处理功能为HT功能时,第二音频信号是基于第一音频信号、第一信号以及第二信号得到的;或者,
当所述目标处理功能为AH功能时,第二音频信号是基于第一音频信号、第五信号以及第四信号得到的,所述第五信号为所述第一信号中的事件信号,所述事件信号满足预设频谱。
在一种可能的实施方式中,所述通信模块1901,还用于接收来自所述终端设备的第一控制指令,所述第一控制指令携带所述目标模式,所述目标模式是所述终端设备根据当前外部环境的场景类型确定的;
将所述目标模式发送给所述获取模块1902。
在一种可能的实施方式中,所述通信模块1901,还用于:
接收来自所述终端设备的第二控制指令,所述第二控制指令携带目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度;
所述第一处理模块1903,具体用于:
根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
在一种可能的实施方式中,还包括:
第二处理模块1904,用于根据所述第一信号确定当前外部环境中事件声音对应的目标事件,根据所述目标事件确定在所述目标模式下的目标处理强度;其中,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度;
所述第一处理模块1903具体用于:
根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
在一种可能的实施方式中,所述耳机还包括骨传导传感器,所述骨传导传感器用于采集所述用户声带震动产生的骨传导信号;
所述第一处理模块1901,具体用于根据所述第一信号和所述骨传导信号确定当前外部环境中事件声音对应的目标事件。
在一种可能的实施方式中,所述目标事件包括啸叫事件、风噪事件、突发事件或者人声事件中的一项。
在一种可能的实施方式中,还包括:
第三处理模块1905,用于根据所述第一信号识别当前外部环境的场景类型为目标场景,根据所述目标场景确定所述耳机采用的所述目标模式,所述目标模式是对应于所述目标场景的处理模式。
在一种可能的实施方式中,所述目标场景包括走路场景、跑步场景、安静场景、多人说话场景、咖啡厅场景、地铁场景、火车场景、候车厅场景、对话场景、办公室场景、户外场景、开车场景、大风场景、飞机场景、警报声场景、鸣笛声场景、哭叫声场景中的一项。
在一种可能的实施方式中,所述通信模块1901,还用于向所述终端设备发送指示信息,所述指示信息携带所述目标模式;
接收来自所述终端设备的第三控制信令,所述第三控制信令包括所述目标模式下的目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度;
所述第一处理模块1903,具体用于:
根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
在一种可能的实施方式中,当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,
当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,
当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
在一种可能的实施方式中,所述耳机为左耳机,或者所述耳机为右耳机。
在一种可能的实施方式中,所述目标模式指示所述耳机实现ANC功能,所述第一处理模 块1903,具体用于:
对所述第一信号进行所述第一滤波处理得到第一滤波信号;
滤除所述第二信号中包括的第一音频信号得到第一经滤除信号;
对所述第一滤波信号、所述经滤除信号进行混音处理得到第三音频信号;
对所述第三音频信号进行所述第三滤波处理得到第四音频信号;
对所述第四音频信号以及所述第一音频信号进行混音处理得到所述第二音频信号。
在一种可能的实施方式中,所述第一滤波处理采用的滤波系数为所述ANC功能下,针对所述第一滤波处理,所述目标处理强度所关联的滤波系数;或者,
所述第三滤波处理采用的滤波系数为所述ANC功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
在一种可能的实施方式中,所述目标模式指示所述耳机实现HT功能,所述第一处理模块1903,具体用于:
对所述第一信号进行第一信号处理得到第一经处理信号,所述第一信号处理包括第二滤波处理;
对所述第一经处理信号和所述第一音频信号进行混音处理得到第五音频信号;
滤波所述第二信号包括的所述第五音频信号得到第二经滤除信号;
对所述第二经滤除信号进行第三滤波处理得到第三经滤除信号;
将所述第三经滤除信号与所述第五音频信号进行混音处理得到所述第二音频信号。
在一种可能的实施方式中,所述第二滤波处理采用的滤波系数为所述HT功能下,针对所述第二滤波处理,所述目标处理强度所关联的滤波系数;或者,
所述第三滤波处理采用的滤波系数为所述HT功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
在一种可能的实施方式中,目标模式指示所述耳机实现AH功能,所述第一处理模块1903,具体用于:
对所述第一信号进行第二滤波处理得到第二滤波信号,并对所述第二滤波信号进行增强处理得到滤波增强信号;
对所述第一信号进行第一滤波处理得到第一滤波信号;
对所述滤波增强信号和所述第一音频信号进行混音处理得到第六音频信号;
滤波所述第二信号包括的所述第六音频信号得到第四经滤除信号;
对所述第四经滤除信号进行第三滤波处理得到第五经滤除信号;
对所述第五经滤除信号、所述第六音频信号以及所述第一滤波信号进行混音处理得到所述第二音频信号。
在一种可能的实施方式中,所述第一滤波处理采用的滤波系数为所述AH功能下,针对所述第一滤波处理,所述目标处理强度所关联的滤波系数;或者,
所述第二滤波处理采用的滤波系数为所述AH功能下,针对所述第二滤波处理,所述目标处理强度所关联的滤波系数;或者,
所述第三滤波处理采用的滤波系数为所述AH功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
可以理解的是,为了实现上述方法实施例中功能,终端设备包括了执行各个功能相应的硬件结构和/或软件模块。本领域技术人员应该很容易意识到,结合本申请中所公开的实施例描述的各示例的模块及方法步骤,本申请能够以硬件或硬件和计算机软件相结合的形式来实 现。某个功能究竟以硬件还是计算机软件驱动硬件的方式来执行,取决于技术方案的特定应用场景和设计约束条件。
基于与上述方法同样的发明构思,如图20所示,本申请实施例还提供一种模式控制装置2000。模式控制装置2000应用于终端设备100,可以用于实现上述方法实施例中终端设备的功能,因此可以实现上述方法实施例所具备的有益效果。
模式控制装置2000包括第一检测模块2001和发送模块2002,还可以包括显示模块2003、第二检测模块2004。
第一检测模块2001,用于识别当前外部环境的场景类型为目标场景时,根据所述目标场景确定目标模式;
其中,所述目标模式为耳机支持的处理模式中的一个,不同的处理模式对应于不同的场景类型,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种。
发送模块2002,用于向所述耳机发送所述目标模式,所述目标模式用于指示所述耳机实现所述目标模式所对应的处理功能。
在一种可能的实施方式中,还包括:
显示模块2003,用于根据所述目标场景确定目标模式时,显示结果提示信息,所述结果提示信息用于提示用户所述耳机实现所述目标模式所对应的处理功能。
在一种可能的实施方式中,显示模块2003,用于在向所述耳机发送第一控制信令之前,显示选择提示信息,所述选择提示信息用于提示用户是否调整所述耳机的处理模式为所述目标模式;
第二检测模块2004,用于检测到用户选择所述耳机的处理模式为所述目标模式的操作。
在一种可能的实施方式中,所述显示模块2003,还用于显示第一控件和第二控件,其中,所述第二控件在所述第一控件上的不同位置用于指示所述目标模式下的不同处理强度;
所述第二检测模块2004,还用于在发送模块2002向所述耳机发送第一控制信令之前,检测用户触控所述第二控件移动到所述第一控件上的第一位置的操作,所述第二控件在所述第一控件上的第一位置指示所述目标模式下的目标处理强度;
所述发送模块2002,还用于还用于向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
在一种可能的实施方式中,所述第一控件的形状为圆环,所述用户触控所述第二控件按顺时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
所述用户触控所述第二控件按逆时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;
在一种可能的实施方式中,所述第一控件的形状为条形,所述用户触控所述第二控件从上向下在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
所述用户触控所述第二控件从下向上在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
所述用户触控所述第二控件从左向右在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
所述用户触控所述第二控件从右向左在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
在一种可能的实施方式中,当所述目标处理功能为ANC功能,所述目标处理强度越大, 所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,
当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,
当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
基于与上述方法同样的发明构思,如图21所示,本申请实施例还提供一种模式控制装置2100。模式控制装置2100应用于终端设备100,可以用于实现上述方法实施例中终端设备的功能,因此可以实现上述方法实施例所具备的有益效果。模式控制装置2100包括处理模块2101、发送模块2102、接收模块2103、显示模块2104和检测模块2105。
处理模块2101,用于获取目标模式;所述目标模式为耳机支持的处理模式中的一个,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;
所述处理模块2101,还用于根据当前外部环境的场景类型确定所述目标模式下的目标处理强度;不同的场景类型对应于目标模式下的不同处理强度;
发送模块2102,用于向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
在一种可能的实施方式中,还包括:
接收模块2103,用于接收所述耳机发送的所述目标模式。
在一种可能的实施方式中,还包括:
显示模块2104,用于显示选择控件,所述选择控件包括耳机支持的处理模式,检测到用户通过所述选择控件在所述耳机的处理模式中选择所述目标模式的操作。
在一种可能的实施方式中,显示模块2104,还用于:
在所述处理模块2101根据当前外部环境的场景类型确定所述目标模式下的目标处理强度之前,所述接收模块2103接收到所述耳机发送的所述目标模式的情况下,显示选择提示信息,所述选择提示信息用于指示所述用户是否调整所述耳机的处理模式为所述目标模式;
还包括:
检测模块2105,用于检测到用户选择调整所述耳机的处理模式为所述目标模式的操作。
在一种可能的实施方式中,当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,
当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,
当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
基于与上述方法同样的发明构思,如图22所示,本申请实施例还提供一种模式控制装置2200。模式控制装置2200应用于终端设备100,可以用于实现上述方法实施例中终端设备的功能,因此可以实现上述方法实施例所具备的有益效果。模式控制装置2100包括显示模块2201、检测模块2202、发送模块2203、处理模块2204、识别模块2205。
显示模块2201,用于第一界面上包括第一选择控件,所述第一选择控件包括第一目标耳机支持的处理模式以及第一目标耳机支持的处理模式对应的处理强度;所述第一目标耳机的 处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;
检测模块2202,检测用户针对所述第一界面执行的第一操作;所述第一操作是用户通过所述第一选择控件在第一目标耳机支持的处理模式中选择第一目标模式以及选择第一目标模式下的处理强度为第一目标处理强度产生的;
发送模块2203,用于向所述第一目标耳机发送所述第一目标模式以及所述第一目标处理强度,所述第一目标模式用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能,所述第一目标处理强度用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能时的处理强度。
在一种可能的实施方式中,所述显示模块2201,还用于在显示第一界面之前,显示选择提示信息,所述选择提示信息用于所述用户选择是否调整所述第一目标耳机的处理模式;
所述检测模块2202,还用于检测用户选择调整所述第一目标耳机的处理模式的操作。
在一种可能的实施方式中,还包括:
识别模块2205,用于在显示模块2201显示所述第一界面之前,识别当前外部环境的场景类型为目标场景,所述目标场景适配需要调整所述第一目标耳机的处理模式的场景类型。
在一种可能的实施方式中,还包括:
识别模块2205,用于在显示模块2201显示所述第一界面之前,识别到所述终端设备触发所述第一目标耳机播放音频。
在一种可能的实施方式中,所述检测模块2202,还用于在显示模块显示所述第一界面之前,检测到所述终端设备与所述第一目标耳机建立连接。
在一种可能的实施方式中,所述检测模块2202,在显示模块2201显示所述第一界面之前,检测到所述终端设备与所述第一目标耳机建立连接的情况下,检测到所述用户在主界面执行的第二操作。
其中,所述主界面包括第一应用的图标,所述第二操作是所述用户触控所述第一应用的图标产生的,所述第一界面为所述第一应用的显示界面。
在一种可能的实施方式中,所述第一选择控件包括第一控件和第二控件,第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的两种不同处理模式,或者,所述第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的的同一种处理模式下的不同处理强度;
所述第一操作是所述用户移动所述第二控件位于所述第一控件上所述第一目标模式对应的区域内的第一位置产生的,所述第一位置对应于所述第一目标模式下的第一目标处理强度。
在一种可能的实施方式中,所述第一控件的形状为圆环,所述圆环包括至少两个圆弧段,所述第二控件位于不同的圆弧段指示第一目标耳机的不同处理模式,所述第二控件位于同一圆弧段的不同位置指示第一目标耳机的同一处理模式的不同处理强度;或者,
所述第一控件的形状为条形,所述条形包括至少两个条形段,所述第二控件位于不同的条形段指示所述第一目标耳机的不同处理模式,所述第二控件位于同一条形段的不同位置指示所述第一目标耳机的同一处理模式的不同处理强度。
在一种可能的实施方式中,所述检测模块2202,还用于检测用户在第一界面执行的第三操作,所述第一界面上还包括第二选择控件,所述第二选择控件包括第二目标耳机支持的处理模式以及第二目标耳机支持的处理模式对应的处理强度,所述第一目标耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种,所述第三操作是所述用户通过所述第二选择控件在第二目标耳机的处理模式中选择第二目标模式 以及选择第二目标模式下的处理强度为第二目标处理强度产生的;所述第一目标耳机为左耳机时,所述第二目标耳机为右耳机,或者,所述第一目标耳机为右耳机,所述第二目标耳机为左耳机;
所述发送模块2203,还用于向所述第二目标耳机发送所述第二目标模式以及所述第二目标处理强度,所述第二目标模式用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能,所述第二目标处理强度用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能时的处理强度。
基于此,本申请实施例中还提供一种终端设备,参见图23所示,终端设备中包括处理器2301、存储器2302、通信接口2303和显示器2304。存储器2302用于存储处理器2301执行的指令或程序,或存储处理器2301运行指令或程序所需要的输入数据,或存储处理器2301运行指令或程序后产生的数据。处理器2301用于运行存储器2302存储的指令或者程序执行上述方法中终端设备所执行的功能。
一种可能的场景中,处理器2301用于执行第一检测模块2001和发送模块2002、显示模块2003、第二检测模块2004的功能。或者,处理器2301用于执行第一检测模块2001和第二检测模块2004的功能。发送模块2002的功能由通信接口2303实现,显示模块2003的功能可以由显示器2304实现。
另一种可能的场景中,处理模块2101、发送模块2102、接收模块2103、显示模块2104和检测模块2105可以由处理器2301实现。或者,处理器2301可以用于执行处理模块2101以及检测模块2105的功能,发送模块2102、接收模块2103的功能可以由通信接口2303实现,显示模块2104的功能可以由显示器2304实现。
又一种可能的场景中,显示模块2201、检测模块2202、发送模块2203、处理模块2204、识别模块2205可以由处理器2301实现。或者,处理模块2204、检测模块2202以及识别模块2205的功能均可以由处理器2301实现。发送模块2203的功能可以由通信接口2303实现,显示模块2201的功能可以由显示器2304实现。
可以理解的是,本申请的实施例中的处理器可以是中央处理模块(central processing unit,CPU),还可以是其它通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(application specific integrated circuit,ASIC)、现场可编程门阵列(field programmable gate array,FPGA)或者其它可编程逻辑器件、晶体管逻辑器件,硬件部件或者其任意组合。通用处理器可以是微处理器,也可以是任何常规的处理器。
结合前述实施例中可选的实施方式,本申请将结合一些具体地应用场景描述一些可选的耳机的使用实施例和信号处理实施例。用户可以结合终端与耳机配合使用。耳机可以至少支持ANC功能、HT功能、AH功能中的一种,当然还可以包括空模式;其中,ANC功能的终极目标是把实际听觉中的噪声消除;HT功能的终极目标是消除耳机对于外部声音进如人耳起到的影响,使得用户通过耳机听到的外部环境信号的效果与用户裸耳听到当前外部环境的声音的效果等效;等效可以指效果一样或者近似一样。
对于终端可以执行如下方法:
S3001,终端设备与耳机建立通信连接;
S3002,显示第一界面;第一界面用于设置耳机的功能;第一界面中可以包括ANC功能、HT功能、AH功能或空模式的使能选项,或其关闭选项。第一界面一种可能的示例如图24所示,耳机支持的噪声控制模式可以包括空模式(关闭)、ANC功能(降噪)、HT功能(透 传)的选项。应理解,第一界面还可以包括更多的设置内容或选项,本申请中的附图并未全部示出。
S3003,当HT功能的选项使能时,激活耳机的HT功能。
可选地,可以进一步地获取HT功能的透传强度,并根据获取的透传强度控制所述耳机的HT功能。
S3004,在第一界面中新增事件声音增强的选项;应注意,可选地,这个选项在HT选项使能之前可以是没有的。事件声音为外部环境中符合预设事件条件的声音;事件声音可以包括人声或者符合预设频谱特性的其它声音,具体还可以参照前述实施例中的相关描述。
一种可能的实现形式可以参见图25,相比于图24,当透传HT功能使能时,界面中新增了人声增强(属于一种事件声音增强)这一选项。需要说明,本申请中图示中的黑框只是为了方便说明进行辅助添加的,并不对界面的真实形态构成任何限定。
S3005,当S3004中的事件声音增强功能的选项使能或者当S3002中的AH功能的选项使能时,终端可以控制耳机针对耳机采集的信号中事件声音的信噪比进行提升;其中,事件声音的信噪比越高,表明信号中事件声音的能量占比越大。本发明实施例中将在后面的实例中详细介绍下耳机的两种针对耳机采集的信号中事件声音的信噪比进行提升的方法。可以参见S4001-S4005和S5001-S5005。
一种可能的实现形式可以参见图26,相比于图25,当HT功能(透传)使能后;可以接收用户点选指令或者获取到预设的人声增强使能指令,使界面中人声增强这一选项使能。
可选地,此时终端可以控制耳机的ANC功能和HT功能均处于开启状态。具体地,可以是在保持HT功能处于开启状态时,再激活耳机的ANC功能。其中,ANC功能和HT功能共同使能用于对耳机采集到的外部环境的声音信号进行处理,所述处理包括针对采集的信号中事件声音的信噪比进行提升。
可选地,此时终端可以开启耳机的ANC功能,并控制耳机结合一些增强、降噪的算法针对耳机采集的信号中事件声音的信噪比进行提升。
可选地,还可以进一步获取事件声音增强的强度,并根据获取的事件声音增强的强度控制耳机的事件声音增强功能。事件声音增强的强度越强,事件声音的信噪比越高。
S3006,当ANC功能的选项使能时,激活所述耳机的ANC功能。
S3007,在第一界面中新增ANC功能的强度选项。应注意,可选地,这个选项在ANC选项使能之前可以是没有出现在第一界面中的,也可以是与ANC、HT功能选项同时出现在第一界面中的。ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,第一稳态ANC强度选项和第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的且稳定的ANC功能强度;自适应ANC强度选项对应的ANC功能强度与终端设备或耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。可选地,当前环境不同的场景类型可以包括第一场景和第二场景。
一种可能的实现形式可以参见图27,相比于图24,当ANC功能(降噪)功能使能时,界面中新增了ANC功能的强度选项(降噪方式)这一选项。具体地,ANC功能的强度选项(降噪方式)进一步可以包括多个选项,可以参见图28,包括但不限于轻度、均衡、深度和智慧动态降噪等降噪方式。其中,轻度等级适合较安静的环境,深度等级适合非常嘈杂的环境,即不属于轻度等级而不属于深度等级的场景都可以归类为均衡等级,也就是普通场景。
S3008,根据强度选项的使能结果进行ANC降噪。
具体地,当用户选择轻度等级时,此时控制耳机ANC功能的降噪强度对应为轻度等级;当用户选择均衡等级时,此时控制耳机ANC的降噪强度对应为中等等级;当用户选择深度等级时,此时控制耳机ANC的降噪强度对应为深度等级。应理解,轻度、均衡、深度这三种等级对应的ANC的降噪强度依次变大,并且各自都是一种稳态的或稳定的降噪强度。当用户选择轻度等级时,此时无论终端或耳机所处的环境如何变化,耳机ANC功能的降噪强度都对应为轻度等级。同理,并且当用户选择均衡等级时,此时无论终端或耳机所处的环境如何变化,耳机ANC功能的降噪强度都对应为中等等级。并且当用户选择深度等级时,此时无论终端或耳机所处的环境如何变化,耳机ANC功能的降噪强度都对应为深度等级。
可选地,举例说明,轻度等级对应的ANC降噪深度可以包括20-28dB,均衡等级对应的ANC降噪深度可以包括30-36dB,深度等级对应的ANC降噪深度可以大于40dB。
作为举例,轻度等级对应的环境场景可以包括但不限于办公室、卧室、安静的客厅等;均衡等级对应的环境场景可以包括但不限于超市、广场、候车室、马路、咖啡厅、商场等;深度等级对应的环境场景可以包括但不限于地铁、高铁、出租车、飞机等。
此外,本申请实施例中还包括一种智慧动态降噪,即一种自适应环境噪声降噪;即可以通过获取终端或耳机所在的环境的场景类型,根据当前环境的场景类型确定ANC强度;根据确定出来的ANC强度控制所述ANC功能。
作为可选地,自适应环境噪声降噪可以包括但不限于上述轻度、均衡、深度等级中的至少一种。例如,自适应环境噪声降噪可以根据所处环境的状态在轻度、均衡、深度三个等级中进行相应等级的ANC降噪。检测当前环境究竟属于轻度、均衡、深度可以由耳机或终端来执行。相比于上述稳态式的ANC降噪,自适应环境噪声降噪可以让用户无需手动操作,而根据环境的变化进行自适应地不同等级的降噪,提升用户体验。
作为可选地,关于耳机对信号的处理,可以参照S6001-S6005。
S3009,当用户想使用ANC或者HT或者人声增强时,可以使能界面中的“关闭”选项。
应理解,上述所提到的ANC、HT、事件声音增强、降噪模式或者关闭选项的使能方法包括但不限于接收到用户针对相应功能项的点选、终端自适应切换、耳机自适应切换,或者快捷方式的触发。例如,接收到用户针对所述事件声音增强功能的选项的选择操作、针对所述HT功能的选项的选择操作或针对所述ANC功能的选项的选择操作;或者;识别到当前环境是与事件声音增强功能对应的场景并激活所述事件声音增强功能的选项,识别到当前环境是与所述HT功能对应的场景并激活所述HT功能的选项,或者识别到当前环境是与所述ANC功能对应的场景并激活所述ANC功能的选项;或者;响应用户针对所述耳机的按压操作,所述按压操作用于在所述事件声音增强功能、所述HT功能或所述ANC功能中的至少两种功能中切换。例如耳机中可以含有压力传感器,压力传感器可以预先定义一些快捷操作,如降噪模式之间的切换。
应理解,本申请所提到的ANC功能的强度可以理解为ANC的降噪强度,或者简称降噪强度;HT功能的强度可以理解为透传强度;AH功能的强度可以理解为增强强度。不同强度将会影响相关的滤波器系数,具体可以参照前述实施例相关描述,此处不再赘述。
对于耳机侧可能的信号处理,可以参见如下两个实例。
耳机包括第一麦克风(参考麦克风)、第二麦克风(误差麦克风)和扬声器;结合前述实施例中的耳机的可能实现形式,耳机可以执行如下方法:
S4001,通过第一麦克风采集第一信号,第一信号用于表征当前外部环境的声音;现有技 术中,参考麦克风所采集到的信号也叫做参考信号
S4002,通过第二麦克风采集第二信号,第二信号用于表征佩戴耳机的用户耳道内部环境的声音。现有技术中,误差麦克风采集到的信号也叫做误差信号。
耳道内部环境的声音可以理解为用户的耳朵带上耳机之后,结合耳机可能播放的声音、耳机正在采用的算法(如降噪、透传等)、人体耳朵环境等因素对环境声音的综合声音感知。
例如,如果耳机不播音频也不开算法,耳道内部环境的声音可以理解为但不限于误差麦克风采集到的环境声音结合人体耳朵环境的综合声音体现。
例如,如果耳机播音频但不开算法,耳道内部环境的声音可以理解为但不限于误差麦克风采集到的环境声音结合耳机麦克风播放声音以及人体耳朵环境的综合声音体现。
例如,如果耳机播放音频又开了算法,耳道内部环境的声音可以理解为但不限于误差麦克风采集到的环境声音结合耳机麦克风播放经算法处理后的声音以及人体耳朵环境的综合声音体现。
应理解,参考信号和误差信号具体的信号分量跟环境相关,变化因素较多,因此难以用量化的概念来详细描述,但对于本领域技术人员是清楚的概念。
S4003,接收事件声音增强的指令;事件声音为外部环境中符合预设事件条件的声音;
S4004,控制ANC功能和HT功能同时处于开启状态;至少利用HT功能、ANC功能对第一信号和第二信号进行目标处理得到目标信号;目标信号中的事件声音的信噪比大于第一信号中的事件声音的信噪比。
具体地,可以参加图8B或8C,利用HT功能透传参考麦克风采集的第一信号得到还原信号C1;针对还原信号C1中的事件声音信号(例如人声)进行增强,并对于还原信号C1中的非事件声音信号进行减弱,得到事件声音增强信号C2;
利用ANC功能对第一信号C1、误差麦克风采集到的信号、所述事件声音增强信号C2进行处理得到目标信号。
可选地,相关的增强处理可以参考但不限于图9的实施方式。
S4005,通过所述扬声器播放所述目标信号。应理解,从用户的听觉来看,扬声器播放所述目标信号几乎可以抵消用户佩戴耳机时原本能够听到的环境噪声,使得用户最终能够听到事件声音信噪比更高。
可选地,耳机还可以执行如下方法:
S5001,通过第一麦克风用于采集第一信号,第一信号用于表征当前外部环境的声音;
S5002,通过第二麦克风用于采集第二信号,第二信号用于表征佩戴耳机的用户耳道内部环境的声音;
S5003,接收事件声音增强的指令;事件声音为外部环境中符合预设事件条件的声音;
S5004,开启所述ANC功能;对第一信号中的事件声音信号进行增强,并对第一信号中的非事件声音信号进行减弱,得到事件声音增强信号;利用ANC功能对第一信号、第二信号和事件声音增强信号进行处理得到目标信号;目标信号中的事件声音的信噪比大于第一信号中的所述事件声音的信噪比;其中,信号中事件声音的信噪比越高,表明信号中事件声音的能量占比越大。
可选地,相关的增强处理可以参考但不限于图9的实施方式。
S5005,通过扬声器播放目标信号。
对于主动降噪ANC的强度调节,耳机至少支持ANC功能,耳机包括第一麦克风、第三麦克风;这里的第一麦克风可以理解为前述实施例中的参考麦克风,更侧重于采集当前外部环境的声音,第三麦克风更侧重于拾音;当用户佩戴耳机时,第三麦克风相比于第一麦克风更靠近用户的嘴巴,因此第三麦克风相比于第一麦克风能够拾取到更清晰的用户的语音信号;耳机还可以执行如下方法:
S6001,所述耳机开启或使能所述ANC功能;
S6002,通过第一麦克风针对当前环境采集第一信号;
S6003,通过第三麦克风针对当前环境采集第二信号;
S6004,根据第一信号和第二信号判断当前场景的噪声等级;其中,不同的噪声等级对应于不同的ANC强度。
可选地,可以利用第一信号和第二信号的相关性特征作语音活动检测;对非语音信号进行噪声跟踪;若噪声能量小于第一阈值,则将当前场景确定为安静场景;或者,若噪声的频谱重心位于低频段,且噪声能量大于第二阈值,则将当前场景确定为重噪场景;或者,若当前场景既不属于安静场景也不属于重噪场景时,则将当前场景确定为普通场景;其中,第二阈值(例如但不限于[-80dB,-65dB]中的某一个值)大于第一阈值(例如但不限于[-40dB,-30dB]中的某一个值)。所述安静场景、所述普通场景、所述重噪场景对应的ANC降噪强度依次变大。
S6005,根据当前的噪声等级控制ANC功能。
可选地,ANC强度可以预设多种强度调节模式。
可选地,确定噪声等级后按照响应的噪声等级对应的ANC降噪强度调整ANC算法的滤波器,控制ANC功能即可。
可以接受终端发送的ANC强度调节指令,根据ANC强度调节指令调整ANC算法的滤波器,控制ANC功能。可选地,控制ANC功能的方式还可以包括如上述S3007和S3008的控制ANC的强度的方法,此处不再赘述。
可选地,ANC场景间切换还可以有一些策略:若检测到当前场景为新的噪声等级且持续预设时长;获取新的噪声等级对应的ANC强度;根据新的噪声等级对应的ANC强度控制ANC功能。
举例说明:1)满足阈值累计1秒后才允许切换,即新检测到的噪声等级持续预设时长时候才切换到新的噪声等级;2)当前降噪等级至少保持10秒钟才允许切换下一个场景,即新切换的状态至少维持第二预设时长;3)重噪、普通或安静噪声等级可能会存在来回切换的情况,若切换太快则会引起不适。因此,可以监测一段时间内的切换频率,若异常则抬升判定等级的阈值。例如,预设时长内,如果切换的次数超过预设次数(例如2分钟4次),则提升偏向于普通模式的阈值,减少模式的频繁切换,提升用户体验。
应理解,本发明实施例较多,相同的术语、表述、或者信号处理可以相互引用或复用,各个方案中的信号处理过程在不违背自然规律的前提下,可以自由组合;本申请中不予以一一列举。
相应地,本申请实施例中还相应提供一种耳机控制装置,该装置应用于终端设备,终端设备与耳机建立通信连接;耳机支持主动降噪ANC功能和环境声透传HT功能;装置包括:
显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包 括事件声音增强的功能选项;所述事件声音为外部环境中符合预设事件条件的声音。
可选地,第一界面中包括用于控制耳机HT功能的选项;当HT功能的选项使能时,处理模块用于耳机的HT功能;所述显示模块还用于:在所述第一界面中新增所述事件声音增强的选项。
处理模块,用于当所述事件声音增强功能的选项使能时,控制所述耳机的ANC功能和HT功能均处于开启状态。
可选地,当所述ANC功能的选项使能时,处理模块还用于激活所述耳机的ANC功能。
相应地,本申请实施例中还相应提供一种耳机控制装置,该装置应用于终端设备,所述终端设备与耳机建立通信连接;所述耳机至少支持主动降噪ANC功能;所述装置包括:
显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机ANC功能的选项;
处理模块,用于当所述ANC功能的选项使能时,激活所述耳机的ANC功能;
所述显示模块还用于在所述ANC功能的选项使能之后在所述第一界面中新增所述ANC功能的强度选项;
所述处理模块还用于根据所述ANC功能的强度选项的使能结果进行ANC降噪;
其中,所述ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,所述第一稳态ANC强度选项和所述第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的且稳定的ANC功能强度;所述自适应ANC强度选项对应的ANC功能强度与所述终端设备或所述耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。
所述处理模块具体用于:当第一稳态ANC强度选项使能时,获取第一稳态的ANC强度选项对应的第一ANC功能强度;根据所述第一ANC功能强度控制所述ANC功能;或者,当第二稳态ANC强度选项使能时,获取第二稳态的ANC强度选项对应的第二ANC功能强度;根据所述第二ANC功能强度控制所述ANC功能;或者,当所述自适应ANC强度选项使能时,获取所述终端设备或所述耳机所处当前环境的场景类型;根据当前环境的场景类型确定ANC强度;根据确定出来的ANC强度控制所述ANC功能。
相应地,本申请实施例中还相应提供一种耳机控制装置,该装置应用于终端设备,所述终端设备与耳机建立通信连接;所述耳机至少支持环境声透传HT功能,所述装置包括:
显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机HT功能的选项;
处理模块,用于当所述HT功能的选项使能时,激活所述耳机的HT功能;
显示模块还用于在所述HT功能的选项使能后,在所述第一界面中新增所述事件声音增强的选项;所述事件声音为外部环境中符合预设事件条件的声音;
所述处理模块还用于当所述事件声音增强功能的选项使能时,控制所述耳机针对所述耳机采集的信号中所述事件声音的信噪比进行提升;其中,所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大。所述处理模块还具体用于:获取ANC功能的第一强度,并根据所述第一强度控制所述耳机的ANC功能;或者,获取HT功能的第二强度,并根据所述第二强度控制所述耳机的HT功能;或者,获取事件声音增强的第三强度,并根据所述第三强度控制所述耳机的事件声音增强功能。
相应地,本申请实施例中还相应提供一种降噪装置,该装置应用于耳机,所述耳机至少支持主动降噪ANC功能和环境声透传HT功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述装置包括:
采集模块,用于通过所述第一麦克风采集第一信号,所述第一信号用于表征当前外部环境的声音;还用于通过所述第二麦克风采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;
接收模块,用于接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;
处理模块,用于控制ANC功能和HT功能均处于开启状态,并至少利用所述HT功能、所述ANC功能对所述第一信号和所述第二信号进行目标处理得到目标信号;所述目标信号中的事件声音的信噪比大于所述第一信号中的事件声音的信噪比。所述处理模块具体用于:利用HT功能透传所述第一信号得到还原信号;针对所述还原信号中的事件声音信号进行增强,并对于所述还原信号中的非事件声音信号进行减弱,得到事件声音增强信号;利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号。
播放模块,通过所述扬声器播放所述目标信号。
相应地,本申请实施例中还相应提供一种降噪装置,该装置应用于耳机,所述耳机至少支持主动降噪ANC功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述装置包括:
采集模块,用于通过所述第一麦克风采集第一信号,所述第一信号用于表征当前外部环境的声音;还用于通过所述第二麦克风采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;
接收模块,用于接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;
处理模块,用于开启所述ANC功能;对所述第一信号中的事件声音信号进行增强,并对所述第一信号中的非事件声音信号进行减弱,得到事件声音增强信号;利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号;所述目标信号中的所述事件声音的信噪比大于所述第一信号中的所述事件声音的信噪比;其中,信号中所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大;
播放模块,通过所述扬声器播放所述目标信号。
相应地,本申请实施例中还相应提供一种降噪装置,该装置应用于耳机,所述耳机至少支持主动降噪ANC功能、环境声透传HT功能和AH功能,所述耳机包括HT滤波器组、反馈滤波器组和前馈滤波器组;所述装置包括:
获取模块,用于获取耳机的工作模式;
调用模块,用于:当所述工作模式为ANC功能时,调用所述反馈滤波器组和前馈滤波器组,实现所述ANC功能;当所述工作模式为HT功能时,调用所述HT滤波器组和所述反馈滤波器组,实现所述HT功能;当所述工作模式为AH功能时,调用所述HT滤波器组、所述前馈滤波器组和所述反馈滤波器组,实现所述AH功能。
相应地,本申请实施例中还相应提供一种降噪装置,该装置应用于耳机,所述耳机至少 支持ANC功能;所述耳机包括第一麦克风、第三麦克风;所述第一麦克风更侧重于采集当前外部环境的声音,所述第三麦克风更侧重于拾音;
采集模块,用于当所述耳机开启所述ANC功能时,通过所述第一麦克风针对当前环境采集第一信号;通过所述第三麦克风针对所述当前环境采集第二信号;
识别模块,用于根据所述第一信号和所述第二信号判断所述当前场景的噪声等级;其中,不同的噪声等级对应于不同的ANC强度。可选地,所述识别模块具体用于:利用第一信号和第二信号的相关性特征作语音活动检测;对非语音信号进行噪声跟踪;若所述噪声能量小于第一阈值,则将当前场景确定为安静场景;或者,若所述噪声的频谱重心位于低频段,且噪声能量大于第二阈值,则将当前场景确定为重噪场景;或者,若所述当前场景既不属于所述安静场景也不属于所述重噪场景时,则将当前场景确定为普通场景;其中,所述第二阈值大于所述第一阈值。
处理模块,用于根据当前的噪声等级控制所述ANC功能。可选地,所述处理模块具体用于:若检测到当前场景为新的噪声等级且持续预设时长;获取所述新的噪声等级对应的ANC强度;根据所述新的噪声等级对应的ANC强度控制所述ANC功能。
应理解,本发明实施例较多,一些处理方法、名词、可选的实例可以在不同的实施例之间进行复用,本申请不予以详细赘述。
本申请的实施例中的方法步骤可以通过硬件的方式来实现,也可以由处理器执行软件指令的方式来实现。软件指令可以由相应的软件模块组成,软件模块可以被存放于随机存取存储器(random access memory,RAM)、闪存、只读存储器(read-only memory,ROM)、可编程只读存储器(programmableROM,PROM)、可擦除可编程只读存储器(erasable PROM,EPROM)、电可擦除可编程只读存储器(electrically ePROM,EEPROM)、寄存器、硬盘、移动硬盘、CD-ROM或者本领域熟知的任何其它形式的存储介质中。一种示例性的存储介质耦合至处理器,从而使处理器能够从该存储介质读取信息,且可向该存储介质写入信息。当然,存储介质也可以是处理器的组成部分。处理器和存储介质可以位于ASIC中。另外,该ASIC可以位于终端设备中。当然,处理器和存储介质也可以作为分立组件存在于终端设备中。
在上述实施例中,可以全部或部分地通过软件、硬件、固件或者其任意组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。所述计算机程序产品包括一个或多个计算机程序或指令。在计算机上加载和执行所述计算机程序或指令时,全部或部分地执行本申请实施例所述的流程或功能。所述计算机可以是通用计算机、专用计算机、计算机网络、用户设备或者其它可编程装置。所述计算机程序或指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,所述计算机程序或指令可以从一个网站站点、计算机、服务器或数据中心通过有线或无线方式向另一个网站站点、计算机、服务器或数据中心进行传输。所述计算机可读存储介质可以是计算机能够存取的任何可用介质或者是集成一个或多个可用介质的服务器、数据中心等数据存储设备。所述可用介质可以是磁性介质,例如,软盘、硬盘、磁带;也可以是光介质,例如,数字视频光盘(digital video disc,DVD);还可以是半导体介质,例如,固态硬盘(solid state drive,SSD)。
在本申请的各个实施例中,如果没有特殊说明以及逻辑冲突,不同的实施例之间的术语和/或描述具有一致性、且可以相互引用,不同的实施例中的技术特征根据其内在的逻辑关系 可以组合形成新的实施例。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元。方法、系统、产品或设备不必限于清楚地列出的那些步骤或单元,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它步骤或单元。
尽管结合具体特征及其实施例对本申请进行了描述,显而易见的,在不脱离本申请的精神和范围的情况下,可对其进行各种修改和组合。相应地,本说明书和附图仅仅是所附权利要求所界定的方案进行示例性说明,且视为已覆盖本申请范围内的任意和所有修改、变化、组合或等同物。
显然,本领域的技术人员可以对本申请进行各种改动和变型而不脱离本申请的范围。这样,倘若本申请实施例的这些修改和变型属于本申请权利要求及其等同技术的范围之内,则本申请也意图包含这些改动和变型在内。

Claims (93)

  1. 一种模式控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:
    当识别当前外部环境的场景类型为目标场景时,根据所述目标场景确定目标模式;
    其中,所述目标模式为耳机支持的处理模式中的一个,不同的处理模式对应于不同的场景类型,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;
    向所述耳机发送所述目标模式,所述目标模式用于指示所述耳机实现所述目标模式所对应的处理功能。
  2. 如权利要求1所述的方法,其特征在于,根据所述目标场景确定目标模式时,所述方法还包括:
    显示结果提示信息,所述结果提示信息用于提示用户所述耳机实现所述目标模式所对应的处理功能。
  3. 如权利要求1所述的方法,其特征在于,向所述耳机发送第一控制信令之前,所述方法还包括:
    显示选择提示信息,所述选择提示信息用于提示用户是否调整所述耳机的处理模式为所述目标模式;
    检测到用户选择所述耳机的处理模式为所述目标模式的操作。
  4. 如权利要求2或3所述的方法,其特征在于,所述方法还包括:
    显示第一控件和第二控件,其中,所述第二控件在所述第一控件上的不同位置用于指示所述目标模式下的不同处理强度;
    向所述耳机发送第一控制信令之前,所述方法还包括:
    响应于用户触控所述第二控件移动到所述第一控件上的第一位置,所述第二控件在所述第一控件上的第一位置指示所述目标模式下的目标处理强度;
    向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
  5. 如权利要求4所述的方法,其特征在于,所述第一控件的形状为圆环,所述用户触控所述第二控件按顺时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
    所述用户触控所述第二控件按逆时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
  6. 如权利要求4所述的方法,其特征在于,所述第一控件的形状为条形,所述用户触控所述第二控件从上向下在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
    所述用户触控所述第二控件从下向上在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
    所述用户触控所述第二控件从左向右在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
    所述用户触控所述第二控件从右向左在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
  7. 如权利要求4-6任一所述的方法,其特征在于:
    当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,
    当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,
    当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
  8. 一种模式控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:
    获取目标模式;所述目标模式为耳机支持的处理模式中的一个,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;
    根据当前外部环境的场景类型确定所述目标模式下的目标处理强度;不同的场景类型对应于目标模式下的不同处理强度;
    向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
  9. 如权利要求8所述的方法,其特征在于,所述获取目标模式,包括:
    接收所述耳机发送的所述目标模式;或者,
    显示选择控件,所述选择控件包括耳机支持的处理模式,检测到用户通过所述选择控件在所述耳机的处理模式中选择所述目标模式的操作。
  10. 如权利要求8或9所述的方法,其特征在于,根据当前外部环境的场景类型确定所述目标模式下的目标处理强度之前,所述方法还包括:
    接收到所述耳机发送的所述目标模式的情况下,显示选择提示信息,所述选择提示信息用于指示所述用户是否调整所述耳机的处理模式为所述目标模式;
    检测到用户选择调整所述耳机的处理模式为所述目标模式的操作。
  11. 如权利要求8-10任一所述的方法,其特征在于:
    当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,
    当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,
    当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
  12. 一种模式控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:
    显示第一界面;第一界面上包括第一选择控件,所述第一选择控件包括第一目标耳机支持的处理模式以及第一目标耳机支持的处理模式对应的处理强度;所述第一目标耳机的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;
    响应于用户针对所述第一界面执行的第一操作;所述第一操作是用户通过所述第一选择控件在第一目标耳机支持的处理模式中选择第一目标模式以及选择第一目标模式下的处理强度为第一目标处理强度产生的;
    向所述第一目标耳机发送所述第一目标模式以及所述第一目标处理强度,所述第一目标模式用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能,所述第一目标处 理强度用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能时的处理强度。
  13. 如权利要求12所述的方法,其特征在于,显示第一界面之前,所述方法还包括:
    显示选择提示信息,所述选择提示信息用于所述用户选择是否调整所述第一目标耳机的处理模式;
    检测到用户选择调整所述第一目标耳机的处理模式的操作。
  14. 如权利要求11或12所述的方法,其特征在于,显示所述第一界面之前,所述方法还包括:
    识别当前外部环境的场景类型为目标场景,所述目标场景适配需要调整所述第一目标耳机的处理模式的场景类型。
  15. 如权利要求11或12所述的方法,其特征在于,显示所述第一界面之前,所述方法还包括:
    识别到所述终端设备触发所述第一目标耳机播放音频。
  16. 如权利要求11或12所述的方法,其特征在于,显示所述第一界面之前,所述方法还包括:
    检测到所述终端设备与所述第一目标耳机建立连接。
  17. 如权利要求11或12所述的方法,其特征在于,显示所述第一界面之前,所述方法还包括:
    检测到所述终端设备与所述第一目标耳机建立连接的情况下,检测到所述用户在主界面执行的第二操作;
    其中,所述主界面包括第一应用的图标,所述第二操作是所述用户触控所述第一应用的图标产生的,所述第一界面为所述第一应用的显示界面。
  18. 如权利要求11-17任一所述的方法,其特征在于,所述第一选择控件包括第一控件和第二控件,第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的两种不同处理模式,或者,所述第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的的同一种处理模式下的不同处理强度;
    所述第一操作是所述用户移动所述第二控件位于所述第一控件上所述第一目标模式对应的区域内的第一位置产生的,所述第一位置对应于所述第一目标模式下的第一目标处理强度。
  19. 如权利要求18所述的方法,其特征在于,所述第一控件的形状为圆环,所述圆环包括至少两个圆弧段,所述第二控件位于不同的圆弧段指示第一目标耳机的不同处理模式,所述第二控件位于同一圆弧段的不同位置指示第一目标耳机的同一处理模式的不同处理强度;或者,
    所述第一控件的形状为条形,所述条形包括至少两个条形段,所述第二控件位于不同的条形段指示所述第一目标耳机的不同处理模式,所述第二控件位于同一条形段的不同位置指示所述第一目标耳机的同一处理模式的不同处理强度。
  20. 如权利要求11-19任一所述的方法,其特征在于,所述方法还包括:
    响应于用户在第一界面执行的第三操作,所述第一界面上还包括第二选择控件,所述第二选择控件包括第二目标耳机支持的处理模式以及第二目标耳机支持的处理模式对应的处理强度,所述第一目标耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种,所述第三操作是所述用户通过所述第二选择控件在第二目标耳机的处理模式中选择第二目标模式以及选择第二目标模式下的处理强度为第二目标处理 强度产生的;所述第一目标耳机为左耳机时,所述第二目标耳机为右耳机,或者,所述第一目标耳机为右耳机,所述第二目标耳机为左耳机;
    向所述第二目标耳机发送所述第二目标模式以及所述第二目标处理强度,所述第二目标模式用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能,所述第二目标处理强度用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能时的处理强度。
  21. 一种模式控制装置,其特征在于,所述装置应用于终端设备,所述装置包括:
    第一检测模块,用于识别当前外部环境的场景类型为目标场景时,根据所述目标场景确定目标模式;
    其中,所述目标模式为耳机支持的处理模式中的一个,不同的处理模式对应于不同的场景类型,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;
    发送模块,用于向所述耳机发送所述目标模式,所述目标模式用于指示所述耳机实现所述目标模式所对应的处理功能。
  22. 如权利要求21所述的装置,其特征在于,还包括:
    显示模块,用于根据所述目标场景确定目标模式时,显示结果提示信息,所述结果提示信息用于提示用户所述耳机实现所述目标模式所对应的处理功能。
  23. 如权利要求21所述的装置,其特征在于,还包括:
    显示模块,用于在向所述耳机发送第一控制信令之前,显示选择提示信息,所述选择提示信息用于提示用户是否调整所述耳机的处理模式为所述目标模式;
    第二检测模块,用于检测到用户选择所述耳机的处理模式为所述目标模式的操作。
  24. 如权利要求23所述的装置,其特征在于,所述显示模块,还用于显示第一控件和第二控件,其中,所述第二控件在所述第一控件上的不同位置用于指示所述目标模式下的不同处理强度;
    所述第二检测模块,还用于在发送模块向所述耳机发送第一控制信令之前,检测用户触控所述第二控件移动到所述第一控件上的第一位置的操作,所述第二控件在所述第一控件上的第一位置指示所述目标模式下的目标处理强度;
    所述发送模块,还用于向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
  25. 如权利要求24所述的装置,其特征在于,所述第一控件的形状为圆环,所述用户触控所述第二控件按顺时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
    所述用户触控所述第二控件按逆时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
  26. 如权利要求24所述的装置,其特征在于,所述第一控件的形状为条形,所述用户触控所述第二控件从上向下在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
    所述用户触控所述第二控件从下向上在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
    所述用户触控所述第二控件从左向右在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,
    所述用户触控所述第二控件从右向左在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
  27. 如权利要求24-26任一所述的装置,其特征在于:
    当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,
    当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,
    当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
  28. 一种模式控制装置,其特征在于,所述装置应用于终端设备,所述装置包括:
    处理模块,用于获取目标模式;所述目标模式为耳机支持的处理模式中的一个,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;
    所述处理模块,还用于根据当前外部环境的场景类型确定所述目标模式下的目标处理强度;不同的场景类型对应于目标模式下的不同处理强度;
    发送模块,用于向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
  29. 如权利要求28所述的装置,其特征在于,还包括:
    接收模块,用于接收所述耳机发送的所述目标模式。
  30. 如权利要求28所述的装置,其特征在于,还包括:
    显示模块,用于显示选择控件,所述选择控件包括耳机支持的处理模式,检测到用户通过所述选择控件在所述耳机的处理模式中选择所述目标模式的操作。
  31. 如权利要求29所述的装置,其特征在于,显示模块,还用于:
    在所述处理模块根据当前外部环境的场景类型确定所述目标模式下的目标处理强度之前,所述接收模块接收到所述耳机发送的所述目标模式的情况下,显示选择提示信息,所述选择提示信息用于指示所述用户是否调整所述耳机的处理模式为所述目标模式;
    还包括:
    检测模块,用于检测到用户选择调整所述耳机的处理模式为所述目标模式的操作。
  32. 如权利要求28-31任一所述的装置,其特征在于:
    当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,
    当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,
    当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
  33. 一种模式控制装置,其特征在于,所述装置应用于终端设备,所述装置包括:
    显示模块,用于第一界面上包括第一选择控件,所述第一选择控件包括第一目标耳机支持的处理模式以及第一目标耳机支持的处理模式对应的处理强度;所述第一目标耳机的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;
    检测模块,检测用户针对所述第一界面执行的第一操作;所述第一操作是用户通过所述第一选择控件在第一目标耳机支持的处理模式中选择第一目标模式以及选择第一目标模式下 的处理强度为第一目标处理强度产生的;
    发送模块,用于向所述第一目标耳机发送所述第一目标模式以及所述第一目标处理强度,所述第一目标模式用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能,所述第一目标处理强度用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能时的处理强度。
  34. 如权利要求33所述的装置法,其特征在于,所述显示模块,还用于在显示第一界面之前,显示选择提示信息,所述选择提示信息用于所述用户选择是否调整所述第一目标耳机的处理模式;
    所述检测模块,还用于检测用户选择调整所述第一目标耳机的处理模式的操作。
  35. 如权利要求33或34所述的装置,其特征在于,还包括:
    识别模块,用于在显示模式显示第一界面之前,识别当前外部环境的场景类型为目标场景,所述目标场景适配需要调整所述第一目标耳机的处理模式的场景类型。
  36. 如权利要求33或34所述的装置,其特征在于,还包括:
    识别模块,用于在显示模块显示所述第一界面之前,识别到所述终端设备触发所述第一目标耳机播放音频。
  37. 如权利要求33或34所述的装置,其特征在于,所述检测模块,还用于在显示模块显示所述第一界面之前,检测到所述终端设备与所述第一目标耳机建立连接。
  38. 如权利要求33或34所述的装置,其特征在于,所述检测模块,在显示模块显示所述第一界面之前,检测到所述终端设备与所述第一目标耳机建立连接的情况下,检测到所述用户在主界面执行的第二操作;
    其中,所述主界面包括第一应用的图标,所述第二操作是所述用户触控所述第一应用的图标产生的,所述第一界面为所述第一应用的显示界面。
  39. 如权利要求33-38任一所述的装置,其特征在于,所述第一选择控件包括第一控件和第二控件,第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的两种不同处理模式,或者,所述第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的的同一种处理模式下的不同处理强度;
    所述第一操作是所述用户移动所述第二控件位于所述第一控件上所述第一目标模式对应的区域内的第一位置产生的,所述第一位置对应于所述第一目标模式下的第一目标处理强度。
  40. 如权利要求39所述的装置,其特征在于,所述第一控件的形状为圆环,所述圆环包括至少两个圆弧段,所述第二控件位于不同的圆弧段指示第一目标耳机的不同处理模式,所述第二控件位于同一圆弧段的不同位置指示第一目标耳机的同一处理模式的不同处理强度;或者,
    所述第一控件的形状为条形,所述条形包括至少两个条形段,所述第二控件位于不同的条形段指示所述第一目标耳机的不同处理模式,所述第二控件位于同一条形段的不同位置指示所述第一目标耳机的同一处理模式的不同处理强度。
  41. 如权利要求33-40任一所述的方法,其特征在于,所述检测模块,还用于响应于用户在第一界面执行的第三操作,所述第一界面上还包括第二选择控件,所述第二选择控件包括第二目标耳机支持的处理模式以及第二目标耳机支持的处理模式对应的处理强度,所述第一目标耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种,所述第三操作是所述用户通过所述第二选择控件在第二目标耳机的处理模式中选择第二目标模式以及选择第二目标模式下的处理强度为第二目标处理强度产生的; 所述第一目标耳机为左耳机时,所述第二目标耳机为右耳机,或者,所述第一目标耳机为右耳机,所述第二目标耳机为左耳机;
    所述发送模块,还用于向所述第二目标耳机发送所述第二目标模式以及所述第二目标处理强度,所述第二目标模式用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能,所述第二目标处理强度用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能时的处理强度。
  42. 一种终端设备,其特征在于,包括存储器、处理器和显示器;
    所述显示器,用于显示界面;
    所述存储器,用于存储程序或指令;
    所述处理器,用于调用所述程序或指令,以使得所述终端设备执行如权利要求1至20任一项所述的方法。
  43. 一种耳机控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:
    所述终端设备与耳机建立通信连接;所述耳机支持主动降噪ANC功能和环境声透传HT功能,
    显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括事件声音增强的功能选项;所述事件声音为外部环境中符合预设事件条件的声音;
    当所述事件声音增强功能的选项使能时,控制所述耳机的ANC功能和HT功能均处于开启状态。
  44. 如权利要求43所述的方法,其特征在于,所述第一界面中包括用于控制所述耳机HT功能的选项;在所述显示第一界面之后,所述方法还包括:
    当所述HT功能的选项使能时,激活所述耳机的HT功能;
    在所述第一界面中新增所述事件声音增强的选项。
  45. 如权利要求44所述的方法,其特征在于,所述控制所述耳机的ANC功能和HT功能均处于开启状态包括:
    控制所述耳机保持所述HT功能处于开启状态;
    激活所述耳机的ANC功能。
  46. 如权利要求43-45中任一项所述的方法,其特征在于,所述第一界面中包括用于控制所述耳机ANC功能的选项;所述方法还包括:
    当所述ANC功能的选项使能时,激活所述耳机的ANC功能。
  47. 如权利要求43-46中任一项所述的方法,其特征在于,所述第一界面中还包括用于关闭所述ANC功能的选项,和/或关闭所述HT功能的选项。
  48. 如权利要求47所述的方法,其特征在于,在激活所述耳机的ANC功能之后,在所述第一界面中新增所述ANC功能的强度选项。
  49. 如权利要求48所述的方法,其特征在于,
    所述ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,所述第一稳态ANC强度选项和所述第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的ANC功能强度;所述自适应ANC强度选项对应的ANC功能强度与所述终端设备或所述耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。
  50. 如权利要求49所述的方法,其特征在于,所述方法还包括,当所述ANC强度自适 应选项被使能时,获取所述终端或所述耳机所处当前环境的场景类型;
    根据所述场景类型匹配出ANC功能的目标强度;
    根据所述目标强度控制所述耳机的ANC功能。
  51. 如权利要求50所述的方法,其特征在于,所述当前环境不同的场景类型包括第一场景和第二场景。
  52. 如权利要求43-51任一项所述的方法,其特征在于,所述事件声音包括人声或者符合预设频谱特性的其它声音。
  53. 如权利要求43-52任一项所述的方法,其特征在于,在激活所述耳机的HT功能之后,所述方法还包括:
    获取HT功能的目标强度;
    根据所述HT功能的目标强度控制所述耳机的HT功能。
  54. 如权利要求43-53任一项所述的方法,其特征在于,所述事件声音增强功能的选项使能、所述HT功能的选项使能或所述ANC功能的选项使能包括:
    响应用户针对相应功能选项的点选、相应功能的自适应切换,或者相应功能的快捷方式触发。
  55. 如权利要求43-54任一项所述的方法,其特征在于,所述ANC功能和HT功能共同开启用于对所述耳机采集到的外部环境的声音信号进行处理,所述处理包括针对采集的信号中所述事件声音的信噪比进行提升;其中,所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大。
  56. 一种降噪方法,其特征在于,所述方法应用于耳机,所述耳机至少支持主动降噪ANC功能和环境声透传HT功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述方法包括:
    通过所述第一麦克风采集第一信号,所述第一信号用于表征当前外部环境的声音;
    通过所述第二麦克风采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;
    接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;
    控制ANC功能和HT功能均处于开启状态;至少利用所述HT功能、所述ANC功能对所述第一信号和所述第二信号进行目标处理得到目标信号;所述目标信号中的事件声音的信噪比大于所述第一信号中的事件声音的信噪比;
    通过所述扬声器播放所述目标信号。
  57. 如权利要求56所述的方法,其特征在于,所述至少利用所述HT功能、所述ANC功能对所述第一信号和所述第二信号进行目标处理得到目标信号包括:
    利用HT功能透传所述第一信号得到还原信号;
    针对所述还原信号中的事件声音信号进行增强,并对于所述还原信号中的非事件声音信号进行减弱,得到事件声音增强信号;
    利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号。
  58. 一种降噪方法,其特征在于,所述方法应用于耳机,所述耳机至少支持主动降噪ANC功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述方法包括:
    通过所述第一麦克风用于采集第一信号,所述第一信号用于表征当前外部环境的声音;
    通过所述第二麦克风用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;
    接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;
    开启所述ANC功能;
    对所述第一信号中的事件声音信号进行增强,并对所述第一信号中的非事件声音信号进行减弱,得到事件声音增强信号;
    利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号;所述目标信号中的所述事件声音的信噪比大于所述第一信号中的所述事件声音的信噪比;其中,信号中所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大;
    通过所述扬声器播放所述目标信号。
  59. 一种信号处理方法,其特征在于,所述方法应用于耳机,所述耳机至少支持主动降噪ANC功能、环境声透传HT功能和AH功能,所述耳机包括HT滤波器组、反馈滤波器组和前馈滤波器组;所述方法包括:
    获取耳机的工作模式;
    当所述工作模式为ANC功能时,调用所述反馈滤波器组和前馈滤波器组,实现所述ANC功能;
    当所述工作模式为HT功能时,调用所述HT滤波器组和所述反馈滤波器组,实现所述HT功能;
    当所述工作模式为AH功能时,调用所述HT滤波器组、所述前馈滤波器组和所述反馈滤波器组,实现所述AH功能。
  60. 一种主动降噪ANC的强度调节方法,其特征在于,所述方法应用于耳机,所述耳机至少支持ANC功能;所述耳机包括第一麦克风、第三麦克风;所述第一麦克风更侧重于采集当前外部环境的声音,所述第三麦克风更侧重于拾音;
    当所述耳机的所述ANC功能处于开启状态时,
    通过所述第一麦克风针对当前环境采集第一信号;
    通过所述第三麦克风针对所述当前环境采集第二信号;
    根据所述第一信号和所述第二信号判断所述当前场景的噪声等级;其中,不同的噪声等级对应于不同的ANC强度;
    根据当前的噪声等级控制所述ANC功能。
  61. 如权利要求60所述的方法,其特征在于,所述根据所述第一信号和所述第二信号判断所述当前场景的噪声等级包括:
    利用第一信号和第二信号的相关性特征作语音活动检测;
    对非语音信号进行噪声跟踪;
    若所述噪声能量小于第一阈值,则将当前场景确定为安静场景;或者,
    若所述噪声的频谱重心位于低频段,且噪声能量大于第二阈值,则将当前场景确定为重噪场景;或者,
    若所述当前场景既不属于所述安静场景也不属于所述重噪场景时,则将当前场景确定为普通场景;其中,所述第二阈值大于所述第一阈值。
  62. 如权利要求61所述的方法,其特征在于,所述安静场景、所述普通场景、所述重噪 场景对应的ANC强度依次变大。
  63. 如权利要求60-62中任一项所述的方法,其特征在于,所述根据当前的噪声等级控制所述ANC功能包括:
    若检测到当前场景为新的噪声等级且持续预设时长;
    获取所述新的噪声等级对应的ANC强度;
    根据所述新的噪声等级对应的ANC强度控制所述ANC功能。
  64. 一种耳机控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:
    所述终端设备与耳机建立通信连接;所述耳机至少支持主动降噪ANC功能;
    显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机ANC功能的选项;
    当所述ANC功能的选项使能时,激活所述耳机的ANC功能;
    在所述第一界面中新增所述ANC功能的强度选项;
    根据所述ANC功能的强度选项的使能结果进行ANC降噪;
    其中,所述ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,所述第一稳态ANC强度选项和所述第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的且稳定的ANC功能强度;所述自适应ANC强度选项对应的ANC功能强度与所述终端设备或所述耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。
  65. 如权利要求64所述的方法,其特征在于,所述根据所述ANC功能的强度选项的使能结果进行ANC降噪包括:
    当第一稳态ANC强度选项使能时,
    获取第一稳态的ANC强度选项对应的第一ANC功能强度;
    根据所述第一ANC功能强度控制所述ANC功能;或者,
    当第二稳态ANC强度选项使能时,
    获取第二稳态的ANC强度选项对应的第二ANC功能强度;
    根据所述第二ANC功能强度控制所述ANC功能。
  66. 如权利要求64或65所述的方法,其特征在于,所述根据所述ANC功能的强度选项的使能结果进行ANC降噪包括:
    当所述自适应ANC强度选项使能时,
    获取所述终端设备或所述耳机所处当前环境的场景类型;
    根据当前环境的场景类型确定ANC强度;
    根据确定出来的ANC强度控制所述ANC功能。
  67. 如权利要求64-66中任一项所述的方法,其特征在于,所述当前环境不同的场景类型包括第一场景和第二场景。
  68. 一种耳机控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:
    所述终端设备与耳机建立通信连接;所述耳机至少支持环境声透传HT功能,
    显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机HT功能的选项;
    当所述HT功能的选项使能时,激活所述耳机的HT功能;
    在所述第一界面中新增所述事件声音增强的选项;所述事件声音为外部环境中符合预设事件条件的声音;
    当所述事件声音增强功能的选项使能时,控制所述耳机针对所述耳机采集的信号中所述事件声音的信噪比进行提升;其中,所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大。
  69. 如权利要求68所述的方法,其特征在于,所述事件声音包括人声或者符合预设频谱特性的其它声音。
  70. 如权利要求68或69所述的方法,其特征在于,所述第一界面中包括用于控制所述耳机ANC功能的选项、用于关闭所述ANC功能的选项,和/或关闭所述HT功能的选项。
  71. 如权利要求70所述的方法,其特征在于,所述方法还包括:
    获取ANC功能的第一强度,并根据所述第一强度控制所述耳机的ANC功能;或者,
    获取HT功能的第二强度,并根据所述第二强度控制所述耳机的HT功能;或者,
    获取事件声音增强的第三强度,并根据所述第三强度控制所述耳机的事件声音增强功能。
  72. 一种耳机控制装置,其特征在于,所述装置应用于终端设备,所述终端设备与耳机建立通信连接;所述耳机支持主动降噪ANC功能和环境声透传HT功能;所述装置包括:
    显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括事件声音增强的功能选项;所述事件声音为外部环境中符合预设事件条件的声音;
    处理模块,用于当所述事件声音增强功能的选项使能时,控制所述耳机的ANC功能和HT功能均处于开启状态。
  73. 如权利要求72所述的装置,其特征在于,所述第一界面中包括用于控制所述耳机HT功能的选项;当所述HT功能的选项使能时,所述处理模块用于激活所述耳机的HT功能;所述显示模块还用于:在所述第一界面中新增所述事件声音增强的选项。
  74. 如权利要求73所述的装置,其特征在于,所述处理模块还用于控制所述耳机保持所述HT功能处于开启状态;激活所述耳机的ANC功能。
  75. 如权利要求72-74中任一项所述的装置,其特征在于,所述第一界面中包括用于控制所述耳机ANC功能的选项;所述处理模块还用于:
    当所述ANC功能的选项使能时,激活所述耳机的ANC功能。
  76. 如权利要求72-75中任一项所述的装置,其特征在于,所述第一界面中还包括用于关闭所述ANC功能的选项,和/或关闭所述HT功能的选项。
  77. 如权利要求72-76中任一项所述的装置,其特征在于,所述事件声音包括人声或者符合预设频谱特性的其它声音。
  78. 一种降噪装置,其特征在于,所述装置应用于耳机,所述耳机至少支持主动降噪ANC功能和环境声透传HT功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述装置包括:
    采集模块,用于通过所述第一麦克风采集第一信号,所述第一信号用于表征当前外部环境的声音;还用于通过所述第二麦克风采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;
    接收模块,用于接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;
    处理模块,用于控制ANC功能和HT功能均处于开启状态,并至少利用所述HT功能、 所述ANC功能对所述第一信号和所述第二信号进行目标处理得到目标信号;所述目标信号中的事件声音的信噪比大于所述第一信号中的事件声音的信噪比;
    播放模块,通过所述扬声器播放所述目标信号。
  79. 如权利要求78所述的装置,其特征在于,所述处理模块具体用于:
    利用HT功能透传所述第一信号得到还原信号;
    针对所述还原信号中的事件声音信号进行增强,并对于所述还原信号中的非事件声音信号进行减弱,得到事件声音增强信号;
    利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号。
  80. 一种降噪装置,其特征在于,所述装置应用于耳机,所述耳机至少支持主动降噪ANC功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述装置包括:
    采集模块,用于通过所述第一麦克风采集第一信号,所述第一信号用于表征当前外部环境的声音;还用于通过所述第二麦克风采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;
    接收模块,用于接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;
    处理模块,用于开启所述ANC功能;对所述第一信号中的事件声音信号进行增强,并对所述第一信号中的非事件声音信号进行减弱,得到事件声音增强信号;利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号;所述目标信号中的所述事件声音的信噪比大于所述第一信号中的所述事件声音的信噪比;其中,信号中所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大;
    播放模块,通过所述扬声器播放所述目标信号。
  81. 一种信号处理装置,其特征在于,所述装置应用于耳机,所述耳机至少支持主动降噪ANC功能、环境声透传HT功能和AH功能,所述耳机包括HT滤波器组、反馈滤波器组和前馈滤波器组;所述装置包括:
    获取模块,用于获取耳机的工作模式;
    调用模块,用于:
    当所述工作模式为ANC功能时,调用所述反馈滤波器组和前馈滤波器组,实现所述ANC功能;
    当所述工作模式为HT功能时,调用所述HT滤波器组和所述反馈滤波器组,实现所述HT功能;
    当所述工作模式为AH功能时,调用所述HT滤波器组、所述前馈滤波器组和所述反馈滤波器组,实现所述AH功能。
  82. 一种主动降噪ANC的强度调节装置,其特征在于,所述装置应用于耳机,所述耳机至少支持ANC功能;所述耳机包括第一麦克风、第三麦克风;所述第一麦克风更侧重于采集当前外部环境的声音,所述第三麦克风更侧重于拾音;
    采集模块,用于当所述耳机的所述ANC功能处于开启状态时,通过所述第一麦克风针对当前环境采集第一信号;通过所述第三麦克风针对所述当前环境采集第二信号;
    识别模块,用于根据所述第一信号和所述第二信号判断所述当前场景的噪声等级;其中,不同的噪声等级对应于不同的ANC强度;
    处理模块,用于根据当前的噪声等级控制所述ANC功能。
  83. 如权利要求82所述的装置,其特征在于,所述识别模块具体用于:
    利用第一信号和第二信号的相关性特征作语音活动检测;
    对非语音信号进行噪声跟踪;
    若所述噪声能量小于第一阈值,则将当前场景确定为安静场景;或者,
    若所述噪声的频谱重心位于低频段,且噪声能量大于第二阈值,则将当前场景确定为重噪场景;或者,
    若所述当前场景既不属于所述安静场景也不属于所述重噪场景时,则将当前场景确定为普通场景;其中,所述第二阈值大于所述第一阈值。
  84. 如权利要求83所述的装置,其特征在于,所述安静场景、所述普通场景、所述重噪场景对应的ANC强度依次变大。
  85. 如权利要求82-84中任一项所述的装置,其特征在于,所述处理模块具体用于:
    若检测到当前场景为新的噪声等级且持续预设时长;
    获取所述新的噪声等级对应的ANC强度;
    根据所述新的噪声等级对应的ANC强度控制所述ANC功能。
  86. 一种耳机控制装置,其特征在于,所述装置应用于终端设备,所述终端设备与耳机建立通信连接;所述耳机至少支持主动降噪ANC功能;所述装置包括:
    显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机ANC功能的选项;
    处理模块,用于当所述ANC功能的选项使能时,激活所述耳机的ANC功能;
    所述显示模块还用于在所述ANC功能的选项使能之后在所述第一界面中新增所述ANC功能的强度选项;
    所述处理模块还用于根据所述ANC功能的强度选项的使能结果进行ANC降噪;
    其中,所述ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,所述第一稳态ANC强度选项和所述第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的且稳定的ANC功能强度;所述自适应ANC强度选项对应的ANC功能强度与所述终端设备或所述耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。
  87. 如权利要求86所述的装置,其特征在于,所述处理模块具体用于:
    当第一稳态ANC强度选项使能时,
    获取第一稳态的ANC强度选项对应的第一ANC功能强度;
    根据所述第一ANC功能强度控制所述ANC功能;或者,
    当第二稳态ANC强度选项使能时,
    获取第二稳态的ANC强度选项对应的第二ANC功能强度;
    根据所述第二ANC功能强度控制所述ANC功能;或者,
    当所述自适应ANC强度选项使能时,
    获取所述终端设备或所述耳机所处当前环境的场景类型;
    根据当前环境的场景类型确定ANC强度;
    根据确定出来的ANC强度控制所述ANC功能。
  88. 一种耳机控制装置,其特征在于,所述装置应用于终端设备,所述终端设备与耳机建立通信连接;所述耳机至少支持环境声透传HT功能,所述装置包括:
    显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机HT功能的选项;
    处理模块,用于当所述HT功能的选项使能时,激活所述耳机的HT功能;
    显示模块还用于在所述HT功能的选项使能后,在所述第一界面中新增所述事件声音增强的选项;所述事件声音为外部环境中符合预设事件条件的声音;
    所述处理模块还用于当所述事件声音增强功能的选项使能时,控制所述耳机针对所述耳机采集的信号中所述事件声音的信噪比进行提升;其中,所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大。
  89. 如权利要求88所述的装置,其特征在于,所述事件声音包括人声或者符合预设频谱特性的其它声音。
  90. 如权利要求88或89所述的装置,其特征在于,所述第一界面中包括用于控制所述耳机ANC功能的选项、用于关闭所述ANC功能的选项,和/或关闭所述HT功能的选项。
  91. 如权利要求90所述的装置,其特征在于,所述处理模块还具体用于:
    获取ANC功能的第一强度,并根据所述第一强度控制所述耳机的ANC功能;或者,
    获取HT功能的第二强度,并根据所述第二强度控制所述耳机的HT功能;或者,
    获取事件声音增强的第三强度,并根据所述第三强度控制所述耳机的事件声音增强功能。
  92. 一种终端设备,其特征在于,所述终端设备包括存储器、处理器、触摸屏、总线;所述存储器、所述触摸屏以及所述处理器通过所述总线相连;
    所述存储器用于存储计算机程序和指令;
    所述触摸屏用于显示界面;
    所述处理器用于调用所述存储器中存储的所述计算机程序、指令,可执行如权利要求43-55,64-67,68-71中任一项所述方法。
  93. 一种耳机,所述耳机包括第一麦克风、第二麦克风、处理器、存储器以及扬声器;所述第一麦克风,用于采集第一信号,所述第一信号用于表征当前外部环境的声音;所述第二麦克风,用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;所述扬声器用于播放音频,所述存储器,用于存储程序或指令;所述处理器,用于调用所述程序或指令,以使得所述耳机执行如权利要求56-57,58,59,60-63中任一项所述方法。
PCT/CN2021/103435 2020-06-30 2021-06-30 一种模式控制方法、装置及终端设备 WO2022002110A1 (zh)

Priority Applications (4)

Application Number Priority Date Filing Date Title
EP21834164.2A EP4171060A4 (en) 2020-06-30 2021-06-30 MODE CONTROL METHOD AND APPARATUS, AND TERMINAL DEVICE
BR112022026923A BR112022026923A2 (pt) 2020-06-30 2021-06-30 Método e aparelho de controle de modo e dispositivo terminal
KR1020237002989A KR20230027296A (ko) 2020-06-30 2021-06-30 모드 제어 방법 및 장치, 그리고 단말 디바이스
US18/148,080 US20230164475A1 (en) 2020-06-30 2022-12-29 Mode Control Method and Apparatus, and Terminal Device

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN202010616084 2020-06-30
CN202010616084.7 2020-06-30
CN202010949885.5A CN113873379B (zh) 2020-06-30 2020-09-10 一种模式控制方法、装置及终端设备
CN202010949885.5 2020-09-10

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US18/148,080 Continuation US20230164475A1 (en) 2020-06-30 2022-12-29 Mode Control Method and Apparatus, and Terminal Device

Publications (1)

Publication Number Publication Date
WO2022002110A1 true WO2022002110A1 (zh) 2022-01-06

Family

ID=78982086

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/103435 WO2022002110A1 (zh) 2020-06-30 2021-06-30 一种模式控制方法、装置及终端设备

Country Status (6)

Country Link
US (1) US20230164475A1 (zh)
EP (1) EP4171060A4 (zh)
KR (1) KR20230027296A (zh)
CN (1) CN113873379B (zh)
BR (1) BR112022026923A2 (zh)
WO (1) WO2022002110A1 (zh)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114640938A (zh) * 2022-05-18 2022-06-17 深圳市听多多科技有限公司 一种基于蓝牙耳机芯片的助听功能实现方法及蓝牙耳机
CN114640937A (zh) * 2022-05-18 2022-06-17 深圳市听多多科技有限公司 一种基于穿戴设备系统的助听功能实现方法及穿戴设备
EP4290885A1 (en) * 2022-06-07 2023-12-13 Starkey Laboratories, Inc. Context-based situational awareness for hearing instruments

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
USD921669S1 (en) * 2019-09-09 2021-06-08 Apple Inc. Display screen or portion thereof with animated graphical user interface
TWI802108B (zh) * 2021-05-08 2023-05-11 英屬開曼群島商意騰科技股份有限公司 降低聲學回音之語音處理裝置及其方法
CN113938787B (zh) * 2021-12-16 2022-03-15 深圳市鑫正宇科技有限公司 基于数字均衡技术的骨传导耳机
CN114466278B (zh) * 2022-04-11 2022-08-16 北京荣耀终端有限公司 一种耳机模式对应的参数确定方法、耳机、终端和系统
CN117751585A (zh) * 2022-06-20 2024-03-22 北京小米移动软件有限公司 智能耳机的控制方法、装置、电子设备和存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140126734A1 (en) * 2012-11-02 2014-05-08 Bose Corporation Providing Ambient Naturalness in ANR Headphones
CN106303839A (zh) * 2016-09-29 2017-01-04 中山市天键电声有限公司 基于手机app的anc降噪控制方法
US20180077482A1 (en) * 2015-05-15 2018-03-15 Huawei Technologies Co., Ltd. Noise Reduction Headset Setting Method, Terminal, and Noise Reduction Headset
US10284955B2 (en) * 2013-05-23 2019-05-07 Comhear, Inc. Headphone audio enhancement system
CN110825446A (zh) * 2019-10-28 2020-02-21 Oppo广东移动通信有限公司 参数配置方法、装置、存储介质及电子设备
CN111107461A (zh) * 2019-12-13 2020-05-05 恒玄科技(北京)有限公司 一种降噪耳机的配置方法、装置及智能终端、降噪耳机

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9716939B2 (en) * 2014-01-06 2017-07-25 Harman International Industries, Inc. System and method for user controllable auditory environment customization
CN104618829A (zh) * 2014-12-29 2015-05-13 歌尔声学股份有限公司 耳机环境声音的调节方法和耳机
JP6588098B2 (ja) * 2015-01-26 2019-10-09 深▲せん▼市冠旭電子股▲ふん▼有限公司Shenzhen Grandsun Electronic Co., Ltd. ヘッドセットの騒音低減を制御する方法及び装置
WO2017101067A1 (zh) * 2015-12-17 2017-06-22 华为技术有限公司 一种对周围环境音的处理方法及设备
US10809972B2 (en) * 2016-09-27 2020-10-20 Sony Corporation Information processing device, information processing method, and program
CN110049403A (zh) * 2018-01-17 2019-07-23 北京小鸟听听科技有限公司 一种基于场景识别的自适应音频控制装置和方法
CN108429963A (zh) * 2018-05-08 2018-08-21 歌尔股份有限公司 一种耳机及降噪方法

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140126734A1 (en) * 2012-11-02 2014-05-08 Bose Corporation Providing Ambient Naturalness in ANR Headphones
US10284955B2 (en) * 2013-05-23 2019-05-07 Comhear, Inc. Headphone audio enhancement system
US20180077482A1 (en) * 2015-05-15 2018-03-15 Huawei Technologies Co., Ltd. Noise Reduction Headset Setting Method, Terminal, and Noise Reduction Headset
CN106303839A (zh) * 2016-09-29 2017-01-04 中山市天键电声有限公司 基于手机app的anc降噪控制方法
CN110825446A (zh) * 2019-10-28 2020-02-21 Oppo广东移动通信有限公司 参数配置方法、装置、存储介质及电子设备
CN111107461A (zh) * 2019-12-13 2020-05-05 恒玄科技(北京)有限公司 一种降噪耳机的配置方法、装置及智能终端、降噪耳机

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP4171060A4 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114640938A (zh) * 2022-05-18 2022-06-17 深圳市听多多科技有限公司 一种基于蓝牙耳机芯片的助听功能实现方法及蓝牙耳机
CN114640937A (zh) * 2022-05-18 2022-06-17 深圳市听多多科技有限公司 一种基于穿戴设备系统的助听功能实现方法及穿戴设备
CN114640938B (zh) * 2022-05-18 2022-08-23 深圳市听多多科技有限公司 一种基于蓝牙耳机芯片的助听功能实现方法及蓝牙耳机
CN114640937B (zh) * 2022-05-18 2022-09-02 深圳市听多多科技有限公司 一种基于穿戴设备系统的助听功能实现方法及穿戴设备
EP4290885A1 (en) * 2022-06-07 2023-12-13 Starkey Laboratories, Inc. Context-based situational awareness for hearing instruments

Also Published As

Publication number Publication date
KR20230027296A (ko) 2023-02-27
CN113873379B (zh) 2023-05-02
US20230164475A1 (en) 2023-05-25
BR112022026923A2 (pt) 2023-03-14
EP4171060A4 (en) 2024-03-13
EP4171060A1 (en) 2023-04-26
CN113873379A (zh) 2021-12-31

Similar Documents

Publication Publication Date Title
WO2022002166A1 (zh) 一种耳机噪声处理方法、装置及耳机
WO2022002110A1 (zh) 一种模式控制方法、装置及终端设备
CN113169760B (zh) 无线短距离音频共享方法及电子设备
WO2021083128A1 (zh) 一种声音处理方法及其装置
WO2021000817A1 (zh) 环境音处理方法及相关装置
CN113571035B (zh) 降噪方法及降噪装置
CN115729511A (zh) 一种播放音频的方法及电子设备
CN114466107A (zh) 音效控制方法、装置、电子设备及计算机可读存储介质
JP2023156321A (ja) 通話方法および装置
CN113141483B (zh) 基于视频通话的共享屏幕方法及移动设备
CN111065020B (zh) 音频数据处理的方法和装置
WO2022257563A1 (zh) 一种音量调节的方法,电子设备和系统
CN114339429A (zh) 音视频播放控制方法、电子设备和存储介质
WO2022089563A1 (zh) 一种声音增强方法、耳机控制方法、装置及耳机
WO2024046416A1 (zh) 一种音量调节方法、电子设备及系统
CN113938556B (zh) 来电提示方法、装置和电子设备
WO2024066933A9 (zh) 扬声器控制方法及设备
WO2023197999A1 (zh) 显示方法及电子设备
WO2024046182A1 (zh) 一种音频播放方法、系统及相关装置
WO2024032035A1 (zh) 一种语音信号的输出方法和电子设备
CN117528370A (zh) 信号处理方法及装置、设备控制方法及装置
CN118098261A (zh) 一种音频处理方法、设备及系统
CN115706752A (zh) 来电音量控制方法及电子设备
CN117692845A (zh) 声场校准方法、电子设备及系统

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21834164

Country of ref document: EP

Kind code of ref document: A1

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112022026923

Country of ref document: BR

ENP Entry into the national phase

Ref document number: 2021834164

Country of ref document: EP

Effective date: 20230120

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 112022026923

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20221228