WO2022002166A1 - 一种耳机噪声处理方法、装置及耳机 - Google Patents
一种耳机噪声处理方法、装置及耳机 Download PDFInfo
- Publication number
- WO2022002166A1 WO2022002166A1 PCT/CN2021/103768 CN2021103768W WO2022002166A1 WO 2022002166 A1 WO2022002166 A1 WO 2022002166A1 CN 2021103768 W CN2021103768 W CN 2021103768W WO 2022002166 A1 WO2022002166 A1 WO 2022002166A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- signal
- processing
- target
- scene
- function
- Prior art date
Links
- 238000003672 processing method Methods 0.000 title claims abstract description 8
- 238000012545 processing Methods 0.000 claims abstract description 946
- 230000006870 function Effects 0.000 claims abstract description 384
- 230000005236 sound signal Effects 0.000 claims abstract description 286
- 210000000613 ear canal Anatomy 0.000 claims abstract description 33
- 238000000034 method Methods 0.000 claims description 238
- 238000001914 filtration Methods 0.000 claims description 200
- 230000008569 process Effects 0.000 claims description 135
- 210000000988 bone and bone Anatomy 0.000 claims description 80
- 230000009467 reduction Effects 0.000 claims description 77
- 230000011664 signaling Effects 0.000 claims description 72
- 230000006854 communication Effects 0.000 claims description 60
- 238000004891 communication Methods 0.000 claims description 59
- 230000015654 memory Effects 0.000 claims description 46
- 238000002156 mixing Methods 0.000 claims description 43
- 230000005540 biological transmission Effects 0.000 claims description 35
- 230000008447 perception Effects 0.000 claims description 27
- 238000001228 spectrum Methods 0.000 claims description 17
- 206010011469 Crying Diseases 0.000 claims description 12
- 210000001260 vocal cord Anatomy 0.000 claims description 4
- 230000000694 effects Effects 0.000 abstract description 71
- 238000001514 detection method Methods 0.000 description 197
- 238000013461 design Methods 0.000 description 75
- 238000010586 diagram Methods 0.000 description 42
- 230000004044 response Effects 0.000 description 22
- 230000003321 amplification Effects 0.000 description 21
- 238000007726 management method Methods 0.000 description 21
- 238000003199 nucleic acid amplification method Methods 0.000 description 21
- 238000003860 storage Methods 0.000 description 20
- 238000004590 computer program Methods 0.000 description 18
- 238000013473 artificial intelligence Methods 0.000 description 16
- 230000008859 change Effects 0.000 description 11
- 238000010295 mobile communication Methods 0.000 description 11
- 238000005516 engineering process Methods 0.000 description 10
- 238000013507 mapping Methods 0.000 description 10
- 238000013528 artificial neural network Methods 0.000 description 8
- 238000004422 calculation algorithm Methods 0.000 description 8
- 230000003044 adaptive effect Effects 0.000 description 6
- 238000010079 rubber tapping Methods 0.000 description 6
- 230000009286 beneficial effect Effects 0.000 description 5
- 238000012986 modification Methods 0.000 description 5
- 230000004048 modification Effects 0.000 description 5
- 238000010183 spectrum analysis Methods 0.000 description 5
- 230000002159 abnormal effect Effects 0.000 description 4
- 208000016354 hearing loss disease Diseases 0.000 description 4
- 229920001621 AMOLED Polymers 0.000 description 3
- 206010011878 Deafness Diseases 0.000 description 3
- 208000006735 Periostitis Diseases 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 3
- 230000006978 adaptation Effects 0.000 description 3
- 230000007613 environmental effect Effects 0.000 description 3
- 230000010370 hearing loss Effects 0.000 description 3
- 231100000888 hearing loss Toxicity 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 210000003460 periosteum Anatomy 0.000 description 3
- 230000001960 triggered effect Effects 0.000 description 3
- 230000004913 activation Effects 0.000 description 2
- 230000003416 augmentation Effects 0.000 description 2
- 210000000721 basilar membrane Anatomy 0.000 description 2
- 230000036772 blood pressure Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000012790 confirmation Methods 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 210000005069 ears Anatomy 0.000 description 2
- 230000001976 improved effect Effects 0.000 description 2
- 230000000977 initiatory effect Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000037361 pathway Effects 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 239000002096 quantum dot Substances 0.000 description 2
- 230000005855 radiation Effects 0.000 description 2
- 238000011946 reduction process Methods 0.000 description 2
- 238000009877 rendering Methods 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 239000007787 solid Substances 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 230000001755 vocal effect Effects 0.000 description 2
- 101100006960 Caenorhabditis elegans let-2 gene Proteins 0.000 description 1
- 101100127891 Caenorhabditis elegans let-4 gene Proteins 0.000 description 1
- 230000003213 activating effect Effects 0.000 description 1
- 238000010009 beating Methods 0.000 description 1
- 230000007175 bidirectional communication Effects 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 238000013529 biological neural network Methods 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 210000003477 cochlea Anatomy 0.000 description 1
- 230000019771 cognition Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 230000003862 health status Effects 0.000 description 1
- 230000001939 inductive effect Effects 0.000 description 1
- 239000010985 leather Substances 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 230000000873 masking effect Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000006403 short-term memory Effects 0.000 description 1
- 230000008054 signal transmission Effects 0.000 description 1
- 230000003238 somatosensory effect Effects 0.000 description 1
- 230000006641 stabilisation Effects 0.000 description 1
- 238000011105 stabilization Methods 0.000 description 1
- 230000008093 supporting effect Effects 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1083—Reduction of ambient noise
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K11/00—Methods or devices for transmitting, conducting or directing sound in general; Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/16—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/175—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound
- G10K11/178—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase
- G10K11/1783—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase handling or detecting of non-standard events or conditions, e.g. changing operating modes under specific operating conditions
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K11/00—Methods or devices for transmitting, conducting or directing sound in general; Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/16—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/175—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound
- G10K11/178—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase
- G10K11/1783—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase handling or detecting of non-standard events or conditions, e.g. changing operating modes under specific operating conditions
- G10K11/17837—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase handling or detecting of non-standard events or conditions, e.g. changing operating modes under specific operating conditions by retaining part of the ambient acoustic environment, e.g. speech or alarm signals that the user needs to hear
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K11/00—Methods or devices for transmitting, conducting or directing sound in general; Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/16—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/175—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound
- G10K11/178—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase
- G10K11/1785—Methods, e.g. algorithms; Devices
- G10K11/17853—Methods, e.g. algorithms; Devices of the filter
- G10K11/17854—Methods, e.g. algorithms; Devices of the filter the filter being an adaptive filter
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K11/00—Methods or devices for transmitting, conducting or directing sound in general; Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/16—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/175—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound
- G10K11/178—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase
- G10K11/1787—General system configurations
- G10K11/17879—General system configurations using both a reference signal and an error signal
- G10K11/17881—General system configurations using both a reference signal and an error signal the reference signal being an acoustic signal, e.g. recorded with a microphone
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1041—Mechanical or electronic switches, or control elements
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K2210/00—Details of active noise control [ANC] covered by G10K11/178 but not provided for in any of its subgroups
- G10K2210/10—Applications
- G10K2210/108—Communication systems, e.g. where useful sound is kept and noise is cancelled
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K2210/00—Details of active noise control [ANC] covered by G10K11/178 but not provided for in any of its subgroups
- G10K2210/10—Applications
- G10K2210/108—Communication systems, e.g. where useful sound is kept and noise is cancelled
- G10K2210/1081—Earphones, e.g. for telephones, ear protectors or headsets
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K2210/00—Details of active noise control [ANC] covered by G10K11/178 but not provided for in any of its subgroups
- G10K2210/30—Means
- G10K2210/301—Computational
- G10K2210/3023—Estimation of noise, e.g. on error signals
- G10K2210/30231—Sources, e.g. identifying noisy processes or components
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K2210/00—Details of active noise control [ANC] covered by G10K11/178 but not provided for in any of its subgroups
- G10K2210/30—Means
- G10K2210/301—Computational
- G10K2210/3024—Expert systems, e.g. artificial intelligence
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K2210/00—Details of active noise control [ANC] covered by G10K11/178 but not provided for in any of its subgroups
- G10K2210/30—Means
- G10K2210/301—Computational
- G10K2210/3035—Models, e.g. of the acoustic system
- G10K2210/30351—Identification of the environment for applying appropriate model characteristics
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K2210/00—Details of active noise control [ANC] covered by G10K11/178 but not provided for in any of its subgroups
- G10K2210/30—Means
- G10K2210/301—Computational
- G10K2210/3039—Nonlinear, e.g. clipping, numerical truncation, thresholding or variable input and output gain
- G10K2210/30391—Resetting of the filter parameters or changing the algorithm according to prevailing conditions
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2460/00—Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
- H04R2460/01—Hearing devices using active noise cancellation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/50—Customised settings for obtaining desired overall acoustical characteristics
- H04R25/505—Customised settings for obtaining desired overall acoustical characteristics using digital signal processing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/70—Adaptation of deaf aid to hearing loss, e.g. initial electronic fitting
Definitions
- Embodiments of the present application relate to the technical field of audio processing, and in particular, to a method and device for processing noise in an earphone, and an earphone.
- headset users In recent years, there have been more and more headset users, and users' functional requirements for headsets have become more and more differentiated. For example, some users do not want to hear external noise when wearing headphones, and can use the active noise control (ANC) function to eliminate the noise in the ears. Some users want to hear the sound outside the earphone, and need to use the ambient sound through (hear through, HT) function to realize the same feeling of the outside sound as without the earphone. Some users may have hearing impairments, and the external signals that the user wants can be delivered to the user through the augmentation hearing (AH) function, and the unwanted signals can be filtered out at the same time.
- ANC active noise control
- HT ambient sound through
- HT ambient sound through
- Some users may have hearing impairments, and the external signals that the user wants can be delivered to the user through the augmentation hearing (AH) function, and the unwanted signals can be filtered out at the same time.
- AH augmentation hearing
- the headset cannot achieve the effect that the user wants to achieve according to the user's needs.
- Embodiments of the present application provide an earphone noise processing method, device, and earphone, so as to achieve the effect that the user wants to achieve according to the user's needs.
- an embodiment of the present application provides a noise processing method for an earphone, the earphone has at least two functions of an active noise reduction ANC function, an ambient sound transparent transmission HT function, or an auditory enhancement AH function, and the earphone includes a first microphone and a second microphone, the first microphone is used to collect the first signal, the first signal is used to characterize the sound of the current external environment, the second microphone is used to collect the second signal, and the second signal is used to characterize the wearing The sound of the environment inside the ear canal of the user of the headset.
- the earphone can be a left earphone or a right earphone.
- the processing modes used by the left earphone and the right earphone can be the same or different.
- the earphone receives the first audio signal from the terminal device; obtains a target mode; wherein, the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the earphone to implement the target processing function, so
- the target processing function is one of the active noise reduction ANC function, the ambient sound transparent transmission HT function or the auditory enhancement AH function; according to the target mode, the first audio signal, the first signal, and the second signal A second audio signal is obtained.
- the target mode is determined according to the scene type of the external environment of the field, and the user's listening effect can be optimized in real time.
- the earphone further includes a speaker, and the speaker is used to play the second audio signal.
- the target processing function is an ANC function
- the second audio signal played by the speaker can reduce the sound of the user's current environment and the sound inside the user's ear canal. Perception of ambient sound; or, the target processing function is an HT function, and the second audio signal played by the speaker can enhance the user's perception of the sound of the environment where the user is currently located; or, the target processing function
- the function is the AH function, and the second audio signal played by the speaker can enhance the user's perception of the event sound; the event sound satisfies a preset frequency spectrum.
- the audio signal played by the speaker of the left earphone can attenuate the sound of the user's left earphone to the current environment of the user (that is, the sound of the current external environment) and the sound inside the ear canal of the user's left earphone. Perception of ambient sound.
- the audio signal played by the speaker of the right earphone can reduce the user's right earphone's perception of the sound of the user's current environment (that is, the sound of the current external environment) and the ambient sound inside the ear canal of the user's right earphone.
- the feeling of the left ear follows the processing mode adopted by the left earphone
- the feeling of the right ear follows the processing mode adopted by the right earphone.
- the target processing function is the ANC function
- the second audio signal is obtained based on the first audio signal, the third signal and the fourth signal
- the third signal is the first signal.
- the fourth signal is an inverted signal of the second signal; or, when the target processing function is the HT function, the second audio signal is based on the first audio signal, the first signal and the second audio signal or, when the target processing function is the AH function, the second audio signal is obtained based on the first audio signal, the fifth signal and the fourth signal, and the fifth signal is one of the first audio signals.
- the event signal is used to represent a specific sound in the current external environment, and the event signal satisfies a preset frequency spectrum.
- the above design provides a simple and effective way of obtaining the signal output by the speaker under different processing modes.
- the acquiring target mode includes:
- a first control instruction from the terminal device is received, where the first control instruction carries the target mode, and the target mode is determined by the terminal device according to the scene type of the current external environment.
- the terminal device determines the target mode according to the scene type of the external environment and indicates to the earphone, which can optimize the user's listening effect in real time.
- a second control instruction from the terminal device, where the second control instruction carries a target processing strength, and the target processing strength is used to instruct the headset to perform the target processing function processing strength; obtaining a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal, including: according to the target mode, the target processing strength, the The first audio signal, the first signal, and the second signal result in a second audio signal.
- the terminal device indicates the processing intensity in the corresponding processing mode of the earphone. Adjusting the processing intensity on the basis of the processing mode further improves the user's hearing experience.
- the target event corresponding to the event sound in the current external environment is determined according to the first signal, and the target processing intensity in the target mode is determined according to the target event; wherein, the target processing intensity It is used to indicate the processing strength when the earphone realizes the target processing function; obtaining a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal, including: A second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal. where different processing intensities correspond to different events.
- the processing intensity can be in one-to-one correspondence with events, or one processing intensity corresponds to multiple events. For example, two events can use the same processing intensity, but different processing intensity cannot be used for the same event.
- the earphone determines the processing intensity according to the event sound in the external environment, realizes different auditory experience in different external environments, can reduce the sense of bottom noise and enhance the noise reduction.
- the headset further includes a bone conduction sensor, and the bone conduction sensor is used to collect bone conduction signals generated by the vibration of the user's vocal cords;
- the first scene at the location includes: identifying the first scene where the user is currently located according to the first signal and the bone conduction signal.
- the target event is a howling event, a wind noise event, a sudden event or a human voice event.
- the acquiring the target mode includes: identifying, according to the first signal, a scene type of the current external environment as a target scene type (referred to as a target scene, or simply referred to as a target type), and according to the target
- the scene determines the target mode adopted by the headset, and the target mode is a processing mode corresponding to the target scene.
- the different processing modes correspond to different scene types.
- the processing modes may be in one-to-one correspondence with scene types, or one processing mode may correspond to multiple scene types, for example, two scene types may adopt the same processing mode.
- the earphone determines the processing mode adopted by the earphone according to the recognized scene type, reduces the time delay, and optimizes the user's hearing experience in real time.
- the target scene is a walking scene, a running scene, a quiet scene, a multi-person talking scene, a cafe scene, a subway scene, a train scene, a waiting room scene, a dialogue scene, an office scene, an outdoor scene, One of the driving scene, windy scene, airplane scene, siren scene, whistle scene, and crying scene.
- the audio signal includes: obtaining a second audio signal according to the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
- the earphone determines the processing mode to be adopted, and instructs the terminal device to adjust the processing intensity and reduce the occupation of processing resources of the earphone.
- the target processing function is an ANC function
- the greater the target processing intensity is, the more the user perceives the sound of the current environment of the user and the environment inside the user's ear canal The weaker the sound; or, when the target processing function is the HT function, the greater the target processing intensity, the greater the intensity of the sound of the environment where the user is currently perceived by the user; or, when the target When the processing function is the AH function, the greater the target processing intensity, the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- the target mode instructs the headset to implement an ANC function, and obtaining a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal, including:
- the second audio signal is obtained by performing mixing processing on the fourth audio signal and the first audio signal.
- FF filtering and FB serial processing are used to implement ANC processing to obtain better noise-reduced signals and enhance the noise-reduction effect.
- the filter coefficient used in the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the ANC function; or, the third The filter coefficient used in the filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the ANC function.
- the target mode instructs the earphone to implement the HT function
- the second audio signal is obtained according to the target mode, the first audio signal, the first signal, and the second signal, including:
- the first signal processing including a second filtering process (such as HT filtering);
- a third filtering process (such as FB filtering) on the second filtered signal to obtain a third filtered signal
- the second audio signal is obtained by performing mixing processing on the third filtered signal and the fifth audio signal.
- filter compensation processing may be performed on the fifth audio signal to reduce hearing loss.
- the HT filter undergoes down-mixing processing and filter compensation processing to further reduce hearing loss.
- performing first signal processing on the first environmental signal to obtain a processed environmental signal includes: performing a second filtering process on the first signal to obtain a second filtered signal;
- the second signal processing includes de-blocking effect processing.
- the second signal processing further includes at least one of the following: noise reduction processing, wind noise reduction processing, gain adjustment processing or frequency response adjustment processing.
- the noise floor and abnormal sounds are reduced, and the user's hearing experience is improved.
- the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the HT function; or,
- the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the HT function.
- the target mode instructs the headset to implement the AH function
- the second audio signal is obtained according to the target mode, the first audio signal, the first signal, and the second signal, including:
- Second filtering processing (such as HT filtering) on the first signal to obtain a second filtering signal
- enhancement processing on the second filtering signal to obtain a filtering enhanced signal
- Performing a first filtering process (such as FF filtering) on the first signal to obtain a first filtered signal
- the second audio signal is obtained by mixing the fifth filtered signal, the sixth audio signal and the first filtered signal.
- the transparently transmitted signal is processed by transparent transmission filtering and enhancement processing, and the transparently transmitted signal is clearer.
- filtering and compensating the sixth audio signal can avoid losses caused by FB filtering and ensure transparency to the greatest extent. Signal transmission without distortion.
- performing enhancement processing on the second filtered signal to obtain a filtered enhanced signal including:
- noise reduction processing includes artificial intelligence AI noise reduction processing and/or wind noise reduction processing
- the filtered enhanced signal is obtained by performing gain amplification processing and frequency response adjustment on the signal obtained after the noise reduction processing.
- the transparently transmitted signal is enhanced. Improve the user's sense of hearing for desired external sounds.
- the headset includes a bone conduction sensor, the bone conduction sensor is used to collect the bone conduction signal of the headset user, and the signal obtained after noise reduction processing is subjected to gain amplification processing, including: : performing harmonic expansion on the bone conduction signal to obtain a harmonically expanded signal; using the first gain coefficient to amplify the signal obtained by the noise reduction process; using the fourth filter coefficient to filter out the signal obtained by the amplification process The harmonically extended signal included in ; wherein the fourth filter coefficient is determined based on the first gain coefficient.
- an amplification method which only amplifies the specific sounds other than the voice of the wearing user, so as to improve the effect of the specific sounds in the transparently transmitted ambient sound.
- the first gain coefficient is a gain coefficient associated with the target processing intensity in the target mode.
- performing enhancement processing on the second filtered signal to obtain a filtered enhanced signal including:
- the audio event signal in the deblocked signal is subjected to gain amplification processing and frequency response adjustment to obtain a filtered enhanced signal.
- the headset further includes a bone conduction sensor, the bone conduction sensor is used to collect the bone conduction signal of the headset user, and perform gain amplification on the audio event signal in the de-occluded signal processing, including: performing harmonic expansion on the bone conduction signal to obtain a harmonically extended signal; amplifying an audio event signal in the de-occluded signal using a second gain coefficient to obtain an amplified signal; using a second filter coefficient Filtering out the harmonically extended signal included in the amplified signal; wherein the second filter coefficient is determined based on the second gain coefficient.
- the bone conduction sensor is used to collect the bone conduction signal of the headset user, and perform gain amplification on the audio event signal in the de-occluded signal processing, including: performing harmonic expansion on the bone conduction signal to obtain a harmonically extended signal; amplifying an audio event signal in the de-occluded signal using a second gain coefficient to obtain an amplified signal; using a second filter coefficient Filtering out the harmonically extended signal included in the amplified signal; where
- the second gain coefficient is a gain coefficient associated with the target processing intensity for the first filtering processing when the first noise processing is performed.
- the second gain coefficient is a gain coefficient associated with the first scene identifier for the first filtering process when the first noise processing is performed.
- the filter coefficient used by the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the AH function; or,
- the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the AH function; or,
- the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the AH function.
- the earphone further includes a bone conduction sensor, and the bone conduction sensor is used to collect the bone conduction signal of the user of the earphone; performing de-occlusion effect processing on the second filtered signal, including:
- the low-frequency component in the signal obtains a third filtered signal, and the high-frequency component in the third filtered signal from which the low-frequency component is removed is amplified.
- an embodiment of the present invention provides a mode control method, the method is applied to a terminal device, and the method includes: when a scene type of the current external environment is identified as a target scene, determining a target mode according to the target scene; Wherein, the target mode is one of the processing modes supported by the headset, and the processing modes supported by the headset include at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode, or the hearing enhancement AH mode; The earphone sends the target mode, where the target mode is used to instruct the earphone to implement the processing function corresponding to the target mode.
- Different processing modes correspond to different scene types, and processing modes may be in one-to-one correspondence with scene types, or one processing mode may correspond to multiple scene types, for example, two scene types may adopt the same processing mode.
- the terminal device controls the processing mode of the headset in real time according to the scene recognition, so as to optimize the user's hearing experience in real time.
- the method when determining the target mode corresponding to the target scene in the processing mode of the headset, the method further includes: displaying result prompt information, where the result prompt information is used to prompt the user that the headset implements the desired Describe the processing function corresponding to the target mode.
- the above design enables the user to determine the current processing mode of the headset in real time.
- the method before sending the first control signaling to the headset, the method further includes: displaying selection prompt information, where the selection prompt information is used to prompt the user whether to adjust the processing mode of the headset to any The target mode is detected; an operation that the user selects to adjust the processing mode of the headset to the target mode is detected.
- the user can determine whether to adjust the processing mode of the earphone according to the needs, so as to improve the user experience.
- a first control and a second control are displayed, wherein different positions of the second control on the first control are used to indicate different processing intensities in the target mode;
- the method further includes: in response to the user touching the second control to move to a first position on the first control, the second control is on the first control
- the first position of the device indicates the target processing strength in the target mode; the target processing strength is sent to the headset, where the target processing strength is used to indicate the processing when the headset implements the processing function corresponding to the target mode strength.
- the user can select the processing intensity of the earphone according to the needs, so as to meet the different needs of the user.
- the shape of the first control is a ring
- the processing intensity in the target mode is Change from small to large
- the shape of the first control is a bar
- the target processing function when the target processing function is an ANC function, the greater the target processing intensity is, the more the user perceives the sound of the current environment of the user and the environment inside the user's ear canal The weaker the sound.
- the target processing function when the target processing function is the HT function, the greater the target processing intensity, the greater the intensity of the sound of the environment where the user is currently located; or, when the target processing function is AH
- the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- the left earphone and the right earphone may adopt the same processing mode and processing intensity, and the perception of the user's left ear and right ear may be the same.
- the left earphone and the right earphone may also adopt different processing modes or different processing intensities, so that the perception of the left ear and the right ear is different.
- an embodiment of the present invention provides a mode control method, the method is applied to a terminal device, and the method includes: acquiring a target mode; the target mode is one of processing modes supported by a headset, and the headset supports The processing mode includes at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode or the auditory enhancement AH mode; the target processing intensity in the target mode is determined according to the scene type of the current external environment; different scene types correspond to different processing intensities in the target mode; sending the target processing intensities to the earphone, where the target processing intensities are used to indicate the processing intensities when the earphones implement the processing functions corresponding to the target modes.
- the acquiring the target mode includes: receiving the target mode sent by the headset; or, displaying a selection control, where the selection control includes a processing mode supported by the headset, and it is detected that the user has passed the A selection control selects the operation of the target mode among the processing modes of the headset.
- the selection control includes the processing mode supported by the headset, or is interpreted as an option for the selection control to provide the processing mode supported by the headset, or the selection control displays the processing mode supported by the headset, and the user can perform in the processing mode supported by the headset. choose.
- the method before determining the target processing intensity in the target mode according to the scene type of the current external environment, the method further includes: when the target mode sent by the headset is received, displaying a selection Prompt information, the selection prompt information is used to indicate whether the user adjusts the processing mode of the headset to the target mode; an operation of the user selecting to adjust the processing mode of the headset to the target mode is detected.
- the target processing function is an ANC function
- the target processing function is the HT function
- the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
- the target processing function is the AH function
- the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- an embodiment of the present application provides a mode control method, the method is applied to a terminal device, the method includes: displaying a first interface; the first interface includes a first selection control, and the first selection control includes The processing mode supported by the first target earphone and the processing intensity corresponding to the processing mode supported by the first target earphone; the processing mode of the first target earphone includes active noise reduction ANC mode, ambient sound transparent transmission HT mode or hearing enhancement AH mode at least two kinds of; in response to a first operation performed by the user on the first interface; the first operation is that the user selects the first target mode from the processing modes supported by the first target headset through the first selection control; and The processing intensity in the first target mode is selected to be generated by the first target processing intensity; the first target mode and the first target processing intensity are sent to the first target earphone, and the first target mode is used to indicate The first target earphone implements the processing function corresponding to the first target mode, and the first target processing intensity is used to indicate the processing when the first target ear
- the first selection control includes the processing mode supported by the first target headset and the processing intensity corresponding to the processing mode supported by the first target headset, which can be interpreted as the first selection control provides the user with multiple processing modes (all are the first target headset. supported) options and adjustments to the processing strength for each processing mode.
- the user can freely switch the processing mode and intensity corresponding to the effect that the earphone needs to achieve through the UI interface, so as to meet the different needs of the user.
- the method before displaying the first interface, the method further includes: displaying selection prompt information, where the selection prompt information is used for the user to select whether to adjust the processing mode of the first target headset; detecting The user selects an operation to adjust the processing mode of the first target headset.
- the method before displaying the first interface, the method further includes: identifying the scene type of the current external environment as a target scene, and the target scene adaptation needs to adjust the processing mode of the first target headset scene type.
- a method is provided to actively pop up the first interface in a specific scenario, so as to reduce the manual operation process of the user.
- the method before displaying the first interface, the method further includes: recognizing that the terminal device triggers the first target headset to play audio. Recognizing that the terminal device triggers the first target earphone to play audio may be interpreted as recognizing that the terminal device starts sending an audio signal to the first target earphone.
- the method before displaying the first interface, the method further includes: detecting that the terminal device establishes a connection with the first target headset.
- the method before displaying the first interface, further includes: in the case of detecting that the terminal device establishes a connection with the first target headset, detecting that the user executes the operation on the main interface the second operation; wherein the main interface includes an icon of a first application, the second operation is generated by the user touching the icon of the first application, and the first interface is the first application display interface.
- the first selection control includes a first control and a second control
- the second control indicates two different processing modes of the first target earphone at any two different positions of the first control
- the second control indicates different processing intensities in the same processing mode of the first target headset at any two different positions of the first control
- the first operation is that the user moves the second control It is generated by a first position in the area corresponding to the first target mode on the first control, where the first position corresponds to the first target processing intensity in the first target mode.
- the shape of the first control is a ring or a bar.
- the shape of the first control is a ring
- the ring includes at least two arc segments
- the second control is located in different arc segments to indicate different processing modes of the first target headset
- the second The different positions of the controls in the same arc segment indicate different processing intensities of the same processing mode of the first target earphone.
- the shape of the first control is a bar
- the bar includes at least two bar segments
- the second controls are located in different bar segments to indicate different processing modes of the first target earphone, so Different positions of the second control in the same bar segment indicate different processing intensities of the same processing mode of the first target earphone.
- the method further includes:
- the first interface further includes a second selection control
- the second selection control includes a processing mode supported by the second target headset and a processing mode supported by the second target headset
- the processing modes supported by the first target headset include at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode, or the hearing enhancement AH mode
- the third operation is that the user
- the second selection control selects the second target mode in the processing mode of the second target earphone and selects the processing intensity in the second target mode to be generated by the second target processing intensity; when the first target earphone is the left earphone, the The second target earphone is a right earphone, or the first target earphone is a right earphone, and the second target earphone is a left earphone;
- the second target processing intensity is used to indicate the processing intensity when the second target earphone implements the processing function corresponding to the second target mode.
- the user can operate the processing mode and processing intensity of the left earphone and the right earphone separately, so as to meet the user's differentiated requirements for the hearing sense of the left ear and the right ear.
- an embodiment of the present application further provides a noise processing device, the device is applied to an earphone, and the earphone has at least two functions of an active noise reduction ANC function, an ambient sound transparent transmission HT function, or an auditory enhancement AH function;
- the headset includes a first microphone and a second microphone; the first microphone is used to collect a first signal, and the first signal is used to represent the sound of the current external environment; the second microphone is used to collect the second signal, The second signal is used to characterize the sound of the internal environment of the ear canal of the user wearing the headset.
- the noise processing apparatus includes corresponding functional modules, which are respectively used to implement the steps in the method of the first aspect above. For details, please refer to the detailed description in the method example, which will not be repeated here.
- the functions can be implemented by hardware, or by executing corresponding software by hardware.
- the hardware or software includes one or more modules corresponding to the above functions.
- the noise processing device includes:
- a communication module for receiving the first audio signal from the terminal device
- an acquisition module configured to acquire a target mode; wherein, the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the headset to implement a target processing function, and the target processing function is active noise reduction One of ANC function, ambient sound transparent transmission HT function or auditory enhancement AH function;
- a first processing module configured to obtain a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal.
- an embodiment of the present application provides a target earphone, including a left earphone and a right earphone, where the left earphone is used to implement the method described in the first aspect or any design of the first aspect, or the right earphone is used to implement the first aspect.
- the left and right earphones use different processing modes.
- an embodiment of the present application provides a target earphone, where the target earphone includes a left earphone and a right earphone.
- the left earphone includes or the right earphone includes a first microphone, a second microphone, a processor, a memory, and a speaker;
- the first microphone is used to collect a first signal, and the first signal is used to represent the sound of the current external environment ;
- the second microphone is used to collect a second signal, and the second signal is used to characterize the sound of the internal environment of the ear canal of the user wearing the headset;
- the memory is used to store programs or instructions;
- the processor used to invoke the program or instruction, so that the electronic device executes the method described in any design of the first aspect to obtain a second audio signal;
- the speaker is used to play the second audio signal.
- an embodiment of the present application provides a mode control apparatus, where the apparatus is applied to a terminal device.
- the apparatus includes corresponding functional modules, which are respectively used to implement the steps in the methods of the second aspect to the fourth aspect.
- the functions can be implemented by hardware, or by executing corresponding software by hardware.
- the hardware or software includes one or more modules corresponding to the above functions.
- an embodiment of the present application provides a terminal device, including a memory, a processor, and a display; the display is used to display an interface; the memory is used to store programs or instructions; the processor is used to call The program or instruction is used to cause the terminal device to execute the steps in the method of the second aspect - the fourth aspect.
- the present application provides a computer-readable storage medium, where computer programs or instructions are stored in the computer-readable storage medium, and when the computer program or instructions are executed by the earphone, the earphone is made to perform the above-mentioned first aspect or the first aspect. methods in any possible design.
- the present application provides a computer-readable storage medium.
- the computer-readable storage medium stores a computer program or instruction.
- the earphone is made to perform the above-mentioned second aspect-fourth aspect. method in any possible design of the aspect.
- the present application provides a computer program product, the computer program product includes a computer program or an instruction, when the computer program or instruction is executed by the headset, the above-mentioned first aspect or any possible implementation manner of the first aspect is realized method in .
- the present application provides a computer program product, the computer program product includes a computer program or an instruction, when the computer program or instruction is executed by a headset, any possible implementation manner of the above-mentioned second aspect to the fourth aspect is realized method in .
- FIG. 1 is a schematic diagram of a hardware structure of a terminal device 100 in an embodiment of the present application
- FIG. 2 is a schematic diagram of a software structure of a terminal device 100 in an embodiment of the present application
- FIG. 3 is a schematic structural diagram of an earphone 200 in an embodiment of the present application.
- FIG. 4 is a schematic diagram of the AHA pathway in the embodiment of the present application.
- 5A is a flowchart of ANC processing in an embodiment of the present application.
- 5B is a schematic diagram of an ANC processing flow in an embodiment of the present application.
- 6A is a flowchart of HT processing in an embodiment of the present application.
- 6B is a schematic diagram of a HT processing flow in an embodiment of the present application.
- 6C is a schematic diagram of another HT processing flow in an embodiment of the present application.
- FIG. 7 is a schematic diagram of a process flow of de-occlusion effect processing in an embodiment of the present application.
- FIG. 8A is a flowchart of AH processing in an embodiment of the present application.
- FIG. 8B is a schematic diagram of an AH processing flow in an embodiment of the present application.
- 8C is a schematic diagram of another AH processing flow in an embodiment of the present application.
- FIG. 9 is a schematic diagram of a noise reduction processing flow diagram in an embodiment of the present application.
- FIG. 10 is a schematic diagram of a process flow of gain amplification in an embodiment of the present application.
- FIG. 11 is a schematic diagram of another gain amplification processing flow in an embodiment of the present application.
- 12A is a schematic diagram of a main interface of a terminal device in an embodiment of the present application.
- 12B is a schematic diagram of a control interface of an earphone application in an embodiment of the present application.
- 12C is a schematic diagram of the control of a terminal device controlling a headset in an ANC mode in an embodiment of the present application
- 12D is a schematic diagram of the control of a terminal device controlling an earphone in an HT mode in an embodiment of the present application
- 12E is a schematic diagram of the control of a terminal device controlling an earphone in an AH mode in an embodiment of the present application
- 12F is a schematic diagram of a selection control in an embodiment of the present application.
- 12G is a schematic diagram of another selection control in an embodiment of the present application.
- 12H is a schematic diagram of triggering a control interface of an earphone in an embodiment of the present application.
- FIG. 13 is a schematic diagram of another selection control in an embodiment of the application.
- FIG. 14A is a schematic diagram of control for enabling a smart scene detection function according to an embodiment of the present application.
- FIG. 14B is a schematic diagram of another kind of smart scene detection function enabling control in an embodiment of the present application.
- 14C is a schematic diagram of an earphone control interface in an embodiment of the present application.
- 15 is a schematic diagram of event detection in an embodiment of the present application.
- 16 is a schematic diagram of an interaction processing mode and processing intensity between a terminal device and a headset in an embodiment of the application;
- FIG. 17A is a schematic diagram of displaying a scene detection result in an embodiment of the present application.
- FIG. 17B is a schematic diagram of another scene detection result display in an embodiment of the present application.
- FIG. 19 is a schematic structural diagram of a noise processing apparatus 1900 in an embodiment of the present application.
- FIG. 20 is a schematic structural diagram of a mode control apparatus 2000 in an embodiment of the present application.
- FIG. 21 is a schematic structural diagram of a mode control device 2100 in an embodiment of the present application.
- FIG. 22 is a schematic structural diagram of a mode control apparatus 2200 in an embodiment of the present application.
- FIG. 23 is a schematic structural diagram of a terminal device 2300 in an embodiment of the present application.
- the application (application, app) involved in the embodiments of the present application is a software program capable of implementing one or more specific functions.
- multiple applications can be installed in a terminal device.
- the applications mentioned below may be system applications that have been installed when the terminal device is shipped from the factory, or may be third-party applications downloaded from the network or obtained from other terminal devices by the user during the use of the terminal device.
- the human auditory system has a masking effect, that is, strong-frequency sounds will hinder the perception of human alignment of nearby weak-frequency sounds, and the basilar membrane of the cochlea has a frequency selection and tuning effect on external sound signals. Therefore, the concept of critical frequency band is introduced, from Perceptually measures sound frequency. It is generally believed that there are 24 critical frequency bands within the hearing threshold of 22 Hz to 22 kHz, which can cause vibrations at different positions on the basilar membrane. Each critical band is called a bark subband.
- VAD Voice endpoint detection
- At least one (item) refers to one (item) or more (item), and “multiple (item)” refers to two (item) or more than two (item).
- “And/or”, which describes the relationship of the associated objects, indicates that there can be three kinds of relationships, for example, A and/or B, it can indicate that A exists alone, A and B exist at the same time, and B exists alone, where A, B can be singular or plural.
- the character “/” generally indicates that the associated objects are an “or” relationship.
- At least one item(s) below” or similar expressions thereof refer to any combination of these items, including any combination of single item(s) or plural items(s).
- At least one (a) of a, b or c may represent: a, b, c, a-b, a-c, b-c or a-b-c, where a, b, c may be single or multiple.
- the symbol "(a, b)" represents an open interval, and the range is greater than a and less than b; "[a, b]” represents a closed interval, and the range is greater than or equal to a and less than or equal to b; "(a ,b]” represents a half-open and half-closed interval with a range greater than a and less than or equal to b; "(a,b]” represents a half-open and half-closed interval with a range greater than a and less than or equal to b.
- first and second mentioned in the embodiments of this application are used to distinguish multiple objects, and are not used to limit the size, content, order, timing, priority or importance of multiple objects. degree, etc.
- first microphone and the second microphone are only used to distinguish different microphones, and do not indicate the difference in size, priority, or importance of the two microphones.
- An embodiment of the present application provides a system, where the system includes a terminal device 100 and an earphone 200 .
- the terminal device 100 is connected 200 to the earphone, and the connection may be a wireless connection or a wired connection.
- the terminal device may be connected to the headset through Bluetooth technology, wireless fidelity (Wi-Fi) technology, infrared IR technology, and ultra-wideband technology.
- the terminal device 100 is a device having a display interface function.
- the terminal device 100 may be, for example, a product with a display interface such as a mobile phone, a monitor, a tablet computer, and a vehicle-mounted device, as well as a smart display wearable product such as a smart watch and a smart bracelet.
- a product with a display interface such as a mobile phone, a monitor, a tablet computer, and a vehicle-mounted device, as well as a smart display wearable product such as a smart watch and a smart bracelet.
- a smart display wearable product such as a smart watch and a smart bracelet.
- the specific form of the above mobile terminal is not particularly limited in this embodiment of the present application.
- the earphone 200 includes two sounding units hanging around the ears.
- the one that fits the left ear can be called the left earphone, and the one that fits the right ear can be called the right earphone.
- the earphone 200 in this embodiment of the present application may be a headphone, an ear-hook earphone, a neck-hung earphone, or an earbud earphone, or the like.
- In-ear headphones also include in-ear headphones (or referred to as ear canal headphones) or semi-in-ear headphones.
- the earphone 200 has at least two of the ANC function, the HT function, or the AH function.
- AHA the convenience of description, ANC, HT, and AH are collectively referred to as AHA in this embodiment of the present application, and of course, other names may also be ordered, which is not limited in this application.
- the left earphone and the right earphone have similar structures. Either the left earphone or the right earphone can adopt the earphone structure described below.
- the earphone structure (left earphone or right earphone) includes a rubber sleeve that can be inserted into the ear canal, an ear bag close to the ear, and an earphone rod suspended on the ear bag. The rubber sleeve guides the sound to the ear canal.
- the ear bag includes devices such as batteries, speakers, and sensors. Microphones and physical buttons can be arranged on the headphone rod.
- the earphone rod can be in the shape of a cylinder, a rectangular parallelepiped, an ellipsoid, or the like.
- the microphone arranged inside the ear can be called the error microphone, and the microphone arranged outside the earphone is called the reference microphone.
- the error microphone is used to pick up the sound of the external environment. Referring to the microphone, when the user wears the earphone, the sound of the internal environment of the ear canal of the user wearing the earphone is collected.
- the two microphones can be either analog microphones or digital microphones. After the user wears the earphone, the positional relationship between the two microphones and the speaker is as follows: the error microphone is inside the ear, close to the earphone rubber sleeve.
- the loudspeaker is located between the error microphone and the reference microphone.
- the reference microphone is close to the outer structure of the ear and can be arranged on the upper part of the earphone stem.
- the tube of the error microphone can either face the loudspeaker or the inside of the ear canal.
- the terminal device 100 is configured to send downlink audio signals and/or control signaling to the headset 200 .
- the control signaling is used to control the processing mode adopted by the headset 200 .
- the processing mode adopted by the headset 200 may include at least two of an empty mode indicating no processing, an ANC mode indicating implementing an ANC function, an HT mode indicating implementing an HT function, or an AH mode indicating implementing an AH function.
- the earphone user's perception of the sound of the current external environment and the ambient sound inside the ear canal of the user wearing the earphone can be weakened.
- the headset adopts the HT mode the user's perception of the sound of the current external environment can be enhanced.
- the headset adopts the AH mode the user's perception of the event sound included in the sound of the current external environment can be enhanced.
- the event sound refers to a preset sound in an external environment, or the event sound satisfies a preset frequency spectrum.
- the event sound includes a station announcement sound or a whistle sound in a railway station; then the event sound satisfies the frequency spectrum of the station announcement sound or the frequency spectrum of the whistle sound in the railway station.
- the event sound may include a notification sound in an airplane terminal, a broadcast sound on an airplane, etc.; another example is the sound of a hotel calling number, and the like.
- the earphone 200 includes a left earphone and a right earphone, and the left earphone and the right earphone may adopt the same processing mode or different processing modes.
- the left earphone and the right earphone use the same processing mode, the user's left ear wearing the left earphone and the right ear wearing the right earphone may have the same auditory perception.
- the left earphone and the right earphone adopt different processing modes, the user's left ear wearing the left earphone and the right ear wearing the right earphone have different auditory perceptions.
- the left earphone using ANC and the right earphone using AH as an example, when the left earphone adopts the ANC mode, the sound of the earphone user's left ear to the current external environment and the environmental sound inside the ear canal of the user wearing the earphone can be reduced. perception.
- the right earphone adopts the AH mode the perception of the user's right ear to the event sound included in the sound of the current external environment can be enhanced.
- any one of the following possible ways can be used to achieve it;
- the terminal device 100 provides a control interface for the user to select the processing mode of the headset 200 according to requirements. For example, under the operation of the user, the terminal device 100 sends control signaling to the headset 200 , and the control signaling is used to indicate the processing mode adopted by the headset 200 .
- the processing modes adopted by the left earphone and the right earphone in the earphone 200 may be the same or different.
- a selection control on the control surface is used to select the same processing mode for the left and right earphones.
- the control interface may include two selection controls, wherein one selection control is used to select a processing mode for the left earphone, and the other selection control is used to select a processing mode for the right earphone. The control interface and selection controls will be described in detail later, and the description will not be repeated here.
- the terminal device identifies the scene type of the user's current external environment.
- the processing modes adopted by the headset 200 are different, that is, the processing functions implemented by the headset are different.
- the earphone 200 recognizes the user's operation, and determines that the earphone 200 selected by the user adopts the ANC mode, the HT mode or the AH mode.
- the user's operation may be an operation of the user tapping the earphone, or buttons are provided on the earphone, and different buttons indicate different processing modes.
- the earphone recognizes the scene type of the external environment of the earphone, and the processing mode adopted by the earphone is different in different scenes.
- FIG. 1 shows a schematic diagram of an optional hardware structure of the terminal device 100 .
- the terminal device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2 , mobile communication module 150, wireless communication module 160, audio module 170, speaker 170A, receiver 170B, microphone 170C, headphone jack 170D, sensor module 180, buttons 190, motor 191, indicator 192, camera 193, display screen 194, and Subscriber identification module (subscriber identification module, SIM) card interface 195 and so on.
- SIM Subscriber identification module
- the sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, and ambient light. Sensor 180L, bone conduction sensor 180M, etc.
- the structures illustrated in the embodiments of the present invention do not constitute a specific limitation on the terminal device 100 .
- the terminal device 100 may include more or less components than those shown in the drawings, or combine some components, or separate some components, or arrange different components.
- the illustrated components may be implemented in hardware, software or a combination of software and hardware.
- the processor 110 may include one or more processing units, for example, the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processor (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), controller, video codec, digital signal processor (digital signal processor, DSP), baseband processor, and/or neural-network processing unit (neural-network processing unit, NPU), etc. Wherein, different processing units may be independent devices, or may be integrated in one or more processors.
- application processor application processor, AP
- modem processor graphics processor
- ISP image signal processor
- controller video codec
- digital signal processor digital signal processor
- baseband processor baseband processor
- neural-network processing unit neural-network processing unit
- the controller can generate an operation control signal according to the instruction operation code and timing signal, and complete the control of fetching and executing instructions.
- a memory may also be provided in the processor 110 for storing instructions and data.
- the memory in processor 110 is cache memory. This memory may hold instructions or data that have just been used or recycled by the processor 110 . If the processor 110 needs to use the instruction or data again, it can be called directly from the memory. Repeated accesses are avoided and the latency of the processor 110 is reduced, thereby increasing the efficiency of the system.
- the processor 110 may include one or more interfaces.
- the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, a universal asynchronous transceiver (universal asynchronous transmitter) receiver/transmitter, UART) interface, mobile industry processor interface (MIPI), general-purpose input/output (GPIO) interface, subscriber identity module (SIM) interface, and / or universal serial bus (universal serial bus, USB) interface, etc.
- I2C integrated circuit
- I2S integrated circuit built-in audio
- PCM pulse code modulation
- PCM pulse code modulation
- UART universal asynchronous transceiver
- MIPI mobile industry processor interface
- GPIO general-purpose input/output
- SIM subscriber identity module
- USB universal serial bus
- the I2C interface is a bidirectional synchronous serial bus that includes a serial data line (SDA) and a serial clock line (SCL).
- the processor 110 may contain multiple sets of I2C buses.
- the processor 110 can be respectively coupled to the touch sensor 180K, the charger, the flash, the camera 193 and the like through different I2C bus interfaces.
- the processor 110 may couple the touch sensor 180K through the I2C interface, so that the processor 110 and the touch sensor 180K communicate with each other through the I2C bus interface, so as to realize the touch function of the terminal device 100 .
- the I2S interface can be used for audio communication.
- the processor 110 may contain multiple sets of I2S buses.
- the processor 110 may be coupled with the audio module 170 through an I2S bus to implement communication between the processor 110 and the audio module 170 .
- the audio module 170 can transmit an audio signal to the wireless communication module 160 through the I2S interface, so as to realize the function of answering a call through the headset 200 (such as a Bluetooth headset).
- the PCM interface can also be used for audio communications, sampling, quantizing and encoding analog signals.
- the audio module 170 and the wireless communication module 160 may be coupled through a PCM bus interface.
- the audio module 170 can also transmit audio signals to the wireless communication module 160 through the PCM interface, so as to realize the function of answering a call through the Bluetooth headset 200 .
- Both the I2S interface and the PCM interface can be used for audio communication.
- the UART interface is a universal serial data bus used for asynchronous communication.
- the bus may be a bidirectional communication bus. It converts the data to be transmitted between serial communication and parallel communication.
- a UART interface is typically used to connect the processor 110 with the wireless communication module 160 .
- the processor 110 communicates with the Bluetooth module in the wireless communication module 160 through the UART interface to implement the Bluetooth function.
- the audio module 170 can transmit audio signals to the wireless communication module 160 through the UART interface, so as to realize the function of playing music through the Bluetooth headset 200 .
- the MIPI interface can be used to connect the processor 110 with peripheral devices such as the display screen 194 and the camera 193 .
- MIPI interfaces include camera serial interface (CSI), display serial interface (DSI), etc.
- the processor 110 communicates with the camera 193 through the CSI interface, so as to realize the shooting function of the terminal device 100 .
- the processor 110 communicates with the display screen 194 through the DSI interface to implement the display function of the terminal device 100 .
- the GPIO interface can be configured by software.
- the GPIO interface can be configured as a control signal or as a data signal.
- the GPIO interface may be used to connect the processor 110 with the camera 193, the display screen 194, the wireless communication module 160, the audio module 170, the sensor module 180, and the like.
- the GPIO interface can also be configured as I2C interface, I2S interface, UART interface, MIPI interface, etc.
- the USB interface 130 is an interface that conforms to the USB standard specification, and may specifically be a Mini USB interface, a Micro USB interface, a USB Type C interface, and the like.
- the USB interface 130 can be used to connect a charger to charge the terminal device 100, and can also be used to transmit data between the terminal device 100 and peripheral devices. It can also be used to connect the headset 200 to play audio through the headset 200 .
- This interface can also be used to connect other terminal devices, such as AR devices.
- the interface connection relationship between the modules illustrated in the embodiment of the present invention is only a schematic illustration, and does not constitute a structural limitation of the terminal device 100 .
- the terminal device 100 may also adopt different interface connection manners in the foregoing embodiments, or a combination of multiple interface connection manners.
- the charging management module 140 is used to receive charging input from the charger.
- the charger may be a wireless charger or a wired charger.
- the charging management module 140 may receive charging input from the wired charger through the USB interface 130 .
- the charging management module 140 may receive wireless charging input through the wireless charging coil of the terminal device 100 . While the charging management module 140 charges the battery 142 , it can also supply power to the terminal device through the power management module 141 .
- the power management module 141 is used for connecting the battery 142 , the charging management module 140 and the processor 110 .
- the power management module 141 receives input from the battery 142 and/or the charging management module 140, and supplies power to the processor 110, the internal memory 121, the display screen 194, the camera 193, and the wireless communication module 160.
- the power management module 141 can also be used to monitor parameters such as battery capacity, battery cycle times, battery health status (leakage, impedance).
- the power management module 141 may also be provided in the processor 110 .
- the power management module 141 and the charging management module 140 may also be provided in the same device.
- the wireless communication function of the terminal device 100 may be implemented by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, the modulation and demodulation processor, the baseband processor, and the like.
- Antenna 1 and Antenna 2 are used to transmit and receive electromagnetic wave signals.
- Each antenna in terminal device 100 may be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve antenna utilization.
- the antenna 1 can be multiplexed as a diversity antenna of the wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
- the mobile communication module 150 may provide a wireless communication solution including 2G/3G/4G/5G, etc. applied on the terminal device 100 .
- the mobile communication module 150 may include at least one filter, switch, power amplifier, low noise amplifier (LNA) and the like.
- the mobile communication module 150 can receive electromagnetic waves from the antenna 1, filter and amplify the received electromagnetic waves, and transmit them to the modulation and demodulation processor for demodulation.
- the mobile communication module 150 can also amplify the signal modulated by the modulation and demodulation processor, and then turn it into an electromagnetic wave for radiation through the antenna 1 .
- at least part of the functional modules of the mobile communication module 150 may be provided in the processor 110 .
- at least part of the functional modules of the mobile communication module 150 may be provided in the same device as at least part of the modules of the processor 110 .
- the modem processor may include a modulator and a demodulator.
- the modulator is used to modulate the low frequency baseband signal to be sent into a medium and high frequency signal.
- the demodulator is used to demodulate the received electromagnetic wave signal into a low frequency baseband signal.
- the demodulator then transmits the demodulated low-frequency baseband signal to the baseband processor for processing.
- the low frequency baseband signal is processed by the baseband processor and passed to the application processor.
- the application processor outputs sound signals through audio devices (not limited to the speaker 170A, the receiver 170B, etc.), or displays images or videos through the display screen 194 .
- the modem processor may be a stand-alone device.
- the modem processor may be independent of the processor 110, and may be provided in the same device as the mobile communication module 150 or other functional modules.
- the wireless communication module 160 can provide applications on the terminal device 100 including wireless local area networks (WLAN) (such as wireless fidelity (Wi-Fi) networks), bluetooth (BT), global navigation satellites Wireless communication solutions such as global navigation satellite system (GNSS), frequency modulation (FM), near field communication (NFC), and infrared technology (IR).
- WLAN wireless local area networks
- BT Bluetooth
- GNSS global navigation satellite system
- FM frequency modulation
- NFC near field communication
- IR infrared technology
- the wireless communication module 160 may be one or more devices integrating at least one communication processing module.
- the wireless communication module 160 receives electromagnetic waves via the antenna 2 , frequency modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 110 .
- the wireless communication module 160 can also receive the signal to be sent from the processor 110 , perform frequency modulation on it, amplify it, and convert it into electromagnetic waves for radiation through the antenna 2 .
- the wireless communication module 160 includes a Bluetooth module, and the terminal device 100 establishes a wireless connection with the headset 200 through Bluetooth.
- the wireless communication module 160 includes an infrared module, and the terminal device 100 can establish a wireless connection with the headset 200 through the infrared module.
- the antenna 1 of the terminal device 100 is coupled with the mobile communication module 150, and the antenna 2 is coupled with the wireless communication module 160, so that the terminal device 100 can communicate with the network and other devices through wireless communication technology.
- the wireless communication technology may include global system for mobile communications (GSM), general packet radio service (GPRS), code division multiple access (CDMA), broadband Code Division Multiple Access (WCDMA), Time Division Code Division Multiple Access (TD-SCDMA), Long Term Evolution (LTE), BT, GNSS, WLAN, NFC , FM, and/or IR technology, etc.
- the GNSS may include a global positioning system (global positioning system, GPS), a global navigation satellite system (GLONASS), a Beidou navigation satellite system (BDS), a quasi-zenith satellite system (quasi -zenith satellite system, QZSS) and/or satellite based augmentation systems (SBAS).
- GPS global positioning system
- GLONASS global navigation satellite system
- BDS Beidou navigation satellite system
- QZSS quasi-zenith satellite system
- SBAS satellite based augmentation systems
- the terminal device 100 implements a display function through a GPU, a display screen 194, an application processor, and the like.
- the GPU is a microprocessor for image processing, and is connected to the display screen 194 and the application processor.
- the GPU is used to perform mathematical and geometric calculations for graphics rendering.
- Processor 110 may include one or more GPUs that execute program instructions to generate or alter display information.
- Display screen 194 is used to display images, videos, and the like.
- Display screen 194 includes a display panel.
- the display panel can be a liquid crystal display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode or an active-matrix organic light-emitting diode (active-matrix organic light).
- LED diode AMOLED
- flexible light-emitting diode flexible light-emitting diode (flex light-emitting diode, FLED), Miniled, MicroLed, Micro-oLed, quantum dot light-emitting diode (quantum dot light emitting diodes, QLED) and so on.
- the terminal device 100 may include 1 or N1 display screens 194 , where N1 is a positive integer greater than 1.
- the terminal device 100 can realize the shooting function through the ISP, the camera 193, the video codec, the GPU, the display screen 194 and the application processor.
- the ISP is used to process the data fed back by the camera 193 .
- the shutter is opened, the light is transmitted to the camera photosensitive element through the lens, the light signal is converted into an electrical signal, and the camera photosensitive element transmits the electrical signal to the ISP for processing, and converts it into an image visible to the naked eye.
- ISP can also perform algorithm optimization on image noise, brightness, and skin tone.
- ISP can also optimize the exposure, color temperature and other parameters of the shooting scene.
- the ISP may be provided in the camera 193 .
- Camera 193 is used to capture still images or video.
- the object is projected through the lens to generate an optical image onto the photosensitive element.
- the photosensitive element may be a charge coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor.
- CMOS complementary metal-oxide-semiconductor
- the photosensitive element converts the optical signal into an electrical signal, and then transmits the electrical signal to the ISP to convert it into a digital image signal.
- the ISP outputs the digital image signal to the DSP for processing.
- DSP converts digital image signals into standard RGB, YUV and other formats of image signals.
- the processor 110 may trigger the startup of the camera 193 according to a program or an instruction in the internal memory 121, so that the camera 193 captures at least one image and performs corresponding processing on the at least one image according to the program or instruction.
- the terminal device 100 may include 1 or N2 cameras 193 , where N2 is a positive integer greater than 1.
- a digital signal processor is used to process digital signals, in addition to processing digital image signals, it can also process other digital signals. For example, when the terminal device 100 selects a frequency point, the digital signal processor is used to perform Fourier transform on the frequency point energy, and the like.
- Video codecs are used to compress or decompress digital video.
- the terminal device 100 may support one or more video codecs.
- the terminal device 100 can play or record videos in various encoding formats, for example, moving picture experts group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4 and so on.
- MPEG moving picture experts group
- the NPU is a neural-network (NN) computing processor.
- NN neural-network
- Applications such as intelligent cognition of the terminal device 100 can be implemented through the NPU, such as image recognition, face recognition, speech recognition, text understanding, and the like.
- the external memory interface 120 can be used to connect an external memory card, such as a Micro SD card, to expand the storage capacity of the terminal device 100 .
- the external memory card communicates with the processor 110 through the external memory interface 120 to realize the data storage function. For example to save files like music, video etc in external memory card.
- Internal memory 121 may be used to store computer executable program code, which includes instructions.
- the internal memory 121 may include a storage program area and a storage data area.
- the program storage area may store an operating system, an application program (such as a camera application) required for at least one function, and the like.
- the storage data area may store data created during the use of the terminal device 100 (such as images captured by a camera, etc.) and the like.
- the internal memory 121 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, universal flash storage (UFS), and the like.
- the processor 110 executes various functional applications and data processing of the terminal device 100 by executing instructions stored in the internal memory 121 and/or instructions stored in a memory provided in the processor.
- the internal memory 121 may also store the downlink audio signal provided by the embodiment of the present application.
- the internal memory 121 may also store codes for implementing the functions of controlling the headset 200 .
- the code stored in the internal memory 121 for implementing the function of controlling the earphone 200 is executed by the processor 110, the earphone 200 is controlled to realize the corresponding function, such as the ANC function, the HT function or the AH function.
- the code for implementing the function of controlling the headset 200 provided by the embodiment of the present application may also be stored in an external memory.
- the processor 110 may run the corresponding data stored in the external memory to realize the function of controlling the earphone 200 through the external memory interface 120, so as to control the earphone 200 to realize the corresponding function.
- the terminal device 100 may implement audio functions through an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, an application processor, and the like. Such as music playback, recording, etc.
- the audio module 170 is used to convert digital audio information to analog audio signal output, and also to convert analog audio input to digital audio signal. Audio module 170 may also be used to encode and decode audio signals. In some embodiments, the audio module 170 may be provided in the processor 110 , or some functional modules of the audio module 170 may be provided in the processor 110 .
- Speaker 170A also referred to as a "speaker" is used to convert audio electrical signals into sound signals.
- the terminal device 100 can listen to music through the speaker 170A, or listen to a hands-free call.
- the receiver 170B also referred to as "earpiece" is used to convert audio electrical signals into sound signals.
- the terminal device 100 answers a call or a voice message, the voice can be answered by placing the receiver 170B close to the human ear.
- the microphone 170C also called “microphone” or “microphone” is used to convert sound signals into electrical signals.
- the user can make a sound by approaching the microphone 170C through a human mouth, and input the sound signal into the microphone 170C.
- the terminal device 100 may be provided with at least one microphone 170C.
- the terminal device 100 may be provided with two microphones 170C, which may implement a noise reduction function in addition to collecting sound signals.
- the terminal device 100 may further be provided with three, four or more microphones 170C to collect sound signals, reduce noise, identify sound sources, and implement directional recording functions.
- the earphone jack 170D is used to connect wired earphones.
- the terminal device 100 is connected to the earphone through the earphone interface 170D.
- the earphone interface 170D may be the USB interface 130, or may be a 3.5mm open mobile terminal platform (OMTP) standard interface, a cellular telecommunications industry association of the USA (CTIA) standard interface.
- OMTP open mobile terminal platform
- CTIA cellular telecommunications industry association of the USA
- the pressure sensor 180A is used to sense pressure signals, and can convert the pressure signals into electrical signals.
- the pressure sensor 180A may be provided on the display screen 194 .
- the capacitive pressure sensor may be comprised of at least two parallel plates of conductive material. When a force is applied to the pressure sensor 180A, the capacitance between the electrodes changes.
- the terminal device 100 determines the intensity of the pressure according to the change in capacitance. When a touch operation acts on the display screen 194, the terminal device 100 detects the intensity of the touch operation according to the pressure sensor 180A.
- the terminal device 100 may also calculate the touched position according to the detection signal of the pressure sensor 180A.
- touch operations acting on the same touch position but with different touch operation intensities may correspond to different operation instructions. For example, when a touch operation whose intensity is less than the first pressure threshold acts on the short message application icon, the instruction for viewing the short message is executed. When a touch operation with a touch operation intensity greater than or equal to the first pressure threshold acts on the short message application icon, the instruction to create a new short message is executed.
- the gyro sensor 180B may be used to determine the motion attitude of the terminal device 100 .
- the angular velocity of the end device 100 about three axes ie, the x, y and z axes
- the gyro sensor 180B can be used for image stabilization.
- the gyro sensor 180B detects the shaking angle of the terminal device 100, calculates the distance to be compensated by the lens module according to the angle, and allows the lens to offset the shaking of the terminal device 100 through reverse motion to achieve anti-shake.
- the gyro sensor 180B can also be used for navigation and somatosensory game scenarios.
- the air pressure sensor 180C is used to measure air pressure.
- the terminal device 100 calculates the altitude through the air pressure value measured by the air pressure sensor 180C to assist in positioning and navigation.
- the magnetic sensor 180D includes a Hall sensor.
- the terminal device 100 can detect the opening and closing of the flip holster using the magnetic sensor 180D.
- the terminal device 100 can detect the opening and closing of the flip according to the magnetic sensor 180D. Further, according to the detected opening and closing state of the leather case or the opening and closing state of the flip cover, characteristics such as automatic unlocking of the flip cover are set.
- the acceleration sensor 180E can detect the magnitude of the acceleration of the terminal device 100 in various directions (generally three axes).
- the magnitude and direction of gravity can be detected when the terminal device 100 is stationary. It can also be used to identify the posture of terminal devices, and can be used in applications such as horizontal and vertical screen switching, pedometers, etc.
- the terminal device 100 can measure the distance through infrared or laser. In some embodiments, when shooting a scene, the terminal device 100 can use the distance sensor 180F to measure the distance to achieve fast focusing.
- Proximity light sensor 180G may include, for example, light emitting diodes (LEDs) and light detectors, such as photodiodes.
- the light emitting diodes may be infrared light emitting diodes.
- the terminal device 100 emits infrared light to the outside through the light emitting diode.
- the terminal device 100 detects infrared reflected light from nearby objects using a photodiode. When sufficient reflected light is detected, it can be determined that there is an object near the terminal device 100 . When insufficient reflected light is detected, the terminal device 100 may determine that there is no object near the terminal device 100 .
- the terminal device 100 can use the proximity light sensor 180G to detect that the user holds the terminal device 100 close to the ear to talk, so as to automatically turn off the screen to save power.
- Proximity light sensor 180G can also be used in holster mode, pocket mode automatically unlocks and locks the screen.
- the ambient light sensor 180L is used to sense ambient light brightness.
- the terminal device 100 may determine the exposure time of the image according to the ambient light brightness sensed by the ambient light sensor 180L.
- the terminal device 100 can adaptively adjust the brightness of the display screen 194 according to the perceived ambient light brightness.
- the ambient light sensor 180L can also be used to automatically adjust the white balance when taking pictures.
- the ambient light sensor 180L can also cooperate with the proximity light sensor 180G to detect whether the terminal device 100 is in a pocket, so as to prevent accidental touch.
- the fingerprint sensor 180H is used to collect fingerprints.
- the terminal device 100 can use the collected fingerprint characteristics to realize fingerprint unlocking, accessing application locks, taking photos with fingerprints, answering incoming calls with fingerprints, and the like.
- the temperature sensor 180J is used to detect the temperature.
- the terminal device 100 uses the temperature detected by the temperature sensor 180J to execute the temperature processing strategy. For example, when the temperature reported by the temperature sensor 180J exceeds a threshold value, the terminal device 100 reduces the performance of the processor located near the temperature sensor 180J, so as to reduce power consumption and implement thermal protection.
- the terminal device 100 when the temperature is lower than another threshold, the terminal device 100 heats the battery 142 to avoid abnormal shutdown of the terminal device 100 caused by the low temperature.
- the terminal device 100 boosts the output voltage of the battery 142 to avoid abnormal shutdown caused by low temperature.
- Touch sensor 180K also called “touch device”.
- the touch sensor 180K may be disposed on the display screen 194 , and the touch sensor 180K and the display screen 194 form a touch screen, also called a “touch screen”.
- the touch sensor 180K is used to detect a touch operation on or near it.
- the touch sensor can pass the detected touch operation to the application processor to determine the type of touch event.
- Visual output related to touch operations may be provided through display screen 194 .
- the touch sensor 180K may also be disposed on the surface of the terminal device 100 , which is different from the position where the display screen 194 is located.
- the bone conduction sensor 180M can acquire vibration signals.
- the bone conduction sensor 180M can acquire the vibration signal of the vibrating bone mass of the human voice.
- the bone conduction sensor 180M can also contact the pulse of the human body and receive the blood pressure beating signal.
- the bone conduction sensor 180M can also be disposed in the earphone, combined with the bone conduction earphone.
- the audio module 170 can analyze the voice signal based on the vibration signal of the vocal vibration bone block obtained by the bone conduction sensor 180M, so as to realize the voice function.
- the application processor can analyze the heart rate information based on the blood pressure beat signal obtained by the bone conduction sensor 180M, and realize the function of heart rate detection.
- the keys 190 include a power-on key, a volume key, and the like. Keys 190 may be mechanical keys. It can also be a touch key.
- the terminal device 100 may receive key input and generate key signal input related to user settings and function control of the terminal device 100 .
- Motor 191 can generate vibrating cues.
- the motor 191 can be used for incoming call vibration alerts, and can also be used for touch vibration feedback.
- touch operations acting on different applications can correspond to different vibration feedback effects.
- the motor 191 can also correspond to different vibration feedback effects for touch operations on different areas of the display screen 194 .
- Different application scenarios for example: time reminder, receiving information, alarm clock, games, etc.
- the touch vibration feedback effect can also support customization.
- the indicator 192 can be an indicator light, which can be used to indicate the charging state, the change of the power, and can also be used to indicate a message, a missed call, a notification, and the like.
- the SIM card interface 195 is used to connect a SIM card.
- the SIM card can be contacted and separated from the terminal device 100 by inserting into the SIM card interface 195 or pulling out from the SIM card interface 195 .
- the terminal device 100 may support 1 or N3 SIM card interfaces, where N3 is a positive integer greater than 1.
- the SIM card interface 195 can support Nano SIM card, Micro SIM card, SIM card and so on. Multiple cards can be inserted into the same SIM card interface 195 at the same time. The types of the plurality of cards may be the same or different.
- the SIM card interface 195 can also be compatible with different types of SIM cards.
- the SIM card interface 195 is also compatible with external memory cards.
- the terminal device 100 interacts with the network through the SIM card to realize functions such as calls and data communication.
- the terminal device 100 adopts an eSIM, that is, an embedded SIM card.
- the eSIM card can be embedded in the terminal device 100 and cannot be separated from the terminal device 100 .
- the software system of the terminal device 100 may adopt a layered architecture, an event-driven architecture, a microkernel architecture, a microservice architecture, or a cloud architecture.
- the embodiments of the present invention take an Android system with a layered architecture as an example to exemplarily describe the software structure of the terminal device 100 .
- FIG. 2 is a block diagram of a software structure of a terminal device 100 according to an embodiment of the present invention.
- the layered architecture divides the software into several layers, and each layer has a clear role and division of labor. Layers communicate with each other through software interfaces.
- the Android system is divided into four layers, which are, from top to bottom, an application layer, an application framework layer, an Android runtime (Android runtime) and a system library, and a kernel layer.
- the application layer can include a series of application packages.
- the application package can include applications such as camera, gallery, calendar, call, map, navigation, WLAN, Bluetooth, music, video, short message and so on.
- the application framework layer provides an application programming interface (application programming interface, API) and a programming framework for applications in the application layer.
- the application framework layer includes some predefined functions.
- the application framework layer may include window managers, content providers, view systems, telephony managers, resource managers, notification managers, and the like.
- a window manager is used to manage window programs.
- the window manager can get the size of the display screen, determine whether there is a status bar, lock the screen, take screenshots, etc.
- Content providers are used to store and retrieve data and make these data accessible to applications.
- the data may include video, images, audio, calls made and received, browsing history and bookmarks, phone book, etc.
- the view system includes visual controls, such as controls for displaying text, controls for displaying pictures, and so on. View systems can be used to build applications.
- a display interface can consist of one or more views.
- the display interface including the short message notification icon may include a view for displaying text and a view for displaying pictures.
- the telephony manager is used to provide the communication function of the terminal device 100 .
- the management of call status including connecting, hanging up, etc.).
- the resource manager provides various resources for the application, such as localization strings, icons, pictures, layout files, video files and so on.
- the notification manager enables applications to display notification information in the status bar, which can be used to convey notification-type messages, and can disappear automatically after a brief pause without user interaction. For example, the notification manager is used to notify download completion, message reminders, etc.
- the notification manager can also display notifications in the status bar at the top of the system in the form of graphs or scroll bar text, such as notifications of applications running in the background, and notifications on the screen in the form of dialog windows. For example, text information is prompted in the status bar, a prompt sound is issued, the terminal device vibrates, and the indicator light flashes.
- Android Runtime includes core libraries and a virtual machine. Android runtime is responsible for scheduling and management of the Android system.
- the core library consists of two parts: one is the function functions that the java language needs to call, and the other is the core library of Android.
- the application layer and the application framework layer run in virtual machines.
- the virtual machine executes the java files of the application layer and the application framework layer as binary files.
- the virtual machine is used to perform functions such as object lifecycle management, stack management, thread management, safety and exception management, and garbage collection.
- a system library can include multiple functional modules. For example: surface manager (surface manager), media library (Media Libraries), 3D graphics processing library (eg: OpenGL ES), 2D graphics engine (eg: SGL), etc.
- surface manager surface manager
- media library Media Libraries
- 3D graphics processing library eg: OpenGL ES
- 2D graphics engine eg: SGL
- the Surface Manager is used to manage the display subsystem and provides a fusion of 2D and 3D layers for multiple applications.
- the media library supports playback and recording of a variety of commonly used audio and video formats, as well as still image files.
- the media library can support a variety of audio and video encoding formats, such as: MPEG4, H.264, MP3, AAC, AMR, JPG, PNG, etc.
- the 3D graphics processing library is used to implement 3D graphics drawing, image rendering, compositing, and layer processing.
- 2D graphics engine is a drawing engine for 2D drawing.
- the kernel layer is the layer between hardware and software.
- the kernel layer includes at least display drivers, camera drivers, audio drivers, headphone drivers, and sensor drivers.
- the workflow of the software and hardware of the terminal device 100 is exemplarily described below in conjunction with the audio capture and playback scene.
- a corresponding hardware interrupt is sent to the kernel layer.
- the kernel layer processes touch operations into raw input events (including touch coordinates, timestamps of touch operations, etc.). Raw input events are stored at the kernel layer.
- the application framework layer obtains the original input event from the kernel layer, and identifies the control corresponding to the input event. Take the touch operation as a touch click operation, and the control corresponding to the click operation is an audio application icon control as an example, the audio application invokes the interface of the application framework layer to start the headphone control application, and then starts the headphone driver by calling the kernel layer, The audio signal is sent to the earphone, and the audio signal is played through the earphone 200 .
- FIG. 3 shows a schematic diagram of an optional hardware structure of the earphone 200 .
- the earphone 200 includes a left earphone and a right earphone.
- the left earphone and the right earphone have similar structures.
- the structures of the earphones include a first microphone 301 , a second microphone 302 and a third microphone 303 .
- a processor 304 and a speaker 305 may also be included in the headset. It should be understood that the earphone described later may be interpreted as the left earphone, and may also be interpreted as the right earphone.
- the first microphone 301 is used to collect the sound of the current external environment, and the first microphone 301 may also be referred to as a reference microphone.
- the first microphone 301 is located outside the earphone, or the second microphone 301 is located outside the ear.
- the second microphone 302 collects the sound of the internal environment of the user's ear canal.
- the second microphone 302 may also be referred to as an error microphone.
- the second microphone 302 is located inside the earphone and close to the ear canal.
- the third microphone 303 is used to collect call signals.
- the third microphone 303 may be located outside the earphone. When the user wears the earphone, the third microphone 303 is closer to the user's mouth than the first microphone 301 .
- the first microphone 301 is used to collect the sound of the current external environment, which can be interpreted as the sound of the external environment where the headset user wears the headset, for example, on a train, the sound of the external environment is the surrounding environment of the user wearing the headset the sound of.
- the first microphone 301 on the left earphone adopts the sound of the external environment of the left earphone.
- the first microphone 301 on the right earphone collects the sound of the external environment of the right earphone.
- the signal collected by the first microphone 301 (reference microphone) is called the first signal
- the signal collected by the second microphone 302 (error microphone) is called the second signal.
- the microphone involved in the embodiments of the present application may be an analog microphone or a digital microphone.
- the microphone is an analog microphone, before filtering the signal collected by the microphone, the analog signal can be converted into a digital signal.
- the first microphone and the second microphone are both digital microphones as an example for description, and the first signal and the second signal are both digital signals.
- the processor 304 is used to process downlink audio signals and/or signals collected by microphones (including the first microphone 301 , the second microphone 302 or the third microphone 303 ), such as performing ANC processing, HT processing, or AH processing.
- the processor 304 may include a main control unit and a noise reduction processing unit.
- the main control unit is used for the user to generate control commands for the operation of the earphone or to receive control commands from the terminal device.
- the noise reduction processing unit is configured to perform ANC processing, HT processing or AH processing on downlink audio signals and signals collected by the microphones (including the first microphone 301 , the second microphone 302 or the three microphones 303 ) according to the control command.
- the left and right earphones may also include memory for storing programs or instructions executed by the processor 304 .
- the processor 304 performs ANC processing, HT processing, or AH processing according to programs or instructions stored in the memory.
- the memory may include random access memory (RAM), flash memory, read-only memory (ROM), programmable read-only memory (PROM), erasable programmable read-only memory (erasable).
- RAM random access memory
- ROM read-only memory
- PROM programmable read-only memory
- PROM erasable programmable read-only memory
- the main control unit for example, can be composed of an ARM processing chip, a central processing unit (CPU), a system on chip (SoC), a digital signal processing (DSP), or a micro controller (micro controller). unit, MCU) in one or more implementations.
- the noise reduction processing unit may include, for example, a codec (coder-decoder, CODEC) chip or a high-fidelity (high-fidelity, HiFi) chip, or the like.
- the codec is hardened with a filter, an equalizer (EQ), a dynamic range controller (DRC), a limiter (limiter), and a gain adjuster ( Gain), mixer (mixer), etc., are mainly used for filtering, mixing, gain adjustment and other processing of signals.
- the noise reduction processing unit may further include a DSP, and the DSP may be used for processing such as scene detection, speech enhancement, and occlusion elimination.
- the headset may further include a wireless communication unit for establishing a communication connection with the terminal device 200 through the wireless communication module 160 in the terminal device 100 .
- the wireless communication unit can provide applications on the headset including wireless local area networks (WLAN) (such as wireless fidelity (Wi-Fi) networks), Bluetooth (bluetooth, BT), short-range wireless communication technology ( near field communication, NFC), infrared technology (infrared, IR) and other wireless communication solutions.
- the wireless communication unit may be one or more devices integrating at least one communication processing module.
- the wireless communication module 160 may be Bluetooth, the wireless communication unit is also Bluetooth, and the headset 200 and the terminal device 100 are connected through Bluetooth.
- an active noise reduction output channel outputs through three channels, respectively, an active noise reduction output channel, an ambient sound transmission output channel, and an auditory enhancement output channel.
- different output channels use different processing methods, as shown in Figure 4.
- the active noise reduction processing in the active noise reduction output path may, but is not limited to, include: performing noise suppression using the inverted signal of the first signal collected by the reference microphone and the inverted signal of the second signal collected by the error microphone.
- the active noise reduction output path includes an inverted signal of the first signal and an inverted signal of the second signal. It should be noted that the phase difference between the first signal and the inverted signal of the first signal is 180°.
- the speaker outputs a signal obtained by superimposing the inverted signal of the first signal and the inverted signal of the second signal, so that the sound of the current external environment played by the speaker and the sound of the external environment actually heard by the ear are cancelled to achieve active denoising. Effect. Therefore, when the headset adopts the ANC mode, the headset user's perception of the sound of the current environment and the ambient sound inside the user's ear canal can be weakened.
- filter compensation can be performed on the downlink audio signal.
- the influence of the downlink audio signal can be removed when the inverted signal of the ambient sound is obtained.
- the first filtering process and the third filtering process may be used.
- the first filtering process may be a feed forward (feed forward, FF) filtering process, which may be implemented by a feed forward filter.
- the third filtering process may be a feedback (feedback, FB) filtering process, which may be implemented by a feedback filter.
- FF filtering and FB filtering use a parallel processing architecture to enhance the noise reduction effect. The processing flow of the ANC will be described in detail later, and the description will not be repeated here.
- the ambient sound transparent transmission processing in the outgoing path of ambient sound transmission may include, but is not limited to, including: performing a third filtering process on the first signal collected by the error microphone to achieve part of the active noise reduction function, and performing a second filtering process on the signal collected by the reference microphone as well as HT enhanced processing.
- the second filtering process may be a transparent transmission (HT) filtering process, which may be implemented by a transparent transmission filter.
- the audio signal played by the speaker is obtained according to the first signal and the second signal, so that after the audio signal played by the speaker, the user can hear the sound in the external environment through the earphone, which is compared with the sound of the external environment heard when the HT processing is not performed.
- the sound intensity is better and the effect is better. Therefore, when the headset adopts the HT mode, the user's perception of the intensity of the sound in the environment where the user is currently located can be enhanced.
- the processing flow of the HT will be described in detail later, and the description will not be repeated here.
- the transparent transmission processing of ambient sound in the auditory enhancement output path may include, but is not limited to, using the signal collected by the error microphone to realize part of the active noise reduction function, performing first filtering processing on the signal collected by the reference microphone, and hearing enhancement processing to make the environment where the user is located.
- the sound of the event in the sound is enhanced, and a second filtering process is performed on the signal collected by the reference microphone.
- the output signal of the speaker is obtained according to the mixed signal of the event signal in the first signal and the inverted signal of the second signal. It should be noted that the phase difference between the second signal and the inverted signal of the second signal is 180°.
- the speaker outputs the inverted signal of the second signal, the inverted signal of the first signal and the signal after the event signal in the first signal is superimposed, so that the signal output by the speaker and the sound in the environment actually heard by the ear are canceled to achieve the initiative.
- the processing flow of the AH will be described in detail later, and the description will not be repeated here.
- the downlink audio signal, the first signal and the second signal may be a uniform frame signal or a signal of a period of time.
- the downlink audio signal, the first signal and the second signal are all one-frame signals
- the downlink audio signal, the first signal and the second signal belong to three signal streams respectively
- the signal frame of the downlink audio signal and the signal of the first signal The frame and the signal frame of the second signal are in the same time period or overlap in time, etc.
- functional processing such as ANC, HT, or AH
- the functional processing is continuously performed for the signal flow where the downlink audio signal is located, the signal flow where the first signal is located, and the signal flow of the second signal.
- the downlink audio signal sent by the terminal device 100 to the earphone 200 is referred to as the first audio signal as an example in the subsequent description.
- the first audio signal may be a call signal, a music signal, or the like.
- the signal collected by the reference microphone is called the first signal
- the signal collected by the error microphone is called the second signal. Headphones are in ANC mode.
- the downlink audio signals sent by the terminal device 100 to the left earphone and the right earphone in the earphone 200 may be the same signal or different signals.
- the terminal device adopts stereo effect, and the terminal device 100 sends different downlink audio signals to the earphone 200 to realize the stereo effect.
- the terminal device can also send the same downlink audio signal to the left earphone and the right earphone, and the left earphone and the right earphone adopt stereo processing, so as to achieve a stereo effect.
- the left earphone or the right earphone may perform the process of FIG. 5A or 5B according to the user's control.
- S501 Perform a first filtering process on a first signal collected by a reference microphone to obtain a first filtered signal.
- the first filtered signal is simply referred to as signal A1 in FIG. 5B .
- the first filtered signal is simply referred to as signal A2 in Figure 5B.
- the first audio signal when filtering out the first audio signal included in the second signal collected by the error microphone, the first audio signal may be filtered and compensated first.
- S503 Perform sound mixing processing on the first filtered signal and the first filtered signal to obtain a third audio signal.
- the third audio signal is simply referred to as the signal A3, that is, the signal A3 is obtained by mixing the signal A1 and the signal A2.
- the fourth audio signal is simply referred to as signal A4 in FIG. 5B .
- S505. Perform sound mixing processing on the fourth audio signal and the first audio signal to obtain the second audio signal.
- the speaker is responsible for playing the second audio signal.
- the second audio signal is simply referred to as A5.
- the signal output by the speaker is the fourth audio frequency that has not been mixed. Signal.
- S502 and S505 need not be performed.
- the first filtering process is an FF filtering process implemented by an FF filter as an example
- the third filtering process is an FB filtering process implemented by an FB filter as an example.
- the reference microphone in the earphone 200 picks up the first signal, which is input to the FF filter to obtain the signal A1 after FF filtering processing.
- the error microphone picks up the second signal and inputs it to the subtractor.
- the downlink audio signal is also input to the subtractor after filtering and compensating.
- the subtracter removes the downlink audio signal after filtering and compensation included in the second signal, so as to eliminate the effect of the downlink audio signal. Affect gets signal A2.
- the signal A3 is obtained after mixing the signal A1 and the signal A2 by the sound mixer, and the signal A4 is input into the FB filter, and the signal A4 is obtained after the FB filtering process. After the signal A4 is mixed with the downlink audio signal, the signal A5 is obtained, and the signal A5 is input to the speaker for playback.
- FF filtering and FB serial processing are used to realize ANC processing, so as to obtain better noise-reduced signals and enhance the noise-reduction effect.
- the quality of the ANC effect may be determined by the processing intensity of the ANC processing.
- the processing strength of the ANC processing depends on the FF filter coefficients used for FF filtering and/or the FB filter coefficients used for FB filtering.
- the default FF filter coefficients in the ANC mode may be used.
- the FF filter coefficient used when the ANC mode was selected last time can be used.
- the earphone determines the FF filter coefficient used in the ANC mode according to the recognized scene.
- the user indicates to the headset the FF filter coefficient used in the ANC mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the ANC mode as the target processing intensity through the UI control provided by the terminal device, and different processing intensities correspond to different FF filter coefficients.
- the default FB filter coefficient in the ANC mode can be used.
- the FB filter coefficient used when the ANC mode was selected last time can be used.
- the earphone determines the adopted FB filter coefficient according to the recognized scene.
- the user indicates to the headset the FF filter coefficient used in the ANC mode through the UI control provided by the terminal device, for example, the user selects the processing intensity in the ANC mode through the UI control provided by the terminal device as the target processing intensity, Different processing strengths correspond to different FB filter coefficients.
- the acquisition of the FF filter coefficient and the FB filter coefficient in the ANC mode may be any combination of the above-mentioned methods.
- the FF filter adopts the default filter coefficient in the ANC mode, and the FB filter coefficient is determined by the headset according to the recognized scene.
- the FB filter adopts the default filter coefficient in the ANC mode, and the FF filter coefficient is determined by the user through the UI control provided by the terminal device.
- the FB filter adopts the default filter coefficient in the ANC mode, and the FF filter coefficient is selected and indicated to the headset by the user through the UI control provided by the terminal device.
- the determination of the processing intensity in the ANC mode will be described in detail later by using a specific example, and the description will not be repeated here.
- FIG. 6A , FIG. 6B and FIG. 6C it is a schematic diagram of a process flow of transparent transmission of ambient sound.
- the downlink audio signal sent by the terminal device 100 to the earphone 200 is referred to as the first audio signal as an example in the subsequent description.
- the first audio signal may be a call signal, a music signal, or the like.
- the signal collected by the reference microphone is called the first signal
- the signal collected by the error microphone is called the second signal.
- the left earphone or the right earphone in the earphone 200 may perform the process of FIG. 6A or FIG. 6B or FIG. 6C according to the user's control.
- S601 Perform first signal processing on the first signal collected by the reference microphone to obtain a first processed signal.
- the first processed signal is referred to as signal B1 in Figures 6B and 6C.
- the first signal processing includes HT filtering.
- S602. Perform sound mixing processing on the first processed signal and the first audio signal to obtain a fifth audio signal.
- the fifth audio signal is referred to as signal B2 in Figures 6B and 6C.
- the signal B2 is obtained by performing mixing processing on the signal B1 and the downlink audio signal (ie, the first audio signal).
- the second filtered signal is referred to as signal B3 in Figures 6B and 6C. That is, the signal B3 is obtained by filtering out the signal B2 included in the second ambient signal.
- filter compensation processing may be performed on the fifth audio signal to reduce hearing loss.
- the third filtered signal is referred to as signal B4 in Figures 6B and 6C. That is, the signal B4 is obtained by FB filtering the signal B3.
- the first processed signal when the first processed signal is obtained by performing the first signal processing on the first signal collected by the reference microphone, it may be implemented in the following manner:
- the HT filtering is performed on the first signal to obtain a second filtered signal.
- the second filtered signal is referred to as signal B5.
- the second signal processing is performed on the second filtered signal to obtain the second processed signal; the second signal processing may also be referred to as low-latency algorithm processing or HT enhancement processing.
- the low-latency algorithm processing includes one or more of de-blocking effect processing, noise reduction processing, wind noise reduction processing, gain adjustment processing or frequency response adjustment processing.
- the HT-filtered signal is further processed by a low-latency algorithm to reduce the noise floor, reduce abnormal sounds, and optimize the user's sense of hearing.
- the HT filtering process may be implemented by a noise reduction processing unit, as shown in FIG. 6B .
- the noise reduction processing unit of the earphone includes a CODEC.
- the CODEC includes an HT filter, an FB filter, a subtractor, a first mixer, a second mixer, and a filter compensation unit.
- FIG. 6B takes as an example that the noise reduction processing unit further includes a DSP.
- the DSP can be used to perform low-latency algorithmic processing.
- the reference microphone in the earphone 200 picks up the first signal, which is input to the HT filter to obtain the signal B5 after HT filter processing.
- the signal B5 is input to the DSP, and the DSP performs low-latency algorithm processing on the signal B5 to obtain the signal B1.
- the signal B1 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal B1 to obtain the signal B2.
- After the signal B2 is processed by the filter compensation unit, it is input to the subtractor.
- the subtractor is used to filter the signal B2 that has undergone filtering and compensation processing included in the second ambient signal picked up by the error microphone to obtain a signal B3.
- the signal B3 is input to the FB filter, and the FB filter performs the FB filtering process on the signal B3 to obtain the signal B4.
- Signal B4 is input to the second mixer.
- the input of the second mixer also includes signal B2.
- the second sound mixer performs sound mixing processing on the signal B2 and the signal B4 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
- the HT filtering process can be implemented by DSP, as shown in FIG. 6C .
- the DSP can be used to perform HT filtering processing as well as low-latency algorithm processing.
- the noise reduction processing unit of the earphone includes an FB filter, a subtractor, a first sound mixer, a second sound mixer and a filter compensation unit.
- the reference microphone in the earphone picks up the first signal and inputs it to the DSP.
- the DSP performs HT filtering processing on the first signal, and performs low-latency algorithm processing to obtain a signal B1.
- the signal B1 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal B1 to obtain the signal B2.
- the signal B2 After the signal B2 is processed by the filter compensation unit, it is input to the subtractor.
- the subtractor is used to filter the signal B2 included in the second signal picked up by the error microphone to obtain the signal B3.
- the signal B3 is input to the FB filter, and the FB filter performs the FB filtering process on the signal B3 to obtain the signal B4.
- Signal B4 is input to the second mixer.
- the input of the second mixer also includes signal B2.
- the second sound mixer performs sound mixing processing on the signal B2 and the signal B4 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
- the low-latency algorithm processing includes de-blocking effect processing.
- the principle of the occlusion effect is explained first.
- Path 1 from the bone conduction to the periosteum, which is only a low-frequency signal.
- Path 2 It propagates from the outside air to the periosteum and is perceived, and the signal contains low frequency signals and medium and high frequency signals.
- the low-frequency signal and the medium-high frequency signal are superimposed, the low-frequency signal is too sufficient, and the low-frequency signal cannot be emitted when the earphone is worn, resulting in low-frequency chaos inside the ear, resulting in an occlusion effect.
- the signal B5 obtained after the HT filtering process is processed by the de-blocking effect, and the specific method may be as follows.
- Mode 1 is shown in FIG. 7 .
- a first voice harmonic signal matching the bone conduction signal in a voice harmonic set where the voice harmonic set includes a plurality of voice harmonic signals.
- the plurality of speech harmonic signals included in the speech harmonic set correspond to different frequencies. Specifically, the frequency of the bone conduction signal can be determined, and the first voice harmonic signal is determined from the set of voice harmonics according to the frequency of the bone conduction signal.
- the speech harmonic signal may also be referred to as the speech harmonic component.
- S702 remove the first voice harmonic signal in the signal B5 obtained after the HT filtering process.
- the signal C1 is obtained by removing the first voice harmonic signal in the signal B5 obtained after the HT filtering process. Since the voice of a person collected by a general bone conduction sensor is generally a low-frequency harmonic component, S702 is to remove the low-frequency harmonic component in the signal B5, thereby obtaining a signal C1 that does not include the low-frequency harmonic component.
- S703 Amplify the high frequency components in the signal B5 from which the first voice harmonic signal is removed. That is, the high frequency components of the signal C1 are amplified.
- the first voice harmonic signal matching the bone conduction signal can be determined, that is, the bone conduction sensor can detect the bone conduction signal, that is, the earphone wearer is currently speaking, such as speaking or singing.
- the signal obtained by boosting the high-frequency components on the basis of the signal C1 includes only mid- and high-frequency components, so that the signal heard by the earphone wearer has relatively no occlusion effect.
- the set of speech harmonics may be pre-stored in the headset.
- the set of speech harmonics may be obtained in an offline manner or in an online manner.
- the bone conduction signals of multiple people can be collected through the bone conduction sensor, and the following processing is performed on the bone conduction signals of each person in the bone conduction signals of the multiple people:
- the first bone conduction signal is transformed into a frequency domain signal by FFT; the fundamental frequency signal in the frequency domain signal is determined by means of a pilot frequency to find the fundamental frequency; the harmonic component of the bone conduction signal is determined according to the fundamental frequency signal, Thereby, the mapping relationship between the frequency of the bone conduction signal and the harmonic component is obtained to obtain a set of speech harmonics.
- the voice harmonic set may include mapping relationships between different frequencies and harmonic components.
- the second bone conduction signal collected within a set period of time can be obtained through the bone conduction sensor in the headset.
- the set period of time there may be multiple people using the headset, or only one person may use the headset, that is, the user;
- the fundamental frequency signal in the frequency domain signal is determined by means of a pilot frequency to find the fundamental frequency. If multiple people use the headset within the set time period, it may be possible to determine multiple fundamental frequency signals corresponding to multiple different time periods within the set time period.
- Multiple harmonic components of the bone conduction signal may be determined according to multiple fundamental frequency signals, so as to obtain a mapping relationship between frequencies and harmonic components to obtain a voice harmonic set.
- the voice harmonic set may include mapping relationships between different frequencies and different harmonic components.
- adaptive filtering can be performed on the signal B5 obtained after the HT filtering process to remove the low frequency component in the signal B5 to obtain the signal C1, that is, the signal of the earphone wearer's voice in the signal B5 is removed.
- the high frequency components in the third filtered signal from which the low frequency components are removed are amplified, that is, the high frequency components in the signal C1 are amplified.
- the signal obtained by boosting the high-frequency components on the basis of the signal C1 includes only mid- and high-frequency components, so that the signal heard by the earphone wearer has relatively no occlusion effect.
- the quality of the HT effect can be determined by the treatment intensity of the HT treatment.
- the processing strength of the HT processing depends on the HT filter coefficients used for the HT filtering and/or the FB filter coefficients used for the FB filtering.
- the default HT filter coefficients in the HT mode may be used.
- the HT filter coefficient used when the HT mode was selected last time can be used.
- the earphone determines the HT filter coefficient used in the HT mode according to the recognized scene.
- the user indicates to the headset the HT filter coefficient used in the HT mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the HT mode as the target processing intensity through the UI control provided by the terminal device, and different processing intensities correspond to different HT filter coefficients.
- the FB filter coefficient in one way, the default FB filter coefficient in the HT mode can be used.
- the FB filter coefficient used when the HT mode was selected last time can be used.
- the earphone determines the adopted FB filter coefficient according to the recognized scene.
- the user indicates to the headset the HT filter coefficient used in the HT mode through the UI control provided by the terminal device, for example, the user selects the processing intensity in the HT mode through the UI control provided by the terminal device as the target processing intensity, Different processing strengths correspond to different FB filter coefficients.
- the acquisition of the HT filter coefficients and the FB filter coefficients in the HT mode may be any combination of the methods provided above.
- the downlink audio signal sent by the terminal device 100 to the earphone 110 is referred to as the first audio signal as an example in the subsequent description.
- the first audio signal may be a call signal, a music signal, a prompt tone, or the like.
- the signal collected by the reference microphone is called the first signal
- the signal collected by the error microphone is called the second signal.
- the left earphone or the right earphone in the earphone 200 may perform the process of FIG. 8A or FIG. 8B or FIG. 8C according to the user's control.
- S801 Perform HT filtering on the first signal collected by the reference microphone to obtain a second filtered signal (signal C1).
- the second filtered signal is referred to as signal C1.
- the first filtered signal is simply referred to as signal C3.
- Step S804 Perform sound mixing processing on the filtered enhanced signal and the first audio signal to obtain a sixth audio signal.
- the sixth audio signal is simply referred to as signal C4.
- Step S804 is to mix the signal C2 and the downlink audio signal to obtain the signal C4.
- Step S805 filtering the sixth audio signal included in the second signal to obtain a fourth filtered signal.
- the fourth filtered signal is simply referred to as signal C5.
- Step S805 is to filter out the signal C4 included in the second environment signal to obtain the signal C5.
- step S805 when step S805 is performed, the signal C4 may be filtered and compensated first to obtain a compensated signal, and then the compensated signal included in the second signal may be filtered to obtain C5.
- Step S806 is to perform FB filtering on the signal C5 to obtain the signal C6.
- Step S807. Perform sound mixing processing on the fifth filtered signal, the sixth audio signal and the first filtered signal to obtain the second audio signal.
- Step S806 is to mix the signal C6, the signal C4 and the signal C3 to obtain the second audio signal.
- Mode 1 is shown in FIG. 9 .
- the manner of performing de-blocking effect processing on the signal C1 may be the same as that for performing the de-blocking effect processing on the signal B5. For details, refer to manners 1 and 2 in scenario 2, which will not be repeated here.
- noise reduction processing is performed on the signal processed by the de-occlusion effect, and the noise reduction processing includes artificial intelligence AI noise reduction processing and/or wind noise reduction processing; in FIG. 9, the noise reduction processing includes AI noise reduction processing and wind noise reduction processing. Take processing as an example.
- a gain amplification process is performed on the signal obtained after the wind noise processing, and a feasible way is to directly amplify the signal obtained after the wind noise processing.
- the method of directly amplifying the signal not only amplifies the external signal, but also amplifies the voice of the wearer.
- the embodiment of the present application provides a gain amplification processing method, which only amplifies the external signal, and does not amplify the voice signal speaking to the wearer. For example, as shown in FIG. 10 , when the signal obtained after the noise reduction processing is subjected to gain amplification processing, it can be implemented in the following manner.
- the voice signal of the wearer's speech is conducted to the periosteum through the bone, and the voice signal is concentrated in the low frequency, which is recorded as the bone conduction signal D1.
- the bone conduction signal D1 is acquired by the bone conduction sensor.
- the bone conduction signal D1 is harmonically extended to obtain a harmonically extended signal, for example, the harmonically extended signal is called D2.
- the harmonic extension may be a harmonic enhancement method or a method in which the harmonics of the bone conduction signal D1 are directly upwardly extended.
- the signal obtained by the noise reduction process is referred to as signal D3.
- the signal D4 is obtained by amplifying the signal D3 by using the first gain coefficient.
- the amplification processing can directly amplify the signal.
- the value of the first gain coefficient may be related to the value of the processing intensity of the AH. For example, a mapping relationship between the first gain coefficient and the value of the processing intensity of AH is stored in the earphone.
- the first filter coefficient is determined based on the first gain coefficient.
- the first gain coefficient gain to adjust the strength of the adaptive filtering can also be called the first filter coefficient, that is, the number of dB used for signal amplification by the first gain coefficient is the same as the number of dB filtered out by the adaptive filter, so that the wearer's voice The signal can remain balanced, neither amplified nor reduced.
- the second method is shown in FIG. 11 .
- the manner of performing de-blocking effect processing on the signal C1 may be the same as that for performing the de-blocking effect processing on the signal B5. For details, refer to manners 1 and 2 in scenario 2, which will not be repeated here.
- S1102 Perform audio event detection on the deblocked signal to obtain an audio event signal (which may be referred to as an event signal for short) in the deblocked signal. Audio event signals, such as station announcements, whistles, etc.
- S1103 Perform gain amplification on the audio event signal in the deblocked signal.
- the audio event signal in the de-blocked signal such as station announcement sound, whistle sound, etc.
- the method of performing gain amplification processing on the audio event signal in the de-blocked signal may be the same as when performing gain amplification processing on the signal obtained after the noise reduction processing, and will not be repeated here.
- the noise reduction processing unit includes a CODEC and a DSP.
- the CODEC of the earphone includes an HT filter, a FB filter, a FF filter, a subtractor, a first mixer, a second mixer, and a filter compensation unit.
- the HT filtering process is implemented by the CODEC.
- DSP can be used to perform enhancement processing.
- the reference microphone in the earphone 110 picks up the first signal, which is input to the HT filter to obtain the signal C1 after HT filter processing.
- the signal C1 is input to the DSP, and the DSP performs enhancement processing on the signal C1 to obtain the signal C2.
- the signal C2 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal C2 to obtain the signal C4.
- the signal C4 is processed by the filter compensation unit, it is input to the subtractor.
- the subtractor is used to filter out the filtered and compensated signal C4 included in the second ambient signal picked up by the error microphone to obtain the signal C5.
- the signal C5 is input to the FB filter, and the FB filter performs the FB filtering process on the signal C5 to obtain the signal C6.
- Signal C6 is input to the second mixer.
- the input of the second mixer also includes signal C4 and signal C3.
- the second audio mixer performs audio mixing processing on the signal C3, the signal C4 and the signal C6 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
- the noise reduction processing unit includes a CODEC and a DSP as an example.
- the DSP can be used to perform HT filtering processing as well as enhancement processing.
- the CODEC of the earphone includes a FB filter, an FF filter, a subtractor, a first mixer, a second mixer, and a filter compensation unit.
- the reference microphone in the earphone 110 picks up the first signal and inputs it to the DSP.
- the DSP performs HT filtering processing on the first signal to obtain a signal C1.
- the DSP performs enhancement processing on the signal C1 to obtain a signal C2.
- the signal C2 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal C2 to obtain the signal C4.
- the signal C4 is processed by the filter compensation unit, it is input to the subtractor.
- the subtractor is used for filtering the filtered and compensated signal C4 included in the second ambient signal picked up by the error microphone to obtain the signal C5.
- the signal C5 is input to the FB filter, and the FB filter performs the FB filtering process on the signal C5 to obtain the signal C6.
- Signal C6 is input to the second mixer.
- the input of the second mixer also includes signal C4 and signal C3.
- the second audio mixer performs audio mixing processing on the signal C3, the signal C4 and the signal C6 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
- the quality of the AH effect can be determined by the treatment intensity of the AH treatment.
- the processing strength of the AH processing depends on at least one of the HT filter coefficient, the FB filter coefficient, or the FF filter coefficient.
- the default FF filter coefficient in the AH mode can be used.
- the FF filter coefficient used when the AH mode was selected last time can be used.
- the earphone determines the FF filter coefficient used in the AH mode according to the recognized scene.
- the user indicates to the headset the FF filter coefficient used in the AH mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the AH mode as the target processing intensity through a UI control provided by the terminal device, and different processing intensities correspond to different FF filter coefficients.
- the default HT filter coefficients in the AH mode may be used.
- the HT filter coefficient used when the AH mode was selected last time can be used.
- the earphone determines the HT filter coefficient used in the AH mode according to the recognized scene.
- the user indicates to the headset the HT filter coefficients used in the AH mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the AH mode as the target processing intensity through the UI control provided by the terminal device, and different processing intensities correspond to different HT filter coefficients.
- the FB filter coefficient in one way, the default FB filter coefficient in the AH mode can be used. In another way, the FB filter coefficient used when the AH mode was selected last time can be used.
- the earphone determines the adopted FB filter coefficient according to the recognized scene.
- the user indicates to the headset the HT filter coefficient used in the AH mode through the UI control provided by the terminal device, for example, the user selects the processing intensity in the AH mode through the UI control provided by the terminal device as the target processing intensity, Different processing strengths correspond to different FB filter coefficients.
- the acquisition of the HT filter coefficient, the FB filter coefficient or the FF filter coefficient in the AH mode may be any combination of the above-mentioned methods.
- the determination of the processing mode adopted by the earphone 200 can be selected by the user through the UI control on the terminal device 100 and indicated to the earphone, or the terminal device can determine and indicate to the earphone according to the adaptively recognized scene, Alternatively, it can also be determined by the headset according to adaptive scene recognition.
- Example 1 a single control controls the left and right earphones.
- the terminal device 100 provides a control interface for the user to select the processing mode of the earphone 200 (including the left earphone and the right earphone) according to requirements: empty mode, ANC mode, HT mode or AH mode. No processing is done in empty mode.
- the processing modes of the earphones for the user to select on the control interface are all processing modes supported by the earphones.
- the left earphone and the right earphone have the same processing function, or support the same processing mode.
- the left earphone supports AHA
- the right earphone also supports AHA.
- the terminal device is installed with an earphone application adapted to the earphone 200, and during the adaptation process, the processing function of the earphone can be known.
- the terminal device can determine the processing function possessed by the headset according to the function parameters.
- the control interface includes user interface (user interface, UI) controls.
- the UI control is used for the user to select the processing mode of the headset 200 , and for the convenience of distinction, the UI control used for the user to select the processing mode of the headset is called a selection control.
- the processing mode includes at least two of ANC mode, HT mode or AH mode.
- the terminal device 100 sends control signaling 1 to the left earphone and the right earphone respectively in response to the user selecting the target mode from the processing modes supported by the earphone through the selection control. Control signaling 1 carries the target mode.
- the selection controls can also be used to select the processing intensity in target mode.
- the shape of the selection control can be a circle, a bar or other shapes.
- the selection controls may include a first control and a second control. Any two different positions of the second control on the first control correspond to different processing modes of the headset or two different positions of the second control on the first control correspond to different processing intensities in the same processing mode of the headset.
- the user selects different processing modes and controls the processing intensity by moving the position of the second control representing the user's selection on the first control on the display screen.
- control of the processing modes of the left earphone and the right earphone is implemented through an earphone application (appliaction, APP).
- the terminal device 100 includes a headset control application for controlling the headset, which is referred to as a headset application.
- a headset application For example, refer to the main interface of the terminal device shown in FIG. 12A .
- the terminal device can respond to the user's operation of clicking the icon 001, start the headset application, and display the control of the headset application on the display screen interface, or it is interpreted that when the headset application is started, the control interface of the headset application pops up.
- the first control in the annular selection control in FIG. 12B includes three circular arc segments, corresponding to the ANC mode, the HT mode, and the AH mode, respectively. If the second control is located in the arc segment of the ANC mode, it is determined to be the ANC mode, and different positions of the second control located in the arc segment of the ANC mode correspond to different processing intensities in the ANC mode.
- the HT mode is determined, and different positions of the second control located in the arc segment of the HT mode correspond to different processing intensities in the HT mode. If the second control is located in the arc segment of the AH mode, it is determined to be the AH mode, and different positions of the second control located in the arc segment of the AH mode correspond to different processing intensities in the AH mode.
- the highlighted black dots on the ring represent the second control for the user to select the processing intensity.
- the user can select different processing modes and control the processing intensity by moving the position of the black dots on the circumference.
- the terminal device 100 (such as a processor) responds to operation 1 performed by the user on the control interface, for example, operation 1 is generated when the user moves the position of the second control on the display screen representing the second control selected by the user.
- the terminal device 100 sends a control instruction 1 to the left earphone and the right earphone respectively, and the control instruction 1 is used to indicate the target mode and the target processing intensity.
- the target mode is the ANC mode.
- control instruction 1 may include an ANC identifier and a parameter value indicating a target processing intensity when performing ANC processing.
- the FB filter coefficients and/or the FF filter coefficients corresponding to different processing intensities are different.
- the control command 1 includes radians.
- the corresponding processing mode can be determined according to the range in which the radian is located. Different radian values correspond to the processing strength in this processing mode. Referring to Fig. 12B, the processing mode corresponding to (0, 180] is the ANC mode, the processing mode corresponding to (180, 270] is the HT mode, and the processing mode corresponding to (270, 360] is the AH mode.
- the left earphone and the right earphone can be Including the mapping relationship between the radian range and the processing mode, and the mapping relationship between the radian value and the filter coefficient. Taking the ANC mode as an example, the FB filter coefficient and the FF filter coefficient corresponding to different radian values are different.
- the user can touch the black dot in the disc to rotate clockwise from 0 degrees to 360 degrees.
- the corresponding FF filter coefficient and FB filter coefficient at 0 degrees make the ANC noise reduction effect the strongest, that is, the user
- the perceived sound of the user's current environment and the ambient sound inside the user's ear canal are weaker.
- the FF filter coefficient and the FB filter coefficient change, so that the active noise reduction effect gradually weakens, and the active noise reduction effect is the weakest when it reaches 180 degrees, similar to the headphones without any noise reduction processing.
- From 180 degrees to 270 degrees is to control the transparent transmission of the ambient sound.
- the corresponding HT filter coefficients and FB filter coefficients at 180 degrees make The transparent transmission effect of ambient sound is the weakest, that is, the lower the sound intensity of the user's current environment perceived by the user, is similar to the use of empty mode after wearing headphones.
- the HT filter coefficient and the FB filter coefficient change, which makes the transparent transmission effect of ambient sound stronger and stronger.
- 270 degrees to 360 degrees is used to control the hearing enhancement, that is, the user touches the black dot in the disc
- the corresponding FF filter coefficient, HT filter coefficient and FB filter coefficient at 180 degrees make the hearing enhancement effect the weakest, and the user perceives The sound of the event included in the sound of the environment where the user is currently located is weaker.
- the FF filter coefficient, HT filter coefficient and FB filter coefficient change, so that the auditory enhancement effect becomes stronger and stronger, that is, the event signal that the user wants to hear becomes stronger and stronger to assist hearing.
- the terminal device 100 sends a control command 1 to the left earphone and the right earphone respectively through Bluetooth in response to the user's operation 1 .
- the control command 1 may include the ANC identification and the parameter value of the target processing intensity.
- the operations performed after the left earphone and the right earphone receive the control command 1 are similar, and the processing of the left earphone is taken as an example in the subsequent description.
- the main control unit of the left earphone acquires the FF filter coefficient and the FB filter coefficient processed by the ANC from the coefficient library according to the ANC identifier and the target processing strength.
- the coefficient library includes the mapping relationship shown in Table 1.
- Table 1 is only an example, and constitutes a specific definition of the mapping relationship.
- the parameter value of the target processing intensity is intensity 1
- the main control unit of the left earphone obtains the FF filter coefficient corresponding to intensity 1 according to Table 1 as the coefficient FF1
- the FB filter coefficient is the coefficient FB1
- the main control unit controls the FF filter to use the coefficient FF1
- a signal A1 is obtained by performing FF filtering processing on the first signal collected by the reference microphone.
- the main control unit controls the FB filter to use the coefficient FB1 to perform FB filtering processing on the signal A3 to obtain the second audio signal.
- the main control unit writes the coefficient FF1 and the coefficient FB1 into the AHA core, so that the AHA core executes the steps of S501-S504 to obtain the second audio signal.
- the terminal device 100 sends a control instruction 1 to the left earphone and the right earphone respectively through Bluetooth in response to the user's operation 1 .
- the control instruction 1 may include an HT identifier and a target processing intensity, where the target processing intensity is used to indicate the processing intensity when the HT processing is performed.
- the operations performed after the left earphone and the right earphone receive the control command 1 are similar, and the processing of the left earphone is taken as an example in the subsequent description.
- the main control unit of the left earphone acquires the HT filter coefficient and/or the FB filter coefficient of the HT processing from the coefficient library according to the HT identifier and the target processing strength.
- the value of the target processing intensity is intensity 5
- the main control unit of the left earphone obtains the HT filter coefficient corresponding to intensity 5 according to Table 1 as the coefficient HT1
- the FB filter coefficient is the coefficient FB5
- the main control unit controls the HT filter.
- the controller uses the coefficient HT1 to perform HT filtering processing on the first signal collected by the reference microphone.
- the main control unit controls the FB filter to use the coefficient FB5 to perform FB filtering processing on the signal B3.
- the main control unit writes the coefficient HT1 and the coefficient FB5 into the AHA core, so that the AHA core executes the steps of S601-S605 to obtain the second audio signal.
- the terminal device 100 sends a control instruction 1 to the left earphone and the right earphone respectively through Bluetooth in response to the user's operation 1 .
- the control instruction 1 may include the AH identifier and the parameter value of the target processing strength.
- the operations performed after the left earphone and the right earphone receive the control command 1 are similar, and the processing of the left earphone is taken as an example in the subsequent description.
- the main control unit of the left earphone obtains the HT filter coefficient, FF filter coefficient and FB filter coefficient of the AH process from the coefficient library according to the HT identifier and the target processing strength.
- the value of the target processing strength is indication 7
- the main control unit of the left earphone obtains the HT filter coefficient corresponding to indication 7 according to Table 1 as the coefficient HT3
- the FB filter coefficient is the coefficient FB7
- the FF filter coefficient is the coefficient FF5.
- the main control unit controls the HT filter to perform HT filter processing on the first signal collected by the reference microphone using the coefficient HT3.
- the main control unit controls the FB filter to use the coefficient FB7 to perform FB filtering processing on the signal C5.
- the main control unit controls the FF filter to use the coefficient FF5 to perform FF filtering processing on the first signal.
- the main control unit writes the coefficient HT3, the coefficient FB7 and the coefficient FF5 into the AHA core, so that the AHA core executes the steps of S801-S807 to obtain the second audio signal.
- the selection control may take the form of a bar.
- the selection controls include a first control and a second control.
- the bar of the first control may be divided into a plurality of bar segments, which are divided according to the number of processing modes supported by the headset.
- the second control indicates different processing modes in different bar segments of the first control.
- the second control indicates different processing intensities of the same processing mode at different positions of the same bar segment of the first control.
- the bar of the first control includes three bar segments.
- the user can touch the black bar to slide left and right.
- the black bar is at the K1 position
- the corresponding FF filter coefficient and FB filter coefficient make the ANC noise reduction effect the strongest.
- the FF filter coefficient and FB filter coefficient are the strongest.
- the filter coefficient changes, so that the active noise reduction effect gradually weakens.
- the K2 position is reached, the active noise reduction effect is the weakest, similar to the headphones without any noise reduction processing.
- the area between K2 and K3 is the part that controls the transparent transmission of ambient sound.
- Position K3 to position K4 is used to control auditory enhancement, that is, the user moves from position K3 to position K4 by touching the black bar, and when the black bar moves to position K3, the corresponding FF filter coefficient, HT filter coefficient and FB filter The coefficient makes the hearing enhancement effect the weakest. Moving from position K3 to position K4, the FF filter coefficient, HT filter coefficient and FB filter coefficient change, making the hearing enhancement effect stronger and stronger, that is, the voice signal that the user wants to hear is getting stronger and stronger. , to assist hearing.
- the selection controls in (a) include buttons corresponding to different processing modes, including an ANC button, an HT button, and an AH button.
- the terminal device 100 displays the display interface shown in FIG. 12G(b) in response to the user's operation of clicking the ANC button.
- the display interface of (b) includes controls 002 for selecting treatment intensity. Users can touch the black bar and slide up and down to determine the processing intensity of ANC, that is, to select the corresponding FF filter coefficient and FB filter coefficient. Black bars slide in the area of L1-L2. When the black bar is located at the L1 position, the corresponding FF filter coefficient and FB filter coefficient make the ANC noise reduction effect the strongest. After sliding down, the FF filter coefficient and the FB filter coefficient change, so that the active noise reduction effect gradually weakens. When it reaches the L2 position Active noise reduction has the weakest effect, similar to wearing headphones without any noise reduction.
- the headset APP when the headset 200 establishes a connection with the terminal device, the headset APP can be triggered to start, and a control interface including selection controls is displayed, such as the control interface illustrated in FIG. 12A , FIG. 12B , FIG. 12F or FIG. 12G .
- the interface displayed by the terminal device is interface 1
- the terminal device recognizes that the headset 200 is connected to the terminal device, and can jump from interface 1 to the control interface.
- the headset APP can be triggered to start, that is, a control interface including selection controls is displayed, such as FIG. 12A , FIG. 12B , FIG. 12C or Figure 12D illustrates the display interface.
- a control interface including selection controls is displayed, such as FIG. 12A , FIG. 12B , FIG. 12C or Figure 12D illustrates the display interface.
- the terminal device triggers the headset to play audio it may be that after the terminal device establishes a connection with the headset, a song is played, and a control interface including selection controls may be displayed.
- a control interface including selection controls can be displayed.
- the identified scene type of the current external environment is the target scene, and the target scene adaptation needs to adjust all settings.
- prompt information can be displayed, and the prompt information is used to prompt the user whether to adjust the processing mode of the headset.
- a control interface including selection controls may be displayed, such as the control interface illustrated in Figures 12A, 12B, 12C, or 12D.
- FIG. 12E takes the example of displaying the control interface of FIG. 12A as an example.
- the terminal device recognizes that the current scene the user is in is a noisy scene.
- the user may need to start the processing mode, so as to display selection prompt information (such as in the form of a prompt box) to prompt the user whether to adjust the processing mode of the headset .
- selection prompt information such as in the form of a prompt box
- the terminal device recognizes that the scene type of the external environment is a noisy scene.
- the user may need to start the processing mode, so as to display a prompt box to prompt the user whether to adjust the processing mode of the headset.
- the scene types that trigger the display of the prompt box may include: noisy scene, terminal building scene, train station scene, bus station scene, road scene, and the like.
- the strength of the identified signal reaches a set threshold, it is considered a noisy scene.
- the specific sound of an aircraft announcing the station is recognized, and the terminal building scene is determined to be recognized.
- the sound of the train time notification is recognized, and the train station scene is determined to be recognized; for another example, the bus ticket broadcast is recognized, and the bus station scene is determined to be recognized.
- the ticking of a signal light or the whistle of a car is recognized, and the road scene is determined to be recognized.
- a control interface including selection controls is displayed according to the identified scene where the user is currently located.
- Example 2 dual controls to control left and right earphones
- the terminal device 100 provides a control interface for the user to select the processing modes of the left earphone and the right earphone respectively according to requirements, and the processing modes of the left earphone and the right earphone may be different.
- the left earphone selects ANC mode
- the right earphone selects HT mode.
- the control interface includes selection controls for the left earphone and selection controls for the right earphone.
- the selection control of the left earphone is called the first selection control
- the selection control of the right earphone is called the second selection control.
- the first selection control is used for the user to select the processing mode of the left earphone
- the second selection control is used for the user to select the processing mode of the right earphone.
- the shape of the first selection control and the second selection control may be a circular ring, a bar or other shapes.
- the shapes of the first selection control and the second selection control may be the same or different.
- the user realizes the selection of different processing modes and the control of processing intensity by moving the position of the control representing the user's selection on the display screen.
- the shapes of the controls adopted by the left earphone and the right earphone may refer to the description in Example 1, and details are not repeated here.
- both the first selection control and the second selection control include a first control and a second control.
- Two different positions of the second control on the first control correspond to different processing modes, or, two different positions of the second control on the first control correspond to different processing intensities in the same processing mode. See the control interface shown in Figure 13 as an example.
- the user can select the different processing modes and control the processing intensity implemented by the left earphone through the position of the second control (black dot) in the first selection control of the left earphone moving on the circumference of the first control.
- the user can select the different processing modes implemented by the right earphone and control the processing intensity by moving the position of the second control on the first control in the second selection control of the right earphone.
- the user can select different processing modes for the left and right earphones, or the same processing intensity in the same processing mode, or different processing intensities in the same processing mode, so as to match the ear differences or match the needs of different applications. need.
- Example 2 the manner of triggering the display of the control interface including the first selection control and the second selection control refers to the description of Example 1, and details are not repeated here.
- Example 3 terminal device smart scene detection.
- the terminal device identifies the scene the user is currently in, and in different scenes, the headset adopts different processing modes.
- the terminal device recognizes that the scene type indication of the current external environment is the first scene, it determines the target mode corresponding to the first scene in the processing mode of the earphone, and sends control signaling 2 to the left earphone and the right earphone respectively.
- Control signaling 2 is used to indicate the target mode. Different target modes correspond to different scene types.
- the terminal device determines which function to instruct the earphone to perform according to the identified scene. Correspond the functions of AHA to the scene type, and select the most suitable function for the scene type, so that users can automatically feel the most needed effect.
- the scene types may include walking scene, running scene, quiet scene, multi-person talking scene, cafe scene, subway scene, train scene, car scene, waiting room scene, conversation scene, office scene, outdoor scene, driving Scene, windy scene, airplane scene, siren scene, whistle scene, cry scene, etc.
- AI models can be built offline and stored on terminal devices. For example, a microphone on a terminal device records a large amount of noise and sensor data and/or video processing unit (vpu) data in different scenarios, and manually labels the data for its corresponding scenarios.
- vpu noise and sensor data and/or video processing unit
- initialize and build an AI model which can be one of convolutional neural networks (CNN)/deep neural networks (DNN)/long short-term memory (LSTM) , can also be a mixture of different models.
- CNN convolutional neural networks
- DNN deep neural networks
- LSTM long short-term memory
- use the labeled data for model training to obtain the corresponding AI model.
- the processing modes suitable for different scene types are exemplified.
- the processing mode corresponding to the scene type is indicated in brackets.
- the noise of the plane is relatively loud when flying, so it is suitable to use the ANC mode.
- the HT mode For another example, in walking scenes, running scenes and quiet scenes, it is suitable to use the HT mode, and you can hear the sound of sudden events. For another example, in the cafe scene, the user needs to be quiet, and the ANC mode can be used. For another example, in a light music scene, the HT mode can be used. For another example, the alarm sound scene (AH), the whistle sound scene (AH), and the crying sound scene (AH) are used to hear the preset sound, and the AH mode is suitable.
- the AH mode is suitable.
- the terminal device 100 when the terminal device 100 recognizes that the scene type of the current external environment is an airplane scene, it can send control signaling 2 to the headset, and the control signaling 2 is used to indicate that the headset needs to implement the ANC function, that is, to indicate Headphones are in ANC mode. After the left earphone and the right earphone receive the control signaling 2, the processes of S501-S504 are performed respectively.
- the terminal device 100 when the terminal device 100 recognizes that the scene type of the current external environment is a walking scene, it can send control signaling 2 to the headset, and the control signaling 2 is used to indicate that the headset needs to implement the HT function, that is, the headset Use HT mode. After the left earphone and the right earphone receive the control signaling 2, the processes of S601-S605 are performed respectively.
- the terminal device 100 when the terminal device 100 recognizes that the scene type of the current external environment is a train station scene, it can send control signaling 2 to the headset, and the control signaling 2 is used to indicate that the headset needs to implement the AH function , that is, the headset adopts AH mode. After the left earphone and the right earphone receive the control signaling 2, the processes of S801-S807 are performed respectively.
- the terminal device After the headset is connected to the terminal device, the terminal device starts scene detection. After the detection of the terminal device is completed, the detection result can also be displayed to the user, so that the user can know the processing mode adopted by the headset. For example, the detection result is displayed to the user in the form of a prompt box.
- the detection result may include the detected scene, and may also include a processing mode corresponding to the detected scene. For example, when the terminal device recognizes that the scene is the first scene, it determines the target mode corresponding to the first scene in the processing mode of the headset, and can display the detection result, that is, the first scene and the target mode to the user. Then send control signaling 2 to the left earphone and the right earphone respectively. Control signaling 2 is used to indicate the target mode.
- a function switch for enabling smart scene detection is configured on the terminal device, and the terminal device triggers scene detection when the terminal device responds to the user initiating the function of smart scene detection.
- the scene in which it is identified is the first scene
- the target mode corresponding to the first scene in the processing mode of the earphone is determined, and then control signaling 2 is sent to the left earphone and the right earphone respectively. Control signaling 2 is used to indicate the target mode.
- the detection result can also be displayed to the user, so that the user can know the processing mode adopted by the headset.
- the detection result may include the detected scene, and may also include a processing mode corresponding to the detected scene.
- the terminal device determines the target mode corresponding to the first scene in the processing mode of the headset, and can display the detection result, that is, the first scene and the target mode to the user. Then send control signaling 2 to the left earphone and the right earphone respectively. Control signaling 2 is used to indicate the target mode.
- the left earphone and the right earphone send control signaling 2 again.
- the function switch for enabling smart scene detection configured by the terminal device may be configured on the control interface of the headset application, or may also be configured in the system setting menu bar of the terminal device.
- the terminal device can control the processing mode adopted by the headset by identifying the scene, and the terminal device can also identify the user operation on the selection control on the control interface to control the processing mode adopted by the headset. processing mode.
- the terminal device can determine whether to enable the smart scene detection function according to requirements. When the smart scene detection function is not enabled, the processing mode adopted by the headset can be manually selected by using Example 1.
- the terminal device 100 identifies the current user's location. Scenes.
- the user turns on the smart scene detection function he can update to another interface based on the interface for manually selecting the processing mode, or display the detection results based on the interface for manually selecting the processing function.
- the processing function selected by the user on the terminal device is the HT function; after the smart scene detection function is enabled, the terminal device recognizes that the current scene the user is in is an airplane scene, and the ANC function is suitable.
- the user starts the headset application, and the control interface of the headset application is displayed on the display screen, taking a circular ring as an example.
- the processing function selected by the user is the HT function, as shown in (a) of FIG. 14A .
- the control interface includes an option control for whether to enable the smart scene detection function.
- the user triggers the smart scene detection function, executes the scene detection to obtain the detection result, and changes the position of the control representing the processing function selected by the user to the area belonging to the ANC function.
- the position of the black dot on the disk may be the default value under the ANC function or the position where the processing intensity selected by the last user selection of the ANC function is located. For example, as shown in (b) of FIG. 14A . In (b) of FIG. 14A , the scene where an airplane is detected is taken as an example.
- the terminal device 100 sends control signaling 2 to the left earphone and the right earphone respectively, and the control signaling 2 is used to indicate the ANC function.
- the user starts the headset application, and the control interface of the headset application is displayed on the display screen, taking a circular ring as an example.
- the processing function selected by the user is the HT function, as shown in (a) of FIG. 14B .
- the control interface includes an option control for whether to enable the smart scene detection function. After the user triggers the option control for enabling the smart scene detection function, the user triggers the smart scene detection function, executes the scene detection to obtain the detection result, and displays the detection result on the detection result interface.
- the detection interface may also include a scene that the terminal device can recognize, and a processing function corresponding to the scene. For example, as shown in (b) of FIG. 14B , the detection result is an airplane scene, and the corresponding processing function is the ANC function.
- the terminal device 100 sends control signaling 2 to the left earphone and the right earphone respectively, and the control signaling 2 is used to indicate the ANC function.
- the target processing intensity in the target mode can be determined by any one of the following methods.
- Method 1 The headset adopts the default target processing intensity in target mode.
- the terminal device After the terminal device sends the control signaling 2 to the left earphone and the right earphone respectively, taking the left earphone as an example, after the left earphone receives the control signaling 2, it determines that the processing mode used is the target mode, and the control signaling 2 does not indicate the target processing intensity. , the headset determines the default target processing strength. Taking the target mode as the ANC mode as an example, after receiving the control signaling 2, the left earphone determines to use the ANC mode, and obtains the default FF filter coefficient and FB filter coefficient in the ANC mode from the left earphone.
- Method 2 Use the processing intensity used when the target mode was used last time as the target processing intensity.
- the target processing strength is determined by the terminal device, and is indicated to the left earphone and the right earphone through control signaling. After the terminal device performs scene detection and determines the target mode according to the detected scene, obtains the processing intensity used when the target mode was used last time as the target processing intensity, and sends control signaling 2 to the left earphone and the right earphone respectively. Let 2 indicate the target mode and target processing strength.
- the headset itself determines the processing intensity in the target mode.
- the terminal device After the terminal device performs scene detection, after determining the target mode according to the detected scene, it sends control signaling 2 to the left earphone and the right earphone respectively, and the control signaling 2 indicates the target mode.
- the left earphone and the right earphone receive the control signaling 2 , determine the adopted processing mode as the target mode, and obtain the saved processing intensity when the target mode was adopted last time as the target processing intensity.
- the target mode as ANC as an example, obtain the saved FF filter coefficients and FB filter coefficients when the ANC mode was last adopted, and perform ANC processing.
- Manner 3 The terminal device determines the target processing intensity according to the recognized scene.
- the function switch for enabling smart scene detection is not configured on the terminal device, after the terminal device recognizes the scene, it can determine the target processing intensity according to the recognized scene.
- the processing modes determined in different scenarios are the same, and the processing intensities corresponding to different scenarios are different.
- the walking scene, the running scene, and the quiet scene are all suitable for using the HT mode.
- the walking scene, the running scene, and the quiet scene have different processing intensities.
- ANC mode is suitable for multi-person talking scenes, cafe scenes, train scenes, airplane scenes, windy scenes, and office scenes.
- the processing intensity corresponding to the scene, the windy scene, and the office scene is different.
- the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene are all suitable for using the AH mode. When the AH mode is used, the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene correspond respectively different processing intensity.
- the terminal device sends control signaling 2 to the left earphone and right earphone according to the stored correspondence between the scene type, the target mode and the processing intensity.
- the control signaling 2 indicates the target mode and the target processing intensity in the target mode. Therefore, after receiving the control signaling 2, the earphone determines to adopt the target mode according to the control signaling 2, and determines the filter coefficient corresponding to the target processing intensity.
- the target mode as AH as an example
- the FF filter coefficient, the FB filter coefficient and the HT filter coefficient are determined according to the target processing intensity, and S801-S807 are performed according to the FF filter coefficient, the FB filter coefficient and the HT filter coefficient.
- the user indicates to the headset the processing intensity used in the target mode through a UI control provided by the terminal device.
- the detection result is displayed on the display interface of the terminal device, and the detection result includes the detected scene and the target mode corresponding to the detected scene.
- the display interface may include a control for selecting the treatment intensity.
- the control for selecting the treatment intensity is called an intensity control.
- Intensity controls can include Control 1 and Control 2. Different positions of control 1 are used to indicate different processing intensity in target mode. Intensity controls can be circles, bars, or other shapes.
- the detected scene is a terminal building scene. Taking the control 1 in the intensity control as a circle and the control 2 as a circular black dot as an example, in response to the user touching the control 2 and moving to the position of the control 1 1.
- Position 1 represents the target processing intensity in the user-selected target mode.
- Send control instruction 2 to the left earphone and right earphone, and control instruction 2 indicates the target mode and the target processing intensity corresponding to position 1.
- the target mode and target processing intensity can be sent to the left and right earphones through different control commands.
- the terminal device After determining the target mode according to the detected scene, the terminal device sends control signaling indicating the target mode to the left earphone and the right earphone.
- the left earphone and the right earphone adopt the default processing intensity in the target mode, that is, adopt the default filter coefficient in the target mode to realize the target processing corresponding to the target mode.
- the control signaling indicating the target processing intensity is sent to the left earphone and the right earphone, and further, the left earphone and the right earphone use the filter coefficient corresponding to the target processing intensity to achieve the target.
- the target processing corresponding to the mode.
- the user triggers the smart scene detection function, executes the scene detection to obtain the detection result, and sets the position of the control representing the processing function selected by the user by the position of the control. Changed to the area belonging to the ANC function.
- the position of the black dot on the disk may be the default value under the ANC function or the position where the processing intensity selected by the last user selection of the ANC function is located.
- the user selects the processing intensity in ANC mode by moving the position of the black dot. And send control signaling 2 to the left earphone and right earphone, and the control signaling 2 indicates the ANC mode and the target processing strength.
- Example 4 scene detection of headphones, different scenes correspond to different processing functions.
- the headset has a scene detection function.
- the headset recognizes the scene the user is currently in.
- the processing functions implemented by the headset are different.
- the left earphone may have the scene detection function, or the right earphone may have the scene detection function, or both the left earphone and the right earphone may have the scene detection function.
- one of the left earphone and the right earphone is used to perform scene detection.
- the left earphone performs scene detection, and sends the detection result to the right earphone, so that both the left earphone and the right earphone are based on the detection result of the left earphone. Executes processing for realizing the processing function corresponding to the detection result.
- both the left earphone and the right earphone execute processing for realizing the processing function corresponding to the detection result according to the detection result of the right earphone.
- both the left earphone and the right earphone perform scene detection, the left earphone performs the processing for realizing the processing function corresponding to the detection result according to the detection result of the left earphone, and the right earphone performs the processing for realizing the detection according to the detection result of the right earphone The result corresponds to the processing of the processing function.
- the enabling of the scene detection function of the headset may be controlled by the user through the headset or through the terminal device.
- the headset is provided with a button for activating the scene detection function.
- the user can turn on or turn off the scene detection function of the headset by touching the button.
- the headset identifies the scene where the current user is located (or the scene where the headset is currently located), the corresponding relationship between the scene and the processing mode determines the processing mode corresponding to the identified scene, and realizes the processing corresponding to the processing mode.
- the user can turn on or turn off the scene detection function of the earphone by tapping on the earphone, such as three consecutive taps.
- the headset responds to the operation of the user tapping the headset three times to enable the scene detection function of the headset.
- the scene detection function of the earphone is enabled, the earphone turns off the scene detection function of the earphone in response to the operation of the user tapping the earphone three times.
- the headset identifies the scene where the current user is located (or the scene where the headset is currently located), the corresponding relationship between the scene and the processing mode determines the processing mode corresponding to the identified scene, and realizes the processing corresponding to the processing mode.
- the activation of the scene detection function of the left earphone or the right earphone is controlled by the terminal device 100 .
- the headphone control interface includes a switch button for the headphone scene detection function.
- the terminal device can determine whether to enable the scene detection function of the headset according to user requirements. When the scene detection function of the headset is not enabled, Example 1 can be used to manually select the processing function to be implemented by the headset.
- the headset identifies the scene type of the current external environment, determines the processing mode corresponding to the identified scene type according to the corresponding relationship between the scene type and the processing mode, and implements the processing function corresponding to the processing mode.
- the terminal device 100 sends control signaling 3 to the headset 200, where the control signaling 3 is used to instruct the headset to start the scene detection function.
- the headset 200 starts to perform scene detection according to the control signaling 3 .
- the headset 200 determines the processing function to be executed according to the detected scene type of the current external environment, such as an ANC function, and the headset 200 executes the ANC processing and executes S501-S504.
- the headset after the headset is connected to the terminal device, the headset starts scene detection; or when the headset receives the downlink audio signal sent by the terminal device, the scene detection starts.
- the detection result may also be sent to the terminal device.
- the detection result may be included in the indication information and sent to the terminal device.
- the detection result may include the detected scene and the processing mode corresponding to the scene.
- the terminal device receives the detection result, it displays the detection result to the user, so that the user knows the processing mode adopted by the headset.
- the detection result is displayed to the user in the form of a prompt box.
- only the detected scene may be included in the detection result.
- the terminal device determines the processing mode corresponding to the scene detected by the headset, and displays the scene detected by the headset and the processing mode corresponding to the scene to the user. For example, when the earphone recognizes that the scene is the first scene, it determines the target mode corresponding to the first scene in the processing mode of the earphone, and the detection result, that is, the first scene and the target mode, can be displayed to the user.
- the processing function of the processing mode corresponding to the scene is not executed immediately, the detection result is sent to the terminal device, and the terminal device displays the detection result to the user.
- the terminal device responds to the user's operation of determining the processing mode, it sends a confirmation instruction to the earphone, and when the earphone receives the confirmation instruction, it adopts the processing mode corresponding to the scene detected by the earphone to execute the processing function.
- the types of scenes that can be recognized by the headset may include walking scenes, running scenes, quiet scenes, multi-person talking scenes, cafe scenes, subway scenes, train scenes, car scenes, waiting room scenes, conversation scenes, office scenes, outdoor scenes.
- the processing modes suitable for different scene types are exemplified.
- the processing mode corresponding to the scene type is indicated in brackets.
- the noise of the plane is relatively loud when flying, so it is suitable to use the ANC mode.
- the HT mode For another example, in walking scenes, running scenes and quiet scenes, it is suitable to use the HT mode, and you can hear the sound of sudden events. For another example, in the cafe scene, the user needs to be quiet, and the ANC mode can be used. For another example, in a light music scene, the HT mode can be used. For another example, the alarm sound scene (AH), the whistle sound scene (AH), and the crying sound scene (AH) are used to hear the preset sound, and the AH mode is suitable.
- the AH mode is suitable.
- the processing of S601-S605 is performed respectively for the left earphone and the right earphone.
- the scene type is the train station scene
- the left earphone and the right earphone perform the processing of S801-S807, respectively.
- the target processing intensity in the target mode may be determined in any of the following manners.
- Method 1 The headset adopts the default target processing intensity in target mode.
- the earphone determines the adopted processing mode as the target mode according to the detected scene, and the left earphone and the right earphone determine to adopt the default target processing intensity.
- the target mode as the ANC mode as an example
- the left earphone and the right earphone obtain the default FF filter coefficient and FB filter coefficient in the ANC mode.
- Method 2 Use the processing intensity used when the target mode was used last time as the target processing intensity.
- the earphone determines the processing intensity in the target mode by itself. After the headset performs scene detection, after determining the target mode according to the detected scene, the saved processing intensity of the last time the target mode was used is obtained as the target processing intensity. Taking the target mode as ANC as an example, obtain the saved FF filter coefficients and FB filter coefficients when the ANC mode was last adopted, and perform ANC processing.
- the target processing strength is determined by the terminal device, and is indicated to the left earphone and the right earphone through control signaling. After the headset performs scene detection, the detection result is sent to the terminal device, so that the terminal device obtains the processing intensity used when the target mode was used last time as the target processing intensity, and sends control signaling 4 to the left headset and the right headset respectively. Let 4 indicate the target processing strength.
- the headset determines the target processing intensity according to the recognized scene.
- the target processing intensity can be determined according to the recognized scene.
- the processing modes determined in different scenarios are the same, and the processing intensities corresponding to different scenarios are different.
- the walking scene, the running scene, and the quiet scene are all suitable for using the HT mode.
- the walking scene, the running scene, and the quiet scene have different processing intensities.
- ANC mode is suitable for multi-person talking scenes, cafe scenes, train scenes, airplane scenes, windy scenes, and office scenes.
- the processing intensity corresponding to the scene, the windy scene, and the office scene is different.
- the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene are all suitable for using the AH mode. When the AH mode is used, the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene correspond respectively different processing intensity.
- the left earphone and the right earphone determine the target mode corresponding to the detected scene and the target processing intensity in the target mode according to the stored correspondence between the scene type, the target mode, and the processing intensity. Thereby, the left earphone and the right earphone obtain the filter coefficient corresponding to the target processing intensity.
- the target mode as AH as an example
- the FF filter coefficient, the FB filter coefficient and the HT filter coefficient are determined according to the target processing intensity
- S801-S807 are performed according to the FF filter coefficient, the FB filter coefficient and the HT filter coefficient.
- the headset may also perform emergency event detection to determine a target event (or referred to as a target event).
- a target event For example, emergent events include one or more of wind noise events, whistling events, emergencies, human voice events, or no emergencies, and different events have different processing intensities.
- the headset performs scene detection, as well as event detection.
- target mode different events correspond to different filter coefficients. Taking ANC as an example, different events correspond to different FF filter coefficients and/or different FB filter coefficients.
- the left earphone can obtain the FF filter coefficient or FB corresponding to the detected event from the coefficient library according to the detection result.
- the coefficient library stores processing modes, events, and the mapping relationship between FF filter coefficients and FB filter coefficients.
- the processing effect of ANC mainly depends on FB filtering and/or FF filtering.
- the filter coefficient of the FF filter is controlled according to the detected scene, and the FB filter coefficient adopts a fixed value.
- the filter coefficient of the FB filter is controlled according to the detected scene, and the FF filter coefficient adopts a fixed value.
- the FF filter coefficient and the FB filter coefficient are controlled according to the detected scene.
- the event includes a howling event, a wind noise event, a sudden event, a human voice event, or no sudden event.
- the earphone 200 detects the event sound in the external environment, and can determine the target event corresponding to the event sound in the external environment according to the signal collected by the reference microphone.
- the signal collected by the reference microphone includes a signal of a preset frequency spectrum, and then an event corresponding to the signal of the preset frequency spectrum is determined.
- the signal collected by the reference microphone includes a wind sound signal, that is, the collected signal includes a signal matching the frequency spectrum of the wind sound, and the event corresponding to the detected event sound in the external environment is determined to be a wind noise event.
- a spectrum matching method or a deep neural network (deep neural networks, DNN) matching method may be used.
- the headset 200 can determine the event in the current environment of the user according to the signal collected by the reference microphone in the following manner, as shown in FIG. 15 .
- Bone conduction sensors are also included in the headset 200 .
- the bone conduction sensor is used to collect the bone conduction signal of the earphone user.
- the bone conduction sensor obtains the bone conduction signal by collecting the bone conduction signal, that is, collecting the periosteal vibration signal generated by the user's speaking.
- the activation of the scene detection function of the left earphone or the right earphone may be controlled by the terminal device 100, or may be controlled by a user's operation on the earphone, such as an operation of tapping the left earphone or the right earphone.
- a bone conduction sensor is included on the earphone, and the user's upper and lower teeth touch to generate a tooth touching sound, so that the bone conduction sensor activates the scene detection function by detecting the audio signal generated by the user's upper and lower teeth touching.
- the third signal collected by the reference microphone is a signal collected by the reference microphone after the headset starts the scene detection function.
- the energy of the bone conduction signal collected by the bone conduction sensor is small.
- the signal AA1 is the third signal.
- the earphone 200 may first determine the energy of the bone conduction signal, and if the energy of the bone conduction signal is less than the set threshold, the filtering operation is not performed, that is, S1501 is not performed, when it is determined that the energy of the bone conduction signal is greater than or equal to the set threshold When the threshold value is reached, execute S1501.
- the earphone 200 performs spectrum analysis on the signal AA1 to obtain the energy characteristic of the signal AA1. For example, the earphone 200 performs spectral analysis on the signal to obtain the energy of the entire frame of the signal AA1 and the energy of each bark subband of the signal AA1, thereby forming the energy characteristic of the signal AA1 represented by a vector.
- S1503 Determine a first energy feature of the energy features included in the energy feature set that matches the energy feature of the filtered signal, where different energy features included in the energy feature set correspond to different event identifiers.
- S1504 Determine that the event identified by the event identifier corresponding to the first energy feature is an event in the current environment of the user, that is, a detection result of event detection.
- the energy feature set may be generated by: performing wind noise detection, burst noise detection, howling detection, and human voice detection on the signals collected by the first microphone, the second microphone, and the third microphone to obtain wind noise signal, burst noise signal and howling signal. Then, spectrum analysis is performed on the wind noise signal, the burst noise signal, the howling signal and the human voice signal, respectively, to obtain the sub-band energy characteristics of the wind noise signal, the sub-band energy characteristics of the burst noise signal, and the sub-band of the howling signal. Energy signature and sub-band energy signature of the vocal signal.
- the sub-band energy features of the wind noise signal, the sub-band energy features of the burst noise signal, the sub-band energy features of the howling signal, and the sub-band energy features of the human voice signal constitute the energy feature set. It should be understood that in a quiet scene, the sub-band energy of the noise is relatively weak.
- a spectrum matching method can be used, and a deep neural network (DNN) can also be used.
- DNN deep neural network
- matching degree between the energy feature of the filtered signal and each energy feature included in the energy feature set can be determined through the DNN, and the event identified by the event identifier corresponding to the first energy feature with the highest matching degree is Test results.
- the main control unit in the headset 200 may implement the determination of an event in the current environment of the user according to the signal collected by the reference microphone.
- the main control unit includes a DSP for executing S1501-S1504.
- the user indicates to the headset the processing intensity used in the target mode through a UI control provided by the terminal device.
- the detection result is sent to the terminal device, and the terminal device displays the detection result to the user.
- the detection result is displayed on the display interface of the terminal device, and the detection result includes the scene detected by the headset and the target mode corresponding to the detected scene.
- the display interface further includes a control for selecting the treatment intensity.
- the control for selecting the treatment intensity is called an intensity control.
- Intensity controls can include Control 1 and Control 2. Different positions of control 1 are used to indicate different processing intensity in target mode. Intensity controls can be circles, bars, or other shapes. Referring to FIG.
- position 2 represents the target processing intensity in the user-selected target mode.
- the target mode is HT as an example.
- the terminal device 100 sends control signaling 3 to the headset 200 in response to the user initiating the function of scene detection in the headset, where the control signaling 3 is used to instruct the headset to activate the scene detection function.
- the headset 200 starts and executes scene detection according to the control signaling 3 to obtain the detection result.
- the headset 200 may send the detection result to the terminal device 100, so that the terminal device 100 displays it to the user, and displays the detected scene corresponding to the processing mode that the headset needs to adopt to the user.
- the user may update the interface to manually select the processing mode to another interface, or display the detection results based on the interface for manually selecting the processing function.
- the processing function selected by the user on the terminal device is the HT function; after the headset scene detection function is enabled, the headset 200 recognizes that the scene the user is currently in is an airplane scene, and the ANC function is suitable for detecting The result, the aircraft scene, the ANC function, is sent to the end device.
- the user starts the headset application, and the control interface of the headset application is displayed on the display screen, taking a circular ring as an example.
- the processing function selected by the user is the HT function, as shown in (a) of FIG. 17A .
- the control interface includes an option control for whether to enable the headset scene detection function.
- the terminal device triggers the headset scene detection function, and sends control signaling 3 to the headset 200, where the control signaling 3 is used to instruct the headset to activate the scene detection function.
- the headset 200 starts and executes scene detection according to the control signaling 3 to obtain the detection result.
- the headset 200 sends the detection result to the terminal device 100, and after receiving the detection result, the terminal device 100 changes the position of the control representing the processing function selected by the user to an area belonging to the ANC function.
- the user selects the processing intensity in ANC mode by moving the position of the black dot on the disc. For example, as shown in (b) of FIG. 17A . In (b) of FIG. 17A , a scene where an airplane is detected is taken as an example.
- the user starts the earphone application, and the control interface of the earphone application is displayed on the display screen, taking a circular ring as an example.
- the processing function selected by the user is the HT function, as shown in (a) of FIG. 17B .
- the control interface includes an option control for whether to enable the headset scene detection function.
- the terminal device triggers the headset scene detection function, and sends control signaling 3 to the headset 200, where the control signaling 3 is used to instruct the headset to activate the scene detection function.
- the headset 200 starts and executes scene detection according to the control signaling 3 to obtain the detection result.
- the headset 200 sends the detection result to the terminal device 100, and after receiving the detection result, the terminal device 100 displays the detection result on the detection result interface.
- the detection interface may also include a scene that the headset can recognize, and a processing mode corresponding to the scene.
- the user selects the processing intensity in ANC mode by moving the position of the black dot on the disc. For example, as shown in (b) of FIG. 17B , the detection result is an airplane scene, and the corresponding processing mode is the ANC mode.
- an artificial intelligence (artificial intelligence, AI) model can be used to perform detection and classification.
- the AI model can be configured in the headset. Another way can determine the scene type according to the signal collected by the reference microphone. For example, the headset 200 can determine the scene where the user is currently located according to the signal collected by the reference microphone in the following manner, as shown in FIG. 18 .
- S1801 Perform spectrum analysis on the first signal collected by the reference microphone, divide the first signal into multiple subbands, and calculate the energy of each subband. For example, the first information collected by the reference microphone is divided into subbands in the frequency domain according to the bark subband division method, and the energy of each subband is calculated.
- S1802 Determine the VAD to obtain the noise segment in the first signal, and obtain the smoothed energy of each subband in the noise segment.
- the VAD is determined as follows:
- S1803 Determine the scene type according to the smoothed energy of each subband in the noise segment.
- the processing is performed for the determined noise segment to determine the scene type:
- Example 5 After determining the processing mode, the headset performs event detection in the processing mode. In the processing mode, the filter coefficients corresponding to different events (ie, the processing intensity in the processing mode) are different.
- the earphone recognizes the user's operation, and determines that the earphone 200 selected by the user needs to implement ANC processing, HT processing or AH processing.
- the processing mode adopted by the earphone 200 is the ANC mode function.
- the user's operation may be an operation of the user tapping the earphone, and the processing mode is determined to be the ANC mode, the HT mode or the AH mode through different operations.
- buttons are provided on the earphone, and different buttons indicate different processing modes. The user selects the processing mode employed by the headset by pressing a button.
- the left earphone and the right earphone perform ANC processing, and specifically perform S501-S504.
- the processing mode to be implemented by the earphone is selected and controlled by the terminal device 100 .
- Either the left earphone or the right earphone can have an event detection function.
- one of the left earphone and the right earphone is used to perform event detection.
- the left earphone performs event detection and sends the detection result to the right earphone, or the right earphone performs event detection and sends the detection result. to the left earphone.
- different events correspond to different FF filter coefficients and FB filter coefficients.
- the left earphone or the right earphone performs event detection, taking the left earphone as an example, the left earphone can obtain the FF filter coefficient or FB filter coefficient corresponding to the detected event in the ANC mode from the coefficient library according to the detection result.
- the coefficient library For example, refer to the content included in the coefficient library shown in Table 2, and take the event including a howling event, a wind noise event, a sudden event or a human voice event as an example.
- the headset includes hardware structures and/or software modules corresponding to executing each function.
- modules and method steps of each example described in conjunction with the embodiments disclosed in the present application can be implemented in the form of hardware or a combination of hardware and computer software. Whether a function is performed by hardware or computer software-driven hardware depends on the specific application scenarios and design constraints of the technical solution.
- an embodiment of the present application further provides a noise processing apparatus 1900 .
- the noise processing device 1900 is applied to an earphone.
- the headset has at least two functions in the ANC function, the HT function or the AH function; the headset includes a first microphone and a second microphone; the first microphone is used to collect a first signal, and the first signal is used to represent the current the sound of the external environment; the second microphone is used to collect a second signal, and the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the earphone.
- the noise processing apparatus 1900 can be used to implement the function of the earphone in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
- the apparatus may include a communication module 1901 , an acquisition module 1902 , and a first processing module 1903 .
- a communication module 1901 configured to receive a first audio signal from the terminal device
- the acquiring module 1902 is configured to acquire a target mode; wherein, the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the earphone to implement a target processing function, and the target processing function is an active reduction.
- the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the earphone to implement a target processing function, and the target processing function is an active reduction.
- the first processing module 1903 is configured to obtain a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal.
- the apparatus further includes: a playing module, configured to play the second audio signal.
- a playing module configured to play the second audio signal.
- the playback module it is not shown in FIG. 19 .
- the target processing function is an ANC function
- the second audio signal played by the playback module can attenuate the user's sound to the user's current environment and the user's ear canal Perception of internal ambient sounds; or,
- the target processing function is the HT function, and the second audio signal played by the playback module can enhance the user's perception of the sound of the environment where the user is currently located; or,
- the target processing function is the AH function, and the second audio signal played by the playing module can enhance the user's perception of the event sound; the event sound satisfies a preset frequency spectrum.
- the target processing function is an ANC function
- the second audio signal is obtained based on the first audio signal, the third signal and the fourth signal, and the third signal is the first signal
- the inverted signal of the fourth signal is the inverted signal of the second signal
- the target processing function is the HT function
- the second audio signal is obtained based on the first audio signal, the first signal and the second signal; or,
- the target processing function is the AH function
- the second audio signal is obtained based on the first audio signal, the fifth signal and the fourth signal
- the fifth signal is an event signal in the first signal
- the The event signal satisfies the preset spectrum.
- the communication module 1901 is further configured to receive a first control instruction from the terminal device, where the first control instruction carries the target mode, and the target mode is the terminal The device is determined according to the scene type of the current external environment;
- the target mode is sent to the obtaining module 1902 .
- the communication module 1901 is further configured to:
- the second control instruction carries a target processing strength, and the target processing strength is used to indicate the processing strength when the headset implements the target processing function;
- the first processing module 1903 is specifically used for:
- a second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
- it also includes:
- the second processing module 1904 is configured to determine the target event corresponding to the event sound in the current external environment according to the first signal, and determine the target processing intensity in the target mode according to the target event; wherein, the target processing intensity used to indicate the processing strength when the headset implements the target processing function;
- the first processing module 1903 is specifically used for:
- a second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
- the headset further includes a bone conduction sensor, and the bone conduction sensor is used to collect bone conduction signals generated by the vibration of the user's vocal cords;
- the first processing module 1901 is specifically configured to determine the target event corresponding to the event sound in the current external environment according to the first signal and the bone conduction signal.
- the target event includes one of a howling event, a wind noise event, a sudden event or a human voice event.
- it also includes:
- the third processing module 1905 is configured to identify the scene type of the current external environment as a target scene according to the first signal, and determine the target mode adopted by the headset according to the target scene, where the target mode corresponds to the target scene. The processing mode of the target scene.
- the target scene includes a walking scene, a running scene, a quiet scene, a multi-person talking scene, a cafe scene, a subway scene, a train scene, a waiting room scene, a dialogue scene, an office scene, and an outdoor scene , driving scene, windy scene, airplane scene, siren scene, whistle scene, cry scene.
- the communication module 1901 is further configured to send indication information to the terminal device, where the indication information carries the target mode;
- the first processing module 1903 is specifically used for:
- a second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
- the target processing function is an ANC function
- the greater the target processing intensity is the more the user perceives the sound of the current environment of the user and the sound inside the user's ear canal.
- the target processing function is the HT function
- the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
- the target processing function is the AH function
- the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- the earphone is a left earphone, or the earphone is a right earphone.
- the target mode instructs the headset to implement the ANC function
- the first processing module 1903 is specifically configured to:
- the second audio signal is obtained by performing mixing processing on the fourth audio signal and the first audio signal.
- the filter coefficient used in the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the ANC function; or,
- the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the ANC function.
- the target mode indicates that the headset implements the HT function
- the first processing module 1903 is specifically configured to:
- the second audio signal is obtained by performing mixing processing on the third filtered signal and the fifth audio signal.
- the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the HT function; or,
- the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the HT function.
- the target mode indicates that the headset implements the AH function
- the first processing module 1903 is specifically configured to:
- the second audio signal is obtained by mixing the fifth filtered signal, the sixth audio signal and the first filtered signal.
- the filter coefficient used in the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the AH function; or,
- the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the AH function; or,
- the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the AH function.
- the terminal device includes corresponding hardware structures and/or software modules for executing each function.
- modules and method steps of each example described in conjunction with the embodiments disclosed in the present application can be implemented in the form of hardware or a combination of hardware and computer software. Whether a function is performed by hardware or computer software-driven hardware depends on the specific application scenarios and design constraints of the technical solution.
- an embodiment of the present application further provides a mode control apparatus 2000 .
- the mode control apparatus 2000 is applied to the terminal device 100, and can be used to implement the functions of the terminal device in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
- the mode control apparatus 2000 includes a first detection module 2001 and a transmission module 2002 , and may further include a display module 2003 and a second detection module 2004 .
- a first detection module 2001 configured to identify a target mode according to the target scene when the scene type of the current external environment is a target scene;
- the target mode is one of the processing modes supported by the headset, different processing modes correspond to different scene types, and the processing modes supported by the headset include active noise reduction ANC mode, ambient sound transparent transmission HT mode or hearing enhancement At least two of the AH modes.
- the sending module 2002 is configured to send the target mode to the earphone, where the target mode is used to instruct the earphone to implement the processing function corresponding to the target mode.
- it also includes:
- the display module 2003 is configured to display result prompt information when the target mode is determined according to the target scene, and the result prompt information is used to prompt the user that the headset implements the processing function corresponding to the target mode.
- the display module 2003 is configured to display selection prompt information before sending the first control signaling to the headset, where the selection prompt information is used to prompt the user whether to adjust the processing mode of the headset is the target mode;
- the second detection module 2004 is configured to detect an operation in which the user selects the processing mode of the earphone as the target mode.
- the display module 2003 is further configured to display a first control and a second control, wherein different positions of the second control on the first control are used to indicate the target Different processing intensity in mode;
- the second detection module 2004 is further configured to detect the operation of the user touching the second control to move to the first position on the first control before the sending module 2002 sends the first control signaling to the headset , the first position of the second control on the first control indicates the target processing intensity in the target mode;
- the sending module 2002 is further configured to send the target processing intensity to the earphone, where the target processing intensity is used to indicate the processing intensity when the earphone implements the processing function corresponding to the target mode.
- the shape of the first control is a ring
- the processing in the target mode The intensity varies from small to large; or,
- the user touches the second control to move on the first control in a counterclockwise direction, and the processing intensity in the target mode changes from small to large;
- the shape of the first control is a bar
- the processing in the target mode The intensity varies from small to large; or
- the target processing function is an ANC function
- the greater the target processing intensity is the more the user perceives the sound of the current environment of the user and the sound inside the user's ear canal.
- the target processing function is the HT function
- the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
- the target processing function is the AH function
- the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- an embodiment of the present application further provides a mode control apparatus 2100 .
- the mode control apparatus 2100 is applied to the terminal device 100, and can be used to implement the functions of the terminal device in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
- the mode control device 2100 includes a processing module 2101 , a sending module 2102 , a receiving module 2103 , a display module 2104 and a detection module 2105 .
- the processing module 2101 is used to obtain a target mode;
- the target mode is one of the processing modes supported by the headset, and the processing modes supported by the headset include active noise reduction ANC mode, ambient sound transparent transmission HT mode or auditory enhancement AH mode at least two of them;
- the processing module 2101 is further configured to determine the target processing intensity in the target mode according to the scene type of the current external environment; different scene types correspond to different processing intensities in the target mode;
- the sending module 2102 is configured to send the target processing strength to the earphone, where the target processing strength is used to indicate the processing strength when the earphone implements the processing function corresponding to the target mode.
- it also includes:
- a receiving module 2103 configured to receive the target mode sent by the headset.
- it also includes:
- the display module 2104 is configured to display a selection control, where the selection control includes a processing mode supported by the headset, and detects an operation of the user selecting the target mode in the processing mode of the headset through the selection control.
- the display module 2104 is further used for:
- the processing module 2101 determines the target processing intensity in the target mode according to the scene type of the current external environment
- the receiving module 2103 receives the target mode sent by the headset
- a selection prompt is displayed,
- the selection prompt information is used to indicate whether the user adjusts the processing mode of the headset to the target mode;
- the detection module 2105 is configured to detect the operation of the user selecting to adjust the processing mode of the earphone to the target mode.
- the target processing function is an ANC function
- the greater the target processing intensity is the more the user perceives the sound of the current environment of the user and the sound inside the user's ear canal.
- the target processing function is the HT function
- the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
- the target processing function is the AH function
- the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- an embodiment of the present application further provides a mode control apparatus 2200 .
- the mode control apparatus 2200 is applied to the terminal device 100, and can be used to implement the functions of the terminal device in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
- the mode control device 2100 includes a display module 2201 , a detection module 2202 , a transmission module 2203 , and an identification module 2204 .
- the display module 2201 is used to include a first selection control on the first interface, the first selection control includes the processing mode supported by the first target headset and the processing intensity corresponding to the processing mode supported by the first target headset; the first target The processing mode of the headset includes at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode, or the hearing enhancement AH mode;
- the detection module 2202 detects the first operation performed by the user on the first interface; the first operation is that the user selects the first target mode and selects the first target mode from the processing modes supported by the first target headset through the first selection control.
- the processing intensity in a target mode is generated by the first target processing intensity;
- a sending module 2203 configured to send the first target mode and the first target processing intensity to the first target earphone, where the first target mode is used to instruct the first target earphone to achieve the first target
- the processing function corresponding to the mode, and the first target processing intensity is used to indicate the processing intensity when the first target earphone implements the processing function corresponding to the first target mode.
- the display module 2201 is further configured to display selection prompt information before displaying the first interface, where the selection prompt information is used for the user to select whether to adjust the first target headset processing mode;
- the detection module 2202 is further configured to detect the user's operation of selecting and adjusting the processing mode of the first target earphone.
- it also includes:
- the identification module 2204 is configured to identify the scene type of the current external environment as the target scene before the display module 2201 displays the first interface, and the target scene is adapted to the scene type that needs to adjust the processing mode of the first target headset.
- it also includes:
- the identification module 2204 is configured to identify that the terminal device triggers the first target headset to play audio before the display module 2201 displays the first interface.
- the detection module 2202 is further configured to detect that the terminal device establishes a connection with the first target headset before the display module displays the first interface.
- the detection module 2202 before the display module 2201 displays the first interface, detects that the terminal device establishes a connection with the first target headset, and detects the The second operation performed by the user on the main interface.
- the main interface includes an icon of a first application, the second operation is generated by the user touching the icon of the first application, and the first interface is a display interface of the first application.
- the first selection control includes a first control and a second control
- the second control indicates two different processing modes of the first target earphone at any two different positions of the first control , or, the second control indicates different processing intensities in the same processing mode of the first target earphone at any two different positions of the first control;
- the first operation is generated by the user moving a first position of the second control in an area corresponding to the first target mode on the first control, where the first position corresponds to the first The first target processing intensity in target mode.
- the shape of the first control is a circular ring, the circular ring includes at least two arc segments, and the second controls are located in different arc segments to indicate that the first target headset is different processing mode, the second controls are located at different positions of the same arc segment to indicate different processing intensities of the same processing mode of the first target earphone; or,
- the shape of the first control is a bar, the bar includes at least two bar segments, the second controls are located in different bar segments to indicate different processing modes of the first target earphone, and the second control is located in different bar segments. Different positions of the controls on the same bar segment indicate different processing intensities of the same processing mode of the first target earphone.
- the detection module 2202 is further configured to detect a third operation performed by the user on the first interface, the first interface further includes a second selection control, and the second selection control includes The processing mode supported by the second target earphone and the processing intensity corresponding to the processing mode supported by the second target earphone, the processing mode supported by the first target earphone includes active noise reduction ANC mode, ambient sound transparent transmission HT mode or auditory enhancement AH mode
- the third operation is that the user selects the second target mode in the processing mode of the second target headset through the second selection control and selects the processing intensity in the second target mode as the second target
- the processing intensity is generated; when the first target earphone is the left earphone, the second target earphone is the right earphone, or, the first target earphone is the right earphone, and the second target earphone is the left earphone;
- the sending module 2203 is further configured to send the second target mode and the second target processing intensity to the second target headset, where the second target mode is used to instruct the second target headset to implement the The processing function corresponding to the second target mode, and the second target processing intensity is used to indicate the processing intensity when the second target earphone implements the processing function corresponding to the second target mode.
- the terminal device includes a processor 2301 , a memory 2302 , a communication interface 2303 and a display 2304 .
- the memory 2302 is used to store instructions or programs executed by the processor 2301, or input data required by the processor 2301 to execute the instructions or programs, or to store data generated after the processor 2301 executes the instructions or programs.
- the processor 2301 is configured to execute the instructions or programs stored in the memory 2302 to perform the functions performed by the terminal device in the above method.
- the processor 2301 is configured to execute the functions of the first detection module 2001 , the sending module 2002 , the display module 2003 , and the second detection module 2004 .
- the processor 2301 is configured to execute the functions of the first detection module 2001 and the second detection module 2004 .
- the function of the sending module 2002 is realized by the communication interface 2303
- the function of the display module 2003 can be realized by the display 2304 .
- the processing module 2101 , the sending module 2102 , the receiving module 2103 , the display module 2104 and the detection module 2105 may be implemented by the processor 2301 .
- the processor 2301 can be used to execute the functions of the processing module 2101 and the detection module 2105
- the functions of the sending module 2102 and the receiving module 2103 can be implemented by the communication interface 2303
- the functions of the display module 2104 can be implemented by the display 2304 .
- the display module 2201 , the detection module 2202 , the sending module 2203 , and the identification module 2204 may be implemented by the processor 2301 .
- the functions of the detection module 2202 and the identification module 2204 may both be implemented by the processor 2301 .
- the function of the sending module 2203 can be implemented by the communication interface 2303
- the function of the display module 2201 can be implemented by the display 2304 .
- the processor in the embodiment of the present application may be a central processing unit (central processing unit, CPU), and may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSP), application-specific integrated circuits (application specific integrated circuit, ASIC), field programmable gate array (field programmable gate array, FPGA) or other programmable logic devices, transistor logic devices, hardware components or any combination thereof.
- a general-purpose processor may be a microprocessor or any conventional processor.
- the method steps in the embodiments of the present application may be implemented in a hardware manner, or may be implemented in a manner in which a processor executes software instructions.
- the software instructions can be composed of corresponding software modules, and the software modules can be stored in random access memory (RAM), flash memory, read-only memory (ROM), programmable read-only memory (programmable ROM, PROM), erasable programmable read-only memory (erasable PROM, EPROM), electrically erasable programmable read-only memory (electrically ePROM, EEPROM), registers, hard disk, removable hard disk, CD-ROM or any known in the art other forms of storage media.
- An exemplary storage medium is coupled to the processor, such that the processor can read information from, and write information to, the storage medium.
- the storage medium can also be an integral part of the processor.
- the processor and storage medium may reside in an ASIC. Alternatively, the ASIC may be located in the end device. Of course, the processor and the storage medium may also exist in the terminal device as
- the above-mentioned embodiments it may be implemented in whole or in part by software, hardware, firmware or any combination thereof.
- software it can be implemented in whole or in part in the form of a computer program product.
- the computer program product includes one or more computer programs or instructions.
- the processes or functions described in the embodiments of the present application are executed in whole or in part.
- the computer may be a general purpose computer, a special purpose computer, a computer network, user equipment, or other programmable apparatus.
- the computer program or instructions may be stored in a computer-readable storage medium or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer program or instructions may be downloaded from a website site, computer, A server or data center transmits by wire or wireless to another website site, computer, server or data center.
- the computer-readable storage medium may be any available medium that can be accessed by a computer, or a data storage device such as a server, data center, or the like that integrates one or more available media.
- the usable medium can be a magnetic medium, such as a floppy disk, a hard disk, and a magnetic tape; it can also be an optical medium, such as a digital video disc (DVD); it can also be a semiconductor medium, such as a solid state drive (solid state drive). , SSD).
- a magnetic medium such as a floppy disk, a hard disk, and a magnetic tape
- an optical medium such as a digital video disc (DVD)
- DVD digital video disc
- it can also be a semiconductor medium, such as a solid state drive (solid state drive). , SSD).
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Health & Medical Sciences (AREA)
- Circuit For Audible Band Transducer (AREA)
- Soundproofing, Sound Blocking, And Sound Damping (AREA)
- Telephone Function (AREA)
- Details Of Audible-Bandwidth Transducers (AREA)
Abstract
Description
Claims (42)
- 一种耳机噪声处理方法,其特征在于,所述耳机具备主动降噪ANC功能、环境声透传HT功能或听觉增强AH功能中至少两个功能,所述耳机包括第一麦克风和第二麦克风,第一麦克风用于采集第一信号,所述第一信号用于表征当前外部环境的声音,所述第二麦克风用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;所述方法包括:接收来自所述终端设备的第一音频信号;获取目标模式;其中,所述目标模式是基于当前外部环境的场景类型确定的,所述目标模式用于指示所述耳机实现目标处理功能,所述目标处理功能为主动降噪ANC功能、环境声透传HT功能或者听觉增强AH功能中一个;根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
- 如权利要求1所述的方法,其特征在于,所述耳机还包括扬声器,所述扬声器用于播放所述第二音频信号。
- 如权利要求2所述的方法,其特征在于:所述目标处理功能为ANC功能,所述扬声器播放的所述第二音频信号,能够减弱用户对所述用户当前所处环境的声音以及所述用户耳道内部的环境声音的感知;或者,所述目标处理功能为HT功能,所述扬声器播放的所述第二音频信号,能够强化用户对所述用户当前所处环境的声音的感知;或者,所述目标处理功能为AH功能,所述扬声器播放的所述第二音频信号,能够强化用户对事件声音的感知;所述事件声音满足预设频谱。
- 如权利要求1-3任一所述的方法,其特征在于:当所述目标处理功能为ANC功能时,第二音频信号是基于第一音频信号、第三信号以及第四信号得到的,所述第三信号为第一信号的反相信号,所述第四信号为所述第二信号的反相信号;或者,当所述目标处理功能为HT功能时,第二音频信号是基于第一音频信号、第一信号以及第二信号得到的;或者,当所述目标处理功能为AH功能时,第二音频信号是基于第一音频信号、第五信号以及第四信号得到的,所述第五信号为所述第一信号中的事件信号,所述事件信号满足预设频谱。
- 如权利要求1-4任一所述的方法,其特征在于,所述获取目标模式,包括:接收来自所述终端设备的第一控制指令,所述第一控制指令携带所述目标模式,所述目标模式是所述终端设备根据当前外部环境的场景类型确定的。
- 如权利要求5所述的方法,其特征在于,还包括:接收来自所述终端设备的第二控制指令,所述第二控制指令携带目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度;根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号,包括:根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所 述第二信号得到第二音频信号。
- 如权利要求5所述的方法,其特征在于,还包括:根据所述第一信号确定当前外部环境中事件声音对应的目标事件,根据所述目标事件确定在所述目标模式下的目标处理强度;其中,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度,不同的处理强度对应于不同的事件;根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号,包括:根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
- 如权利要求7所述的方法,其特征在于,所述耳机还包括骨传导传感器,所述骨传导传感器用于采集所述用户声带震动产生的骨传导信号;所述根据所述第一信号确定当前外部环境中事件声音对应的目标事件,包括:根据所述第一信号和所述骨传导信号确定当前外部环境中事件声音对应的目标事件。
- 如权利要求8所述的方法,其特征在于,所述目标事件包括啸叫事件、风噪事件、突发事件或者人声事件中的一项。
- 如权利要求1-4任一所述的方法,其特征在于,所述获取目标模式,包括:根据所述第一信号识别当前外部环境的场景类型为目标场景,根据所述目标场景确定所述耳机采用的所述目标模式,所述目标模式是对应于所述目标场景的处理模式,不同的处理模式对应于不同的场景类型。
- 如权利要求10所述的方法,其特征在于,所述目标场景包括走路场景、跑步场景、安静场景、多人说话场景、咖啡厅场景、地铁场景、火车场景、候车厅场景、对话场景、办公室场景、户外场景、开车场景、大风场景、飞机场景、警报声场景、鸣笛声场景、哭叫声场景中的一项。
- 如权利要求10或11所述的方法,其特征在于,还包括:向所述终端设备发送指示信息,所述指示信息携带所述目标模式;接收来自所述终端设备的第三控制信令,所述第三控制信令包括所述目标模式下的目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度;根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号,包括:根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
- 如权利要求6、7或12所述的方法,其特征在于:当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
- 如权利要求1-13任一项所述的方法,其特征在于,所述耳机为左耳机,或者所述耳机为右耳机。
- 如权利要求1-14任一所述的方法,其特征在于,所述目标模式指示所述耳机实现ANC功能,根据所述目标模式、所述第一音频信号、第一信号、以及第二信号得到第二音频信号,包括:对所述第一信号进行所述第一滤波处理得到第一滤波信号;滤除所述第二信号中包括的第一音频信号得到第一经滤除信号;对所述第一滤波信号、所述经滤除信号进行混音处理得到第三音频信号;对所述第三音频信号进行所述第三滤波处理得到第四音频信号;对所述第四音频信号以及所述第一音频信号进行混音处理得到所述第二音频信号。
- 如权利要求15所述的方法,其特征在于:所述第一滤波处理采用的滤波系数为所述ANC功能下,针对所述第一滤波处理,所述目标处理强度所关联的滤波系数;或者,所述第三滤波处理采用的滤波系数为所述ANC功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
- 如权利要求1-14任一所述的方法,其特征在于,所述目标模式指示所述耳机实现HT功能,根据所述目标模式、所述第一音频信号、第一信号、以及第二信号得到第二音频信号,包括:对所述第一信号进行第一信号处理得到第一经处理信号,所述第一信号处理包括第二滤波处理;对所述第一经处理信号和所述第一音频信号进行混音处理得到第五音频信号;滤波所述第二信号包括的所述第五音频信号得到第二经滤除信号;对所述第二经滤除信号进行第三滤波处理得到第三经滤除信号;将所述第三经滤除信号与所述第五音频信号进行混音处理得到所述第二音频信号。
- 如权利要求17所述的方法,其特征在于:所述第二滤波处理采用的滤波系数为所述HT功能下,针对所述第二滤波处理,所述目标处理强度所关联的滤波系数;或者,所述第三滤波处理采用的滤波系数为所述HT功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
- 如权利要求1-14任一所述的方法,其特征在于,目标模式指示所述耳机实现AH功能,根据所述目标模式、所述第一音频信号、第一信号、以及第二信号得到第二音频信号,包括:对所述第一信号进行第二滤波处理得到第二滤波信号,并对所述第二滤波信号进行增强处理得到滤波增强信号;对所述第一信号进行第一滤波处理得到第一滤波信号;对所述滤波增强信号和所述第一音频信号进行混音处理得到第六音频信号;滤波所述第二信号包括的所述第六音频信号得到第四经滤除信号;对所述第四经滤除信号进行第三滤波处理得到第五经滤除信号;对所述第五经滤除信号、所述第六音频信号以及所述第一滤波信号进行混音处理得到所述第二音频信号。
- 如权利要求19所述的方法,其特征在于:所述第一滤波处理采用的滤波系数为所述AH功能下,针对所述第一滤波处理,所述 目标处理强度所关联的滤波系数;或者,所述第二滤波处理采用的滤波系数为所述AH功能下,针对所述第二滤波处理,所述目标处理强度所关联的滤波系数;或者,所述第三滤波处理采用的滤波系数为所述AH功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
- 一种噪声处理装置,其特征在于,所述装置应用于耳机,所述耳机具备主动降噪ANC功能、环境声透传HT功能或听觉增强AH功能中至少两个功能;所述耳机包括第一麦克风、第二麦克风;第一麦克风,用于采集第一信号,所述第一信号用于表征当前外部环境的声音;所述第二麦克风,用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;所述装置包括:通信模块,用于接收来自所述终端设备的第一音频信号;获取模块,用于获取目标模式;其中,所述目标模式是基于当前外部环境的场景类型确定的,所述目标模式用于指示所述耳机实现目标处理功能,所述目标处理功能为主动降噪ANC功能、环境声透传HT功能或者听觉增强AH功能中一个;第一处理模块,用于根据所述目标模式、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
- 如权利要求21所述的装置,其特征在于,所述装置还包括:播放模块,用于播放所述第二音频信号。
- 如权利要求22所述的装置,其特征在于:所述目标处理功能为ANC功能,所述播放模块播放的所述第二音频信号,能够减弱用户对所述用户当前所处环境的声音以及所述用户耳道内部的环境声音的感知;或者,所述目标处理功能为HT功能,所述播放模块播放的所述第二音频信号,能够强化用户对所述用户当前所处环境的声音的感知;或者,所述目标处理功能为AH功能,所述播放模块播放的所述第二音频信号,能够强化用户对事件声音的感知;所述事件声音为满足预设频谱。
- 如权利要求21-23任一所述的装置,其特征在于:当所述目标处理功能为ANC功能时,第二音频信号是基于第一音频信号、第三信号以及第四信号得到的,所述第三信号为第一信号的反相信号,所述第四信号为所述第二信号的反相信号;或者,当所述目标处理功能为HT功能时,第二音频信号是基于第一音频信号、第一信号以及第二信号得到的;或者,当所述目标处理功能为AH功能时,第二音频信号是基于第一音频信号、第五信号以及第四信号得到的,所述第五信号为所述第一信号中的事件信号,所述事件信号满足预设频谱。
- 如权利要求21-24任一所述的装置,其特征在于,所述通信模块,还用于接收来自所述终端设备的第一控制指令,所述第一控制指令携带所述目标模式,所述目标模式是所述终端设备根据当前外部环境的场景类型确定的;将所述目标模式发送给所述获取模块。
- 如权利要求25所述的装置,其特征在于,所述通信模块,还用于:接收来自所述终端设备的第二控制指令,所述第二控制指令携带目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度;所述第一处理模块,具体用于:根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
- 如权利要求26所述的装置,其特征在于,还包括:第二处理模块,用于根据所述第一信号确定当前外部环境中事件声音对应的目标事件,根据所述目标事件确定在所述目标模式下的目标处理强度;其中,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度,不同的处理强度对应于不同的事件;所述第一处理模块具体用于:根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
- 如权利要求27所述的装置,其特征在于,所述耳机还包括骨传导传感器,所述骨传导传感器用于采集所述用户声带震动产生的骨传导信号;所述第一处理模块,具体用于根据所述第一信号和所述骨传导信号确定当前外部环境中事件声音对应的目标事件。
- 如权利要求28所述的装置,其特征在于,所述目标事件为啸叫事件、风噪事件、突发事件或者人声事件。
- 如权利要求21-24任一所述的装置,其特征在于,还包括:第三处理模块,用于根据所述第一信号识别当前外部环境的场景类型为目标场景,根据所述目标场景确定所述耳机采用的所述目标模式,所述目标模式是对应于所述目标场景的处理模式,不同的处理模式对应于不同的场景类型。
- 如权利要求30所述的装置,其特征在于,所述目标场景为走路场景、跑步场景、安静场景、多人说话场景、咖啡厅场景、地铁场景、火车场景、候车厅场景、对话场景、办公室场景、户外场景、开车场景、大风场景、飞机场景、警报声场景、鸣笛声场景、哭叫声场景中的一项。
- 如权利要求30或31所述的装置,其特征在于,所述通信模块,还用于向所述终端设备发送指示信息,所述指示信息携带所述目标模式;接收来自所述终端设备的第三控制信令,所述第三控制信令包括所述目标模式下的目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标处理功能时的处理强度;所述第一处理模块,具体用于:根据所述目标模式、所述目标处理强度、所述第一音频信号、所述第一信号、以及所述第二信号得到第二音频信号。
- 如权利要求26、27或32所述的装置,其特征在于:当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
- 如权利要求21-33任一项所述的装置,其特征在于,所述耳机为左耳机,或者所述耳机为右耳机。
- 如权利要求21-34任一所述的装置,其特征在于,所述目标模式指示所述耳机实现ANC功能,所述第一处理模块,具体用于:对所述第一信号进行所述第一滤波处理得到第一滤波信号;滤除所述第二信号中包括的第一音频信号得到第一经滤除信号;对所述第一滤波信号、所述经滤除信号进行混音处理得到第三音频信号;对所述第三音频信号进行所述第三滤波处理得到第四音频信号;对所述第四音频信号以及所述第一音频信号进行混音处理得到所述第二音频信号。
- 如权利要求35所述的装置,其特征在于:所述第一滤波处理采用的滤波系数为所述ANC功能下,针对所述第一滤波处理,所述目标处理强度所关联的滤波系数;或者,所述第三滤波处理采用的滤波系数为所述ANC功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
- 如权利要求21-34任一所述的装置,其特征在于,所述目标模式指示所述耳机实现HT功能,所述第一处理模块,具体用于:对所述第一信号进行第一信号处理得到第一经处理信号,所述第一信号处理包括第二滤波处理;对所述第一经处理信号和所述第一音频信号进行混音处理得到第五音频信号;滤波所述第二信号包括的所述第五音频信号得到第二经滤除信号;对所述第二经滤除信号进行第三滤波处理得到第三经滤除信号;将所述第三经滤除信号与所述第五音频信号进行混音处理得到所述第二音频信号。
- 如权利要求37所述的装置,其特征在于:所述第二滤波处理采用的滤波系数为所述HT功能下,针对所述第二滤波处理,所述目标处理强度所关联的滤波系数;或者,所述第三滤波处理采用的滤波系数为所述HT功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
- 如权利要求21-34任一所述的装置,其特征在于,目标模式指示所述耳机实现AH功能,所述第一处理模块,具体用于:对所述第一信号进行第二滤波处理得到第二滤波信号,并对所述第二滤波信号进行增强处理得到滤波增强信号;对所述第一信号进行第一滤波处理得到第一滤波信号;对所述滤波增强信号和所述第一音频信号进行混音处理得到第六音频信号;滤波所述第二信号包括的所述第六音频信号得到第四经滤除信号;对所述第四经滤除信号进行第三滤波处理得到第五经滤除信号;对所述第五经滤除信号、所述第六音频信号以及所述第一滤波信号进行混音处理得到所述第二音频信号。
- 如权利要求39所述的装置,其特征在于:所述第一滤波处理采用的滤波系数为所述AH功能下,针对所述第一滤波处理,所述目标处理强度所关联的滤波系数;或者,所述第二滤波处理采用的滤波系数为所述AH功能下,针对所述第二滤波处理,所述目标处理强度所关联的滤波系数;或者,所述第三滤波处理采用的滤波系数为所述AH功能下,针对所述第三滤波处理,所述目标处理强度所关联的滤波系数。
- 一种目标耳机,其特征在于,所述目标耳机包括左耳机和右耳机;所述左耳机用于实现权利要求1-20任一所述的方法,或者,所述右耳机用于实现权利要求1-20任一所述的方法。
- 一种目标耳机,其特征在于,包括第一麦克风、第二麦克风、处理器、存储器以及扬声器;所述第一麦克风,用于采集第一信号,所述第一信号用于表征当前外部环境的声音;所述第二麦克风,用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;所述存储器,用于存储程序或指令;所述处理器,用于调用所述程序或指令,以使得所述目标耳机针对所述第一信号和第二信号执行如权利要求1至20任一项所述的方法得到第二音频信号;所述扬声器,用于播放所述第二音频信号。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP21834462.0A EP4167590A4 (en) | 2020-06-30 | 2021-06-30 | EARPHONE NOISE PROCESSING METHOD AND DEVICE, AND EARPHONE |
JP2022580546A JP2023532078A (ja) | 2020-06-30 | 2021-06-30 | ヘッドセットノイズ処理方法、装置及びヘッドセット |
US18/148,116 US20230134787A1 (en) | 2020-06-30 | 2022-12-29 | Headset Noise Processing Method, Apparatus, and Headset |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010623983.X | 2020-06-30 | ||
CN202010623983.XA CN113873378B (zh) | 2020-06-30 | 2020-06-30 | 一种耳机噪声处理方法、装置及耳机 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/148,116 Continuation US20230134787A1 (en) | 2020-06-30 | 2022-12-29 | Headset Noise Processing Method, Apparatus, and Headset |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022002166A1 true WO2022002166A1 (zh) | 2022-01-06 |
Family
ID=78982040
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/103768 WO2022002166A1 (zh) | 2020-06-30 | 2021-06-30 | 一种耳机噪声处理方法、装置及耳机 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20230134787A1 (zh) |
EP (1) | EP4167590A4 (zh) |
JP (1) | JP2023532078A (zh) |
CN (1) | CN113873378B (zh) |
WO (1) | WO2022002166A1 (zh) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114466278A (zh) * | 2022-04-11 | 2022-05-10 | 荣耀终端有限公司 | 一种耳机模式对应的参数确定方法、耳机、终端和系统 |
CN115835079A (zh) * | 2022-11-21 | 2023-03-21 | 荣耀终端有限公司 | 透传模式的切换方法和切换装置 |
CN116156385A (zh) * | 2023-04-19 | 2023-05-23 | 深圳市汇顶科技股份有限公司 | 滤波方法、滤波装置、芯片和耳机 |
WO2023160286A1 (zh) * | 2022-02-28 | 2023-08-31 | 荣耀终端有限公司 | 降噪参数适配方法和装置 |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113507662B (zh) * | 2021-06-29 | 2023-01-10 | 锐迪科微电子(上海)有限公司 | 降噪处理方法、装置、设备、存储介质及程序 |
CN114630239B (zh) * | 2022-02-23 | 2024-08-13 | 北京小米移动软件有限公司 | 降低耳机闭塞效应方法、装置及存储介质 |
CN116709116A (zh) * | 2022-02-28 | 2023-09-05 | 北京荣耀终端有限公司 | 声音信号的处理方法及耳机设备 |
CN115002595B (zh) * | 2022-04-21 | 2023-03-28 | 广东华冠智联科技有限公司 | 耳机的节能控制方法、装置、设备及存储介质 |
CN114640938B (zh) * | 2022-05-18 | 2022-08-23 | 深圳市听多多科技有限公司 | 一种基于蓝牙耳机芯片的助听功能实现方法及蓝牙耳机 |
CN115206278A (zh) * | 2022-06-08 | 2022-10-18 | 荣耀终端有限公司 | 一种声音降噪的方法和装置 |
WO2023245390A1 (zh) * | 2022-06-20 | 2023-12-28 | 北京小米移动软件有限公司 | 智能耳机的控制方法、装置、电子设备和存储介质 |
CN115134708B (zh) * | 2022-06-30 | 2024-07-30 | 歌尔股份有限公司 | 耳机模式切换方法、装置、电子设备及可读存储介质 |
CN117528370A (zh) * | 2022-07-30 | 2024-02-06 | 华为技术有限公司 | 信号处理方法及装置、设备控制方法及装置 |
CN117676406A (zh) * | 2022-08-30 | 2024-03-08 | 博通集成电路(上海)股份有限公司 | 主动降噪耳机 |
CN115379340A (zh) * | 2022-09-13 | 2022-11-22 | 湖南炬神电子有限公司 | 一种消噪运动蓝牙耳机 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140169603A1 (en) * | 2012-12-19 | 2014-06-19 | Starkey Laboratories, Inc. | Hearing assistance device vent valve |
CN206698392U (zh) * | 2017-02-14 | 2017-12-01 | 歌尔股份有限公司 | 降噪耳机以及电子设备 |
CN107517415A (zh) * | 2016-06-16 | 2017-12-26 | 峰范(北京)科技有限公司 | 智能耳机 |
CN108475502A (zh) * | 2015-12-30 | 2018-08-31 | 美商楼氏电子有限公司 | 语音增强感知模式 |
CN112770214A (zh) * | 2021-01-28 | 2021-05-07 | 歌尔科技有限公司 | 耳机的控制方法、装置及耳机 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DK1439734T3 (da) * | 2004-02-10 | 2009-07-27 | Phonak Ag | Akustisk modforvrænger for höreapparater |
AU2015395527B2 (en) * | 2015-05-15 | 2018-10-18 | Huawei Technologies Co., Ltd. | Noise reduction headset setting method, terminal, and noise reduction headset |
US10978041B2 (en) * | 2015-12-17 | 2021-04-13 | Huawei Technologies Co., Ltd. | Ambient sound processing method and device |
EP3188495B1 (en) * | 2015-12-30 | 2020-11-18 | GN Audio A/S | A headset with hear-through mode |
KR101756674B1 (ko) * | 2016-05-27 | 2017-07-25 | 주식회사 이엠텍 | 보청기 기능의 능동 소음 제거 헤드셋 장치 |
CN106792315B (zh) * | 2017-01-05 | 2023-11-21 | 歌尔科技有限公司 | 一种抵消环境噪声的方法和装置及一种主动降噪耳机 |
US10129633B1 (en) * | 2017-10-13 | 2018-11-13 | Bose Corporation | Automated awareness for ANR systems |
CN108391206A (zh) * | 2018-03-30 | 2018-08-10 | 广东欧珀移动通信有限公司 | 信号处理方法、装置、终端、耳机及可读存储介质 |
-
2020
- 2020-06-30 CN CN202010623983.XA patent/CN113873378B/zh active Active
-
2021
- 2021-06-30 WO PCT/CN2021/103768 patent/WO2022002166A1/zh active Application Filing
- 2021-06-30 JP JP2022580546A patent/JP2023532078A/ja active Pending
- 2021-06-30 EP EP21834462.0A patent/EP4167590A4/en active Pending
-
2022
- 2022-12-29 US US18/148,116 patent/US20230134787A1/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140169603A1 (en) * | 2012-12-19 | 2014-06-19 | Starkey Laboratories, Inc. | Hearing assistance device vent valve |
CN108475502A (zh) * | 2015-12-30 | 2018-08-31 | 美商楼氏电子有限公司 | 语音增强感知模式 |
CN107517415A (zh) * | 2016-06-16 | 2017-12-26 | 峰范(北京)科技有限公司 | 智能耳机 |
CN206698392U (zh) * | 2017-02-14 | 2017-12-01 | 歌尔股份有限公司 | 降噪耳机以及电子设备 |
CN112770214A (zh) * | 2021-01-28 | 2021-05-07 | 歌尔科技有限公司 | 耳机的控制方法、装置及耳机 |
Non-Patent Citations (1)
Title |
---|
See also references of EP4167590A4 |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2023160286A1 (zh) * | 2022-02-28 | 2023-08-31 | 荣耀终端有限公司 | 降噪参数适配方法和装置 |
CN114466278A (zh) * | 2022-04-11 | 2022-05-10 | 荣耀终端有限公司 | 一种耳机模式对应的参数确定方法、耳机、终端和系统 |
CN114466278B (zh) * | 2022-04-11 | 2022-08-16 | 北京荣耀终端有限公司 | 一种耳机模式对应的参数确定方法、耳机、终端和系统 |
CN115835079A (zh) * | 2022-11-21 | 2023-03-21 | 荣耀终端有限公司 | 透传模式的切换方法和切换装置 |
CN115835079B (zh) * | 2022-11-21 | 2023-08-08 | 荣耀终端有限公司 | 透传模式的切换方法和切换装置 |
CN116156385A (zh) * | 2023-04-19 | 2023-05-23 | 深圳市汇顶科技股份有限公司 | 滤波方法、滤波装置、芯片和耳机 |
CN116156385B (zh) * | 2023-04-19 | 2023-07-07 | 深圳市汇顶科技股份有限公司 | 滤波方法、滤波装置、芯片和耳机 |
Also Published As
Publication number | Publication date |
---|---|
EP4167590A4 (en) | 2023-12-13 |
JP2023532078A (ja) | 2023-07-26 |
CN113873378B (zh) | 2023-03-10 |
CN113873378A (zh) | 2021-12-31 |
EP4167590A1 (en) | 2023-04-19 |
US20230134787A1 (en) | 2023-05-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022002166A1 (zh) | 一种耳机噪声处理方法、装置及耳机 | |
WO2022002110A1 (zh) | 一种模式控制方法、装置及终端设备 | |
CN113169760B (zh) | 无线短距离音频共享方法及电子设备 | |
WO2021083128A1 (zh) | 一种声音处理方法及其装置 | |
CN112119641B (zh) | 通过转发模式连接的多tws耳机实现自动翻译的方法及装置 | |
US20230041696A1 (en) | Image Syntheis Method, Electronic Device, and Non-Transitory Computer-Readable Storage Medium | |
WO2021000817A1 (zh) | 环境音处理方法及相关装置 | |
WO2024027259A1 (zh) | 信号处理方法及装置、设备控制方法及装置 | |
CN113571035B (zh) | 降噪方法及降噪装置 | |
CN114466107A (zh) | 音效控制方法、装置、电子设备及计算机可读存储介质 | |
CN111065020B (zh) | 音频数据处理的方法和装置 | |
WO2022206825A1 (zh) | 一种调节音量的方法、系统及电子设备 | |
WO2022257563A1 (zh) | 一种音量调节的方法,电子设备和系统 | |
CN114339429A (zh) | 音视频播放控制方法、电子设备和存储介质 | |
WO2022089563A1 (zh) | 一种声音增强方法、耳机控制方法、装置及耳机 | |
WO2024046416A1 (zh) | 一种音量调节方法、电子设备及系统 | |
CN113938556B (zh) | 来电提示方法、装置和电子设备 | |
WO2024046182A1 (zh) | 一种音频播放方法、系统及相关装置 | |
WO2023020420A1 (zh) | 音量显示方法、电子设备及存储介质 | |
WO2024032035A9 (zh) | 一种语音信号的输出方法和电子设备 | |
WO2024066933A9 (zh) | 扬声器控制方法及设备 | |
CN118466894A (zh) | 音效控制的方法、装置和电子设备 | |
CN115378303A (zh) | 驱动波形的调整方法及装置、电子设备、可读存储介质 | |
CN118098261A (zh) | 一种音频处理方法、设备及系统 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21834462 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2022580546 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 202327001384 Country of ref document: IN |
|
ENP | Entry into the national phase |
Ref document number: 2021834462 Country of ref document: EP Effective date: 20230113 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |