WO2022002110A1 - 一种模式控制方法、装置及终端设备 - Google Patents
一种模式控制方法、装置及终端设备 Download PDFInfo
- Publication number
- WO2022002110A1 WO2022002110A1 PCT/CN2021/103435 CN2021103435W WO2022002110A1 WO 2022002110 A1 WO2022002110 A1 WO 2022002110A1 CN 2021103435 W CN2021103435 W CN 2021103435W WO 2022002110 A1 WO2022002110 A1 WO 2022002110A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- target
- processing
- function
- signal
- anc
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 289
- 238000012545 processing Methods 0.000 claims abstract description 1069
- 230000000694 effects Effects 0.000 claims abstract description 78
- 230000006870 function Effects 0.000 claims description 691
- 230000005236 sound signal Effects 0.000 claims description 214
- 238000001514 detection method Methods 0.000 claims description 211
- 230000009467 reduction Effects 0.000 claims description 159
- 230000008569 process Effects 0.000 claims description 101
- 230000011664 signaling Effects 0.000 claims description 72
- 230000006854 communication Effects 0.000 claims description 71
- 238000004891 communication Methods 0.000 claims description 70
- 230000005540 biological transmission Effects 0.000 claims description 64
- 230000015654 memory Effects 0.000 claims description 52
- 210000000613 ear canal Anatomy 0.000 claims description 44
- 230000003044 adaptive effect Effects 0.000 claims description 37
- 230000004044 response Effects 0.000 claims description 28
- 238000004590 computer program Methods 0.000 claims description 20
- 230000003595 spectral effect Effects 0.000 claims description 11
- 230000002708 enhancing effect Effects 0.000 claims description 7
- 238000003672 processing method Methods 0.000 claims description 7
- 230000003213 activating effect Effects 0.000 claims description 6
- 230000005484 gravity Effects 0.000 claims description 5
- 230000001976 improved effect Effects 0.000 claims description 5
- 230000001960 triggered effect Effects 0.000 claims description 5
- 230000003313 weakening effect Effects 0.000 claims description 2
- 238000001914 filtration Methods 0.000 description 145
- 238000013461 design Methods 0.000 description 91
- 210000000988 bone and bone Anatomy 0.000 description 72
- 238000010586 diagram Methods 0.000 description 47
- 238000002156 mixing Methods 0.000 description 31
- 230000008447 perception Effects 0.000 description 22
- 230000003321 amplification Effects 0.000 description 21
- 238000007726 management method Methods 0.000 description 21
- 238000003199 nucleic acid amplification method Methods 0.000 description 21
- 238000003860 storage Methods 0.000 description 20
- 238000013473 artificial intelligence Methods 0.000 description 16
- 238000004422 calculation algorithm Methods 0.000 description 14
- 238000001228 spectrum Methods 0.000 description 14
- 230000008859 change Effects 0.000 description 11
- 238000010295 mobile communication Methods 0.000 description 11
- 206010011469 Crying Diseases 0.000 description 10
- 238000005516 engineering process Methods 0.000 description 10
- 230000007613 environmental effect Effects 0.000 description 10
- 238000013507 mapping Methods 0.000 description 9
- 238000013528 artificial neural network Methods 0.000 description 8
- 230000002829 reductive effect Effects 0.000 description 6
- 238000010079 rubber tapping Methods 0.000 description 6
- 238000012986 modification Methods 0.000 description 5
- 230000004048 modification Effects 0.000 description 5
- 238000010183 spectrum analysis Methods 0.000 description 5
- 230000001755 vocal effect Effects 0.000 description 5
- 230000002159 abnormal effect Effects 0.000 description 4
- 230000009286 beneficial effect Effects 0.000 description 4
- 230000037361 pathway Effects 0.000 description 4
- 229920001621 AMOLED Polymers 0.000 description 3
- 208000006735 Periostitis Diseases 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 3
- 230000006978 adaptation Effects 0.000 description 3
- 208000016354 hearing loss disease Diseases 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 210000003460 periosteum Anatomy 0.000 description 3
- 238000003825 pressing Methods 0.000 description 3
- 206010011878 Deafness Diseases 0.000 description 2
- 230000004913 activation Effects 0.000 description 2
- 238000013459 approach Methods 0.000 description 2
- 230000003416 augmentation Effects 0.000 description 2
- 210000000721 basilar membrane Anatomy 0.000 description 2
- 230000036772 blood pressure Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000012790 confirmation Methods 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 210000005069 ears Anatomy 0.000 description 2
- 230000014509 gene expression Effects 0.000 description 2
- 230000010370 hearing loss Effects 0.000 description 2
- 231100000888 hearing loss Toxicity 0.000 description 2
- 230000001965 increasing effect Effects 0.000 description 2
- 230000000977 initiatory effect Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000033001 locomotion Effects 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 239000002096 quantum dot Substances 0.000 description 2
- 230000005855 radiation Effects 0.000 description 2
- 238000011946 reduction process Methods 0.000 description 2
- 238000009877 rendering Methods 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 230000019491 signal transduction Effects 0.000 description 2
- 239000007787 solid Substances 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 210000001260 vocal cord Anatomy 0.000 description 2
- 101100006960 Caenorhabditis elegans let-2 gene Proteins 0.000 description 1
- 101100127891 Caenorhabditis elegans let-4 gene Proteins 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 230000002238 attenuated effect Effects 0.000 description 1
- 238000010009 beating Methods 0.000 description 1
- 230000007175 bidirectional communication Effects 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 238000013529 biological neural network Methods 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 210000003477 cochlea Anatomy 0.000 description 1
- 230000019771 cognition Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000003862 health status Effects 0.000 description 1
- 230000001939 inductive effect Effects 0.000 description 1
- 239000010985 leather Substances 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 230000000873 masking effect Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000011002 quantification Methods 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000006403 short-term memory Effects 0.000 description 1
- 230000008054 signal transmission Effects 0.000 description 1
- 230000003238 somatosensory effect Effects 0.000 description 1
- 230000006641 stabilisation Effects 0.000 description 1
- 238000011105 stabilization Methods 0.000 description 1
- 230000008093 supporting effect Effects 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 230000029305 taxis Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1041—Mechanical or electronic switches, or control elements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1016—Earpieces of the intra-aural type
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K11/00—Methods or devices for transmitting, conducting or directing sound in general; Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/16—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general
- G10K11/175—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound
- G10K11/178—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase
- G10K11/1781—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase characterised by the analysis of input or output signals, e.g. frequency range, modes, transfer functions
- G10K11/17821—Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase characterised by the analysis of input or output signals, e.g. frequency range, modes, transfer functions characterised by the analysis of the input signals only
- G10K11/17823—Reference signals, e.g. ambient acoustic environment
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/08—Mouthpieces; Microphones; Attachments therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1083—Reduction of ambient noise
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/43—Electronic input selection or mixing based on input signal analysis, e.g. mixing or selection between microphone and telecoil or between microphones with different directivity characteristics
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/55—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired
- H04R25/558—Remote control, e.g. of amplification, frequency
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/005—Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R5/00—Stereophonic arrangements
- H04R5/04—Circuit arrangements, e.g. for selective connection of amplifier inputs/outputs to loudspeakers, for loudspeaker detection, or for adaptation of settings to personal preferences or hearing impairments
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2201/00—Details of transducers, loudspeakers or microphones covered by H04R1/00 but not provided for in any of its subgroups
- H04R2201/10—Details of earpieces, attachments therefor, earphones or monophonic headphones covered by H04R1/10 but not provided for in any of its subgroups
- H04R2201/109—Arrangements to adapt hands free headphones for use on both ears
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2225/00—Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
- H04R2225/41—Detection or adaptation of hearing aid parameters or programs to listening situation, e.g. pub, forest
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2225/00—Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
- H04R2225/43—Signal processing in hearing aids to enhance the speech intelligibility
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2410/00—Microphones
- H04R2410/05—Noise reduction with a separate noise microphone
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2420/00—Details of connection covered by H04R, not provided for in its groups
- H04R2420/01—Input selection or mixing for amplifiers or loudspeakers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2430/00—Signal processing covered by H04R, not provided for in its groups
- H04R2430/01—Aspects of volume control, not necessarily automatic, in sound systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2430/00—Signal processing covered by H04R, not provided for in its groups
- H04R2430/03—Synergistic effects of band splitting and sub-band processing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2460/00—Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
- H04R2460/01—Hearing devices using active noise cancellation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2460/00—Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
- H04R2460/03—Aspects of the reduction of energy consumption in hearing devices
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2460/00—Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
- H04R2460/05—Electronic compensation of the occlusion effect
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2460/00—Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
- H04R2460/13—Hearing devices using bone conduction transducers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/50—Customised settings for obtaining desired overall acoustical characteristics
- H04R25/505—Customised settings for obtaining desired overall acoustical characteristics using digital signal processing
Definitions
- the embodiments of the present application relate to the technical field of audio processing, and in particular, to a mode control method, apparatus, and terminal device.
- headset users In recent years, there have been more and more headset users, and users' functional requirements for headsets have become more and more differentiated. For example, some users do not want to hear external noise when wearing headphones, and can use the active noise control (ANC) function to eliminate the noise in the ears. Some users want to hear the sound outside the earphone, and need to use the ambient sound through (hear through, HT) function to realize the same feeling of the outside sound as without the earphone. Some users may have hearing impairments, and the external signals that the user wants can be delivered to the user through the augmentation hearing (AH) function, and the unwanted signals can be filtered out at the same time.
- ANC active noise control
- HT ambient sound through
- HT ambient sound through
- Some users may have hearing impairments, and the external signals that the user wants can be delivered to the user through the augmentation hearing (AH) function, and the unwanted signals can be filtered out at the same time.
- AH augmentation hearing
- the headset cannot achieve the effect that the user wants to achieve according to the user's needs.
- Embodiments of the present application provide an earphone noise processing method, device, and earphone, so as to achieve the effect that the user wants to achieve according to the user's needs.
- an embodiment of the present application provides a noise processing method for an earphone, the earphone has at least two functions of an active noise reduction ANC function, an ambient sound transparent transmission HT function, or an auditory enhancement AH function, and the earphone includes a first microphone and a second microphone, the first microphone is used to collect the first signal, the first signal is used to characterize the sound of the current external environment, the second microphone is used to collect the second signal, and the second signal is used to characterize the wearing The sound of the environment inside the ear canal of the user of the headset.
- the earphone can be a left earphone or a right earphone.
- the processing modes used by the left earphone and the right earphone can be the same or different.
- the earphone receives the first audio signal from the terminal device; obtains a target mode; wherein, the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the earphone to implement the target processing function, so
- the target processing function is one of the active noise reduction ANC function, the ambient sound transparent transmission HT function or the auditory enhancement AH function; according to the target mode, the first audio signal, the first signal, and the second signal A second audio signal is obtained.
- the target mode is determined according to the scene type of the external environment of the field, and the user's listening effect can be optimized in real time.
- the earphone further includes a speaker, and the speaker is used to play the second audio signal.
- the target processing function is an ANC function
- the second audio signal played by the speaker can reduce the sound of the user's current environment and the sound inside the user's ear canal. Perception of ambient sound; or, the target processing function is an HT function, and the second audio signal played by the speaker can enhance the user's perception of the sound of the environment where the user is currently located; or, the target processing function
- the function is the AH function, and the second audio signal played by the speaker can enhance the user's perception of the event sound; the event sound satisfies a preset frequency spectrum.
- the audio signal played by the speaker of the left earphone can attenuate the sound of the user's left earphone to the current environment of the user (that is, the sound of the current external environment) and the sound inside the ear canal of the user's left earphone. Perception of ambient sound.
- the audio signal played by the speaker of the right earphone can weaken the perception of the user's right earphone to the sound of the user's current environment (that is, the sound of the current external environment) and the ambient sound inside the ear canal of the user's right earphone.
- the feeling of the left ear follows the processing mode adopted by the left earphone
- the feeling of the right ear follows the processing mode adopted by the right earphone.
- the target processing function is the ANC function
- the second audio signal is obtained based on the first audio signal, the third signal and the fourth signal
- the third signal is the first signal.
- the fourth signal is an inverted signal of the second signal; or, when the target processing function is the HT function, the second audio signal is based on the first audio signal, the first signal and the second audio signal or, when the target processing function is the AH function, the second audio signal is obtained based on the first audio signal, the fifth signal and the fourth signal, and the fifth signal is one of the first audio signals.
- the event signal is used to represent a specific sound in the current external environment, and the event signal satisfies a preset frequency spectrum.
- the above design provides a simple and effective way of obtaining the signal output by the speaker under different processing modes.
- the acquiring target mode includes:
- a first control instruction from the terminal device is received, where the first control instruction carries the target mode, and the target mode is determined by the terminal device according to the scene type of the current external environment.
- the terminal device determines the target mode according to the scene type of the external environment and indicates to the earphone, which can optimize the user's listening effect in real time.
- a second control instruction from the terminal device, where the second control instruction carries a target processing strength, and the target processing strength is used to instruct the headset to perform the target processing function processing strength; obtaining a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal, including: according to the target mode, the target processing strength, the The first audio signal, the first signal, and the second signal result in a second audio signal.
- the terminal device indicates the processing intensity in the corresponding processing mode of the earphone. Adjusting the processing intensity on the basis of the processing mode further improves the user's hearing experience.
- the target event corresponding to the event sound in the current external environment is determined according to the first signal, and the target processing intensity in the target mode is determined according to the target event; wherein, the target processing intensity It is used to indicate the processing strength when the earphone realizes the target processing function; obtaining a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal, including: A second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal. where different processing intensities correspond to different events.
- the processing intensity can be in one-to-one correspondence with events, or one processing intensity corresponds to multiple events. For example, two events can use the same processing intensity, but different processing intensity cannot be used for the same event.
- the earphone determines the processing intensity according to the event sound in the external environment, realizes different auditory experience in different external environments, can reduce the sense of bottom noise and enhance the noise reduction.
- the headset further includes a bone conduction sensor, and the bone conduction sensor is used to collect bone conduction signals generated by the vibration of the user's vocal cords;
- the first scene at the location includes: identifying the first scene where the user is currently located according to the first signal and the bone conduction signal.
- the target event is a howling event, a wind noise event, a sudden event or a human voice event.
- the acquiring the target mode includes: identifying, according to the first signal, the scene type of the current external environment as the target scene type (target scene for short, or target type for short), and according to the target scene
- the target mode adopted by the headset is determined, where the target mode is a processing mode corresponding to the target scene.
- the different processing modes correspond to different scene types.
- the processing modes may be in one-to-one correspondence with scene types, or one processing mode may correspond to multiple scene types, for example, two scene types may adopt the same processing mode.
- the earphone determines the processing mode adopted by the earphone according to the recognized scene type, reduces the time delay, and optimizes the user's hearing experience in real time.
- the target scene is a walking scene, a running scene, a quiet scene, a multi-person talking scene, a cafe scene, a subway scene, a train scene, a waiting room scene, a dialogue scene, an office scene, an outdoor scene, One of the driving scene, windy scene, airplane scene, siren scene, whistle scene, and crying scene.
- the audio signal includes: obtaining a second audio signal according to the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
- the earphone determines the processing mode to be adopted, and instructs the terminal device to adjust the processing intensity and reduce the occupation of processing resources of the earphone.
- the target processing function is an ANC function
- the greater the target processing intensity is, the more the user perceives the sound of the current environment of the user and the environment inside the user's ear canal The weaker the sound; or, when the target processing function is the HT function, the greater the target processing intensity, the greater the intensity of the sound of the environment where the user is currently perceived by the user; or, when the target When the processing function is the AH function, the greater the target processing intensity, the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- the target mode instructs the headset to implement an ANC function, and obtaining a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal, including:
- the second audio signal is obtained by performing mixing processing on the fourth audio signal and the first audio signal.
- FF filtering and FB serial processing are used to implement ANC processing to obtain better noise-reduced signals and enhance the noise-reduction effect.
- the filter coefficient used in the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the ANC function; or, the third The filter coefficient used in the filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the ANC function.
- the target mode instructs the earphone to implement the HT function
- the second audio signal is obtained according to the target mode, the first audio signal, the first signal, and the second signal, including:
- the first signal processing including a second filtering process (such as HT filtering);
- a third filtering process (such as FB filtering) on the second filtered signal to obtain a third filtered signal
- the second audio signal is obtained by performing mixing processing on the third filtered signal and the fifth audio signal.
- filtering and supplementary processing may be performed on the fifth audio signal to reduce hearing loss.
- the HT filter undergoes down-mixing processing and filter compensation processing to further reduce hearing loss.
- performing first signal processing on the first environmental signal to obtain a processed environmental signal includes: performing a second filtering process on the first signal to obtain a second filtered signal;
- the second signal processing includes de-blocking effect processing.
- the second signal processing further includes at least one of the following: noise reduction processing, wind noise reduction processing, gain adjustment processing or frequency response adjustment processing.
- the noise floor and abnormal sounds are reduced, and the user's hearing experience is improved.
- the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the HT function; or,
- the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the HT function.
- the target mode instructs the headset to implement the AH function
- the second audio signal is obtained according to the target mode, the first audio signal, the first signal, and the second signal, including:
- Second filtering processing (such as HT filtering) on the first signal to obtain a second filtering signal
- enhancement processing on the second filtering signal to obtain a filtering enhanced signal
- Performing a first filtering process (such as FF filtering) on the first signal to obtain a first filtered signal
- the second audio signal is obtained by mixing the fifth filtered signal, the sixth audio signal and the first filtered signal.
- the transparently transmitted signal is processed by transparent transmission filtering and enhancement processing, and the transparently transmitted signal is clearer.
- filtering and compensating the sixth audio signal can avoid losses caused by FB filtering and ensure transparency to the greatest extent. Signal transmission without distortion.
- performing enhancement processing on the second filtered signal to obtain a filtered enhanced signal including:
- noise reduction processing includes artificial intelligence AI noise reduction processing and/or wind noise reduction processing
- the filtered enhanced signal is obtained by performing gain amplification processing and frequency response adjustment on the signal obtained after the noise reduction processing.
- the transparently transmitted signal is enhanced. Improve the user's sense of hearing for desired external sounds.
- the headset includes a bone conduction sensor, the bone conduction sensor is used to collect the bone conduction signal of the headset user, and the signal obtained after noise reduction processing is subjected to gain amplification processing, including: : performing harmonic expansion on the bone conduction signal to obtain a harmonically expanded signal; using the first gain coefficient to amplify the signal obtained by the noise reduction process; using the fourth filter coefficient to filter out the signal obtained by the amplification process The harmonically extended signal included in ; wherein the fourth filter coefficient is determined based on the first gain coefficient.
- an amplification method which only amplifies the specific sound other than the voice of the wearing user, and improves the effect of the specific sound in the transparently transmitted ambient sound.
- the first gain coefficient is a gain coefficient associated with the target processing intensity in the target mode.
- performing enhancement processing on the second filtered signal to obtain a filtered enhanced signal including:
- the audio event signal in the deblocked signal is subjected to gain amplification processing and frequency response adjustment to obtain a filtered enhanced signal.
- the headset further includes a bone conduction sensor, the bone conduction sensor is used to collect the bone conduction signal of the headset user, and perform gain amplification on the audio event signal in the de-occluded signal processing, including: performing harmonic expansion on the bone conduction signal to obtain a harmonically extended signal; amplifying an audio event signal in the de-occluded signal using a second gain coefficient to obtain an amplified signal; using a second filter coefficient Filtering out the harmonically extended signal included in the amplified signal; wherein the second filter coefficient is determined based on the second gain coefficient.
- the bone conduction sensor is used to collect the bone conduction signal of the headset user, and perform gain amplification on the audio event signal in the de-occluded signal processing, including: performing harmonic expansion on the bone conduction signal to obtain a harmonically extended signal; amplifying an audio event signal in the de-occluded signal using a second gain coefficient to obtain an amplified signal; using a second filter coefficient Filtering out the harmonically extended signal included in the amplified signal; where
- the second gain coefficient is a gain coefficient associated with the target processing intensity for the first filtering processing when the first noise processing is performed.
- the second gain coefficient is a gain coefficient associated with the first scene identifier for the first filtering process when the first noise processing is performed.
- the filter coefficient used by the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the AH function; or,
- the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the AH function; or,
- the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the AH function.
- the earphone further includes a bone conduction sensor, and the bone conduction sensor is used to collect the bone conduction signal of the user of the earphone; performing de-occlusion effect processing on the second filtered signal, including:
- the low-frequency component in the signal obtains a third filtered signal, and the high-frequency component in the third filtered signal from which the low-frequency component is removed is amplified.
- an embodiment of the present invention provides a mode control method, the method is applied to a terminal device, and the method includes: when a scene type of the current external environment is identified as a target scene, determining a target mode according to the target scene; Wherein, the target mode is one of the processing modes supported by the headset, and the processing modes supported by the headset include at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode, or the hearing enhancement AH mode; The earphone sends the target mode, where the target mode is used to instruct the earphone to implement the processing function corresponding to the target mode.
- Different processing modes correspond to different scene types, and processing modes may be in one-to-one correspondence with scene types, or one processing mode may correspond to multiple scene types, for example, two scene types may adopt the same processing mode.
- the terminal device controls the processing mode of the headset in real time according to the scene recognition, so as to optimize the user's hearing experience in real time.
- the method when determining the target mode corresponding to the target scene in the processing mode of the headset, the method further includes: displaying result prompt information, where the result prompt information is used to prompt the user that the headset implements the desired Describe the processing function corresponding to the target mode.
- the above design enables the user to determine the current processing mode of the headset in real time.
- the method before sending the first control signaling to the headset, the method further includes: displaying selection prompt information, where the selection prompt information is used to prompt the user whether to adjust the processing mode of the headset to any The target mode is detected; an operation that the user selects to adjust the processing mode of the headset to the target mode is detected.
- the user can determine whether to adjust the processing mode of the earphone according to the needs, so as to improve the user experience.
- a first control and a second control are displayed, wherein different positions of the second control on the first control are used to indicate different processing intensities in the target mode;
- the method further includes: in response to the user touching the second control to move to a first position on the first control, the second control is on the first control
- the first position of the device indicates the target processing strength in the target mode; the target processing strength is sent to the headset, where the target processing strength is used to indicate the processing when the headset implements the processing function corresponding to the target mode strength.
- the user can select the processing intensity of the earphone according to the needs, so as to meet the different needs of the user.
- the shape of the first control is a ring
- the processing intensity in the target mode is Change from small to large
- the shape of the first control is a bar
- the target processing function when the target processing function is an ANC function, the greater the target processing intensity is, the more the user perceives the sound of the current environment of the user and the environment inside the user's ear canal The weaker the sound.
- the target processing function when the target processing function is the HT function, the greater the target processing intensity, the greater the intensity of the sound of the environment where the user is currently located; or, when the target processing function is AH
- the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- the left earphone and the right earphone may adopt the same processing mode and processing intensity, and the perception of the user's left ear and right ear may be the same.
- the left earphone and the right earphone may also adopt different processing modes or different processing intensities, so that the perception of the left ear and the right ear is different.
- an embodiment of the present invention provides a mode control method, wherein the method is applied to a terminal device, and the method includes: acquiring a target mode; the target mode is one of the processing modes supported by the headset, The processing modes supported by the headset include at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode or the auditory enhancement AH mode; the target processing intensity in the target mode is determined according to the scene type of the current external environment; different The scene types corresponding to different processing intensities in the target mode; the target processing intensities are sent to the earphone, where the target processing intensities are used to indicate the processing intensities when the earphones implement the processing functions corresponding to the target modes.
- the acquiring the target mode includes: receiving the target mode sent by the headset; or, displaying a selection control, where the selection control includes a processing mode supported by the headset, and it is detected that the user has passed the A selection control selects the operation of the target mode among the processing modes of the headset.
- the selection control includes the processing mode supported by the headset, or is interpreted as an option for the selection control to provide the processing mode supported by the headset, or the selection control displays the processing mode supported by the headset, and the user can perform in the processing mode supported by the headset. choose. .
- the method before determining the target processing intensity in the target mode according to the scene type of the current external environment, the method further includes: when the target mode sent by the headset is received, displaying a selection Prompt information, the selection prompt information is used to indicate whether the user adjusts the processing mode of the headset to the target mode; an operation of the user selecting to adjust the processing mode of the headset to the target mode is detected.
- the target processing function is an ANC function
- the target processing function is the HT function
- the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
- the target processing function is the AH function
- the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- an embodiment of the present application provides a mode control method, the method is applied to a terminal device, the method includes: displaying a first interface; the first interface includes a first selection control, and the first selection control includes The processing mode supported by the first target earphone and the processing intensity corresponding to the processing mode supported by the first target earphone; the processing mode of the first target earphone includes active noise reduction ANC mode, ambient sound transparent transmission HT mode or hearing enhancement AH mode at least two kinds of; in response to a first operation performed by the user on the first interface; the first operation is that the user selects the first target mode from the processing modes supported by the first target headset through the first selection control; and The processing intensity in the first target mode is selected to be generated by the first target processing intensity; the first target mode and the first target processing intensity are sent to the first target earphone, and the first target mode is used to indicate The first target earphone implements the processing function corresponding to the first target mode, and the first target processing intensity is used to indicate the processing when the first target ear
- the first selection control includes the processing mode supported by the first target headset and the processing intensity corresponding to the processing mode supported by the first target headset, which can be interpreted as the first selection control provides the user with multiple processing modes (all are the first target headset. supported) options and adjustments to the processing strength for each processing mode.
- the user can freely switch the processing mode and intensity corresponding to the effect that the earphone needs to achieve through the UI interface, so as to meet the different needs of the user.
- the method before displaying the first interface, the method further includes: displaying selection prompt information, where the selection prompt information is used for the user to select whether to adjust the processing mode of the first target headset; detecting The user selects an operation to adjust the processing mode of the first target headset.
- the method before displaying the first interface, the method further includes: identifying the scene type of the current external environment as a target scene, and the target scene adaptation needs to adjust the processing mode of the first target headset scene type.
- a method is provided to actively pop up the first interface in a specific scenario, so as to reduce the manual operation process of the user.
- the method before displaying the first interface, the method further includes: recognizing that the terminal device triggers the first target headset to play audio. Recognizing that the terminal device triggers the first target earphone to play audio may be interpreted as recognizing that the terminal device starts sending an audio signal to the first target earphone.
- the method before displaying the first interface, the method further includes: detecting that the terminal device establishes a connection with the first target headset.
- the method before displaying the first interface, further includes: in the case of detecting that the terminal device establishes a connection with the first target headset, detecting that the user executes the operation on the main interface the second operation; wherein, the main interface includes an icon of a first application, the second operation is generated by the user touching the icon of the first application, and the first interface is the first application display interface.
- the first selection control includes a first control and a second control
- the second control indicates two different processing modes of the first target earphone at any two different positions of the first control
- the second control indicates different processing intensities of the first target earphone in the same processing mode at any two different positions of the first control
- the first operation is that the user moves the second control
- the control is generated at a first position on the first control in an area corresponding to the first target mode, where the first position corresponds to the first target processing intensity in the first target mode.
- the shape of the first control is a ring or a bar.
- the shape of the first control is a ring
- the ring includes at least two arc segments
- the second control is located in different arc segments to indicate different processing modes of the first target headset
- the second The different positions of the controls in the same arc segment indicate different processing intensities of the same processing mode of the first target earphone.
- the shape of the first control is a bar
- the bar includes at least two bar segments
- the second controls are located in different bar segments to indicate different processing modes of the first target earphone, so Different positions of the second control in the same bar segment indicate different processing intensities of the same processing mode of the first target earphone.
- the method further includes:
- the first interface further includes a second selection control
- the second selection control includes a processing mode supported by the second target headset and a processing mode supported by the second target headset
- the processing modes supported by the first target earphone include at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode, or the hearing enhancement AH mode
- the third operation is that the user
- the second selection control selects the second target mode in the processing mode of the second target earphone and selects the processing intensity in the second target mode to be generated by the second target processing intensity; when the first target earphone is the left earphone, the The second target earphone is a right earphone, or the first target earphone is a right earphone, and the second target earphone is a left earphone;
- the second target processing intensity is used to indicate the processing intensity when the second target earphone implements the processing function corresponding to the second target mode.
- the user can operate the processing mode and processing intensity of the left earphone and the right earphone separately, so as to meet the user's differentiated requirements for the hearing sense of the left ear and the right ear.
- an embodiment of the present application further provides a headset control method, the method is applied to a terminal device, and the method includes: establishing a communication connection between the terminal device and the headset; displaying a first interface; the first interface is used to set the headset function; the first interface includes an event sound enhancement function option; the event sound is a sound that meets preset event conditions in the external environment; when the event sound enhancement function option is enabled, control the headset's sound Both the ANC function and the HT function are turned on.
- an embodiment of the present application further provides an earphone control device, the device is applied to a terminal device, and the device includes:
- a display module for displaying a first interface; the first interface is used to set the function of the headset; the first interface includes function options for event sound enhancement; the event sound is in the external environment that meets preset event conditions sound;
- the processing module is configured to control both the ANC function and the HT function of the earphone to be in an enabled state when the option of the event sound enhancement function is enabled.
- the first interface includes an option for controlling the HT function of the earphone; when the option of the HT function is enabled, the HT of the earphone is activated function; and an option of enhancing the sound of the event is added in the first interface.
- the target strength of the HT function can also be acquired; the HT function of the earphone is controlled according to the target strength of the HT function. This step can be performed in cooperation with the display module and the processing module.
- controlling both the ANC function and the HT function of the earphone to be in an on state includes: keeping the HT function in an on state; and activating the ANC function of the earphone. This step may be performed by a processing module.
- the first interface includes an option for controlling the ANC function of the headset; when the option of the ANC function is enabled, the headset is activated ANC function. Further optionally, a strength option of the ANC function is added in the first interface. This step may be performed by a processing module.
- the first interface further includes an option for disabling the ANC function, and/or an option for disabling the HT function.
- the strength options of the ANC function include at least a first steady-state ANC strength option, a second steady-state ANC strength option, and an adaptive ANC strength option; wherein, the The first steady-state ANC strength option and the second steady-state ANC strength option correspond to the first scenario and the second scenario, respectively, and correspond to different ANC functional strengths; the ANC functional strength corresponding to the adaptive ANC strength option It is related to the scene type of the current environment where the terminal device or the headset is located; different scene types of the current environment correspond to different ANC strengths.
- the scene type of the current environment where the terminal or the headset is located is acquired; according to the scene The type matches the target strength of the ANC function; the ANC function of the headset is controlled according to the target strength.
- the different scene types of the current environment include a first scene and a second scene.
- the event sound includes a human voice or other sounds conforming to preset spectral characteristics.
- the option enablement of the event sound enhancement function, the option enablement of the HT function, or the option enablement of the ANC function includes: responding to the user's request for The click of the corresponding function option, the adaptive switching of the corresponding function, or the shortcut of the corresponding function is triggered.
- an embodiment of the present application further provides a noise reduction method, the method is applied to an earphone, the earphone supports at least an active noise reduction ANC function, and can also support an ambient sound transparent transmission HT function, and the earphone includes a first a microphone, a second microphone and a speaker; the method includes: collecting a first signal through the first microphone, where the first signal is used to represent the sound of the current external environment; collecting a second signal through the second microphone, the The second signal is used to characterize the sound of the internal environment of the ear canal of the user wearing the headset; receive an instruction for event sound enhancement; the event sound is the sound that meets the preset event conditions in the external environment; control the ANC function to be turned on, even if the Or turn on the ANC function, at least use the ANC function to perform target processing on the first signal and the second signal to obtain a target signal; the signal-to-noise ratio of the event sound in the target signal is greater than that in the first signal.
- an embodiment of the present application further provides a noise reduction device.
- the device is applied to an earphone.
- the earphone supports at least an active noise reduction ANC function, and can also support an ambient sound transparent transmission HT function.
- the earphone includes a first microphone, a second microphone and a speaker; the device includes: a collection module configured to collect a first signal through the first microphone, where the first signal is used to represent the sound of the current external environment; further configured to use the second microphone Collecting a second signal, the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the headset; the receiving module is used to receive an instruction of event sound enhancement; the event sound is in the external environment and conforms to a preset event conditional sound; a processing module for enabling the ANC function, and at least using the ANC function to perform target processing on the first signal and the second signal to obtain a target signal; the information of the event sound in the target signal
- the noise ratio is greater than the signal-to-noise ratio
- control ANC function and the HT function are both turned on; the HT function is used to transparently transmit the first signal to obtain a restored signal; for events in the restored signal The sound signal is enhanced, and the non-event sound signal in the restored signal is weakened to obtain an event sound enhancement signal; the first signal, the second signal and the event sound enhancement signal are processed by using the ANC function The target signal is obtained.
- the target signal is obtained by processing the first signal, the second signal and the event sound enhancement signal by using an ANC function.
- the earphone supports at least the active noise reduction ANC function, the ambient sound transparent transmission HT function and the AH function, and the earphone includes an HT filter bank, a feedback filter bank and Feedforward filter bank; obtain the working mode of the headset; when the working mode is the ANC function, call the feedback filter bank and the feedforward filter bank to realize the ANC function; when the working mode is the HT function When the HT filter bank and the feedback filter bank are called to realize the HT function; when the working mode is the AH function, the HT filter bank, the feedforward filter bank and the all The feedback filter bank is used to realize the AH function.
- an embodiment of the present application further provides a noise reduction method, the method is applied to an earphone, and the earphone supports at least an ANC function; the earphone includes a first microphone and a third microphone; the first microphone is more focused In order to collect the sound of the current external environment, the third microphone focuses more on sound pickup; when the headset enables the ANC function, the first microphone is used to collect the first signal for the current environment; The microphone collects a second signal for the current environment; judges the noise level of the current scene according to the first signal and the second signal; wherein, different noise levels correspond to different ANC intensities; according to the current noise level Controls the ANC function.
- an embodiment of the present application further provides a noise reduction device, the device is applied to an earphone, and the earphone supports at least an ANC function; the earphone includes a first microphone and a third microphone; the first microphone is more focused In order to collect the sound of the current external environment, the third microphone focuses more on sound pickup; the collection module is used to collect the first microphone for the current environment through the first microphone when the ANC function of the headset is in an on state. signal; collecting a second signal for the current environment through the third microphone; an identification module for judging the noise level of the current scene according to the first signal and the second signal; wherein different noise levels Corresponding to different ANC strengths; the processing module is used to control the ANC function according to the current noise level.
- the correlation feature of the first signal and the second signal for voice activity detection; perform noise tracking on non-voice signals; if the noise energy is less than the first signal threshold, the current scene is determined to be a quiet scene; or, if the spectral center of the noise is in a low frequency band and the noise energy is greater than the second threshold, the current scene is determined to be a heavy noise scene; or, if the current scene is both When it does not belong to the quiet scene or the heavy noise scene, the current scene is determined to be an ordinary scene; wherein the second threshold is greater than the first threshold.
- the ANC strengths corresponding to the quiet scene, the normal scene, and the heavy noise scene sequentially increase.
- the ninth aspect or the tenth aspect in a possible design, if it is detected that the current scene is a new noise level and lasts for a preset period of time; obtain the ANC intensity corresponding to the new noise level; The ANC strength corresponding to the noise level controls the ANC function.
- an embodiment of the present application further provides a method for controlling an earphone.
- the method is applied to a terminal device, and the method includes: establishing a communication connection between the terminal device and the earphone; and the earphone supports at least an active noise reduction (ANC) function.
- ANC active noise reduction
- the first interface is used to set the function of the headset; the first interface includes an option for controlling the ANC function of the headset; when the option of the ANC function is enabled, activate the ANC function of the headset; adding a strength option of the ANC function in the first interface; performing ANC noise reduction according to the enabling result of the strength option of the ANC function; wherein, the strength option of the ANC function at least includes A first steady-state ANC strength option, a second steady-state ANC strength option, and an adaptive ANC strength option; wherein the first steady-state ANC strength option and the second steady-state ANC strength option correspond to the first scenario and The second scenario corresponds to different and stable ANC functional strengths respectively; the ANC functional strength corresponding to the adaptive ANC strength option is related to the scene type of the current environment where the terminal device or the headset is located; the current environment Different scene types correspond to different ANC strengths.
- the different scene types of the current environment include a first scene and a second scene.
- an embodiment of the present application further provides a headset control device, the terminal device establishes a communication connection with the headset; the headset supports at least an active noise reduction ANC function; the device includes: a display module for displaying the first an interface; the first interface is used to set the function of the headset; the first interface includes an option for controlling the ANC function of the headset; a processing module is used to activate the ANC function when the option of the ANC function is enabled The ANC function of the headset; the display module is further configured to add a strength option of the ANC function in the first interface after the option of the ANC function is enabled; the processing module is also configured to ANC noise reduction is performed on the result of enabling the strength option of the ANC function; wherein, the strength option of the ANC function includes at least a first steady-state ANC strength option, a second steady-state ANC strength option, and an adaptive ANC strength option; wherein, The first steady-state ANC strength option and the second steady-state ANC strength option correspond to the
- the first steady state ANC strength option when the first steady state ANC strength option is enabled, the first ANC function strength corresponding to the first steady state ANC strength option is obtained; The first ANC function strength controls the ANC function; or, when the second steady state ANC strength option is enabled, obtain the second ANC function strength corresponding to the second steady state ANC strength option; according to the second ANC function Intensity controls the ANC function.
- the adaptive ANC strength option when the adaptive ANC strength option is enabled, the scene type of the current environment where the terminal device or the headset is located is obtained; the ANC strength is determined according to the scene type of the current environment; the ANC function is controlled according to the determined ANC strength .
- an embodiment of the present application further provides a method for controlling an earphone.
- the method is applied to a terminal device, and the method includes: establishing a communication connection between the terminal device and an earphone; and the earphone supports at least ambient sound transparent transmission HT function, display the first interface; the first interface is used to set the function of the headset; the first interface includes options for controlling the HT function of the headset; when the option of the HT function is enabled, activate all The HT function of the headset; the option of the event sound enhancement is added in the first interface; the event sound is the sound that meets the preset event conditions in the external environment; when the event sound enhancement option is enabled When , the earphone is controlled to improve the signal-to-noise ratio of the event sound in the signal collected by the earphone; wherein, the higher the signal-to-noise ratio of the event sound, the higher the energy ratio of the event sound in the signal big.
- an embodiment of the present application further provides an earphone control device, the device is applied to a terminal device, and the terminal device establishes a communication connection with the earphone; the earphone supports at least the HT function of transparent transmission of ambient sound, and the device It includes: a display module for displaying a first interface; the first interface is used to set the function of the headset; the first interface includes options for controlling the HT function of the headset; When the option of the HT function is enabled, the HT function of the headset is activated; the display module is further configured to add the option of the event sound enhancement in the first interface after the option of the HT function is enabled; The event sound is a sound that meets preset event conditions in the external environment; the processing module is further configured to control the earphone for the event in the signal collected by the earphone when the option of the event sound enhancement function is enabled The signal-to-noise ratio of the sound is improved; wherein, the higher the signal-to-noise ratio of the event sound, the greater the
- the event sound includes human voice or other sounds that conform to preset spectral characteristics.
- the first interface includes an option for controlling the ANC function of the headset, an option for turning off the ANC function, and/or an option for turning off the HT function.
- the first intensity of the ANC function is obtained, and the ANC function of the earphone is controlled according to the first intensity; or, the second intensity of the HT function is obtained.
- an embodiment of the present application further provides a noise processing device, the device is applied to an earphone, and the earphone has at least two functions of an active noise reduction ANC function, an ambient sound transparent transmission HT function, or an auditory enhancement AH function
- the headset includes a first microphone and a second microphone; the first microphone is used to collect a first signal, and the first signal is used to characterize the sound of the current external environment; the second microphone is used to collect the second signal , the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the earphone.
- the noise processing apparatus includes corresponding functional modules, which are respectively used to implement the steps in the method of the first aspect above. For details, please refer to the detailed description in the method example, which will not be repeated here.
- the functions can be implemented by hardware, or by executing corresponding software by hardware.
- the hardware or software includes one or more modules corresponding to the above functions.
- the noise processing device includes:
- a communication module for receiving the first audio signal from the terminal device
- an acquisition module configured to acquire a target mode; wherein, the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the headset to implement a target processing function, and the target processing function is active noise reduction One of ANC function, ambient sound transparent transmission HT function or auditory enhancement AH function;
- a first processing module configured to obtain a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal.
- an embodiment of the present application provides a target earphone, including a left earphone and a right earphone, where the left earphone is used to implement any of the above possible design methods for earphones, or the right earphone is used to implement any of the above possible design approaches regarding headphones.
- the left and right earphones use different processing modes.
- an embodiment of the present application provides a target earphone, where the target earphone includes a left earphone and a right earphone.
- the left earphone includes or the right earphone includes a first microphone, a second microphone, a processor, a memory, and a speaker;
- the first microphone is used to collect a first signal, and the first signal is used to represent the sound of the current external environment ;
- the second microphone is used to collect a second signal, and the second signal is used to characterize the sound of the internal environment of the ear canal of the user wearing the headset;
- the memory is used to store programs or instructions;
- the processor used to call the program or instruction, so that the electronic device executes any possible earphone-related method to obtain the second audio signal or the target signal;
- the speaker is used to play the second audio signal or the target signal Signal.
- an embodiment of the present application provides a mode control apparatus, where the apparatus is applied to a terminal device.
- the apparatus includes corresponding functional modules, which are respectively used to implement any of the above possible terminal-related steps.
- the functions can be implemented by hardware, or by executing corresponding software by hardware.
- the hardware or software includes one or more modules corresponding to the above functions.
- an embodiment of the present application provides a terminal device, including a memory, a processor, and a display; the display is used for displaying an interface; the memory is used for storing programs or instructions; the processor is used for The program or instruction is invoked to cause the terminal device to execute steps in any possible terminal-related method.
- the present application provides a computer-readable storage medium, where computer programs or instructions are stored in the computer-readable storage medium, and when the computer program or instructions are executed by the headset, the headset can execute any of the above possible related Headphone design method.
- the present application provides a computer-readable storage medium.
- the computer-readable storage medium stores a computer program or instruction.
- the headset can perform any of the above possible functions.
- the present application provides a computer program product, the computer program product includes a computer program or an instruction, when the computer program or instruction is executed by the headset, the method in any of the above possible implementations of the headset is implemented .
- the present application provides a computer program product, the computer program product includes a computer program or instruction, when the computer program or instruction is executed by the headset, the method in any of the possible headset implementations described above is implemented.
- FIG. 1 is a schematic diagram of a hardware structure of a terminal device 100 in an embodiment of the present application
- FIG. 2 is a schematic diagram of a software structure of a terminal device 100 in an embodiment of the present application
- FIG. 3 is a schematic structural diagram of an earphone 200 in an embodiment of the present application.
- FIG. 4 is a schematic diagram of the AHA pathway in the embodiment of the present application.
- 5A is a flowchart of ANC processing in an embodiment of the present application.
- 5B is a schematic diagram of an ANC processing flow in an embodiment of the present application.
- 6A is a flowchart of HT processing in an embodiment of the present application.
- 6B is a schematic diagram of a HT processing flow in an embodiment of the present application.
- 6C is a schematic diagram of another HT processing flow in an embodiment of the present application.
- FIG. 7 is a schematic diagram of a process flow of de-occlusion effect processing in an embodiment of the present application.
- FIG. 8A is a flowchart of AH processing in an embodiment of the present application.
- FIG. 8B is a schematic diagram of an AH processing flow in an embodiment of the present application.
- 8C is a schematic diagram of another AH processing flow in an embodiment of the present application.
- FIG. 9 is a schematic diagram of a noise reduction processing flow diagram in an embodiment of the present application.
- FIG. 10 is a schematic diagram of a process flow of gain amplification in an embodiment of the present application.
- FIG. 11 is a schematic diagram of another gain amplification processing flow in an embodiment of the present application.
- 12A is a schematic diagram of a main interface of a terminal device in an embodiment of the present application.
- 12B is a schematic diagram of a control interface of an earphone application in an embodiment of the present application.
- 12C is a schematic diagram of the control of a terminal device controlling a headset in an ANC mode in an embodiment of the present application
- 12D is a schematic diagram of the control of a terminal device controlling an earphone in an HT mode in an embodiment of the present application
- 12E is a schematic diagram of the control of a terminal device controlling an earphone in an AH mode in an embodiment of the present application
- 12F is a schematic diagram of a selection control in an embodiment of the present application.
- 12G is a schematic diagram of another selection control in an embodiment of the present application.
- 12H is a schematic diagram of triggering a control interface of an earphone in an embodiment of the present application.
- FIG. 13 is a schematic diagram of another selection control in an embodiment of the application.
- FIG. 14A is a schematic diagram of control for enabling a smart scene detection function according to an embodiment of the present application.
- FIG. 14B is a schematic diagram of another kind of smart scene detection function enabling control in an embodiment of the present application.
- 14C is a schematic diagram of an earphone control interface in an embodiment of the present application.
- 15 is a schematic diagram of event detection in an embodiment of the present application.
- 16 is a schematic diagram of an interaction processing mode and processing intensity between a terminal device and a headset in an embodiment of the application;
- FIG. 17A is a schematic diagram of displaying a scene detection result in an embodiment of the present application.
- FIG. 17B is a schematic diagram of another scene detection result display in an embodiment of the present application.
- FIG. 19 is a schematic structural diagram of a noise processing apparatus 1900 in an embodiment of the present application.
- FIG. 20 is a schematic structural diagram of a mode control apparatus 2000 in an embodiment of the present application.
- FIG. 21 is a schematic structural diagram of a mode control device 2100 in an embodiment of the present application.
- FIG. 22 is a schematic structural diagram of a mode control apparatus 2200 in an embodiment of the present application.
- FIG. 23 is a schematic structural diagram of a terminal device 2300 in an embodiment of the present application.
- FIG. 24 is a schematic diagram of an earphone control interface in a terminal device in an embodiment of the application.
- FIG. 25 is a schematic diagram of a headset control interface in a terminal device in an embodiment of the application.
- FIG. 26 is a schematic diagram of an earphone control interface in a terminal device according to an embodiment of the application.
- FIG. 27 is a schematic diagram of a headset control interface in a terminal device according to an embodiment of the application.
- FIG. 28 is a schematic diagram of an earphone control interface in a terminal device according to an embodiment of the present application.
- the application (application, app) involved in the embodiments of the present application is a software program capable of implementing one or more specific functions.
- multiple applications can be installed in a terminal device.
- the applications mentioned below may be system applications that have been installed when the terminal device is shipped from the factory, or may be third-party applications downloaded from the network or obtained from other terminal devices by the user during the use of the terminal device.
- the human auditory system has a masking effect, that is, strong-frequency sounds will hinder human perception of the nearby weak-frequency sounds, and the basilar membrane of the cochlea has a frequency selection and tuning effect on external sound signals. Therefore, the concept of critical frequency band is introduced, Perceptually measure sound frequency. It is generally believed that there are 24 critical frequency bands within the hearing threshold of 22 Hz to 22 kHz, which can cause vibrations at different positions on the basilar membrane. Each critical band is called a bark subband.
- VAD Voice endpoint detection
- At least one (item) refers to one (item) or more (item), and “multiple (item)” refers to two (item) or more than two (item).
- “And/or”, which describes the association relationship of the associated objects means that there can be three kinds of relationships, for example, A and/or B, it can mean that A exists alone, A and B exist at the same time, and B exists alone, where A, B can be singular or plural.
- the character “/” generally indicates that the associated objects are an "or” relationship.
- At least one item(s) below” or similar expressions thereof refer to any combination of these items, including any combination of single item(s) or plural items(s).
- At least one (a) of a, b, or c may represent: a, b, c, a-b, a-c, b-c, or a-b-c, where a, b, and c may be single or multiple.
- the symbol "(a, b)" represents an open interval, and the range is greater than a and less than b; "[a, b]” represents a closed interval, and the range is greater than or equal to a and less than or equal to b; "(a ,b]” represents a half-open and half-closed interval with a range greater than a and less than or equal to b; "(a,b]” represents a half-open and half-closed interval with a range greater than a and less than or equal to b.
- first and second mentioned in the embodiments of this application are used to distinguish multiple objects, and are not used to limit the size, content, order, timing, priority or importance of multiple objects. degree, etc.
- first microphone and the second microphone are only used to distinguish different microphones, and do not indicate the difference in size, priority, or importance of the two microphones.
- An embodiment of the present application provides a system, where the system includes a terminal device 100 and an earphone 200 .
- the terminal device 100 is connected 200 to the earphone, and the connection may be a wireless connection or a wired connection.
- the terminal device may be connected to the headset through Bluetooth technology, wireless fidelity (Wi-Fi) technology, infrared IR technology, and ultra-wideband technology.
- the terminal device 100 is a device having a display interface function.
- the terminal device 100 may be, for example, a product with a display interface such as a mobile phone, a monitor, a tablet computer, and a vehicle-mounted device, as well as a smart display wearable product such as a smart watch and a smart bracelet.
- a product with a display interface such as a mobile phone, a monitor, a tablet computer, and a vehicle-mounted device, as well as a smart display wearable product such as a smart watch and a smart bracelet.
- a smart display wearable product such as a smart watch and a smart bracelet.
- the specific form of the above mobile terminal is not particularly limited in this embodiment of the present application.
- the earphone 200 includes two sounding units that can be hung around the ears.
- the one that fits the left ear can be called the left earphone, and the one that fits the right ear can be called the right earphone.
- the earphone 200 in this embodiment of the present application may be a headphone, an ear-hook earphone, a neck-hung earphone, or an earbud earphone, or the like.
- In-ear headphones also include in-ear headphones (or referred to as ear canal headphones) or semi-in-ear headphones.
- the earphone 200 has at least two of the ANC function, the HT function, or the AH function.
- AHA for the convenience of description, ANC, HT, and AH are collectively referred to as AHA in this embodiment of the present application, and of course, other names may also be ordered, which is not limited in this application.
- the left earphone and the right earphone have similar structures. Either the left earphone or the right earphone can adopt the earphone structure described below.
- the earphone structure (left earphone or right earphone) includes a rubber sleeve that can be inserted into the ear canal, an ear bag close to the ear, and an earphone rod suspended on the ear bag. The rubber sleeve guides the sound to the ear canal.
- the ear bag includes devices such as batteries, speakers, and sensors. Microphones and physical buttons can be arranged on the headphone rod.
- the earphone rod can be in the shape of a cylinder, a rectangular parallelepiped, an ellipsoid, or the like.
- the microphone arranged inside the ear can be called the error microphone, and the microphone arranged outside the earphone is called the reference microphone.
- the error microphone is used to pick up the sound of the external environment. Referring to the microphone, when the user wears the earphone, the sound of the internal environment of the ear canal of the user wearing the earphone is collected.
- the two microphones can be either analog microphones or digital microphones. After the user wears the earphone, the positional relationship between the two microphones and the speaker is as follows: the error microphone is inside the ear, close to the earphone rubber sleeve.
- the loudspeaker is located between the error microphone and the reference microphone.
- the reference microphone is close to the outer structure of the ear and can be arranged on the upper part of the earphone stem.
- the tube of the error microphone can either face the loudspeaker or the inside of the ear canal.
- the terminal device 100 is configured to send downlink audio signals and/or control signaling to the headset 200 .
- the control signaling is used to control the processing mode adopted by the headset 200 .
- the processing mode adopted by the headset 200 may include at least two of an empty mode indicating no processing, an ANC mode indicating implementing an ANC function, an HT mode indicating implementing an HT function, or an AH mode indicating implementing an AH function.
- the earphone user's perception of the sound of the current external environment and the ambient sound inside the ear canal of the user wearing the earphone can be weakened.
- the headset adopts the HT mode the user's perception of the sound of the current external environment can be enhanced.
- the headset adopts the AH mode the user's perception of the event sound included in the sound of the current external environment can be enhanced.
- the event sound refers to a preset sound in an external environment, or the event sound satisfies a preset frequency spectrum.
- the event sound includes a station announcement sound or a whistle sound in a railway station; then the event sound satisfies the frequency spectrum of the station announcement sound or the frequency spectrum of the whistle sound in the railway station.
- the event sound may include a notification sound in an airplane terminal, a broadcast sound on an airplane, etc.; another example is the sound of a hotel calling number, and the like. It should be understood that both the terminal and the headset can recognize the event sound.
- the earphone 200 includes a left earphone and a right earphone, and the left earphone and the right earphone may adopt the same processing mode or different processing modes.
- the left earphone and the right earphone use the same processing mode, the user's left ear wearing the left earphone and the right ear wearing the right earphone may have the same auditory perception.
- the left earphone and the right earphone adopt different processing modes, the user's left ear wearing the left earphone and the right ear wearing the right earphone have different auditory perceptions.
- the left earphone using ANC and the right earphone using AH as an example, when the left earphone adopts the ANC mode, the sound of the earphone user's left ear to the current external environment and the environmental sound inside the ear canal of the user wearing the earphone can be reduced. perception.
- the right earphone adopts the AH mode the perception of the user's right ear to the event sound included in the sound of the current external environment can be enhanced.
- any one of the following possible ways can be used to achieve it;
- the terminal device 100 provides a control interface for the user to select the processing mode of the headset 200 according to requirements. For example, under the operation of the user, the terminal device 100 sends control signaling to the headset 200 , and the control signaling is used to indicate the processing mode adopted by the headset 200 .
- the processing modes adopted by the left earphone and the right earphone in the earphone 200 may be the same or different.
- a selection control on the control surface is used to select the same processing mode for the left and right earphones.
- the control interface may include two selection controls, wherein one selection control is used to select a processing mode for the left earphone, and the other selection control is used to select a processing mode for the right earphone. The control interface and selection controls will be described in detail later, and the description will not be repeated here.
- the terminal device identifies the scene type of the user's current external environment.
- the processing modes adopted by the headset 200 are different, that is, the processing functions implemented by the headset are different.
- the earphone 200 recognizes the user's operation, and determines that the earphone 200 selected by the user adopts the ANC mode, the HT mode or the AH mode.
- the user's operation may be an operation of the user tapping the earphone, or buttons are provided on the earphone, and different buttons indicate different processing modes.
- the earphone recognizes the scene type of the external environment of the earphone, and the processing mode adopted by the earphone is different in different scenes.
- FIG. 1 shows a schematic diagram of an optional hardware structure of the terminal device 100 .
- the terminal device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2 , mobile communication module 150, wireless communication module 160, audio module 170, speaker 170A, receiver 170B, microphone 170C, headphone jack 170D, sensor module 180, buttons 190, motor 191, indicator 192, camera 193, display screen 194, and Subscriber identification module (subscriber identification module, SIM) card interface 195 and so on.
- SIM Subscriber identification module
- the sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, and ambient light. Sensor 180L, bone conduction sensor 180M, etc.
- the structures illustrated in the embodiments of the present invention do not constitute a specific limitation on the terminal device 100 .
- the terminal device 100 may include more or less components than shown, or some components are combined, or some components are separated, or different components are arranged.
- the illustrated components may be implemented in hardware, software, or a combination of software and hardware.
- the processor 110 may include one or more processing units, for example, the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processor (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), controller, video codec, digital signal processor (digital signal processor, DSP), baseband processor, and/or neural-network processing unit (neural-network processing unit, NPU), etc. Wherein, different processing units may be independent devices, or may be integrated in one or more processors.
- application processor application processor, AP
- modem processor graphics processor
- ISP image signal processor
- controller video codec
- digital signal processor digital signal processor
- baseband processor baseband processor
- neural-network processing unit neural-network processing unit
- the controller can generate an operation control signal according to the instruction operation code and timing signal, and complete the control of fetching and executing instructions.
- a memory may also be provided in the processor 110 for storing instructions and data.
- the memory in processor 110 is cache memory. This memory may hold instructions or data that have just been used or recycled by the processor 110 . If the processor 110 needs to use the instruction or data again, it can be called directly from the memory. Repeated accesses are avoided and the latency of the processor 110 is reduced, thereby increasing the efficiency of the system.
- the processor 110 may include one or more interfaces.
- the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, a universal asynchronous transceiver (universal asynchronous transmitter) receiver/transmitter, UART) interface, mobile industry processor interface (MIPI), general-purpose input/output (GPIO) interface, subscriber identity module (SIM) interface, and / or universal serial bus (universal serial bus, USB) interface, etc.
- I2C integrated circuit
- I2S integrated circuit built-in audio
- PCM pulse code modulation
- PCM pulse code modulation
- UART universal asynchronous transceiver
- MIPI mobile industry processor interface
- GPIO general-purpose input/output
- SIM subscriber identity module
- USB universal serial bus
- the I2C interface is a bidirectional synchronous serial bus that includes a serial data line (SDA) and a serial clock line (SCL).
- the processor 110 may contain multiple sets of I2C buses.
- the processor 110 can be respectively coupled to the touch sensor 180K, the charger, the flash, the camera 193 and the like through different I2C bus interfaces.
- the processor 110 may couple the touch sensor 180K through the I2C interface, so that the processor 110 and the touch sensor 180K communicate with each other through the I2C bus interface, so as to realize the touch function of the terminal device 100 .
- the I2S interface can be used for audio communication.
- the processor 110 may contain multiple sets of I2S buses.
- the processor 110 may be coupled with the audio module 170 through an I2S bus to implement communication between the processor 110 and the audio module 170 .
- the audio module 170 can transmit an audio signal to the wireless communication module 160 through the I2S interface, so as to realize the function of answering a call through the headset 200 (such as a Bluetooth headset).
- the PCM interface can also be used for audio communications, sampling, quantizing and encoding analog signals.
- the audio module 170 and the wireless communication module 160 may be coupled through a PCM bus interface.
- the audio module 170 can also transmit audio signals to the wireless communication module 160 through the PCM interface, so as to realize the function of answering a call through the Bluetooth headset 200 .
- Both the I2S interface and the PCM interface can be used for audio communication.
- the UART interface is a universal serial data bus used for asynchronous communication.
- the bus may be a bidirectional communication bus. It converts the data to be transmitted between serial communication and parallel communication.
- a UART interface is typically used to connect the processor 110 with the wireless communication module 160 .
- the processor 110 communicates with the Bluetooth module in the wireless communication module 160 through the UART interface to implement the Bluetooth function.
- the audio module 170 can transmit audio signals to the wireless communication module 160 through the UART interface, so as to realize the function of playing music through the Bluetooth headset 200 .
- the MIPI interface can be used to connect the processor 110 with peripheral devices such as the display screen 194 and the camera 193 .
- MIPI interfaces include camera serial interface (CSI), display serial interface (DSI), etc.
- the processor 110 communicates with the camera 193 through the CSI interface, so as to realize the shooting function of the terminal device 100 .
- the processor 110 communicates with the display screen 194 through the DSI interface to implement the display function of the terminal device 100 .
- the GPIO interface can be configured by software.
- the GPIO interface can be configured as a control signal or as a data signal.
- the GPIO interface may be used to connect the processor 110 with the camera 193, the display screen 194, the wireless communication module 160, the audio module 170, the sensor module 180, and the like.
- the GPIO interface can also be configured as I2C interface, I2S interface, UART interface, MIPI interface, etc.
- the USB interface 130 is an interface that conforms to the USB standard specification, and may specifically be a Mini USB interface, a Micro USB interface, a USB Type C interface, and the like.
- the USB interface 130 can be used to connect a charger to charge the terminal device 100, and can also be used to transmit data between the terminal device 100 and peripheral devices. It can also be used to connect the headset 200 to play audio through the headset 200 .
- This interface can also be used to connect other terminal devices, such as AR devices.
- the interface connection relationship between the modules illustrated in the embodiment of the present invention is only a schematic illustration, and does not constitute a structural limitation of the terminal device 100 .
- the terminal device 100 may also adopt different interface connection manners in the foregoing embodiments, or a combination of multiple interface connection manners.
- the charging management module 140 is used to receive charging input from the charger.
- the charger may be a wireless charger or a wired charger.
- the charging management module 140 may receive charging input from the wired charger through the USB interface 130 .
- the charging management module 140 may receive wireless charging input through the wireless charging coil of the terminal device 100 . While the charging management module 140 charges the battery 142 , it can also supply power to the terminal device through the power management module 141 .
- the power management module 141 is used for connecting the battery 142 , the charging management module 140 and the processor 110 .
- the power management module 141 receives input from the battery 142 and/or the charging management module 140, and supplies power to the processor 110, the internal memory 121, the display screen 194, the camera 193, and the wireless communication module 160.
- the power management module 141 can also be used to monitor parameters such as battery capacity, battery cycle times, battery health status (leakage, impedance).
- the power management module 141 may also be provided in the processor 110 .
- the power management module 141 and the charging management module 140 may also be provided in the same device.
- the wireless communication function of the terminal device 100 may be implemented by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, the modulation and demodulation processor, the baseband processor, and the like.
- Antenna 1 and Antenna 2 are used to transmit and receive electromagnetic wave signals.
- Each antenna in terminal device 100 may be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve antenna utilization.
- the antenna 1 can be multiplexed as a diversity antenna of the wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
- the mobile communication module 150 may provide a wireless communication solution including 2G/3G/4G/5G etc. applied on the terminal device 100 .
- the mobile communication module 150 may include at least one filter, switch, power amplifier, low noise amplifier (LNA), and the like.
- the mobile communication module 150 can receive electromagnetic waves from the antenna 1, filter and amplify the received electromagnetic waves, and transmit them to the modulation and demodulation processor for demodulation.
- the mobile communication module 150 can also amplify the signal modulated by the modulation and demodulation processor, and then turn it into an electromagnetic wave for radiation through the antenna 1 .
- at least part of the functional modules of the mobile communication module 150 may be provided in the processor 110 .
- at least part of the functional modules of the mobile communication module 150 may be provided in the same device as at least part of the modules of the processor 110 .
- the modem processor may include a modulator and a demodulator.
- the modulator is used to modulate the low frequency baseband signal to be sent into a medium and high frequency signal.
- the demodulator is used to demodulate the received electromagnetic wave signal into a low frequency baseband signal. Then the demodulator transmits the demodulated low-frequency baseband signal to the baseband processor for processing.
- the low frequency baseband signal is processed by the baseband processor and passed to the application processor.
- the application processor outputs sound signals through audio devices (not limited to the speaker 170A, the receiver 170B, etc.), or displays images or videos through the display screen 194 .
- the modem processor may be a stand-alone device.
- the modem processor may be independent of the processor 110, and may be provided in the same device as the mobile communication module 150 or other functional modules.
- the wireless communication module 160 can provide applications on the terminal device 100 including wireless local area networks (WLAN) (such as wireless fidelity (Wi-Fi) networks), bluetooth (BT), global navigation satellites Wireless communication solutions such as global navigation satellite system (GNSS), frequency modulation (FM), near field communication (NFC), and infrared technology (IR).
- WLAN wireless local area networks
- BT Bluetooth
- GNSS global navigation satellite system
- FM frequency modulation
- NFC near field communication
- IR infrared technology
- the wireless communication module 160 may be one or more devices integrating at least one communication processing module.
- the wireless communication module 160 receives electromagnetic waves via the antenna 2 , frequency modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 110 .
- the wireless communication module 160 can also receive the signal to be sent from the processor 110 , perform frequency modulation on it, amplify it, and convert it into electromagnetic waves for radiation through the antenna 2 .
- the wireless communication module 160 includes a Bluetooth module, and the terminal device 100 establishes a wireless connection with the headset 200 through Bluetooth.
- the wireless communication module 160 includes an infrared module, and the terminal device 100 can establish a wireless connection with the headset 200 through the infrared module.
- the antenna 1 of the terminal device 100 is coupled with the mobile communication module 150, and the antenna 2 is coupled with the wireless communication module 160, so that the terminal device 100 can communicate with the network and other devices through wireless communication technology.
- the wireless communication technology may include global system for mobile communications (GSM), general packet radio service (GPRS), code division multiple access (CDMA), broadband Code Division Multiple Access (WCDMA), Time Division Code Division Multiple Access (TD-SCDMA), Long Term Evolution (LTE), BT, GNSS, WLAN, NFC , FM, and/or IR technology, etc.
- the GNSS may include a global positioning system (global positioning system, GPS), a global navigation satellite system (GLONASS), a Beidou navigation satellite system (BDS), a quasi-zenith satellite system (quasi -zenith satellite system, QZSS) and/or satellite based augmentation systems (SBAS).
- GPS global positioning system
- GLONASS global navigation satellite system
- BDS Beidou navigation satellite system
- QZSS quasi-zenith satellite system
- SBAS satellite based augmentation systems
- the terminal device 100 implements a display function through a GPU, a display screen 194, an application processor, and the like.
- the GPU is a microprocessor for image processing, and is connected to the display screen 194 and the application processor.
- the GPU is used to perform mathematical and geometric calculations for graphics rendering.
- Processor 110 may include one or more GPUs that execute program instructions to generate or alter display information.
- Display screen 194 is used to display images, videos, and the like.
- Display screen 194 includes a display panel.
- the display panel can be a liquid crystal display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode or an active-matrix organic light-emitting diode (active-matrix organic light).
- LED diode AMOLED
- flexible light-emitting diode flexible light-emitting diode (flex light-emitting diode, FLED), Miniled, MicroLed, Micro-oLed, quantum dot light-emitting diode (quantum dot light emitting diodes, QLED) and so on.
- the terminal device 100 may include 1 or N1 display screens 194 , where N1 is a positive integer greater than 1.
- the terminal device 100 can realize the shooting function through the ISP, the camera 193, the video codec, the GPU, the display screen 194 and the application processor.
- the ISP is used to process the data fed back by the camera 193 .
- the shutter is opened, the light is transmitted to the camera photosensitive element through the lens, the light signal is converted into an electrical signal, and the camera photosensitive element transmits the electrical signal to the ISP for processing, and converts it into an image visible to the naked eye.
- ISP can also perform algorithm optimization on image noise, brightness, and skin tone.
- ISP can also optimize the exposure, color temperature and other parameters of the shooting scene.
- the ISP may be provided in the camera 193 .
- Camera 193 is used to capture still images or video.
- the object is projected through the lens to generate an optical image onto the photosensitive element.
- the photosensitive element may be a charge coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor.
- CMOS complementary metal-oxide-semiconductor
- the photosensitive element converts the optical signal into an electrical signal, and then transmits the electrical signal to the ISP to convert it into a digital image signal.
- the ISP outputs the digital image signal to the DSP for processing.
- DSP converts digital image signals into standard RGB, YUV and other formats of image signals.
- the processor 110 may trigger the startup of the camera 193 according to a program or an instruction in the internal memory 121, so that the camera 193 captures at least one image and performs corresponding processing on the at least one image according to the program or instruction.
- the terminal device 100 may include 1 or N2 cameras 193 , where N2 is a positive integer greater than 1.
- a digital signal processor is used to process digital signals, in addition to processing digital image signals, it can also process other digital signals. For example, when the terminal device 100 selects a frequency point, the digital signal processor is used to perform Fourier transform on the frequency point energy, and the like.
- Video codecs are used to compress or decompress digital video.
- the terminal device 100 may support one or more video codecs.
- the terminal device 100 can play or record videos in various encoding formats, for example, moving picture experts group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4 and so on.
- MPEG moving picture experts group
- the NPU is a neural-network (NN) computing processor.
- NN neural-network
- Applications such as intelligent cognition of the terminal device 100 can be implemented through the NPU, such as image recognition, face recognition, speech recognition, text understanding, and the like.
- the external memory interface 120 can be used to connect an external memory card, such as a Micro SD card, to expand the storage capacity of the terminal device 100 .
- the external memory card communicates with the processor 110 through the external memory interface 120 to realize the data storage function. For example to save files like music, video etc in external memory card.
- Internal memory 121 may be used to store computer executable program code, which includes instructions.
- the internal memory 121 may include a storage program area and a storage data area.
- the program storage area may store an operating system, an application program (such as a camera application) required for at least one function, and the like.
- the storage data area may store data created during the use of the terminal device 100 (such as images captured by a camera, etc.) and the like.
- the internal memory 121 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, universal flash storage (UFS), and the like.
- the processor 110 executes various functional applications and data processing of the terminal device 100 by executing instructions stored in the internal memory 121 and/or instructions stored in a memory provided in the processor.
- the internal memory 121 may also store the downlink audio signal provided by the embodiment of the present application.
- the internal memory 121 may also store codes for implementing the functions of controlling the headset 200 .
- the code stored in the internal memory 121 for implementing the function of controlling the earphone 200 is executed by the processor 110, the earphone 200 is controlled to realize the corresponding function, such as the ANC function, the HT function or the AH function.
- the code for implementing the function of controlling the headset 200 provided by the embodiment of the present application may also be stored in an external memory.
- the processor 110 may run the corresponding data stored in the external memory to realize the function of controlling the earphone 200 through the external memory interface 120, so as to control the earphone 200 to realize the corresponding function.
- the terminal device 100 may implement audio functions through an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, an application processor, and the like. Such as music playback, recording, etc.
- the audio module 170 is used for converting digital audio information into analog audio signal output, and also for converting analog audio input into digital audio signal. Audio module 170 may also be used to encode and decode audio signals. In some embodiments, the audio module 170 may be provided in the processor 110 , or some functional modules of the audio module 170 may be provided in the processor 110 .
- Speaker 170A also referred to as a "speaker" is used to convert audio electrical signals into sound signals.
- the terminal device 100 can listen to music through the speaker 170A, or listen to a hands-free call.
- the receiver 170B also referred to as "earpiece" is used to convert audio electrical signals into sound signals.
- the terminal device 100 answers a call or a voice message, the voice can be answered by placing the receiver 170B close to the human ear.
- the microphone 170C also called “microphone” or “microphone” is used to convert sound signals into electrical signals. When making a call or sending a voice message, the user can make a sound near the microphone 170C through the human mouth, and input the sound signal into the microphone 170C.
- the terminal device 100 may be provided with at least one microphone 170C. In other embodiments, the terminal device 100 may be provided with two microphones 170C, which may implement a noise reduction function in addition to collecting sound signals. In other embodiments, the terminal device 100 may further be provided with three, four or more microphones 170C to collect sound signals, reduce noise, identify sound sources, and implement directional recording functions.
- the earphone jack 170D is used to connect wired earphones.
- the terminal device 100 is connected to the earphone through the earphone interface 170D.
- the earphone interface 170D can be a USB interface 130, or a 3.5mm open mobile terminal platform (open mobile terminal platform, OMTP) standard interface, a cellular telecommunications industry association of the USA (CTIA) standard interface.
- OMTP open mobile terminal platform
- CTIA cellular telecommunications industry association of the USA
- the pressure sensor 180A is used to sense pressure signals, and can convert the pressure signals into electrical signals.
- the pressure sensor 180A may be provided on the display screen 194 .
- the capacitive pressure sensor may be comprised of at least two parallel plates of conductive material. When a force is applied to the pressure sensor 180A, the capacitance between the electrodes changes.
- the terminal device 100 determines the intensity of the pressure according to the change in capacitance. When a touch operation acts on the display screen 194, the terminal device 100 detects the intensity of the touch operation according to the pressure sensor 180A.
- the terminal device 100 may also calculate the touched position according to the detection signal of the pressure sensor 180A.
- touch operations acting on the same touch position but with different touch operation intensities may correspond to different operation instructions. For example, when a touch operation whose intensity is less than the first pressure threshold acts on the short message application icon, the instruction for viewing the short message is executed. When a touch operation with a touch operation intensity greater than or equal to the first pressure threshold acts on the short message application icon, the instruction to create a new short message is executed.
- the gyro sensor 180B may be used to determine the motion attitude of the terminal device 100 .
- the angular velocity of the end device 100 about three axes ie, the x, y and z axes
- the gyro sensor 180B can be used for image stabilization.
- the gyro sensor 180B detects the shaking angle of the terminal device 100, calculates the distance to be compensated by the lens module according to the angle, and allows the lens to offset the shaking of the terminal device 100 through reverse motion to achieve anti-shake.
- the gyro sensor 180B can also be used for navigation and somatosensory game scenarios.
- the air pressure sensor 180C is used to measure air pressure.
- the terminal device 100 calculates the altitude through the air pressure value measured by the air pressure sensor 180C to assist in positioning and navigation.
- the magnetic sensor 180D includes a Hall sensor.
- the terminal device 100 can detect the opening and closing of the flip holster using the magnetic sensor 180D.
- the terminal device 100 can detect the opening and closing of the flip according to the magnetic sensor 180D. Further, according to the detected opening and closing state of the leather case or the opening and closing state of the flip cover, characteristics such as automatic unlocking of the flip cover are set.
- the acceleration sensor 180E can detect the magnitude of the acceleration of the terminal device 100 in various directions (generally three axes).
- the magnitude and direction of gravity can be detected when the terminal device 100 is stationary. It can also be used to identify the posture of terminal devices, and can be used in applications such as horizontal and vertical screen switching, pedometers, etc.
- the terminal device 100 can measure the distance through infrared or laser. In some embodiments, when shooting a scene, the terminal device 100 can use the distance sensor 180F to measure the distance to achieve fast focusing.
- Proximity light sensor 180G may include, for example, light emitting diodes (LEDs) and light detectors, such as photodiodes.
- the light emitting diodes may be infrared light emitting diodes.
- the terminal device 100 emits infrared light to the outside through the light emitting diode.
- the terminal device 100 detects infrared reflected light from nearby objects using a photodiode. When sufficient reflected light is detected, it can be determined that there is an object near the terminal device 100 . When insufficient reflected light is detected, the terminal device 100 may determine that there is no object near the terminal device 100 .
- the terminal device 100 can use the proximity light sensor 180G to detect that the user holds the terminal device 100 close to the ear to talk, so as to automatically turn off the screen to save power.
- Proximity light sensor 180G can also be used in holster mode, pocket mode automatically unlocks and locks the screen.
- the ambient light sensor 180L is used to sense ambient light brightness.
- the terminal device 100 may determine the exposure time of the image according to the ambient light brightness sensed by the ambient light sensor 180L.
- the terminal device 100 can adaptively adjust the brightness of the display screen 194 according to the perceived ambient light brightness.
- the ambient light sensor 180L can also be used to automatically adjust the white balance when taking pictures.
- the ambient light sensor 180L can also cooperate with the proximity light sensor 180G to detect whether the terminal device 100 is in a pocket, so as to prevent accidental touch.
- the fingerprint sensor 180H is used to collect fingerprints.
- the terminal device 100 can use the collected fingerprint characteristics to realize fingerprint unlocking, accessing application locks, taking photos with fingerprints, answering incoming calls with fingerprints, and the like.
- the temperature sensor 180J is used to detect the temperature.
- the terminal device 100 uses the temperature detected by the temperature sensor 180J to execute the temperature processing strategy. For example, when the temperature reported by the temperature sensor 180J exceeds a threshold value, the terminal device 100 reduces the performance of the processor located near the temperature sensor 180J, so as to reduce power consumption and implement thermal protection.
- the terminal device 100 when the temperature is lower than another threshold, the terminal device 100 heats the battery 142 to avoid abnormal shutdown of the terminal device 100 caused by the low temperature.
- the terminal device 100 boosts the output voltage of the battery 142 to avoid abnormal shutdown caused by low temperature.
- Touch sensor 180K also called “touch device”.
- the touch sensor 180K may be disposed on the display screen 194 , and the touch sensor 180K and the display screen 194 form a touch screen, also called a “touch screen”.
- the touch sensor 180K is used to detect a touch operation on or near it.
- the touch sensor can pass the detected touch operation to the application processor to determine the type of touch event.
- Visual output related to touch operations may be provided through display screen 194 .
- the touch sensor 180K may also be disposed on the surface of the terminal device 100 , which is different from the position where the display screen 194 is located.
- the bone conduction sensor 180M can acquire vibration signals.
- the bone conduction sensor 180M can acquire the vibration signal of the vibrating bone mass of the human voice.
- the bone conduction sensor 180M can also contact the pulse of the human body and receive the blood pressure beating signal.
- the bone conduction sensor 180M can also be disposed in the earphone, combined with the bone conduction earphone.
- the audio module 170 can analyze the voice signal based on the vibration signal of the vocal vibration bone block obtained by the bone conduction sensor 180M, so as to realize the voice function.
- the application processor can analyze the heart rate information based on the blood pressure beat signal obtained by the bone conduction sensor 180M, and realize the function of heart rate detection.
- the keys 190 include a power-on key, a volume key, and the like. Keys 190 may be mechanical keys. It can also be a touch key.
- the terminal device 100 may receive key input and generate key signal input related to user settings and function control of the terminal device 100 .
- Motor 191 can generate vibrating cues.
- the motor 191 can be used for vibrating alerts for incoming calls, and can also be used for touch vibration feedback.
- touch operations acting on different applications can correspond to different vibration feedback effects.
- the motor 191 can also correspond to different vibration feedback effects for touch operations on different areas of the display screen 194 .
- Different application scenarios for example: time reminder, receiving information, alarm clock, games, etc.
- the touch vibration feedback effect can also support customization.
- the indicator 192 can be an indicator light, which can be used to indicate the charging state, the change of the power, and can also be used to indicate a message, a missed call, a notification, and the like.
- the SIM card interface 195 is used to connect a SIM card.
- the SIM card can be contacted and separated from the terminal device 100 by inserting into the SIM card interface 195 or pulling out from the SIM card interface 195 .
- the terminal device 100 may support 1 or N3 SIM card interfaces, where N3 is a positive integer greater than 1.
- the SIM card interface 195 can support Nano SIM card, Micro SIM card, SIM card and so on. Multiple cards can be inserted into the same SIM card interface 195 at the same time. The types of the plurality of cards may be the same or different.
- the SIM card interface 195 can also be compatible with different types of SIM cards.
- the SIM card interface 195 is also compatible with external memory cards.
- the terminal device 100 interacts with the network through the SIM card to realize functions such as calls and data communication.
- the terminal device 100 adopts an eSIM, that is, an embedded SIM card.
- the eSIM card can be embedded in the terminal device 100 and cannot be separated from the terminal device 100 .
- the software system of the terminal device 100 may adopt a layered architecture, an event-driven architecture, a microkernel architecture, a microservice architecture, or a cloud architecture.
- the embodiments of the present invention take an Android system with a layered architecture as an example to exemplarily describe the software structure of the terminal device 100 .
- FIG. 2 is a block diagram of a software structure of a terminal device 100 according to an embodiment of the present invention.
- the layered architecture divides the software into several layers, and each layer has a clear role and division of labor. Layers communicate with each other through software interfaces.
- the Android system is divided into four layers, which are, from top to bottom, an application layer, an application framework layer, an Android runtime (Android runtime) and a system library, and a kernel layer.
- the application layer can include a series of application packages.
- the application package can include applications such as camera, gallery, calendar, call, map, navigation, WLAN, Bluetooth, music, video, short message and so on.
- the application framework layer provides an application programming interface (application programming interface, API) and a programming framework for applications in the application layer.
- the application framework layer includes some predefined functions.
- the application framework layer may include window managers, content providers, view systems, telephony managers, resource managers, notification managers, and the like.
- a window manager is used to manage window programs.
- the window manager can get the size of the display screen, determine whether there is a status bar, lock the screen, take screenshots, etc.
- Content providers are used to store and retrieve data and make these data accessible to applications.
- the data may include video, images, audio, calls made and received, browsing history and bookmarks, phone book, etc.
- the view system includes visual controls, such as controls for displaying text, controls for displaying pictures, and so on. View systems can be used to build applications.
- a display interface can consist of one or more views.
- the display interface including the short message notification icon may include a view for displaying text and a view for displaying pictures.
- the telephony manager is used to provide the communication function of the terminal device 100 .
- the management of call status including connecting, hanging up, etc.).
- the resource manager provides various resources for the application, such as localization strings, icons, pictures, layout files, video files and so on.
- the notification manager enables applications to display notification information in the status bar, which can be used to convey notification-type messages, and can disappear automatically after a brief pause without user interaction. For example, the notification manager is used to notify download completion, message reminders, etc.
- the notification manager can also display notifications in the status bar at the top of the system in the form of graphs or scroll bar text, such as notifications of applications running in the background, and notifications on the screen in the form of dialog windows. For example, text information is prompted in the status bar, a prompt sound is issued, the terminal device vibrates, and the indicator light flashes.
- Android Runtime includes core libraries and a virtual machine. Android runtime is responsible for scheduling and management of the Android system.
- the core library consists of two parts: one is the function functions that the java language needs to call, and the other is the core library of Android.
- the application layer and the application framework layer run in virtual machines.
- the virtual machine executes the java files of the application layer and the application framework layer as binary files.
- the virtual machine is used to perform functions such as object lifecycle management, stack management, thread management, safety and exception management, and garbage collection.
- a system library can include multiple functional modules. For example: surface manager (surface manager), media library (Media Libraries), 3D graphics processing library (eg: OpenGL ES), 2D graphics engine (eg: SGL), etc.
- surface manager surface manager
- media library Media Libraries
- 3D graphics processing library eg: OpenGL ES
- 2D graphics engine eg: SGL
- the Surface Manager is used to manage the display subsystem and provides a fusion of 2D and 3D layers for multiple applications.
- the media library supports playback and recording of a variety of commonly used audio and video formats, as well as still image files.
- the media library can support a variety of audio and video encoding formats, such as: MPEG4, H.264, MP3, AAC, AMR, JPG, PNG, etc.
- the 3D graphics processing library is used to implement 3D graphics drawing, image rendering, compositing, and layer processing.
- 2D graphics engine is a drawing engine for 2D drawing.
- the kernel layer is the layer between hardware and software.
- the kernel layer includes at least display drivers, camera drivers, audio drivers, headphone drivers, and sensor drivers.
- the workflow of the software and hardware of the terminal device 100 is exemplarily described below in conjunction with the audio capture and playback scene.
- a corresponding hardware interrupt is sent to the kernel layer.
- the kernel layer processes touch operations into raw input events (including touch coordinates, timestamps of touch operations, etc.). Raw input events are stored at the kernel layer.
- the application framework layer obtains the original input event from the kernel layer, and identifies the control corresponding to the input event. Take the touch operation as a touch click operation, and the control corresponding to the click operation is an audio application icon control as an example, the audio application invokes the interface of the application framework layer to start the headphone control application, and then starts the headphone driver by calling the kernel layer, The audio signal is sent to the earphone, and the audio signal is played through the earphone 200 .
- FIG. 3 shows a schematic diagram of an optional hardware structure of the earphone 200 .
- the earphone 200 includes a left earphone and a right earphone.
- the left earphone and the right earphone have similar structures.
- the structures of the earphones include a first microphone 301 , a second microphone 302 and a third microphone 303 .
- a processor 304 and a speaker 305 may also be included in the headset. It should be understood that the earphone described later may be interpreted as the left earphone, and may also be interpreted as the right earphone.
- the first microphone 301 is used to collect the sound of the current external environment, and the first microphone 301 may also be referred to as a reference microphone.
- the first microphone 301 is located outside the earphone, or the second microphone 301 is located outside the ear.
- the second microphone 302 collects the sound of the internal environment of the user's ear canal.
- the second microphone 302 may also be referred to as an error microphone.
- the second microphone 302 is located inside the earphone and close to the ear canal.
- the third microphone 303 is used to collect call signals.
- the third microphone 303 may be located outside the earphone. When the user wears the earphone, the third microphone 303 is closer to the user's mouth than the first microphone 301 .
- the first microphone 301 is used to collect the sound of the current external environment, which can be interpreted as the sound of the external environment where the headset user wears the headset, for example, on a train, the sound of the external environment is the surrounding environment of the user wearing the headset the sound of.
- the first microphone 301 on the left earphone adopts the sound of the external environment of the left earphone.
- the first microphone 301 on the right earphone collects the sound of the external environment of the right earphone.
- the signal collected by the first microphone 301 (reference microphone) is called the first signal
- the signal collected by the second microphone 302 (error microphone) is called the second signal.
- the microphone involved in the embodiments of the present application may be an analog microphone or a digital microphone.
- the microphone is an analog microphone, before filtering the signal collected by the microphone, the analog signal can be converted into a digital signal.
- the first microphone and the second microphone are both digital microphones as an example for description, and the first signal and the second signal are both digital signals.
- the processor 304 is used to process downlink audio signals and/or signals collected by the microphones (including the first microphone 301 , the second microphone 302 and the third microphone 303 ), such as performing ANC processing, HT processing, or AH processing.
- the processor 304 may include a main control unit and a noise reduction processing unit.
- the main control unit is used for the user to generate control commands for the operation of the earphone or to receive control commands from the terminal device.
- the noise reduction processing unit is configured to perform ANC processing, HT processing or AH processing on downlink audio signals and signals collected by the microphones (including the first microphone 301 , the second microphone 302 and the third microphone 303 ) according to the control command.
- the left and right earphones may also include memory for storing programs or instructions executed by the processor 304 .
- the processor 304 performs ANC processing, HT processing, or AH processing according to programs or instructions stored in the memory.
- the memory may include random access memory (RAM), flash memory, read-only memory (ROM), programmable read-only memory (PROM), erasable programmable read-only memory (erasable).
- RAM random access memory
- ROM read-only memory
- PROM programmable read-only memory
- PROM erasable programmable read-only memory
- the main control unit for example, can be composed of an ARM processing chip, a central processing unit (CPU), a system on chip (SoC), a digital signal processing (DSP), or a micro controller (micro controller). unit, MCU) in one or more implementations.
- the noise reduction processing unit may include, for example, a codec (coder-decoder, CODEC) chip or a high-fidelity (high-fidelity, HiFi) chip, or the like.
- the codec is hardened with a filter, an equalizer (EQ), a dynamic range controller (DRC), a limiter (limiter), and a gain adjuster ( Gain), mixer (mixer), etc., are mainly used for filtering, mixing, gain adjustment and other processing of signals.
- the noise reduction processing unit may further include a DSP, and the DSP may be used for processing such as scene detection, speech enhancement, and occlusion elimination.
- the headset may further include a wireless communication unit for establishing a communication connection with the terminal device 200 through the wireless communication module 160 in the terminal device 100 .
- the wireless communication unit can provide applications on the headset including wireless local area networks (WLAN) (such as wireless fidelity (Wi-Fi) networks), Bluetooth (bluetooth, BT), short-range wireless communication technology ( near field communication, NFC), infrared technology (infrared, IR) and other wireless communication solutions.
- the wireless communication unit may be one or more devices integrating at least one communication processing module.
- the wireless communication module 160 may be Bluetooth, the wireless communication unit is also Bluetooth, and the headset 200 and the terminal device 100 are connected through Bluetooth.
- the ANC function when the ANC function is turned on, it means that the signal path of the ANC function is in an active state, and correspondingly, each functional module in the ANC pathway is also in an active state; the turning on of the HT function means that the signal path of the HT function is in an active state. In the activated state, correspondingly, each functional module in the HT pathway is also in the activated state.
- both the ANC function and the HT function of the earphone are in an on state, it indicates that the signal pathway of the ANC function is in an active state, and the signal pathway of the HT function is also in an active state. That is, it can indicate the working state of the headset, and is not limited to specific operations at a certain moment or changes in functional controls.
- the active noise reduction processing in the active noise reduction output path may include, but is not limited to: performing noise suppression using the inverted signal of the first signal collected by the reference microphone and the inverted signal of the second signal collected by the error microphone.
- the active noise reduction output path includes an inverted signal of the first signal and an inverted signal of the second signal. It should be noted that the phase difference between the first signal and the inverted signal of the first signal is 180°.
- the speaker outputs a signal obtained by superimposing the inverted signal of the first signal and the inverted signal of the second signal, so that the sound of the current external environment played by the speaker and the sound of the external environment actually heard by the ear are cancelled to achieve active denoising. Effect. Therefore, when the headset adopts the ANC mode, the headset user's perception of the sound of the current environment and the ambient sound inside the user's ear canal can be weakened.
- filter compensation can be performed on the downlink audio signal.
- the influence of the downlink audio signal can be removed when the inverted signal of the ambient sound is obtained.
- the first filtering process and the third filtering process may be used.
- the first filtering process may be a feed forward (FF) filtering process, which may be implemented by a feed forward filter.
- the third filtering process may be a feedback (feedback, FB) filtering process, which may be implemented by a feedback filter.
- FF filtering and FB filtering use a parallel processing architecture to enhance the noise reduction effect. The processing flow of the ANC will be described in detail later, and the description will not be repeated here.
- the ambient sound transparent transmission processing in the outgoing path of ambient sound transmission may include, but is not limited to, including: performing a third filtering process on the first signal collected by the error microphone to achieve part of the active noise reduction function, and performing a second filtering process on the signal collected by the reference microphone as well as HT enhanced processing.
- the second filtering process may be a transparent transmission (HT) filtering process, which may be implemented by a transparent transmission filter.
- the audio signal played by the speaker is obtained according to the first signal and the second signal, so that after the audio signal played by the speaker, the user can hear the sound in the external environment through the earphone, which is compared with the sound of the external environment heard when the HT processing is not performed.
- the sound intensity is better and the effect is better. Therefore, when the headset adopts the HT mode, the user's perception of the intensity of the sound in the environment where the user is currently located can be enhanced.
- the processing flow of the HT will be described in detail later, and the description will not be repeated here.
- the transparent transmission processing of ambient sound in the auditory enhancement output path may include, but is not limited to, using the signal collected by the error microphone to realize part of the active noise reduction function, performing first filtering processing on the signal collected by the reference microphone, and hearing enhancement processing to make the environment where the user is located.
- the sound of the event in the sound is enhanced, and a second filtering process is performed on the signal collected by the reference microphone.
- the output signal of the speaker is obtained according to the mixed signal of the event signal in the first signal and the inverted signal of the second signal. It should be noted that the phase difference between the second signal and the inverted signal of the second signal is 180°.
- the speaker outputs the inverted signal of the second signal, the inverted signal of the first signal and the signal after the event signal in the first signal is superimposed, so that the signal output by the speaker and the sound in the environment actually heard by the ear are canceled to achieve the initiative.
- the processing flow of the AH will be described in detail later, and the description will not be repeated here.
- the downlink audio signal, the first signal and the second signal may be a frame of signals or signals of a period of time.
- the downlink audio signal, the first signal and the second signal are all one-frame signals
- the downlink audio signal, the first signal and the second signal belong to three signal streams respectively
- the signal frame of the downlink audio signal and the signal of the first signal The frame and the signal frame of the second signal are in the same time period or overlap in time, etc.
- functional processing such as ANC, HT, or AH
- the functional processing is continuously performed for the signal flow where the downlink audio signal is located, the signal flow where the first signal is located, and the signal flow of the second signal.
- the downlink audio signal sent by the terminal device 100 to the headset 200 may also be referred to as the first audio signal in the subsequent description.
- the first audio signal may be a call signal, a music signal, or the like.
- the signal collected by the reference microphone is called the first signal
- the signal collected by the error microphone is called the second signal. Headphones are in ANC mode.
- the downlink audio signals sent by the terminal device 100 to the left earphone and the right earphone in the earphone 200 may be the same signal or different signals.
- the terminal device adopts stereo effect, and the terminal device 100 sends different downlink audio signals to the earphone 200 to realize the stereo effect.
- the terminal device can also send the same downlink audio signal to the left earphone and the right earphone, and the left earphone and the right earphone adopt stereo processing, so as to achieve a stereo effect.
- the left earphone or the right earphone may perform the process of FIG. 5A or FIG. 5B according to the user's control.
- S501 Perform a first filtering process on a first signal collected by a reference microphone to obtain a first filtered signal.
- the first filtered signal is simply referred to as signal A1 in FIG. 5B .
- the first filtered signal is simply referred to as signal A2 in Figure 5B.
- S503 Perform sound mixing processing on the first filtered signal and the first filtered signal to obtain a third audio signal.
- the third audio signal is simply referred to as the signal A3, that is, the signal A3 is obtained by mixing the signal A1 and the signal A2.
- the fourth audio signal is simply referred to as signal A4 in FIG. 5B .
- S505. Perform sound mixing processing on the fourth audio signal and the first audio signal to obtain the second audio signal.
- the speaker is responsible for playing the second audio signal.
- the second audio signal is simply referred to as A5.
- the signal output by the speaker is the fourth audio frequency that has not been mixed. Signal.
- S502 and S505 need not be performed.
- the first filtering process is an FF filtering process implemented by an FF filter as an example
- the third filtering process is an FB filtering process implemented by an FB filter as an example.
- the reference microphone in the earphone 200 picks up the first signal, which is input to the FF filter to obtain the signal A1 after FF filtering processing.
- the error microphone picks up the second signal and inputs it to the subtractor.
- the downlink audio signal is also input to the subtractor after filtering and compensating.
- the subtracter removes the downlink audio signal after filtering and compensation included in the second signal, so as to eliminate the effect of the downlink audio signal. Affect gets signal A2.
- the signal A3 is obtained after mixing the signal A1 and the signal A2 by the sound mixer, and the signal A4 is input into the FB filter, and the signal A4 is obtained after the FB filtering process. After the signal A4 is mixed with the downlink audio signal, the signal A5 is obtained, and the signal A5 is input to the speaker for playback.
- the quality of the ANC effect may be determined by the processing intensity of the ANC processing.
- the processing strength of the ANC processing depends on the FF filter coefficients used for FF filtering and/or the FB filter coefficients used for FB filtering.
- the default FF filter coefficients in the ANC mode may be used.
- the FF filter coefficient used when the ANC mode was selected last time can be used.
- the earphone determines the FF filter coefficient used in the ANC mode according to the recognized scene.
- the user indicates to the headset the FF filter coefficient used in the ANC mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the ANC mode as the target processing intensity through the UI control provided by the terminal device, and different processing intensities correspond to different FF filter coefficients.
- the default FB filter coefficient in the ANC mode can be used.
- the FB filter coefficient used when the ANC mode was selected last time can be used.
- the earphone determines the adopted FB filter coefficient according to the recognized scene.
- the user indicates to the headset the FF filter coefficient used in the ANC mode through the UI control provided by the terminal device, for example, the user selects the processing intensity in the ANC mode through the UI control provided by the terminal device as the target processing intensity, Different processing strengths correspond to different FB filter coefficients.
- the acquisition of the FF filter coefficient and the FB filter coefficient in the ANC mode may be any combination of the above-mentioned methods.
- the FF filter adopts the default filter coefficient in the ANC mode, and the FB filter coefficient is determined by the headset according to the recognized scene.
- the FB filter adopts the default filter coefficient in the ANC mode, and the FF filter coefficient is determined by the user through the UI control provided by the terminal device.
- the FB filter adopts the default filter coefficient in the ANC mode, and the FF filter coefficient is selected and indicated to the headset by the user through the UI control provided by the terminal device.
- the determination of the processing intensity in the ANC mode will be described in detail later by using a specific example, and the description will not be repeated here.
- FIG. 6A , FIG. 6B and FIG. 6C it is a schematic diagram of a process flow of transparent transmission of ambient sound.
- the downlink audio signal sent by the terminal device 100 to the earphone 200 is described below by taking the first audio signal as an example.
- the first audio signal may be a call signal, a music signal, or the like.
- the signal collected by the reference microphone is called the first signal
- the signal collected by the error microphone is called the second signal.
- the left earphone or the right earphone in the earphone 200 may perform the process of FIG. 6A or FIG. 6B or FIG. 6C according to the user's control.
- S601 Perform first signal processing on the first signal collected by the reference microphone to obtain a first processed signal.
- the first processed signal is referred to as signal B1 in Figures 6B and 6C.
- the first signal processing includes HT filtering.
- S602. Perform sound mixing processing on the first processed signal and the first audio signal to obtain a fifth audio signal.
- the fifth audio signal is referred to as signal B2 in Figures 6B and 6C.
- the signal B2 is obtained by performing mixing processing on the signal B1 and the downlink audio signal (ie, the first audio signal).
- the second filtered signal is referred to as signal B3 in Figures 6B and 6C. That is, the signal B3 is obtained by filtering out the signal B2 included in the second ambient signal.
- the third filtered signal is referred to as signal B4 in Figures 6B and 6C. That is, the signal B4 is obtained by FB filtering the signal B3.
- the first processed signal when the first processed signal is obtained by performing the first signal processing on the first signal collected by the reference microphone, it may be implemented in the following manner:
- the HT filtering is performed on the first signal to obtain a second filtered signal.
- the second filtered signal is referred to as signal B5.
- the second signal processing is performed on the second filtered signal to obtain the second processed signal; the second signal processing may also be referred to as low-latency algorithm processing.
- the low-latency algorithm processing includes one or more of de-blocking effect processing, noise reduction processing, wind noise reduction processing, gain adjustment processing or frequency response adjustment processing.
- the HT filtering process may be implemented by a noise reduction processing unit, as shown in FIG. 6B .
- the noise reduction processing unit of the earphone includes a CODEC.
- the CODEC includes an HT filter, an FB filter, a subtractor, a first mixer, a second mixer, and a filter compensation unit.
- FIG. 6B takes as an example that the noise reduction processing unit further includes a DSP.
- the DSP can be used to perform low-latency algorithmic processing.
- the reference microphone in the earphone 200 picks up the first signal, which is input to the HT filter to obtain the signal B5 after HT filter processing.
- the signal B5 is input to the DSP, and the DSP performs low-latency algorithm processing on the signal B5 to obtain the signal B1.
- the signal B1 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal B1 to obtain the signal B2.
- After the signal B2 is processed by the filter compensation unit, it is input to the subtractor.
- the subtractor is used to filter the signal B2 that has undergone filtering and compensation processing included in the second ambient signal picked up by the error microphone to obtain a signal B3.
- the signal B3 is input to the FB filter, and the FB filter performs the FB filtering process on the signal B3 to obtain the signal B4.
- Signal B4 is input to the second mixer.
- the input of the second mixer also includes signal B2.
- the second sound mixer performs sound mixing processing on the signal B2 and the signal B4 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
- the HT filtering process can be implemented by DSP, as shown in FIG. 6C .
- the DSP can be used to perform HT filtering processing as well as low-latency algorithm processing.
- the noise reduction processing unit of the earphone includes an FB filter, a subtractor, a first sound mixer, a second sound mixer and a filter compensation unit.
- the reference microphone in the earphone picks up the first signal and inputs it to the DSP.
- the DSP performs HT filtering processing on the first signal, and performs low-latency algorithm processing to obtain a signal B1.
- the signal B1 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal B1 to obtain the signal B2.
- the signal B2 After the signal B2 is processed by the filter compensation unit, it is input to the subtractor.
- the subtractor is used to filter the signal B2 included in the second signal picked up by the error microphone to obtain the signal B3.
- the signal B3 is input to the FB filter, and the FB filter performs the FB filtering process on the signal B3 to obtain the signal B4.
- Signal B4 is input to the second mixer.
- the input of the second mixer also includes signal B2.
- the second sound mixer performs sound mixing processing on the signal B2 and the signal B4 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
- the low-latency algorithm processing includes de-blocking effect processing.
- the principle of the occlusion effect is explained first.
- Path 1 from the bone conduction to the periosteum, which is only a low-frequency signal.
- Path 2 It propagates from the outside air to the periosteum and is perceived, and the signal contains low frequency signals and medium and high frequency signals.
- the low-frequency signal and the medium-high frequency signal are superimposed, the low-frequency signal is too sufficient, and the low-frequency signal cannot be emitted when the earphone is worn, resulting in low-frequency chaos inside the ear, resulting in an occlusion effect.
- the signal B5 obtained after the HT filtering process is processed by the de-blocking effect, and the specific method may be as follows.
- Mode 1 is shown in FIG. 7 .
- a first voice harmonic signal matching the bone conduction signal in a voice harmonic set where the voice harmonic set includes a plurality of voice harmonic signals.
- the plurality of speech harmonic signals included in the speech harmonic set correspond to different frequencies. Specifically, the frequency of the bone conduction signal can be determined, and the first voice harmonic signal is determined from the set of voice harmonics according to the frequency of the bone conduction signal.
- the speech harmonic signal may also be referred to as the speech harmonic component.
- S702 remove the first voice harmonic signal in the signal B5 obtained after the HT filtering process.
- the signal C1 is obtained by removing the first voice harmonic signal in the signal B5 obtained after the HT filtering process. Since the voice of a person collected by a general bone conduction sensor is generally a low-frequency harmonic component, S702 is to remove the low-frequency harmonic component in the signal B5, thereby obtaining a signal C1 that does not include the low-frequency harmonic component.
- S703 Amplify the high frequency components in the signal B5 from which the first voice harmonic signal is removed. That is, the high frequency components of the signal C1 are amplified.
- the first voice harmonic signal matching the bone conduction signal can be determined, that is, the bone conduction sensor can detect the bone conduction signal, that is, the earphone wearer is currently speaking, such as speaking or singing.
- the signal obtained by boosting the high-frequency components on the basis of the signal C1 includes only mid- and high-frequency components, so that the signal heard by the earphone wearer has relatively no occlusion effect.
- the set of speech harmonics may be pre-stored in the headset.
- the set of speech harmonics may be obtained in an offline manner or in an online manner.
- the bone conduction signals of multiple people can be collected through the bone conduction sensor, and the following processing is performed on the bone conduction signals of each person in the bone conduction signals of the multiple people:
- the first bone conduction signal is transformed into a frequency domain signal by FFT; the fundamental frequency signal in the frequency domain signal is determined by means of a pilot frequency to find the fundamental frequency; the harmonic component of the bone conduction signal is determined according to the fundamental frequency signal, Thereby, the mapping relationship between the frequency of the bone conduction signal and the harmonic component is obtained to obtain a set of speech harmonics.
- the voice harmonic set may include mapping relationships between different frequencies and harmonic components.
- the second bone conduction signal collected within a set period of time can be obtained through the bone conduction sensor in the headset.
- the set period of time there may be multiple people using the headset, or only one person may use the headset, that is, the user;
- the fundamental frequency signal in the frequency domain signal is determined by means of a pilot frequency to find the fundamental frequency. If multiple people use the headset within the set time period, it may be possible to determine multiple fundamental frequency signals corresponding to multiple different time periods within the set time period.
- Multiple harmonic components of the bone conduction signal may be determined according to multiple fundamental frequency signals, so as to obtain a mapping relationship between frequencies and harmonic components to obtain a voice harmonic set.
- the voice harmonic set may include mapping relationships between different frequencies and different harmonic components.
- adaptive filtering can be performed on the signal B5 obtained after the HT filtering process to remove the low frequency component in the signal B5 to obtain the signal C1, that is, the signal of the earphone wearer's voice in the signal B5 is removed.
- the high frequency components in the third filtered signal from which the low frequency components are removed are amplified, that is, the high frequency components in the signal C1 are amplified.
- the signal obtained by boosting the high-frequency components on the basis of the signal C1 includes only mid- and high-frequency components, so that the signal heard by the earphone wearer has relatively no occlusion effect.
- the quality of the HT effect can be determined by the treatment intensity of the HT treatment.
- the processing strength of the HT processing depends on the HT filter coefficients used for the HT filtering and/or the FB filter coefficients used for the FB filtering.
- the default HT filter coefficients in the HT mode may be used.
- the HT filter coefficient used when the HT mode was selected last time can be used.
- the earphone determines the HT filter coefficient used in the HT mode according to the recognized scene.
- the user indicates to the headset the HT filter coefficient used in the HT mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the HT mode as the target processing intensity through the UI control provided by the terminal device, and different processing intensities correspond to different HT filter coefficients.
- the FB filter coefficient in one way, the default FB filter coefficient in the HT mode can be used.
- the FB filter coefficient used when the HT mode was selected last time can be used.
- the earphone determines the adopted FB filter coefficient according to the recognized scene.
- the user indicates to the headset the HT filter coefficient used in the HT mode through the UI control provided by the terminal device, for example, the user selects the processing intensity in the HT mode through the UI control provided by the terminal device as the target processing intensity, Different processing strengths correspond to different FB filter coefficients.
- the acquisition of the HT filter coefficients and the FB filter coefficients in the HT mode may be any combination of the methods provided above.
- the downlink audio signal sent by the terminal device 100 to the earphone 110 is referred to as the first audio signal as an example in the subsequent description.
- the first audio signal may be a call signal, a music signal, a prompt tone, or the like.
- the signal collected by the reference microphone is called the first signal
- the signal collected by the error microphone is called the second signal.
- the left earphone or the right earphone in the earphone 200 may perform the process of FIG. 8A or FIG. 8B or FIG. 8C according to the user's control.
- S801 Perform HT filtering on the first signal collected by the reference microphone to obtain a second filtered signal (signal C1).
- the second filtered signal is referred to as signal C1.
- the first filtered signal is simply referred to as signal C3.
- Step S804 Perform sound mixing processing on the filtered enhanced signal and the first audio signal to obtain a sixth audio signal.
- the sixth audio signal is simply referred to as signal C4.
- Step S804 is to mix the signal C2 and the downlink audio signal to obtain the signal C4.
- Step S805 filtering the sixth audio signal included in the second signal to obtain a fourth filtered signal.
- the fourth filtered signal is simply referred to as signal C5.
- Step S805 is to filter out the signal C4 included in the second environment signal to obtain the signal C5.
- step S805 when step S805 is performed, the signal C4 may be filtered and compensated first to obtain a compensated signal, and then the compensated signal included in the second signal may be filtered to obtain C5.
- Step S806 is to perform FB filtering on the signal C5 to obtain the signal C6.
- Step S807 performing mixing processing on the fifth filtered signal, the sixth audio signal and the first filtered signal to obtain the second audio signal.
- Step S806 is to mix the signal C6, the signal C4 and the signal C3 to obtain the second audio signal.
- Mode 1 is shown in FIG. 9 .
- the manner of performing de-blocking effect processing on the signal C1 may be the same as that for performing the de-blocking effect processing on the signal B5. For details, refer to manners 1 and 2 in scenario 2, which will not be repeated here.
- noise reduction processing is performed on the signal processed by the de-occlusion effect, and the noise reduction processing includes artificial intelligence AI noise reduction processing and/or wind noise reduction processing; in FIG. 9, the noise reduction processing includes AI noise reduction processing and wind noise reduction processing. Take processing as an example.
- a gain amplification process is performed on the signal obtained after the wind noise processing, and a feasible way is to directly amplify the signal obtained after the wind noise processing.
- the method of directly amplifying the signal not only amplifies the external signal, but also amplifies the voice of the wearer.
- the embodiment of the present application provides a gain amplification processing method, which only amplifies the external signal, and does not amplify the voice signal speaking to the wearer. For example, as shown in FIG. 10 , when the signal obtained after the noise reduction processing is subjected to gain amplification processing, it can be implemented in the following manner.
- the voice signal of the wearer's speech is conducted to the periosteum through the bone, and the voice signal is concentrated in the low frequency, which is recorded as the bone conduction signal D1.
- the bone conduction signal D1 is acquired by the bone conduction sensor.
- the bone conduction signal D1 is harmonically extended to obtain a harmonically extended signal, for example, the harmonically extended signal is called D2.
- the harmonic extension may be a harmonic enhancement method or a method in which the harmonics of the bone conduction signal D1 are directly upwardly extended.
- the signal obtained by the noise reduction process is referred to as signal D3.
- the signal D4 is obtained by amplifying the signal D3 by using the first gain coefficient.
- the amplification processing here can directly amplify the signal.
- the first filter coefficient is determined based on the first gain coefficient.
- the first gain coefficient gain to adjust the strength of the adaptive filtering can also be called the first filter coefficient, that is, the number of dB used for signal amplification by the first gain coefficient is the same as the number of dB filtered out by the adaptive filter, so that the wearer's voice The signal can remain balanced, neither amplified nor reduced.
- the second method is shown in FIG. 11 .
- the manner of performing de-blocking effect processing on the signal C1 may be the same as that for performing the de-blocking effect processing on the signal B5. For details, refer to manners 1 and 2 in scenario 2, which will not be repeated here.
- S1102 Perform audio event detection on the deblocked signal to obtain an audio event signal (which may be referred to as an event signal for short) in the deblocked signal. Audio event signals, such as station announcements, whistles, etc.
- S1103 Perform gain amplification on the audio event signal in the deblocked signal.
- the audio event signal in the de-blocked signal such as station announcement sound, whistle sound, etc.
- the method of performing gain amplification processing on the audio event signal in the de-blocked signal may be the same as when performing gain amplification processing on the signal obtained after the noise reduction processing, and will not be repeated here.
- the noise reduction processing unit includes a CODEC and a DSP.
- the CODEC of the earphone includes an HT filter, a FB filter, a FF filter, a subtractor, a first mixer, a second mixer, and a filter compensation unit.
- the HT filtering process is implemented by the CODEC.
- DSP can be used to perform enhancement processing.
- the reference microphone in the earphone 110 picks up the first signal, which is input to the HT filter to obtain the signal C1 after HT filter processing.
- the signal C1 is input to the DSP, and the DSP performs enhancement processing on the signal C1 to obtain the signal C2.
- the signal C2 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal C2 to obtain the signal C4.
- the signal C4 is processed by the filter compensation unit, it is input to the subtractor.
- the subtractor is used for filtering the filtered and compensated signal C4 included in the second ambient signal picked up by the error microphone to obtain the signal C5.
- the signal C5 is input to the FB filter, and the FB filter performs the FB filtering process on the signal C5 to obtain the signal C6.
- Signal C6 is input to the second mixer.
- the input of the second mixer also includes signal C4 and signal C3.
- the second audio mixer performs audio mixing processing on the signal C3, the signal C4 and the signal C6 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
- the noise reduction processing unit includes a CODEC and a DSP as an example.
- the DSP can be used to perform HT filtering processing as well as enhancement processing.
- the CODEC of the earphone includes a FB filter, an FF filter, a subtractor, a first mixer, a second mixer, and a filter compensation unit.
- the reference microphone in the earphone 110 picks up the first signal and inputs it to the DSP.
- the DSP performs HT filtering processing on the first signal to obtain a signal C1.
- the DSP performs enhancement processing on the signal C1 to obtain a signal C2.
- the signal C2 is input to the first sound mixer, and the first sound mixer performs sound mixing processing on the downlink audio signal and the signal C2 to obtain the signal C4.
- the signal C4 is processed by the filter compensation unit, it is input to the subtractor.
- the subtractor is used for filtering the filtered and compensated signal C4 included in the second ambient signal picked up by the error microphone to obtain the signal C5.
- the signal C5 is input to the FB filter, and the FB filter performs the FB filtering process on the signal C5 to obtain the signal C6.
- Signal C6 is input to the second mixer.
- the input of the second mixer also includes signal C4 and signal C3.
- the second audio mixer performs audio mixing processing on the signal C3, the signal C4 and the signal C6 to obtain a second audio signal, and the second audio signal is input to the speaker for playback.
- the quality of the AH effect can be determined by the treatment intensity of the AH treatment.
- the processing strength of the AH processing depends on at least one of the HT filter coefficient, the FB filter coefficient, or the FF filter coefficient.
- the default FF filter coefficient in the AH mode can be used.
- the FF filter coefficient used when the AH mode was selected last time can be used.
- the earphone determines the FF filter coefficient used in the AH mode according to the recognized scene.
- the user indicates to the headset the FF filter coefficient used in the AH mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the AH mode as the target processing intensity through a UI control provided by the terminal device, and different processing intensities correspond to different FF filter coefficients.
- the default HT filter coefficients in the AH mode may be used.
- the HT filter coefficient used when the AH mode was selected last time can be used.
- the earphone determines the HT filter coefficient used in the AH mode according to the recognized scene.
- the user indicates to the headset the HT filter coefficients used in the AH mode through a UI control provided by the terminal device. For example, the user selects the processing intensity in the AH mode as the target processing intensity through the UI control provided by the terminal device, and different processing intensities correspond to different HT filter coefficients.
- the FB filter coefficient in one way, the default FB filter coefficient in the AH mode can be used. In another way, the FB filter coefficient used when the AH mode was selected last time can be used.
- the earphone determines the adopted FB filter coefficient according to the recognized scene.
- the user indicates to the headset the HT filter coefficient used in the AH mode through the UI control provided by the terminal device, for example, the user selects the processing intensity in the AH mode through the UI control provided by the terminal device as the target processing intensity, Different processing strengths correspond to different FB filter coefficients.
- the acquisition of the HT filter coefficient, the FB filter coefficient or the FF filter coefficient in the AH mode may be any combination of the above-mentioned methods.
- the determination of the processing mode adopted by the earphone 200 can be selected by the user through the UI control on the terminal device 100 and indicated to the earphone, or the terminal device can determine and indicate to the earphone according to the adaptively recognized scene, Alternatively, it can also be determined by the headset according to adaptive scene recognition.
- Example 1 a single control controls the left and right earphones.
- the terminal device 100 provides a control interface for the user to select the processing mode of the earphone 200 (including the left earphone and the right earphone) according to requirements: empty mode, ANC mode, HT mode or AH mode. No processing is done in empty mode.
- the processing modes of the earphones for the user to select on the control interface are all processing modes supported by the earphones.
- the left earphone and the right earphone have the same processing function, or support the same processing mode.
- the left earphone supports AHA
- the right earphone also supports AHA.
- the terminal device is installed with an earphone application adapted to the earphone 200, and during the adaptation process, the processing function of the earphone can be known.
- the terminal device can determine the processing function possessed by the headset according to the function parameters.
- the control interface includes user interface (user interface, UI) controls.
- the UI control is used for the user to select the processing mode of the headset 200 , and for the convenience of distinction, the UI control used for the user to select the processing mode of the headset is called a selection control.
- the processing mode includes at least two of ANC mode, HT mode or AH mode.
- the terminal device 100 sends control signaling 1 to the left earphone and the right earphone respectively in response to the user selecting the target mode from the processing modes supported by the earphone through the selection control. Control signaling 1 carries the target mode.
- the selection controls can also be used to select the processing intensity in target mode.
- the shape of the selection control can be a circle, a bar or other shapes.
- the selection controls may include a first control and a second control. Any two different positions of the second control on the first control correspond to different processing modes of the headset or two different positions of the second control on the first control correspond to different processing intensities in the same processing mode of the headset.
- the user selects different processing modes and controls the processing intensity by moving the position of the second control representing the user's selection on the first control on the display screen.
- control of the processing modes of the left earphone and the right earphone is implemented through an earphone application (appliaction, APP).
- the terminal device 100 includes a headset control application for controlling the headset, which is referred to as a headset application.
- a headset application For example, refer to the main interface of the terminal device shown in FIG. 12A .
- the terminal device can respond to the user's operation of clicking the icon 001, start the headset application, and display the control of the headset application on the display screen interface, or it is interpreted that when the headset application is started, the control interface of the headset application pops up.
- the name of the application can be called Audio Assistant, and this control function can also be integrated in the settings options in the terminal system.
- the first control in the annular selection control in FIG. 12B includes three circular arc segments, corresponding to the ANC mode, the HT mode, and the AH mode, respectively. If the second control is located in the arc segment of the ANC mode, it is determined to be the ANC mode, and different positions of the second control located in the arc segment of the ANC mode correspond to different processing intensities in the ANC mode.
- the HT mode is determined, and different positions of the second control located in the arc segment of the HT mode correspond to different processing intensities in the HT mode.
- the second control is located in the arc segment of the AH mode, then it is determined to be the AH mode, and the different positions of the second control located in the arc segment of the AH mode correspond to different processing intensities in the AH mode.
- the strongest noise reduction position of ANC is connected to the position of the weakest HT transparent transmission strength, and the auditory effect can be smoothly transitioned; can also transition smoothly.
- the highlighted black dots on the ring represent the second control for the user to select the processing intensity.
- the user can select different processing modes and control the processing intensity by moving the position of the black dots on the circumference.
- the terminal device 100 (such as a processor) responds to operation 1 performed by the user on the control interface, for example, operation 1 is generated when the user moves the position of the second control on the display screen representing the second control selected by the user.
- the terminal device 100 sends a control instruction 1 to the left earphone and the right earphone respectively, and the control instruction 1 is used to indicate the target mode and the target processing intensity.
- the target mode is the ANC mode.
- control instruction 1 may include an ANC identifier and a parameter value indicating a target processing intensity when performing ANC processing.
- the FB filter coefficients and/or the FF filter coefficients corresponding to different processing intensities are different.
- the control command 1 includes radians.
- the corresponding processing mode can be determined according to the range in which the radian is located. Different radian values correspond to the processing strength in this processing mode. Referring to Fig. 12B, the processing mode corresponding to (0, 180] is the ANC mode, the processing mode corresponding to (180, 270] is the HT mode, and the processing mode corresponding to (270, 360] is the AH mode.
- the left earphone and the right earphone can be Including the mapping relationship between the radian range and the processing mode, and the mapping relationship between the radian value and the filter coefficient. Taking the ANC mode as an example, the FB filter coefficient and the FF filter coefficient corresponding to different radian values are different.
- the user can touch the black dot in the disc to rotate clockwise from 0 degrees to 360 degrees.
- the corresponding FF filter coefficient and FB filter coefficient at 0 degrees make the ANC noise reduction effect the strongest, that is, the user
- the perceived sound of the user's current environment and the ambient sound inside the user's ear canal are weaker.
- the FF filter coefficient and the FB filter coefficient change, so that the active noise reduction effect gradually weakens, and the active noise reduction effect is the weakest when it reaches 180 degrees, similar to the headphones without any noise reduction processing.
- From 180 degrees to 270 degrees is to control the transparent transmission of the ambient sound.
- the corresponding HT filter coefficients and FB filter coefficients at 180 degrees make The transparent transmission effect of ambient sound is the weakest, that is, the lower the sound intensity of the user's current environment perceived by the user, is similar to the use of empty mode after wearing headphones.
- the HT filter coefficient and the FB filter coefficient change, which makes the transparent transmission effect of ambient sound stronger and stronger.
- 270 degrees to 360 degrees is used to control the hearing enhancement, that is, the user touches the black dot in the disc
- the corresponding FF filter coefficient, HT filter coefficient and FB filter coefficient at 180 degrees make the hearing enhancement effect the weakest, and the user perceives The sound of the event included in the sound of the environment where the user is currently located is weaker.
- the FF filter coefficient, HT filter coefficient and FB filter coefficient change, so that the auditory enhancement effect becomes stronger and stronger, that is, the event signal that the user wants to hear becomes stronger and stronger to assist hearing.
- the terminal device 100 sends a control command 1 to the left earphone and the right earphone respectively through Bluetooth in response to the user's operation 1 .
- the control command 1 may include the ANC identification and the parameter value of the target processing strength.
- the operations performed after the left earphone and the right earphone receive the control command 1 are similar, and the processing of the left earphone is taken as an example in the subsequent description.
- the main control unit of the left earphone acquires the FF filter coefficient and the FB filter coefficient processed by the ANC from the coefficient library according to the ANC identifier and the target processing strength.
- the coefficient library includes the mapping relationship shown in Table 1.
- Table 1 is only an example, and constitutes a specific definition of the mapping relationship.
- the parameter value of the target processing intensity is intensity 1
- the main control unit of the left earphone obtains the FF filter coefficient corresponding to intensity 1 according to Table 1 as the coefficient FF1
- the FB filter coefficient is the coefficient FB1
- the main control unit controls the FF filter to use the coefficient FF1
- a signal A1 is obtained by performing FF filtering processing on the first signal collected by the reference microphone.
- the main control unit controls the FB filter to use the coefficient FB1 to perform FB filtering processing on the signal A3 to obtain the second audio signal.
- the main control unit writes the coefficient FF1 and the coefficient FB1 into the AHA core, so that the AHA core executes the steps of S501-S504 to obtain the second audio signal.
- the terminal device 100 sends a control instruction 1 to the left earphone and the right earphone respectively through Bluetooth in response to the user's operation 1 .
- the control instruction 1 may include an HT identifier and a target processing intensity, where the target processing intensity is used to indicate the processing intensity when the HT processing is performed.
- the operations performed after the left earphone and the right earphone receive the control command 1 are similar, and the processing of the left earphone is taken as an example in the subsequent description.
- the main control unit of the left earphone acquires the HT filter coefficient and/or the FB filter coefficient of the HT processing from the coefficient library according to the HT identifier and the target processing strength.
- the value of the target processing intensity is intensity 5
- the main control unit of the left earphone obtains the HT filter coefficient corresponding to intensity 5 according to Table 1 as the coefficient HT1
- the FB filter coefficient is the coefficient FB5
- the main control unit controls the HT filter.
- the controller uses the coefficient HT1 to perform HT filtering processing on the first signal collected by the reference microphone.
- the main control unit controls the FB filter to use the coefficient FB5 to perform FB filtering processing on the signal B3.
- the main control unit writes the coefficient HT1 and the coefficient FB5 into the AHA core, so that the AHA core executes the steps of S601-S605 to obtain the second audio signal.
- the terminal device 100 sends a control instruction 1 to the left earphone and the right earphone respectively through Bluetooth in response to the user's operation 1 .
- the control instruction 1 may include the AH identifier and the parameter value of the target processing strength.
- the operations performed after the left earphone and the right earphone receive the control command 1 are similar, and the processing of the left earphone is taken as an example in the subsequent description.
- the main control unit of the left earphone obtains the HT filter coefficient, FF filter coefficient and FB filter coefficient of the AH process from the coefficient library according to the HT identifier and the target processing strength.
- the value of the target processing strength is indication 7
- the main control unit of the left earphone obtains the HT filter coefficient corresponding to indication 7 according to Table 1 as the coefficient HT3
- the FB filter coefficient is the coefficient FB7
- the FF filter coefficient is the coefficient FF5.
- the main control unit controls the HT filter to perform HT filter processing on the first signal collected by the reference microphone using the coefficient HT3.
- the main control unit controls the FB filter to use the coefficient FB7 to perform FB filtering processing on the signal C5.
- the main control unit controls the FF filter to use the coefficient FF5 to perform FF filtering processing on the first signal.
- the main control unit writes the coefficient HT3, the coefficient FB7 and the coefficient FF5 into the AHA core, so that the AHA core executes the steps of S801-S807 to obtain the second audio signal.
- the selection control may take the form of a bar.
- the selection controls include a first control and a second control.
- the bar of the first control may be divided into a plurality of bar segments, which are divided according to the number of processing modes supported by the headset.
- the second control indicates different processing modes in different bar segments of the first control.
- the second control indicates different processing intensities of the same processing mode at different positions of the same bar segment of the first control.
- the bar of the first control includes three bar segments.
- the user can touch the black bar to slide left and right.
- the black bar is at the K1 position
- the corresponding FF filter coefficient and FB filter coefficient make the ANC noise reduction effect the strongest.
- the FF filter coefficient and FB filter coefficient are the strongest.
- the filter coefficient changes, so that the active noise reduction effect gradually weakens.
- the K2 position is reached, the active noise reduction effect is the weakest, similar to the headphones without any noise reduction processing.
- the area between K2 and K3 is the part that controls the transparent transmission of ambient sound.
- Position K3 to position K4 is used to control auditory enhancement, that is, the user moves from position K3 to position K4 by touching the black bar, and when the black bar moves to position K3, the corresponding FF filter coefficient, HT filter coefficient and FB filter The coefficient makes the hearing enhancement effect the weakest. Moving from position K3 to position K4, the FF filter coefficient, HT filter coefficient and FB filter coefficient change, making the hearing enhancement effect stronger and stronger, that is, the voice signal that the user wants to hear is getting stronger and stronger. , to assist hearing.
- the selection controls in (a) include buttons corresponding to different processing modes, including an ANC button, an HT button, and an AH button.
- the terminal device 100 displays the display interface shown in FIG. 12G(b) in response to the user's operation of clicking the ANC button.
- the display interface of (b) includes controls 002 for selecting treatment intensity. Users can touch the black bar and slide up and down to determine the processing intensity of ANC, that is, to select the corresponding FF filter coefficient and FB filter coefficient. Black bars slide in the area of L1-L2. When the black bar is located at the L1 position, the corresponding FF filter coefficient and FB filter coefficient make the ANC noise reduction effect the strongest. After sliding down, the FF filter coefficient and the FB filter coefficient change, so that the active noise reduction effect gradually weakens. When it reaches the L2 position Active noise reduction has the weakest effect, similar to wearing headphones without any noise reduction.
- the headset APP when the headset 200 establishes a connection with the terminal device, the headset APP can be triggered to start, and a control interface including selection controls is displayed, such as the control interface illustrated in FIG. 12A , FIG. 12B , FIG. 12F or FIG. 12G .
- the interface displayed by the terminal device is interface 1
- the terminal device recognizes that the headset 200 is connected to the terminal device, and can jump from interface 1 to the control interface.
- the headset APP can be triggered to start, that is, a control interface including selection controls is displayed, such as FIG. 12A , FIG. 12B , FIG. 12C or Figure 12D illustrates the display interface.
- a control interface including selection controls is displayed, such as FIG. 12A , FIG. 12B , FIG. 12C or Figure 12D illustrates the display interface.
- the terminal device triggers the headset to play audio it may be that after the terminal device establishes a connection with the headset, a song is played, and a control interface including selection controls may be displayed.
- a control interface including selection controls can be displayed.
- the identified scene type of the current external environment is the target scene, and the target scene adaptation needs to adjust all settings.
- prompt information can be displayed, and the prompt information is used to prompt the user whether to adjust the processing mode of the headset.
- a control interface including selection controls may be displayed, such as the control interface illustrated in Figures 12A, 12B, 12C, or 12D.
- FIG. 12E takes the example of displaying the control interface of FIG. 12A as an example.
- the terminal device recognizes that the current scene the user is in is a noisy scene.
- the user may need to start the processing mode, so as to display selection prompt information (such as in the form of a prompt box) to prompt the user whether to adjust the processing mode of the headset .
- selection prompt information such as in the form of a prompt box
- the terminal device recognizes that the scene type of the external environment is a noisy scene.
- the user may need to start the processing mode, so as to display a prompt box to prompt the user whether to adjust the processing mode of the headset.
- the scene types that trigger the display of the prompt box may include: noisy scene, terminal building scene, train station scene, bus station scene, road scene, and the like.
- the strength of the identified signal reaches a set threshold, it is considered a noisy scene.
- the specific sound of an aircraft announcing the station is recognized, and the terminal building scene is determined to be recognized.
- the sound of the train time notification is recognized, and the train station scene is determined to be recognized; for another example, the bus ticket broadcast is recognized, and the bus station scene is determined to be recognized.
- the ticking of a signal light or the whistle of a car is recognized, and the road scene is determined to be recognized.
- a control interface including selection controls is displayed according to the identified scene where the user is currently located.
- Example 2 dual controls to control left and right earphones
- the terminal device 100 provides a control interface for the user to select the processing modes of the left earphone and the right earphone respectively according to requirements, and the processing modes of the left earphone and the right earphone may be different.
- the left earphone selects ANC mode
- the right earphone selects HT mode.
- the control interface includes selection controls for the left earphone and selection controls for the right earphone.
- the selection control of the left earphone is called the first selection control
- the selection control of the right earphone is called the second selection control.
- the first selection control is used for the user to select the processing mode of the left earphone
- the second selection control is used for the user to select the processing mode of the right earphone.
- the shape of the first selection control and the second selection control may be a circular ring, a bar or other shapes.
- the shapes of the first selection control and the second selection control may be the same or different.
- the user realizes the selection of different processing modes and the control of processing intensity by moving the position of the control representing the user's selection on the display screen.
- the shapes of the controls adopted by the left earphone and the right earphone may refer to the description in Example 1, and details are not repeated here.
- both the first selection control and the second selection control include a first control and a second control.
- Two different positions of the second control on the first control correspond to different processing modes, or, two different positions of the second control on the first control correspond to different processing intensities in the same processing mode. See the control interface shown in Figure 13 as an example.
- the user can select the different processing modes and control the processing intensity implemented by the left earphone through the position of the second control (black dot) in the first selection control of the left earphone moving on the circumference of the first control.
- the user can select the different processing modes implemented by the right earphone and control the processing intensity by moving the position of the second control on the first control in the second selection control of the right earphone.
- the user can select different processing modes for the left and right earphones, or the same processing intensity in the same processing mode, or different processing intensities in the same processing mode, so as to match the ear differences or match the needs of different applications. need.
- Example 2 the manner of triggering the display of the control interface including the first selection control and the second selection control refers to the description of Example 1, and details are not repeated here.
- Example 3 terminal device smart scene detection.
- the terminal device identifies the scene the user is currently in, and in different scenes, the headset adopts different processing modes.
- the terminal device recognizes that the scene type indication of the current external environment is the first scene, it determines the target mode corresponding to the first scene in the processing mode of the earphone, and sends control signaling 2 to the left earphone and the right earphone respectively.
- Control signaling 2 is used to indicate the target mode. Different target modes correspond to different scene types.
- the terminal device determines which function to instruct the earphone to perform according to the identified scene. Correspond the functions of AHA to the scene type, and select the most suitable function for the scene type, so that users can automatically feel the most needed effect.
- the scene types may include walking scene, running scene, quiet scene, multi-person talking scene, cafe scene, subway scene, train scene, car scene, waiting room scene, conversation scene, office scene, outdoor scene, driving Scene, windy scene, airplane scene, siren scene, whistle scene, cry scene, etc.
- AI models can be built offline and stored on terminal devices. For example, a microphone on a terminal device records a large amount of noise and sensor data and/or video processing unit (vpu) data in different scenarios, and manually labels the data for its corresponding scenarios.
- vpu noise and sensor data and/or video processing unit
- initialize and build an AI model which can be one of convolutional neural networks (CNN)/deep neural networks (DNN)/long short-term memory (LSTM) , can also be a mixture of different models.
- CNN convolutional neural networks
- DNN deep neural networks
- LSTM long short-term memory
- use the labeled data for model training to obtain the corresponding AI model.
- the processing modes suitable for different scene types are exemplified.
- the processing mode corresponding to the scene type is indicated in brackets.
- the noise of the plane is relatively loud when flying, so it is suitable to use the ANC mode.
- the HT mode For another example, in walking scenes, running scenes and quiet scenes, it is suitable to use the HT mode, and you can hear the sound of sudden events. For another example, in the cafe scene, the user needs to be quiet, and the ANC mode can be used. For another example, in a light music scene, the HT mode can be used. For another example, the alarm sound scene (AH), the whistle sound scene (AH), and the crying sound scene (AH) are used to hear the preset sound, and the AH mode is suitable.
- the AH mode is suitable.
- the terminal device 100 when the terminal device 100 recognizes that the scene type of the current external environment is an airplane scene, it can send control signaling 2 to the headset, and the control signaling 2 is used to indicate that the headset needs to implement the ANC function, that is, to indicate Headphones are in ANC mode. After the left earphone and the right earphone receive the control signaling 2, the processes of S501-S504 are performed respectively.
- the terminal device 100 when the terminal device 100 recognizes that the scene type of the current external environment is a walking scene, it can send control signaling 2 to the headset, and the control signaling 2 is used to indicate that the headset needs to implement the HT function, that is, the headset Use HT mode. After the left earphone and the right earphone receive the control signaling 2, the processes of S601-S605 are performed respectively.
- the terminal device 100 when the terminal device 100 recognizes that the scene type of the current external environment is a train station scene, it can send control signaling 2 to the headset, and the control signaling 2 is used to indicate that the headset needs to implement the AH function , that is, the headset adopts AH mode. After the left earphone and the right earphone receive the control signaling 2, the processes of S801-S807 are performed respectively.
- the terminal device After the headset is connected to the terminal device, the terminal device starts scene detection. After the detection of the terminal device is completed, the detection result can also be displayed to the user, so that the user can know the processing mode adopted by the headset. For example, the detection result is displayed to the user in the form of a prompt box.
- the detection result may include the detected scene, and may also include a processing mode corresponding to the detected scene. For example, when the terminal device recognizes that the scene is the first scene, it determines the target mode corresponding to the first scene in the processing mode of the headset, and can display the detection result, that is, the first scene and the target mode to the user. Then send control signaling 2 to the left earphone and the right earphone respectively. Control signaling 2 is used to indicate the target mode.
- a function switch for enabling smart scene detection is configured on the terminal device, and the terminal device triggers scene detection when the terminal device responds to the user initiating the function of smart scene detection.
- the scene in which it is identified is the first scene
- the target mode corresponding to the first scene in the processing mode of the earphone is determined, and then control signaling 2 is sent to the left earphone and the right earphone respectively. Control signaling 2 is used to indicate the target mode.
- the detection result can also be displayed to the user, so that the user can know the processing mode adopted by the headset.
- the detection result may include the detected scene, and may also include a processing mode corresponding to the detected scene.
- the terminal device determines the target mode corresponding to the first scene in the processing mode of the headset, and can display the detection result, that is, the first scene and the target mode to the user. Then send control signaling 2 to the left earphone and the right earphone respectively. Control signaling 2 is used to indicate the target mode.
- the left earphone and the right earphone send control signaling 2 again.
- the function switch for enabling smart scene detection configured by the terminal device may be configured on the control interface of the headset application, or may also be configured in the system setting menu bar of the terminal device.
- the terminal device can control the processing mode adopted by the headset by identifying the scene, and the terminal device can also identify the user operation on the selection control on the control interface to control the processing mode adopted by the headset. processing mode.
- the terminal device can determine whether the smart scene detection function is enabled according to requirements. When the smart scene detection function is not enabled, the processing mode adopted by the headset can be manually selected by using Example 1.
- the terminal device 100 identifies the current user scene at.
- the user enables the smart scene detection function, he can update to another interface based on the interface for manually selecting the processing mode, or display the detection results based on the interface for manually selecting the processing function.
- the processing function selected by the user on the terminal device is the HT function; after the smart scene detection function is enabled, the terminal device recognizes that the current scene the user is in is an airplane scene, and the ANC function is suitable.
- the user starts the headset application, and the control interface of the headset application is displayed on the display screen, taking a circular ring as an example.
- the processing function selected by the user is the HT function, as shown in (a) of FIG. 14A .
- the control interface includes an option control for whether to enable the smart scene detection function.
- the user triggers the smart scene detection function, executes the scene detection to obtain the detection result, and changes the position of the control representing the processing function selected by the user to the area belonging to the ANC function.
- the position of the black dot on the disk may be the default value under the ANC function or the position where the processing intensity selected by the last user selection of the ANC function is located. For example, as shown in (b) of FIG. 14A . In (b) of FIG. 14A , the scene where an airplane is detected is taken as an example.
- the terminal device 100 sends control signaling 2 to the left earphone and the right earphone respectively, and the control signaling 2 is used to indicate the ANC function.
- the user starts the headset application, and the control interface of the headset application is displayed on the display screen, taking a circular ring as an example.
- the processing function selected by the user is the HT function, as shown in (a) of FIG. 14B .
- the control interface includes an option control for whether to enable the smart scene detection function. After the user triggers the option control for enabling the smart scene detection function, the user triggers the smart scene detection function, executes the scene detection to obtain the detection result, and displays the detection result on the detection result interface.
- the detection interface interface may further include a scene that the terminal device can recognize, and a processing function corresponding to the scene. For example, as shown in (b) of FIG. 14B , the detection result is an airplane scene, and the corresponding processing function is the ANC function.
- the terminal device 100 sends control signaling 2 to the left earphone and the right earphone respectively, and the control signaling 2 is used to indicate the ANC function.
- the target processing intensity in the target mode can be determined by any one of the following methods.
- Method 1 The headset adopts the default target processing intensity in target mode.
- the terminal device After the terminal device sends the control signaling 2 to the left earphone and the right earphone respectively, taking the left earphone as an example, after the left earphone receives the control signaling 2, it determines that the processing mode used is the target mode, and the control signaling 2 does not indicate the target processing intensity. , the headset determines the default target processing strength. Taking the target mode as the ANC mode as an example, after receiving the control signaling 2, the left earphone determines to use the ANC mode, and obtains the default FF filter coefficient and FB filter coefficient in the ANC mode from the left earphone.
- Method 2 Use the processing intensity used when the target mode was used last time as the target processing intensity.
- the target processing strength is determined by the terminal device, and is indicated to the left earphone and the right earphone through control signaling. After the terminal device performs scene detection and determines the target mode according to the detected scene, obtains the processing intensity used when the target mode was used last time as the target processing intensity, and sends control signaling 2 to the left earphone and the right earphone respectively. Let 2 indicate the target mode and target processing strength.
- the headset itself determines the processing intensity in the target mode.
- the terminal device After the terminal device performs scene detection, after determining the target mode according to the detected scene, it sends control signaling 2 to the left earphone and the right earphone respectively, and the control signaling 2 indicates the target mode.
- the left earphone and the right earphone receive the control signaling 2 , determine the adopted processing mode as the target mode, and obtain the saved processing intensity of the last time the target mode was adopted as the target processing intensity.
- the target mode as ANC as an example, obtain the saved FF filter coefficients and FB filter coefficients when the ANC mode was last adopted, and perform ANC processing.
- Manner 3 The terminal device determines the target processing intensity according to the recognized scene.
- the function switch for enabling smart scene detection is not configured on the terminal device, after the terminal device recognizes the scene, it can determine the target processing intensity according to the recognized scene.
- the processing modes determined in different scenarios are the same, and the processing intensities corresponding to different scenarios are different.
- the walking scene, the running scene, and the quiet scene are all suitable for using the HT mode.
- the walking scene, the running scene, and the quiet scene have different processing intensities.
- ANC mode is suitable for multi-person talking scenes, cafe scenes, train scenes, airplane scenes, windy scenes, and office scenes.
- the processing intensity corresponding to the scene, the windy scene, and the office scene is different.
- the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene are all suitable for using the AH mode. When the AH mode is used, the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene correspond respectively different processing intensity.
- the terminal device sends control signaling 2 to the left earphone and right earphone according to the stored correspondence between the scene type, the target mode and the processing intensity.
- the control signaling 2 indicates the target mode and the target processing intensity in the target mode. Therefore, after receiving the control signaling 2, the earphone determines to adopt the target mode according to the control signaling 2, and determines the filter coefficient corresponding to the target processing intensity.
- the target mode as AH as an example
- the FF filter coefficient, the FB filter coefficient and the HT filter coefficient are determined according to the target processing intensity, and S801-S807 are performed according to the FF filter coefficient, the FB filter coefficient and the HT filter coefficient.
- the user indicates to the headset the processing intensity used in the target mode through a UI control provided by the terminal device.
- the detection result is displayed on the display interface of the terminal device, and the detection result includes the detected scene and the target mode corresponding to the detected scene.
- the display interface may include a control for selecting the treatment intensity.
- the control for selecting the treatment intensity is called an intensity control.
- Intensity controls can include Control 1 and Control 2. Different positions of control 1 are used to indicate different processing intensity in target mode. Intensity controls can be circles, bars, or other shapes.
- the detected scene is a terminal building scene. Taking the control 1 in the intensity control as a circle and the control 2 as a circular black dot as an example, in response to the user touching the control 2 and moving to the position of the control 1 1.
- Position 1 represents the target processing intensity in the user-selected target mode.
- Send control instruction 2 to the left earphone and right earphone, and control instruction 2 indicates the target mode and the target processing intensity corresponding to position 1.
- the target mode and target processing intensity can be sent to the left and right earphones through different control commands.
- the terminal device After determining the target mode according to the detected scene, the terminal device sends control signaling indicating the target mode to the left earphone and the right earphone.
- the left earphone and the right earphone adopt the default processing intensity in the target mode, that is, adopt the default filter coefficient in the target mode to realize the target processing corresponding to the target mode.
- the control signaling indicating the target processing intensity is sent to the left earphone and the right earphone, and further, the left earphone and the right earphone use the filter coefficient corresponding to the target processing intensity to achieve the target.
- the target processing corresponding to the mode.
- the user triggers the smart scene detection function, executes the scene detection to obtain the detection result, and sets the position of the control representing the processing function selected by the user by the position of the control. Changed to the area belonging to the ANC function.
- the position of the black dot on the disk may be the default value under the ANC function or the position where the processing intensity selected by the last user selection of the ANC function is located.
- the user selects the processing intensity in ANC mode by moving the position of the black dot. And send control signaling 2 to the left earphone and right earphone, and the control signaling 2 indicates the ANC mode and the target processing strength.
- Example 4 scene detection of headphones, different scenes correspond to different processing functions.
- the headset has a scene detection function.
- the headset recognizes the scene the user is currently in.
- the processing functions implemented by the headset are different.
- the left earphone may have the scene detection function, or the right earphone may have the scene detection function, or both the left earphone and the right earphone may have the scene detection function.
- one of the left earphone and the right earphone is used to perform scene detection.
- the left earphone performs scene detection, and sends the detection result to the right earphone, so that both the left earphone and the right earphone are based on the detection result of the left earphone. Executes processing for realizing the processing function corresponding to the detection result.
- both the left earphone and the right earphone execute processing for realizing the processing function corresponding to the detection result according to the detection result of the right earphone.
- both the left earphone and the right earphone perform scene detection, the left earphone performs the processing for realizing the processing function corresponding to the detection result according to the detection result of the left earphone, and the right earphone performs the processing for realizing the detection according to the detection result of the right earphone The result corresponds to the processing of the processing function.
- the enabling of the scene detection function of the headset may be controlled by the user through the headset or through the terminal device.
- the headset is provided with a button for activating the scene detection function.
- the user can turn on or turn off the scene detection function of the headset by touching the button.
- the headset identifies the scene where the current user is located (or the scene where the headset is currently located), the corresponding relationship between the scene and the processing mode determines the processing mode corresponding to the identified scene, and realizes the processing corresponding to the processing mode.
- the user can turn on or turn off the scene detection function of the earphone by tapping on the earphone, such as three consecutive taps.
- the headset responds to the operation of the user tapping the headset three times to enable the scene detection function of the headset.
- the scene detection function of the earphone is enabled, the earphone turns off the scene detection function of the earphone in response to the operation of the user tapping the earphone three times.
- the headset identifies the scene where the current user is located (or the scene where the headset is currently located), the corresponding relationship between the scene and the processing mode determines the processing mode corresponding to the identified scene, and realizes the processing corresponding to the processing mode.
- the activation of the scene detection function of the left earphone or the right earphone is controlled by the terminal device 100 .
- the headphone control interface includes a switch button for the headphone scene detection function.
- the terminal device can determine whether to enable the scene detection function of the headset according to user requirements. When the scene detection function of the headset is not enabled, Example 1 can be used to manually select the processing function to be implemented by the headset.
- the headset identifies the scene type of the current external environment, determines the processing mode corresponding to the identified scene type according to the corresponding relationship between the scene type and the processing mode, and implements the processing function corresponding to the processing mode.
- the terminal device 100 sends control signaling 3 to the headset 200, where the control signaling 3 is used to instruct the headset to start the scene detection function.
- the headset 200 starts to perform scene detection according to the control signaling 3 .
- the headset 200 determines the processing function to be executed according to the detected scene type of the current external environment, such as an ANC function, and the headset 200 executes the ANC processing and executes S501-S504.
- the headset after the headset is connected to the terminal device, the headset starts scene detection; or when the headset receives the downlink audio signal sent by the terminal device, the scene detection starts.
- the detection result may also be sent to the terminal device.
- the detection result may be included in the indication information and sent to the terminal device.
- the detection result may include the detected scene and the processing mode corresponding to the scene.
- the terminal device receives the detection result, it displays the detection result to the user, so that the user knows the processing mode adopted by the headset.
- the detection result is displayed to the user in the form of a prompt box.
- only the detected scene may be included in the detection result.
- the terminal device determines the processing mode corresponding to the scene detected by the headset, and displays the scene detected by the headset and the processing mode corresponding to the scene to the user. For example, when the earphone recognizes that the scene is the first scene, it determines the target mode corresponding to the first scene in the processing mode of the earphone, and can display the detection result, that is, the first scene and the target mode to the user.
- the processing function of the processing mode corresponding to the scene is not executed immediately, the detection result is sent to the terminal device, and the terminal device displays the detection result to the user.
- the terminal device responds to the user's operation of determining the processing mode, it sends a confirmation instruction to the earphone, and when the earphone receives the confirmation instruction, it adopts the processing mode corresponding to the scene detected by the earphone to execute the processing function.
- the types of scenes that can be recognized by the headset may include walking scenes, running scenes, quiet scenes, multi-person talking scenes, cafe scenes, subway scenes, train scenes, car scenes, waiting room scenes, conversation scenes, office scenes, outdoor scenes.
- the processing modes suitable for different scene types are exemplified.
- the processing mode corresponding to the scene type is indicated in brackets.
- the noise of the plane is relatively loud when flying, so it is suitable to use the ANC mode.
- the HT mode For another example, in walking scenes, running scenes and quiet scenes, it is suitable to use the HT mode, and you can hear the sound of sudden events. For another example, in the cafe scene, the user needs to be quiet, and the ANC mode can be used. For another example, in a light music scene, the HT mode can be used. For another example, the alarm sound scene (AH), the whistle sound scene (AH), and the crying sound scene (AH) are used to hear the preset sound, and the AH mode is suitable.
- the AH mode is suitable.
- the processing of S601-S605 is performed respectively for the left earphone and the right earphone.
- the scene type is the train station scene
- the left earphone and the right earphone perform the processing of S801-S807, respectively.
- the target processing intensity in the target mode may be determined in any of the following manners.
- Method 1 The headset adopts the default target processing intensity in target mode.
- the earphone determines the adopted processing mode as the target mode according to the detected scene, and the left earphone and the right earphone determine to adopt the default target processing intensity.
- the target mode as the ANC mode as an example
- the left earphone and the right earphone obtain the default FF filter coefficient and FB filter coefficient in the ANC mode.
- Method 2 Use the processing intensity used when the target mode was used last time as the target processing intensity.
- the earphone determines the processing intensity in the target mode by itself. After the headset performs scene detection, after determining the target mode according to the detected scene, the saved processing intensity of the last time the target mode was used is obtained as the target processing intensity. Taking the target mode as ANC as an example, obtain the saved FF filter coefficients and FB filter coefficients when the ANC mode was last adopted, and perform ANC processing.
- the target processing strength is determined by the terminal device, and is indicated to the left earphone and the right earphone through control signaling.
- the headset After the headset performs scene detection, it sends the detection result to the terminal device, so that the terminal device obtains the processing intensity used when the target mode was used last time as the target processing intensity, and sends control signaling 4 to the left headset and the right headset respectively. Let 4 indicate the target processing strength.
- the headset determines the target processing intensity according to the recognized scene.
- the target processing intensity can be determined according to the recognized scene.
- the processing modes determined in different scenarios are the same, and the processing intensities corresponding to different scenarios are different.
- the walking scene, the running scene, and the quiet scene are all suitable for using the HT mode.
- the walking scene, the running scene, and the quiet scene have different processing intensities.
- ANC mode is suitable for multi-person talking scenes, cafe scenes, train scenes, airplane scenes, windy scenes, and office scenes.
- the processing intensity corresponding to the scene, the windy scene, and the office scene is different.
- the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene are all suitable for using the AH mode. When the AH mode is used, the dialogue scene, the alarm sound scene, the whistle sound scene, and the crying sound scene correspond respectively different processing intensity.
- the left earphone and the right earphone determine the target mode corresponding to the detected scene and the target processing intensity in the target mode according to the stored correspondence between the scene type, the target mode, and the processing intensity. Thereby, the left earphone and the right earphone obtain the filter coefficient corresponding to the target processing intensity.
- the target mode as AH as an example
- the FF filter coefficient, the FB filter coefficient and the HT filter coefficient are determined according to the target processing intensity
- S801-S807 are performed according to the FF filter coefficient, the FB filter coefficient and the HT filter coefficient.
- the headset may also perform emergency event detection to determine a target event (or referred to as a target event scenario).
- a target event scenario For example, emergent events include one or more of wind noise events, whistling events, emergencies, human voice events, or no emergencies, and different events have different processing intensities.
- the headset performs scene detection, as well as event detection.
- target mode different events correspond to different filter coefficients. Taking ANC as an example, different events correspond to different FF filter coefficients and/or different FB filter coefficients.
- the left earphone can obtain the FF filter coefficient or FB corresponding to the detected event from the coefficient library according to the detection result.
- the coefficient library stores processing modes, events, and the mapping relationship between FF filter coefficients and FB filter coefficients.
- the processing effect of ANC mainly depends on FB filtering and/or FF filtering.
- the filter coefficient of the FF filter is controlled according to the detected scene, and the FB filter coefficient adopts a fixed value.
- the filter coefficient of the FB filter is controlled according to the detected scene, and the FF filter coefficient adopts a fixed value.
- the FF filter coefficient and the FB filter coefficient are controlled according to the detected scene.
- the event includes a howling event, a wind noise event, a sudden event, a human voice event, or no sudden event.
- the earphone 200 detects the event sound in the external environment, and can determine the target event corresponding to the event sound in the external environment according to the signal collected by the reference microphone.
- the signal collected by the reference microphone includes a signal of a preset frequency spectrum, and then an event corresponding to the signal of the preset frequency spectrum is determined.
- the signal collected by the reference microphone includes a wind sound signal, that is, the collected signal includes a signal matching the frequency spectrum of the wind sound, and the event corresponding to the detected event sound in the external environment is determined to be a wind noise event.
- a spectrum matching method or a deep neural network (deep neural networks, DNN) matching method may be used.
- the headset 200 can determine the event in the current environment of the user according to the signal collected by the reference microphone in the following manner, as shown in FIG. 15 .
- Bone conduction sensors are also included in the headset 200 .
- the bone conduction sensor is used to collect the bone conduction signal of the earphone user.
- the bone conduction sensor obtains the bone conduction signal by collecting the bone conduction signal, that is, collecting the periosteal vibration signal generated by the user's speaking.
- the activation of the scene detection function of the left earphone or the right earphone may be controlled by the terminal device 100, or may be controlled by a user's operation on the earphone, such as an operation of tapping the left earphone or the right earphone.
- a bone conduction sensor is included on the earphone, and the user's upper and lower teeth touch to generate a tooth touching sound, so that the bone conduction sensor activates the scene detection function by detecting the audio signal generated by the user's upper and lower teeth touching.
- the third signal collected by the reference microphone is a signal collected by the reference microphone after the headset starts the scene detection function.
- the energy of the bone conduction signal collected by the bone conduction sensor is small.
- the signal AA1 is the third signal.
- the earphone 200 may first determine the energy of the bone conduction signal, and if the energy of the bone conduction signal is less than the set threshold, the filtering operation is not performed, that is, S1501 is not performed, when it is determined that the energy of the bone conduction signal is greater than or equal to the set threshold When the threshold value is reached, execute S1501.
- the earphone 200 performs spectrum analysis on the signal AA1 to obtain the energy characteristic of the signal AA1. For example, the earphone 200 performs spectral analysis on the signal to obtain the energy of the entire frame of the signal AA1 and the energy of each bark subband of the signal AA1, thereby forming the energy characteristic of the signal AA1 represented by a vector.
- S1503 Determine a first energy feature of the energy features included in the energy feature set that matches the energy feature of the filtered signal, where different energy features included in the energy feature set correspond to different event identifiers.
- S1504 Determine that the event identified by the event identifier corresponding to the first energy feature is an event in the current environment of the user, that is, a detection result of event detection.
- the energy feature set may be generated by: performing wind noise detection, burst noise detection, howling detection, and human voice detection on the signals collected by the first microphone, the second microphone, and the third microphone to obtain wind noise signal, burst noise signal and howling signal. Then, spectrum analysis is performed on the wind noise signal, the burst noise signal, the howling signal and the human voice signal, respectively, to obtain the sub-band energy characteristics of the wind noise signal, the sub-band energy characteristics of the burst noise signal, and the sub-band of the howling signal. Energy signature and sub-band energy signature of the vocal signal.
- the sub-band energy features of the wind noise signal, the sub-band energy features of the burst noise signal, the sub-band energy features of the howling signal, and the sub-band energy features of the human voice signal constitute the energy feature set. It should be understood that in a quiet scene, the sub-band energy of the noise is relatively weak.
- a spectrum matching method can be used, and a deep neural network (DNN) can also be used.
- DNN deep neural network
- matching degree between the energy feature of the filtered signal and each energy feature included in the energy feature set can be determined through the DNN, and the event identified by the event identifier corresponding to the first energy feature with the highest matching degree is Test results.
- the main control unit in the headset 200 may implement the determination of an event in the current environment of the user according to the signal collected by the reference microphone.
- the main control unit includes a DSP for executing S1501-S1504.
- the user indicates to the headset the processing intensity used in the target mode through a UI control provided by the terminal device.
- the detection result is sent to the terminal device, and the terminal device displays the detection result to the user.
- the detection result is displayed on the display interface of the terminal device, and the detection result includes the scene detected by the headset and the target mode corresponding to the detected scene.
- the display interface further includes a control for selecting the treatment intensity.
- the control for selecting the treatment intensity is called an intensity control.
- Intensity controls can include Control 1 and Control 2. Different positions of control 1 are used to indicate different processing intensity in target mode. Intensity controls can be circles, bars, or other shapes. Referring to FIG.
- position 2 represents the target processing intensity in the user-selected target mode.
- the target mode is HT as an example.
- the terminal device 100 sends control signaling 3 to the headset 200 in response to the user initiating the function of scene detection in the headset, where the control signaling 3 is used to instruct the headset to activate the scene detection function.
- the headset 200 starts and executes scene detection according to the control signaling 3 to obtain the detection result.
- the headset 200 may send the detection result to the terminal device 100, so that the terminal device 100 displays it to the user, and displays the detected scene corresponding to the processing mode that the headset needs to adopt to the user.
- the user may update the interface to manually select the processing mode to another interface, or display the detection results based on the interface for manually selecting the processing function.
- the processing function selected by the user on the terminal device is the HT function; after the headset scene detection function is enabled, the headset 200 recognizes that the scene the user is currently in is an airplane scene, and the ANC function is suitable for detecting The result, the aircraft scene, the ANC function, is sent to the end device.
- the user starts the headset application, and the control interface of the headset application is displayed on the display screen, taking a circular ring as an example.
- the processing function selected by the user is the HT function, as shown in (a) of FIG. 17A .
- the control interface includes an option control for whether to enable the headset scene detection function.
- the terminal device triggers the headset scene detection function, and sends control signaling 3 to the headset 200, where the control signaling 3 is used to instruct the headset to activate the scene detection function.
- the headset 200 starts and executes scene detection according to the control signaling 3 to obtain the detection result.
- the headset 200 sends the detection result to the terminal device 100, and after receiving the detection result, the terminal device 100 changes the position of the control representing the processing function selected by the user to an area belonging to the ANC function.
- the user selects the processing intensity in ANC mode by moving the position of the black dot on the disc. For example, as shown in (b) of FIG. 17A . In (b) of FIG. 17A , a scene where an airplane is detected is taken as an example.
- the user starts the earphone application, and the control interface of the earphone application is displayed on the display screen, taking a circular ring as an example.
- the processing function selected by the user is the HT function, as shown in (a) of FIG. 17B .
- the control interface includes an option control for whether to enable the headset scene detection function.
- the terminal device triggers the headset scene detection function, and sends control signaling 3 to the headset 200, where the control signaling 3 is used to instruct the headset to activate the scene detection function.
- the headset 200 starts and executes scene detection according to the control signaling 3 to obtain the detection result.
- the headset 200 sends the detection result to the terminal device 100, and after receiving the detection result, the terminal device 100 displays the detection result on the detection result interface.
- the detection interface interface may also include a scene that the headset can recognize, and a processing mode corresponding to the scene.
- the user selects the processing intensity in ANC mode by moving the position of the black dot on the disc. For example, as shown in (b) of FIG. 17B , the detection result is an airplane scene, and the corresponding processing mode is the ANC mode.
- an artificial intelligence (artificial intelligence, AI) model can be used to perform detection and classification.
- the AI model can be configured in the headset. Another way can determine the scene type according to the signal collected by the reference microphone. For example, the headset 200 can determine the scene where the user is currently located according to the signal collected by the reference microphone in the following manner, as shown in FIG. 18 .
- S1801 Perform spectrum analysis on the first signal collected by the reference microphone, divide the first signal into multiple subbands, and calculate the energy of each subband. For example, the first information collected by the reference microphone is divided into subbands in the frequency domain according to the bark subband division method, and the energy of each subband is calculated.
- S1802 Determine the VAD to obtain the noise segment in the first signal, and obtain the smoothed energy of each subband in the noise segment.
- the VAD is determined as follows:
- S1803 Determine the scene type according to the smoothed energy of each subband in the noise segment.
- the processing is performed for the determined noise segment to determine the scene type:
- Example 5 After determining the processing mode, the headset performs event detection in the processing mode. In the processing mode, the filter coefficients corresponding to different events (ie, the processing intensity in the processing mode) are different.
- the earphone recognizes the user's operation, and determines that the earphone 200 selected by the user needs to implement ANC processing, HT processing or AH processing.
- the processing mode adopted by the earphone 200 is the ANC mode function.
- the user's operation may be an operation of the user tapping the earphone, and the processing mode is determined to be the ANC mode, the HT mode, or the AH mode through different operations.
- buttons are provided on the earphone, and different buttons indicate different processing modes. The user selects the processing mode employed by the headset by pressing a button.
- the left earphone and the right earphone perform ANC processing, and specifically perform S501-S504.
- the processing mode to be implemented by the earphone is selected and controlled by the terminal device 100 .
- Either the left earphone or the right earphone can have an event detection function.
- one of the left earphone and the right earphone is used to perform event detection.
- the left earphone performs event detection and sends the detection result to the right earphone, or the right earphone performs event detection and sends the detection result. to the left earphone.
- different events correspond to different FF filter coefficients and FB filter coefficients.
- the left earphone or the right earphone performs event detection, taking the left earphone as an example, the left earphone can obtain the FF filter coefficient or FB filter coefficient corresponding to the detected event in the ANC mode from the coefficient library according to the detection result.
- the coefficient library For example, refer to the content included in the coefficient library shown in Table 2, and take an event including a howling event, a wind noise event, a sudden event or a human voice event as an example.
- the headset includes hardware structures and/or software modules corresponding to executing each function.
- modules and method steps of each example described in conjunction with the embodiments disclosed in the present application can be implemented in the form of hardware or a combination of hardware and computer software. Whether a function is performed by hardware or computer software-driven hardware depends on the specific application scenarios and design constraints of the technical solution.
- an embodiment of the present application further provides a noise processing apparatus 1900 .
- the noise processing device 1900 is applied to an earphone.
- the headset has at least two functions in the ANC function, the HT function or the AH function; the headset includes a first microphone and a second microphone; the first microphone is used to collect a first signal, and the first signal is used to represent the current the sound of the external environment; the second microphone is used to collect a second signal, and the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the earphone.
- the noise processing apparatus 1900 can be used to implement the function of the earphone in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
- the apparatus may include a communication module 1901 , an acquisition module 1902 , and a first processing module 1903 .
- a communication module 1901 configured to receive a first audio signal from the terminal device
- the acquiring module 1902 is configured to acquire a target mode; wherein, the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the earphone to implement a target processing function, and the target processing function is an active reduction.
- the target mode is determined based on the scene type of the current external environment, and the target mode is used to instruct the earphone to implement a target processing function, and the target processing function is an active reduction.
- the first processing module 1903 is configured to obtain a second audio signal according to the target mode, the first audio signal, the first signal, and the second signal.
- the apparatus further includes: a playing module, configured to play the second audio signal.
- a playing module configured to play the second audio signal.
- the playback module it is not shown in FIG. 19 .
- the target processing function is an ANC function
- the second audio signal played by the playback module can attenuate the user's sound to the user's current environment and the user's ear canal Perception of internal ambient sounds; or,
- the target processing function is the HT function, and the second audio signal played by the playback module can enhance the user's perception of the sound of the environment where the user is currently located; or,
- the target processing function is the AH function, and the second audio signal played by the playing module can enhance the user's perception of the event sound; the event sound satisfies a preset frequency spectrum.
- the target processing function is an ANC function
- the second audio signal is obtained based on the first audio signal, the third signal and the fourth signal, and the third signal is the first signal
- the inverted signal of the fourth signal is the inverted signal of the second signal
- the target processing function is the HT function
- the second audio signal is obtained based on the first audio signal, the first signal and the second signal; or,
- the target processing function is the AH function
- the second audio signal is obtained based on the first audio signal, the fifth signal and the fourth signal
- the fifth signal is an event signal in the first signal
- the The event signal satisfies the preset spectrum.
- the communication module 1901 is further configured to receive a first control instruction from the terminal device, where the first control instruction carries the target mode, and the target mode is the terminal The device is determined according to the scene type of the current external environment;
- the target mode is sent to the obtaining module 1902 .
- the communication module 1901 is further configured to:
- the second control instruction carries a target processing strength, and the target processing strength is used to indicate the processing strength when the headset implements the target processing function;
- the first processing module 1903 is specifically used for:
- a second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
- it also includes:
- the second processing module 1904 is configured to determine the target event corresponding to the event sound in the current external environment according to the first signal, and determine the target processing intensity in the target mode according to the target event; wherein, the target processing intensity used to indicate the processing strength when the headset implements the target processing function;
- the first processing module 1903 is specifically used for:
- a second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
- the headset further includes a bone conduction sensor, and the bone conduction sensor is used to collect bone conduction signals generated by the vibration of the user's vocal cords;
- the first processing module 1901 is specifically configured to determine the target event corresponding to the event sound in the current external environment according to the first signal and the bone conduction signal.
- the target event includes one of a howling event, a wind noise event, a sudden event or a human voice event.
- it also includes:
- the third processing module 1905 is configured to identify the scene type of the current external environment as a target scene according to the first signal, and determine the target mode adopted by the headset according to the target scene, where the target mode corresponds to the target scene. The processing mode of the target scene.
- the target scene includes a walking scene, a running scene, a quiet scene, a multi-person talking scene, a cafe scene, a subway scene, a train scene, a waiting room scene, a dialogue scene, an office scene, and an outdoor scene , driving scene, windy scene, airplane scene, siren scene, whistle scene, cry scene.
- the communication module 1901 is further configured to send indication information to the terminal device, where the indication information carries the target mode;
- the first processing module 1903 is specifically used for:
- a second audio signal is derived from the target mode, the target processing strength, the first audio signal, the first signal, and the second signal.
- the target processing function is an ANC function
- the greater the target processing intensity is the more the user perceives the sound of the current environment of the user and the sound inside the user's ear canal.
- the target processing function is the HT function
- the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
- the target processing function is the AH function
- the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- the earphone is a left earphone, or the earphone is a right earphone.
- the target mode instructs the headset to implement the ANC function
- the first processing module 1903 is specifically configured to:
- the second audio signal is obtained by performing mixing processing on the fourth audio signal and the first audio signal.
- the filter coefficient used in the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the ANC function; or,
- the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the ANC function.
- the target mode indicates that the headset implements the HT function
- the first processing module 1903 is specifically configured to:
- the second audio signal is obtained by performing mixing processing on the third filtered signal and the fifth audio signal.
- the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the HT function; or,
- the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the HT function.
- the target mode indicates that the headset implements the AH function
- the first processing module 1903 is specifically configured to:
- the second audio signal is obtained by mixing the fifth filtered signal, the sixth audio signal and the first filtered signal.
- the filter coefficient used in the first filtering process is the filter coefficient associated with the target processing intensity for the first filtering process under the AH function; or,
- the filter coefficient used in the second filtering process is the filter coefficient associated with the target processing intensity for the second filtering process under the AH function; or,
- the filter coefficient used in the third filtering process is the filter coefficient associated with the target processing intensity for the third filtering process under the AH function.
- the terminal device includes corresponding hardware structures and/or software modules for executing each function.
- modules and method steps of each example described in conjunction with the embodiments disclosed in the present application can be implemented in the form of hardware or a combination of hardware and computer software. Whether a function is performed by hardware or computer software-driven hardware depends on the specific application scenarios and design constraints of the technical solution.
- an embodiment of the present application further provides a mode control apparatus 2000 .
- the mode control apparatus 2000 is applied to the terminal device 100, and can be used to implement the functions of the terminal device in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
- the mode control apparatus 2000 includes a first detection module 2001 and a transmission module 2002 , and may further include a display module 2003 and a second detection module 2004 .
- a first detection module 2001 configured to identify a target mode according to the target scene when the scene type of the current external environment is a target scene;
- the target mode is one of the processing modes supported by the headset, different processing modes correspond to different scene types, and the processing modes supported by the headset include active noise reduction ANC mode, ambient sound transparent transmission HT mode or hearing enhancement At least two of the AH modes.
- the sending module 2002 is configured to send the target mode to the earphone, where the target mode is used to instruct the earphone to implement the processing function corresponding to the target mode.
- it also includes:
- the display module 2003 is configured to display result prompt information when the target mode is determined according to the target scene, and the result prompt information is used to prompt the user that the headset implements the processing function corresponding to the target mode.
- the display module 2003 is configured to display selection prompt information before sending the first control signaling to the headset, where the selection prompt information is used to prompt the user whether to adjust the processing mode of the headset is the target mode;
- the second detection module 2004 is configured to detect an operation in which the user selects the processing mode of the earphone as the target mode.
- the display module 2003 is further configured to display a first control and a second control, wherein different positions of the second control on the first control are used to indicate the target Different processing intensity in mode;
- the second detection module 2004 is further configured to detect the operation of the user touching the second control to move to the first position on the first control before the sending module 2002 sends the first control signaling to the headset , the first position of the second control on the first control indicates the target processing intensity in the target mode;
- the sending module 2002 is further configured to send the target processing intensity to the earphone, where the target processing intensity is used to indicate the processing intensity when the earphone implements the processing function corresponding to the target mode.
- the shape of the first control is a ring
- the processing in the target mode The intensity varies from small to large; or,
- the user touches the second control to move on the first control in a counterclockwise direction, and the processing intensity in the target mode changes from small to large;
- the shape of the first control is a bar
- the processing in the target mode The intensity varies from small to large; or
- the target processing function is an ANC function
- the greater the target processing intensity is the more the user perceives the sound of the current environment where the user is located and the sound inside the user's ear canal.
- the target processing function is the HT function
- the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
- the target processing function is the AH function
- the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- an embodiment of the present application further provides a mode control apparatus 2100 .
- the mode control apparatus 2100 is applied to the terminal device 100, and can be used to implement the functions of the terminal device in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
- the mode control device 2100 includes a processing module 2101 , a sending module 2102 , a receiving module 2103 , a display module 2104 and a detection module 2105 .
- the processing module 2101 is used to obtain a target mode;
- the target mode is one of the processing modes supported by the headset, and the processing modes supported by the headset include active noise reduction ANC mode, ambient sound transparent transmission HT mode or auditory enhancement AH mode at least two of them;
- the processing module 2101 is further configured to determine the target processing intensity in the target mode according to the scene type of the current external environment; different scene types correspond to different processing intensities in the target mode;
- the sending module 2102 is configured to send the target processing strength to the earphone, where the target processing strength is used to indicate the processing strength when the earphone implements the processing function corresponding to the target mode.
- it also includes:
- a receiving module 2103 configured to receive the target mode sent by the headset.
- it also includes:
- the display module 2104 is configured to display a selection control, where the selection control includes a processing mode supported by the headset, and detects an operation of the user selecting the target mode in the processing mode of the headset through the selection control.
- the display module 2104 is further used for:
- the processing module 2101 determines the target processing intensity in the target mode according to the scene type of the current external environment
- the receiving module 2103 receives the target mode sent by the headset
- a selection prompt is displayed,
- the selection prompt information is used to indicate whether the user adjusts the processing mode of the headset to the target mode;
- the detection module 2105 is configured to detect the operation of the user selecting to adjust the processing mode of the earphone to the target mode.
- the target processing function is an ANC function
- the greater the target processing intensity is the more the user perceives the sound of the current environment of the user and the sound inside the user's ear canal.
- the target processing function is the HT function
- the greater the target processing intensity the greater the intensity of the sound of the environment where the user is currently located; or,
- the target processing function is the AH function
- the greater the target processing intensity the stronger the event sound included in the sound of the environment where the user is currently located as perceived by the user.
- an embodiment of the present application further provides a mode control apparatus 2200 .
- the mode control apparatus 2200 is applied to the terminal device 100, and can be used to implement the functions of the terminal device in the above method embodiments, so the beneficial effects of the above method embodiments can be achieved.
- the mode control device 2100 includes a display module 2201 , a detection module 2202 , a transmission module 2203 , a processing module 2204 , and an identification module 2205 .
- the display module 2201 is used to include a first selection control on the first interface, the first selection control includes the processing mode supported by the first target headset and the processing intensity corresponding to the processing mode supported by the first target headset; the first target The processing mode of the headset includes at least two of the active noise reduction ANC mode, the ambient sound transparent transmission HT mode, or the hearing enhancement AH mode;
- the detection module 2202 detects the first operation performed by the user on the first interface; the first operation is that the user selects the first target mode and selects the first target mode from the processing modes supported by the first target headset through the first selection control.
- the processing intensity in a target mode is generated by the first target processing intensity;
- a sending module 2203 configured to send the first target mode and the first target processing intensity to the first target earphone, where the first target mode is used to instruct the first target earphone to achieve the first target
- the processing function corresponding to the mode, and the first target processing intensity is used to indicate the processing intensity when the first target earphone implements the processing function corresponding to the first target mode.
- the display module 2201 is further configured to display selection prompt information before displaying the first interface, where the selection prompt information is used for the user to choose whether to adjust the first target headset processing mode;
- the detection module 2202 is further configured to detect the user's operation of selecting and adjusting the processing mode of the first target earphone.
- it also includes:
- the identification module 2205 is configured to identify the scene type of the current external environment as the target scene before the display module 2201 displays the first interface, and the target scene is adapted to the scene type that needs to adjust the processing mode of the first target headset.
- it also includes:
- the identification module 2205 is configured to identify that the terminal device triggers the first target headset to play audio before the display module 2201 displays the first interface.
- the detection module 2202 is further configured to detect that the terminal device establishes a connection with the first target headset before the display module displays the first interface.
- the detection module 2202 before the display module 2201 displays the first interface, detects that the terminal device establishes a connection with the first target headset, and detects the The second operation performed by the user on the main interface.
- the main interface includes an icon of a first application, the second operation is generated by the user touching the icon of the first application, and the first interface is a display interface of the first application.
- the first selection control includes a first control and a second control
- the second control indicates two different processing modes of the first target earphone at any two different positions of the first control , or, the second control indicates different processing intensities in the same processing mode of the first target earphone at any two different positions of the first control;
- the first operation is generated by the user moving a first position of the second control in an area corresponding to the first target mode on the first control, where the first position corresponds to the first The first target processing intensity in target mode.
- the shape of the first control is a circular ring, the circular ring includes at least two arc segments, and the second controls are located in different arc segments to indicate that the first target headset is different processing mode, the second controls are located at different positions of the same arc segment to indicate different processing intensities of the same processing mode of the first target earphone; or,
- the shape of the first control is a bar, the bar includes at least two bar segments, the second controls are located in different bar segments to indicate different processing modes of the first target earphone, and the second control is located in different bar segments. Different positions of the controls on the same bar segment indicate different processing intensities of the same processing mode of the first target earphone.
- the detection module 2202 is further configured to detect a third operation performed by the user on the first interface, the first interface further includes a second selection control, and the second selection control includes The processing mode supported by the second target earphone and the processing intensity corresponding to the processing mode supported by the second target earphone, the processing mode supported by the first target earphone includes active noise reduction ANC mode, ambient sound transparent transmission HT mode or auditory enhancement AH mode
- the third operation is that the user selects the second target mode in the processing mode of the second target headset through the second selection control and selects the processing intensity in the second target mode as the second target
- the processing intensity is generated; when the first target earphone is the left earphone, the second target earphone is the right earphone, or, the first target earphone is the right earphone, and the second target earphone is the left earphone;
- the sending module 2203 is further configured to send the second target mode and the second target processing intensity to the second target headset, where the second target mode is used to instruct the second target headset to implement the The processing function corresponding to the second target mode, and the second target processing intensity is used to indicate the processing intensity when the second target earphone implements the processing function corresponding to the second target mode.
- the terminal device includes a processor 2301 , a memory 2302 , a communication interface 2303 and a display 2304 .
- the memory 2302 is used to store instructions or programs executed by the processor 2301, or input data required by the processor 2301 to execute the instructions or programs, or to store data generated after the processor 2301 executes the instructions or programs.
- the processor 2301 is configured to execute the instructions or programs stored in the memory 2302 to perform the functions performed by the terminal device in the above method.
- the processor 2301 is configured to execute the functions of the first detection module 2001 , the sending module 2002 , the display module 2003 , and the second detection module 2004 .
- the processor 2301 is configured to execute the functions of the first detection module 2001 and the second detection module 2004 .
- the function of the sending module 2002 is realized by the communication interface 2303
- the function of the display module 2003 can be realized by the display 2304 .
- the processing module 2101 , the sending module 2102 , the receiving module 2103 , the display module 2104 and the detection module 2105 may be implemented by the processor 2301 .
- the processor 2301 can be used to execute the functions of the processing module 2101 and the detection module 2105
- the functions of the sending module 2102 and the receiving module 2103 can be implemented by the communication interface 2303
- the functions of the display module 2104 can be implemented by the display 2304 .
- the display module 2201 , the detection module 2202 , the sending module 2203 , the processing module 2204 , and the identification module 2205 may be implemented by the processor 2301 .
- the functions of the processing module 2204 , the detection module 2202 and the identification module 2205 can all be implemented by the processor 2301 .
- the function of the sending module 2203 can be implemented by the communication interface 2303
- the function of the display module 2201 can be implemented by the display 2304 .
- the processor in the embodiment of the present application may be a central processing unit (central processing unit, CPU), and may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSP), application-specific integrated circuits (application specific integrated circuit, ASIC), field programmable gate array (field programmable gate array, FPGA) or other programmable logic devices, transistor logic devices, hardware components or any combination thereof.
- a general-purpose processor may be a microprocessor or any conventional processor.
- this application will describe some optional earphone usage examples and signal processing examples with reference to some specific application scenarios.
- Users can use it in conjunction with the terminal and the headset.
- the headset can support at least one of the ANC function, the HT function, and the AH function, and of course it can also include an empty mode; among them, the ultimate goal of the ANC function is to eliminate the noise in actual hearing; the ultimate goal of the HT function is to eliminate the external
- the effect of the sound entering the human ear makes the effect of the external environment signal heard by the user through the earphone is equivalent to the effect of the user's naked ear hearing the sound of the current external environment; equivalence can refer to the same or approximately the same effect.
- the terminal device establishes a communication connection with the headset
- the first interface is used to set the function of the headset; the first interface may include an enabling option of an ANC function, an HT function, an AH function or an empty mode, or a disabling option thereof.
- a possible example of the first interface is shown in Figure 24.
- the noise control mode supported by the headset may include options of empty mode (off), ANC function (noise reduction), and HT function (transparent transmission). It should be understood that the first interface may further include more setting contents or options, which are not all shown in the drawings in this application.
- the transparent transmission strength of the HT function can be further acquired, and the HT function of the earphone can be controlled according to the acquired transparent transmission strength.
- an option of event sound enhancement is added in the first interface; it should be noted that, optionally, this option may not be available before the HT option is enabled.
- the event sound is a sound in the external environment that meets the preset event conditions; the event sound may include human voice or other sounds that meet the preset frequency spectrum characteristics, and for details, refer to the relevant descriptions in the foregoing embodiments.
- the terminal can control the earphone to improve the signal-to-noise ratio of the event sound in the signal collected by the earphone; wherein, the event sound The higher the signal-to-noise ratio of , the greater the proportion of the energy of the event sound in the signal.
- two methods for improving the signal-to-noise ratio of the event sound in the signal collected by the earphone will be described in detail in the following examples. See S4001-S4005 and S5001-S5005.
- FIG. 26 A possible implementation form can be seen in Fig. 26.
- the HT function transparent transmission
- the user's click instruction can be received or the preset vocal enhancement enabling instruction can be obtained, so that the Vocal Enhancement This option is enabled.
- the terminal can control both the ANC function and the HT function of the headset to be in an enabled state.
- the ANC function of the earphone may be activated again when the HT function is kept in an on state.
- the ANC function and the HT function are jointly enabled to process the sound signal of the external environment collected by the earphone, and the processing includes improving the signal-to-noise ratio of the event sound in the collected signal.
- the terminal may enable the ANC function of the earphone, and control the earphone to improve the signal-to-noise ratio of the event sound in the signal collected by the earphone in combination with some enhancement and noise reduction algorithms.
- the intensity of event sound enhancement may be further acquired, and the event sound enhancement function of the headset is controlled according to the acquired event sound enhancement intensity.
- the stronger the intensity of the event sound enhancement the higher the signal-to-noise ratio of the event sound.
- a strength option of the ANC function is added in the first interface. It should be noted that, optionally, this option may not appear in the first interface before the ANC option is enabled, or may appear in the first interface simultaneously with the ANC and HT function options.
- the strength options of the ANC function include at least a first steady-state ANC strength option, a second steady-state ANC strength option, and an adaptive ANC strength option; wherein the first steady-state ANC strength option and the second steady-state ANC strength option respectively correspond to the first steady-state ANC strength option and the second steady-state ANC strength option.
- the first scene and the second scene correspond to different and stable ANC function strengths respectively; the ANC function strength corresponding to the adaptive ANC strength option is related to the scene type of the current environment in which the terminal device or headset is located; scenes with different current environments Types correspond to different ANC strengths.
- the different scene types of the current environment may include the first scene and the second scene.
- the option of the intensity option (noise reduction mode) of the ANC function is added to the interface.
- the strength option (noise reduction mode) of the ANC function may further include multiple options, as shown in FIG. 28 , including but not limited to noise reduction modes such as light, balanced, deep, and intelligent dynamic noise reduction.
- the light level is suitable for quieter environments
- the depth level is suitable for very noisy environments, that is, scenes that do not belong to the light level but do not belong to the depth level can be classified as balanced levels, that is, ordinary scenes.
- the noise reduction intensity of the ANC control function of the headset corresponds to the light level
- the noise reduction intensity of the control headset ANC corresponds to the medium level
- the noise reduction intensity of the control headphone ANC corresponds to the depth level.
- the noise reduction intensity of the ANC function of the headset corresponds to the medium level.
- the noise reduction intensity of the ANC function of the headset corresponds to the depth level.
- the ANC noise reduction depth corresponding to the light level may include 20-28dB
- the ANC noise reduction depth corresponding to the balanced level may include 30-36dB
- the ANC noise reduction depth corresponding to the depth level may be greater than 40dB.
- the environmental scenes corresponding to the light level may include but are not limited to offices, bedrooms, quiet living rooms, etc.; the environmental scenes corresponding to the balanced level may include but are not limited to supermarkets, squares, waiting rooms, roads, cafes, shopping malls, etc.;
- the environmental scenes corresponding to the depth level may include but are not limited to subways, high-speed trains, taxis, and airplanes.
- the embodiment of the present application also includes a kind of intelligent dynamic noise reduction, that is, a kind of adaptive environmental noise reduction; that is, the ANC intensity can be determined according to the scene type of the current environment by acquiring the scene type of the environment where the terminal or the headset is located; The ANC function is controlled according to the determined ANC strength.
- a kind of intelligent dynamic noise reduction that is, a kind of adaptive environmental noise noise reduction
- the ANC intensity can be determined according to the scene type of the current environment by acquiring the scene type of the environment where the terminal or the headset is located
- the ANC function is controlled according to the determined ANC strength.
- the adaptive ambient noise noise reduction may include, but is not limited to, at least one of the above-mentioned light, balanced, and deep levels.
- adaptive environmental noise reduction can perform corresponding levels of ANC noise reduction in three levels: light, balanced, and deep according to the state of the environment. Detecting whether the current environment is light, balanced, or deep can be performed by earphones or terminals.
- adaptive ambient noise noise reduction allows users to perform adaptive noise reduction at different levels according to changes in the environment without manual operation, improving user experience.
- enabling methods for ANC, HT, event sound enhancement, noise reduction mode or closing options include, but are not limited to, receiving user clicks on corresponding function items, terminal adaptive switching, and headset adaptive switching. , or the trigger of the shortcut.
- a user's selection operation for the option of the event sound enhancement function, the selection operation for the option for the HT function, or the selection operation for the option of the ANC function is received; or; it is recognized that the current environment is related to the event sound Enhance the scene corresponding to the function and activate the option of the event sound enhancement function, recognize that the current environment is a scene corresponding to the HT function and activate the option of the HT function, or recognize that the current environment is corresponding to the ANC function scene and activate the option of the ANC function; or; in response to a user's pressing operation on the headset, the pressing operation is used for at least two of the event sound enhancement function, the HT function or the ANC function switch between functions.
- a pressure sensor may be included in the headset, and the pressure sensor may predefine some shortcut operations, such as switching between noise reduction modes.
- the strength of the ANC function mentioned in this application can be understood as the noise reduction strength of ANC, or simply the noise reduction strength; the strength of the HT function can be understood as the transparent transmission strength; the strength of the AH function can be understood as the enhancement strength. Different intensities will affect the relevant filter coefficients. For details, reference may be made to the relevant descriptions of the foregoing embodiments, which will not be repeated here.
- the earphone includes a first microphone (reference microphone), a second microphone (error microphone) and a speaker; in combination with the possible implementation forms of the earphone in the foregoing embodiment, the earphone can perform the following methods:
- S4001 collect a first signal through a first microphone, and the first signal is used to characterize the sound of the current external environment; in the prior art, the signal collected by the reference microphone is also called a reference signal
- S4002 Collect a second signal through the second microphone, where the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the headset.
- the signal collected by the error microphone is also called an error signal.
- the sound of the internal environment of the ear canal can be understood as the comprehensive sound of the ambient sound combined with the sound that the earphone may play, the algorithm being used by the earphone (such as noise reduction, transparent transmission, etc.), the human ear environment and other factors after the user puts the earphone on perception.
- the sound of the internal environment of the ear canal can be understood as, but not limited to, the comprehensive sound representation of the environmental sound collected by the error microphone combined with the human ear environment.
- the sound of the internal environment of the ear canal can be understood as, but not limited to, the environmental sound collected by the error microphone combined with the sound played by the headset microphone and the comprehensive sound representation of the human ear environment.
- the sound of the internal environment of the ear canal can be understood as, but not limited to, the ambient sound collected by the error microphone combined with the sound processed by the earphone microphone to play the algorithm-processed sound and the comprehensive sound representation of the human ear environment.
- the event sound is a sound that meets preset event conditions in the external environment
- control the ANC function and the HT function to be in an open state at the same time; at least use the HT function and the ANC function to perform target processing on the first signal and the second signal to obtain the target signal; the signal-to-noise ratio of the event sound in the target signal is greater than that in the first signal. The signal-to-noise ratio of the event sound.
- the HT function is used to transparently transmit the first signal collected by the reference microphone to obtain the restored signal C1; the event sound signal (for example, human voice) in the restored signal C1 is enhanced, and for the restored signal C1 The non-event sound signal is attenuated, and the event sound enhancement signal C2 is obtained;
- the event sound signal for example, human voice
- the target signal is obtained by processing the first signal C1, the signal collected by the error microphone, and the event sound enhancement signal C2 by using the ANC function.
- the related enhancement processing may refer to but not limited to the embodiment of FIG. 9 .
- the headset can also perform the following methods:
- the first microphone is used to collect a first signal, and the first signal is used to represent the sound of the current external environment;
- the second microphone is used to collect a second signal, and the second signal is used to represent the sound of the internal environment of the ear canal of the user wearing the headset;
- the event sound is a sound that meets preset event conditions in the external environment
- S5004 enable the ANC function; enhance the event sound signal in the first signal, and attenuate the non-event sound signal in the first signal to obtain an event sound enhancement signal; use the ANC function to The signal and the event sound enhancement signal are processed to obtain the target signal; the signal-to-noise ratio of the event sound in the target signal is greater than the signal-to-noise ratio of the event sound in the first signal; wherein, the higher the signal-to-noise ratio of the event sound in the signal, the Indicates that the greater the proportion of the energy of the event sound in the signal.
- the related enhancement processing may refer to but not limited to the embodiment of FIG. 9 .
- the earphone supports at least the ANC function, and the earphone includes a first microphone and a third microphone;
- the first microphone here can be understood as the reference microphone in the foregoing embodiment, which is more focused on collecting the sound of the current external environment , the third microphone is more focused on sound pickup; when the user wears the headset, the third microphone is closer to the user's mouth than the first microphone, so the third microphone can pick up the user's voice more clearly than the first microphone. signal; the headset can also perform the following methods:
- the headset enables or enables the ANC function
- S6004 Determine the noise level of the current scene according to the first signal and the second signal; wherein, different noise levels correspond to different ANC intensities.
- the correlation feature of the first signal and the second signal can be used for voice activity detection; noise tracking is performed on non-voice signals; if the noise energy is less than the first threshold, the current scene is determined as a quiet scene; or, if The spectral gravity center of the noise is located in the low frequency band, and the noise energy is greater than the second threshold, the current scene is determined to be a heavy noise scene; or, if the current scene is neither a quiet scene nor a heavy noise scene, the current scene is determined to be Ordinary scene; wherein, the second threshold (such as but not limited to a certain value in [-80dB, -65dB]) is greater than the first threshold (such as but not limited to a certain value in [-40dB, -30dB]).
- the ANC noise reduction intensity corresponding to the quiet scene, the normal scene, and the heavy noise scene increases sequentially.
- the ANC intensity can be preset with various intensity adjustment modes.
- the filter of the ANC algorithm is adjusted according to the ANC noise reduction intensity corresponding to the noise level of the response, and the ANC function can be controlled.
- the manner of controlling the ANC function may further include the methods for controlling the strength of the ANC as described in the above S3007 and S3008, which will not be repeated here.
- Switching is allowed after the threshold is met for 1 second, that is, the newly detected noise level lasts for a preset period of time before switching to the new noise level; 2) The current noise reduction level is maintained for at least 10 seconds before switching is allowed A scenario, that is, the newly switched state is maintained for at least a second preset time; 3) There may be switching back and forth between heavy noise, normal or quiet noise levels, and if the switching is too fast, it will cause discomfort. Therefore, the switching frequency within a period of time can be monitored, and if abnormal, the threshold value of the judgment level can be raised. For example, within the preset time period, if the number of switching exceeds the preset number (for example, 4 times in 2 minutes), the threshold value favoring the normal mode is increased, the frequent switching of the mode is reduced, and the user experience is improved.
- the preset number for example, 4 times in 2 minutes
- the embodiments of the present application also provide a headset control device, which is applied to terminal equipment, and the terminal equipment establishes a communication connection with the headset; the headset supports the active noise reduction ANC function and the ambient sound transparent transmission HT function; the device includes:
- a display module for displaying a first interface; the first interface is used to set the function of the headset; the first interface includes function options for event sound enhancement; the event sound is in the external environment that meets preset event conditions sound.
- the first interface includes an option for controlling the HT function of the earphone; when the option of the HT function is enabled, the processing module is used for the HT function of the earphone; the display module is further used for: in the first interface Added an option to enhance the sound of said events in .
- the processing module is configured to control both the ANC function and the HT function of the earphone to be in an enabled state when the option of the event sound enhancement function is enabled.
- the processing module is further configured to activate the ANC function of the headset.
- the embodiments of the present application also provide an earphone control device, which is applied to a terminal device, and the terminal device establishes a communication connection with the earphone; the earphone supports at least the active noise reduction ANC function; the device includes:
- a display module for displaying a first interface; the first interface is used to set the function of the headset; the first interface includes options for controlling the ANC function of the headset;
- the display module is further configured to add a strength option of the ANC function in the first interface after the option of the ANC function is enabled;
- the processing module is further configured to perform ANC noise reduction according to the enabling result of the strength option of the ANC function;
- the strength options of the ANC function include at least a first steady state ANC strength option, a second steady state ANC strength option, and an adaptive ANC strength option; wherein the first steady state ANC strength option and the second steady state ANC strength option
- the state ANC strength options correspond to the first scene and the second scene respectively, and respectively correspond to different and stable ANC function strengths; the ANC function strength corresponding to the adaptive ANC strength option is the same as that of the terminal device or the headset. It is related to the scene type of the current environment at the location; different scene types of the current environment correspond to different ANC strengths.
- the processing module is specifically configured to: when the first steady state ANC strength option is enabled, obtain the first ANC function strength corresponding to the first steady state ANC strength option; control the ANC function according to the first ANC function strength Or, when the second steady state ANC strength option is enabled, obtain the second ANC function strength corresponding to the second steady state ANC strength option; control the ANC function according to the second ANC function strength;
- the adaptive ANC strength option is enabled, the scene type of the current environment where the terminal device or the headset is located is obtained; the ANC strength is determined according to the scene type of the current environment; the ANC function is controlled according to the determined ANC strength.
- an embodiment of the present application also provides an earphone control device, the device is applied to a terminal device, and the terminal device establishes a communication connection with the earphone; the earphone supports at least an ambient sound transparent transmission HT function, and the device includes: :
- a display module for displaying a first interface; the first interface is used to set the function of the headset; the first interface includes options for controlling the HT function of the headset;
- a processing module configured to activate the HT function of the headset when the option of the HT function is enabled
- the display module is further configured to add an option for enhancing the event sound in the first interface after the option of the HT function is enabled; the event sound is a sound that meets preset event conditions in the external environment;
- the processing module is further configured to control the earphone to improve the signal-to-noise ratio of the event sound in the signal collected by the earphone when the option of the event sound enhancement function is enabled; A higher signal-to-noise ratio indicates a greater proportion of the energy of the event sound in the signal.
- the processing module is further specifically configured to: obtain the first strength of the ANC function, and control the ANC function of the earphone according to the first strength; or obtain the second strength of the HT function, and control according to the second strength The HT function of the earphone; or, acquiring a third intensity of event sound enhancement, and controlling the event sound enhancement function of the earphone according to the third intensity.
- the embodiment of the present application also provides a noise reduction device, which is applied to an earphone, the earphone supports at least the active noise reduction ANC function and the ambient sound transparent transmission HT function, and the earphone includes a first microphone, a third Two microphones and speakers; the device includes:
- the acquisition module is used to collect a first signal through the first microphone, where the first signal is used to represent the sound of the current external environment; and is also used to collect a second signal through the second microphone, and the second signal is used for sound to characterize the internal environment of the ear canal of the user wearing the headset;
- a receiving module for receiving an instruction of event sound enhancement;
- the event sound is a sound that meets preset event conditions in the external environment;
- a processing module configured to control both the ANC function and the HT function to be on, and at least use the HT function and the ANC function to perform target processing on the first signal and the second signal to obtain a target signal; the target The signal-to-noise ratio of the event sound in the signal is greater than the signal-to-noise ratio of the event sound in the first signal.
- the processing module is specifically configured to: transparently transmit the first signal by using the HT function to obtain a restored signal; enhance the event sound signal in the restored signal, and attenuate the non-event sound signal in the restored signal, Obtain the event sound enhancement signal; use the ANC function to process the first signal, the second signal and the event sound enhancement signal to obtain the target signal.
- Playing module playing the target signal through the speaker.
- the embodiments of the present application also provide a noise reduction device, which is applied to an earphone, the earphone supports at least an active noise reduction ANC function, and the earphone includes a first microphone, a second microphone and a speaker; the The device includes:
- the acquisition module is used to collect a first signal through the first microphone, where the first signal is used to represent the sound of the current external environment; and is also used to collect a second signal through the second microphone, and the second signal is used for sound to characterize the internal environment of the ear canal of the user wearing the headset;
- a receiving module for receiving an instruction of event sound enhancement;
- the event sound is a sound that meets preset event conditions in the external environment;
- a processing module for enabling the ANC function; enhancing the event sound signal in the first signal, and weakening the non-event sound signal in the first signal to obtain an event sound enhancement signal; using the ANC function
- the target signal is obtained by processing the first signal, the second signal and the event sound enhancement signal; the signal-to-noise ratio of the event sound in the target signal is greater than that of all the event sounds in the first signal.
- the signal-to-noise ratio of the event sound wherein, the higher the signal-to-noise ratio of the event sound in the signal, the greater the energy ratio of the event sound in the signal;
- Playing module playing the target signal through the speaker.
- the embodiments of the present application also provide a noise reduction device, which is applied to an earphone, the earphone supports at least the active noise reduction ANC function, the ambient sound transparent transmission HT function and the AH function, and the earphone includes an HT filter A filter bank, a feedback filter bank and a feedforward filter bank; the device includes:
- the acquisition module is used to acquire the working mode of the headset
- a calling module for: when the working mode is the ANC function, calling the feedback filter bank and the feedforward filter bank to realize the ANC function; when the working mode is the HT function, calling the HT function The filter bank and the feedback filter bank realize the HT function; when the working mode is the AH function, the HT filter bank, the feedforward filter bank and the feedback filter bank are called, Implement the AH function.
- an embodiment of the present application also provides a noise reduction device, which is applied to an earphone, and the earphone supports at least the ANC function; the earphone includes a first microphone and a third microphone; the first microphone is more focused In order to collect the sound of the current external environment, the third microphone focuses more on sound pickup;
- a collection module configured to collect a first signal for the current environment through the first microphone when the headset enables the ANC function; collect a second signal for the current environment through the third microphone;
- the identification module is configured to judge the noise level of the current scene according to the first signal and the second signal; wherein, different noise levels correspond to different ANC strengths.
- the identification module is specifically configured to: use the correlation feature of the first signal and the second signal for voice activity detection; perform noise tracking on non-speech signals; if the noise energy is less than the first threshold, the current The scene is determined to be a quiet scene; or, if the spectral center of gravity of the noise is located in a low frequency band, and the noise energy is greater than the second threshold, the current scene is determined to be a heavy noise scene; or, if the current scene does not belong to the quiet scene When the scene does not belong to the heavy noise scene, the current scene is determined to be an ordinary scene; wherein, the second threshold is greater than the first threshold.
- the processing module is used for controlling the ANC function according to the current noise level.
- the processing module is specifically configured to: if it is detected that the current scene is a new noise level and lasts for a preset period of time; obtain the ANC intensity corresponding to the new noise level; according to the ANC corresponding to the new noise level Intensity controls the ANC function.
- the method steps in the embodiments of the present application may be implemented in a hardware manner, or may be implemented in a manner in which a processor executes software instructions.
- the software instructions can be composed of corresponding software modules, and the software modules can be stored in random access memory (RAM), flash memory, read-only memory (ROM), programmable read-only memory (programmable ROM, PROM), erasable programmable read-only memory (erasable PROM, EPROM), electrically erasable programmable read-only memory (electrically ePROM, EEPROM), registers, hard disk, removable hard disk, CD-ROM or any known in the art other forms of storage media.
- An exemplary storage medium is coupled to the processor, such that the processor can read information from, and write information to, the storage medium.
- the storage medium can also be an integral part of the processor.
- the processor and storage medium may reside in an ASIC. Alternatively, the ASIC may be located in the end device. Of course, the processor and the storage medium may also exist in the terminal device as
- the above-mentioned embodiments it may be implemented in whole or in part by software, hardware, firmware or any combination thereof.
- software it can be implemented in whole or in part in the form of a computer program product.
- the computer program product includes one or more computer programs or instructions.
- the processes or functions described in the embodiments of the present application are executed in whole or in part.
- the computer may be a general purpose computer, a special purpose computer, a computer network, user equipment, or other programmable apparatus.
- the computer program or instructions may be stored in a computer-readable storage medium or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer program or instructions may be downloaded from a website site, computer, A server or data center transmits by wire or wireless to another website site, computer, server or data center.
- the computer-readable storage medium may be any available medium that can be accessed by a computer, or a data storage device such as a server, data center, or the like that integrates one or more available media.
- the usable medium can be a magnetic medium, such as a floppy disk, a hard disk, and a magnetic tape; it can also be an optical medium, such as a digital video disc (DVD); it can also be a semiconductor medium, such as a solid state drive (solid state drive). , SSD).
- a magnetic medium such as a floppy disk, a hard disk, and a magnetic tape
- an optical medium such as a digital video disc (DVD)
- DVD digital video disc
- it can also be a semiconductor medium, such as a solid state drive (solid state drive). , SSD).
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Otolaryngology (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Neurosurgery (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Computer Networks & Wireless Communication (AREA)
- Telephone Function (AREA)
- Circuit For Audible Band Transducer (AREA)
- User Interface Of Digital Computer (AREA)
- Soundproofing, Sound Blocking, And Sound Damping (AREA)
Abstract
Description
Claims (93)
- 一种模式控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:当识别当前外部环境的场景类型为目标场景时,根据所述目标场景确定目标模式;其中,所述目标模式为耳机支持的处理模式中的一个,不同的处理模式对应于不同的场景类型,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;向所述耳机发送所述目标模式,所述目标模式用于指示所述耳机实现所述目标模式所对应的处理功能。
- 如权利要求1所述的方法,其特征在于,根据所述目标场景确定目标模式时,所述方法还包括:显示结果提示信息,所述结果提示信息用于提示用户所述耳机实现所述目标模式所对应的处理功能。
- 如权利要求1所述的方法,其特征在于,向所述耳机发送第一控制信令之前,所述方法还包括:显示选择提示信息,所述选择提示信息用于提示用户是否调整所述耳机的处理模式为所述目标模式;检测到用户选择所述耳机的处理模式为所述目标模式的操作。
- 如权利要求2或3所述的方法,其特征在于,所述方法还包括:显示第一控件和第二控件,其中,所述第二控件在所述第一控件上的不同位置用于指示所述目标模式下的不同处理强度;向所述耳机发送第一控制信令之前,所述方法还包括:响应于用户触控所述第二控件移动到所述第一控件上的第一位置,所述第二控件在所述第一控件上的第一位置指示所述目标模式下的目标处理强度;向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
- 如权利要求4所述的方法,其特征在于,所述第一控件的形状为圆环,所述用户触控所述第二控件按顺时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件按逆时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
- 如权利要求4所述的方法,其特征在于,所述第一控件的形状为条形,所述用户触控所述第二控件从上向下在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件从下向上在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件从左向右在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件从右向左在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
- 如权利要求4-6任一所述的方法,其特征在于:当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
- 一种模式控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:获取目标模式;所述目标模式为耳机支持的处理模式中的一个,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;根据当前外部环境的场景类型确定所述目标模式下的目标处理强度;不同的场景类型对应于目标模式下的不同处理强度;向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
- 如权利要求8所述的方法,其特征在于,所述获取目标模式,包括:接收所述耳机发送的所述目标模式;或者,显示选择控件,所述选择控件包括耳机支持的处理模式,检测到用户通过所述选择控件在所述耳机的处理模式中选择所述目标模式的操作。
- 如权利要求8或9所述的方法,其特征在于,根据当前外部环境的场景类型确定所述目标模式下的目标处理强度之前,所述方法还包括:接收到所述耳机发送的所述目标模式的情况下,显示选择提示信息,所述选择提示信息用于指示所述用户是否调整所述耳机的处理模式为所述目标模式;检测到用户选择调整所述耳机的处理模式为所述目标模式的操作。
- 如权利要求8-10任一所述的方法,其特征在于:当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
- 一种模式控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:显示第一界面;第一界面上包括第一选择控件,所述第一选择控件包括第一目标耳机支持的处理模式以及第一目标耳机支持的处理模式对应的处理强度;所述第一目标耳机的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;响应于用户针对所述第一界面执行的第一操作;所述第一操作是用户通过所述第一选择控件在第一目标耳机支持的处理模式中选择第一目标模式以及选择第一目标模式下的处理强度为第一目标处理强度产生的;向所述第一目标耳机发送所述第一目标模式以及所述第一目标处理强度,所述第一目标模式用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能,所述第一目标处 理强度用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能时的处理强度。
- 如权利要求12所述的方法,其特征在于,显示第一界面之前,所述方法还包括:显示选择提示信息,所述选择提示信息用于所述用户选择是否调整所述第一目标耳机的处理模式;检测到用户选择调整所述第一目标耳机的处理模式的操作。
- 如权利要求11或12所述的方法,其特征在于,显示所述第一界面之前,所述方法还包括:识别当前外部环境的场景类型为目标场景,所述目标场景适配需要调整所述第一目标耳机的处理模式的场景类型。
- 如权利要求11或12所述的方法,其特征在于,显示所述第一界面之前,所述方法还包括:识别到所述终端设备触发所述第一目标耳机播放音频。
- 如权利要求11或12所述的方法,其特征在于,显示所述第一界面之前,所述方法还包括:检测到所述终端设备与所述第一目标耳机建立连接。
- 如权利要求11或12所述的方法,其特征在于,显示所述第一界面之前,所述方法还包括:检测到所述终端设备与所述第一目标耳机建立连接的情况下,检测到所述用户在主界面执行的第二操作;其中,所述主界面包括第一应用的图标,所述第二操作是所述用户触控所述第一应用的图标产生的,所述第一界面为所述第一应用的显示界面。
- 如权利要求11-17任一所述的方法,其特征在于,所述第一选择控件包括第一控件和第二控件,第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的两种不同处理模式,或者,所述第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的的同一种处理模式下的不同处理强度;所述第一操作是所述用户移动所述第二控件位于所述第一控件上所述第一目标模式对应的区域内的第一位置产生的,所述第一位置对应于所述第一目标模式下的第一目标处理强度。
- 如权利要求18所述的方法,其特征在于,所述第一控件的形状为圆环,所述圆环包括至少两个圆弧段,所述第二控件位于不同的圆弧段指示第一目标耳机的不同处理模式,所述第二控件位于同一圆弧段的不同位置指示第一目标耳机的同一处理模式的不同处理强度;或者,所述第一控件的形状为条形,所述条形包括至少两个条形段,所述第二控件位于不同的条形段指示所述第一目标耳机的不同处理模式,所述第二控件位于同一条形段的不同位置指示所述第一目标耳机的同一处理模式的不同处理强度。
- 如权利要求11-19任一所述的方法,其特征在于,所述方法还包括:响应于用户在第一界面执行的第三操作,所述第一界面上还包括第二选择控件,所述第二选择控件包括第二目标耳机支持的处理模式以及第二目标耳机支持的处理模式对应的处理强度,所述第一目标耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种,所述第三操作是所述用户通过所述第二选择控件在第二目标耳机的处理模式中选择第二目标模式以及选择第二目标模式下的处理强度为第二目标处理 强度产生的;所述第一目标耳机为左耳机时,所述第二目标耳机为右耳机,或者,所述第一目标耳机为右耳机,所述第二目标耳机为左耳机;向所述第二目标耳机发送所述第二目标模式以及所述第二目标处理强度,所述第二目标模式用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能,所述第二目标处理强度用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能时的处理强度。
- 一种模式控制装置,其特征在于,所述装置应用于终端设备,所述装置包括:第一检测模块,用于识别当前外部环境的场景类型为目标场景时,根据所述目标场景确定目标模式;其中,所述目标模式为耳机支持的处理模式中的一个,不同的处理模式对应于不同的场景类型,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;发送模块,用于向所述耳机发送所述目标模式,所述目标模式用于指示所述耳机实现所述目标模式所对应的处理功能。
- 如权利要求21所述的装置,其特征在于,还包括:显示模块,用于根据所述目标场景确定目标模式时,显示结果提示信息,所述结果提示信息用于提示用户所述耳机实现所述目标模式所对应的处理功能。
- 如权利要求21所述的装置,其特征在于,还包括:显示模块,用于在向所述耳机发送第一控制信令之前,显示选择提示信息,所述选择提示信息用于提示用户是否调整所述耳机的处理模式为所述目标模式;第二检测模块,用于检测到用户选择所述耳机的处理模式为所述目标模式的操作。
- 如权利要求23所述的装置,其特征在于,所述显示模块,还用于显示第一控件和第二控件,其中,所述第二控件在所述第一控件上的不同位置用于指示所述目标模式下的不同处理强度;所述第二检测模块,还用于在发送模块向所述耳机发送第一控制信令之前,检测用户触控所述第二控件移动到所述第一控件上的第一位置的操作,所述第二控件在所述第一控件上的第一位置指示所述目标模式下的目标处理强度;所述发送模块,还用于向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
- 如权利要求24所述的装置,其特征在于,所述第一控件的形状为圆环,所述用户触控所述第二控件按顺时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件按逆时针方向在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
- 如权利要求24所述的装置,其特征在于,所述第一控件的形状为条形,所述用户触控所述第二控件从上向下在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件从下向上在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件从左向右在所述第一控件上移动,所述目标模式下的处理强度由小到大变化;或者,所述用户触控所述第二控件从右向左在所述第一控件上移动,所述目标模式下的处理强度由小到大变化。
- 如权利要求24-26任一所述的装置,其特征在于:当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
- 一种模式控制装置,其特征在于,所述装置应用于终端设备,所述装置包括:处理模块,用于获取目标模式;所述目标模式为耳机支持的处理模式中的一个,所述耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;所述处理模块,还用于根据当前外部环境的场景类型确定所述目标模式下的目标处理强度;不同的场景类型对应于目标模式下的不同处理强度;发送模块,用于向所述耳机发送所述目标处理强度,所述目标处理强度用于指示所述耳机实现所述目标模式所对应的处理功能时的处理强度。
- 如权利要求28所述的装置,其特征在于,还包括:接收模块,用于接收所述耳机发送的所述目标模式。
- 如权利要求28所述的装置,其特征在于,还包括:显示模块,用于显示选择控件,所述选择控件包括耳机支持的处理模式,检测到用户通过所述选择控件在所述耳机的处理模式中选择所述目标模式的操作。
- 如权利要求29所述的装置,其特征在于,显示模块,还用于:在所述处理模块根据当前外部环境的场景类型确定所述目标模式下的目标处理强度之前,所述接收模块接收到所述耳机发送的所述目标模式的情况下,显示选择提示信息,所述选择提示信息用于指示所述用户是否调整所述耳机的处理模式为所述目标模式;还包括:检测模块,用于检测到用户选择调整所述耳机的处理模式为所述目标模式的操作。
- 如权利要求28-31任一所述的装置,其特征在于:当所述目标处理功能为ANC功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音以及所述用户耳道内部的环境声音越弱;或者,当所述目标处理功能为HT功能,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音的强度越大;或者,当所述目标处理功能为AH功能时,所述目标处理强度越大,所述用户感知到的所述用户当前所处环境的声音包括的事件声音越强。
- 一种模式控制装置,其特征在于,所述装置应用于终端设备,所述装置包括:显示模块,用于第一界面上包括第一选择控件,所述第一选择控件包括第一目标耳机支持的处理模式以及第一目标耳机支持的处理模式对应的处理强度;所述第一目标耳机的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种;检测模块,检测用户针对所述第一界面执行的第一操作;所述第一操作是用户通过所述第一选择控件在第一目标耳机支持的处理模式中选择第一目标模式以及选择第一目标模式下 的处理强度为第一目标处理强度产生的;发送模块,用于向所述第一目标耳机发送所述第一目标模式以及所述第一目标处理强度,所述第一目标模式用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能,所述第一目标处理强度用于指示所述第一目标耳机实现所述第一目标模式所对应的处理功能时的处理强度。
- 如权利要求33所述的装置法,其特征在于,所述显示模块,还用于在显示第一界面之前,显示选择提示信息,所述选择提示信息用于所述用户选择是否调整所述第一目标耳机的处理模式;所述检测模块,还用于检测用户选择调整所述第一目标耳机的处理模式的操作。
- 如权利要求33或34所述的装置,其特征在于,还包括:识别模块,用于在显示模式显示第一界面之前,识别当前外部环境的场景类型为目标场景,所述目标场景适配需要调整所述第一目标耳机的处理模式的场景类型。
- 如权利要求33或34所述的装置,其特征在于,还包括:识别模块,用于在显示模块显示所述第一界面之前,识别到所述终端设备触发所述第一目标耳机播放音频。
- 如权利要求33或34所述的装置,其特征在于,所述检测模块,还用于在显示模块显示所述第一界面之前,检测到所述终端设备与所述第一目标耳机建立连接。
- 如权利要求33或34所述的装置,其特征在于,所述检测模块,在显示模块显示所述第一界面之前,检测到所述终端设备与所述第一目标耳机建立连接的情况下,检测到所述用户在主界面执行的第二操作;其中,所述主界面包括第一应用的图标,所述第二操作是所述用户触控所述第一应用的图标产生的,所述第一界面为所述第一应用的显示界面。
- 如权利要求33-38任一所述的装置,其特征在于,所述第一选择控件包括第一控件和第二控件,第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的两种不同处理模式,或者,所述第二控件在所述第一控件的任意两个不同位置指示第一目标耳机的的同一种处理模式下的不同处理强度;所述第一操作是所述用户移动所述第二控件位于所述第一控件上所述第一目标模式对应的区域内的第一位置产生的,所述第一位置对应于所述第一目标模式下的第一目标处理强度。
- 如权利要求39所述的装置,其特征在于,所述第一控件的形状为圆环,所述圆环包括至少两个圆弧段,所述第二控件位于不同的圆弧段指示第一目标耳机的不同处理模式,所述第二控件位于同一圆弧段的不同位置指示第一目标耳机的同一处理模式的不同处理强度;或者,所述第一控件的形状为条形,所述条形包括至少两个条形段,所述第二控件位于不同的条形段指示所述第一目标耳机的不同处理模式,所述第二控件位于同一条形段的不同位置指示所述第一目标耳机的同一处理模式的不同处理强度。
- 如权利要求33-40任一所述的方法,其特征在于,所述检测模块,还用于响应于用户在第一界面执行的第三操作,所述第一界面上还包括第二选择控件,所述第二选择控件包括第二目标耳机支持的处理模式以及第二目标耳机支持的处理模式对应的处理强度,所述第一目标耳机支持的处理模式包括主动降噪ANC模式、环境声透传HT模式或者听觉增强AH模式中的至少两种,所述第三操作是所述用户通过所述第二选择控件在第二目标耳机的处理模式中选择第二目标模式以及选择第二目标模式下的处理强度为第二目标处理强度产生的; 所述第一目标耳机为左耳机时,所述第二目标耳机为右耳机,或者,所述第一目标耳机为右耳机,所述第二目标耳机为左耳机;所述发送模块,还用于向所述第二目标耳机发送所述第二目标模式以及所述第二目标处理强度,所述第二目标模式用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能,所述第二目标处理强度用于指示所述第二目标耳机实现所述第二目标模式所对应的处理功能时的处理强度。
- 一种终端设备,其特征在于,包括存储器、处理器和显示器;所述显示器,用于显示界面;所述存储器,用于存储程序或指令;所述处理器,用于调用所述程序或指令,以使得所述终端设备执行如权利要求1至20任一项所述的方法。
- 一种耳机控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:所述终端设备与耳机建立通信连接;所述耳机支持主动降噪ANC功能和环境声透传HT功能,显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括事件声音增强的功能选项;所述事件声音为外部环境中符合预设事件条件的声音;当所述事件声音增强功能的选项使能时,控制所述耳机的ANC功能和HT功能均处于开启状态。
- 如权利要求43所述的方法,其特征在于,所述第一界面中包括用于控制所述耳机HT功能的选项;在所述显示第一界面之后,所述方法还包括:当所述HT功能的选项使能时,激活所述耳机的HT功能;在所述第一界面中新增所述事件声音增强的选项。
- 如权利要求44所述的方法,其特征在于,所述控制所述耳机的ANC功能和HT功能均处于开启状态包括:控制所述耳机保持所述HT功能处于开启状态;激活所述耳机的ANC功能。
- 如权利要求43-45中任一项所述的方法,其特征在于,所述第一界面中包括用于控制所述耳机ANC功能的选项;所述方法还包括:当所述ANC功能的选项使能时,激活所述耳机的ANC功能。
- 如权利要求43-46中任一项所述的方法,其特征在于,所述第一界面中还包括用于关闭所述ANC功能的选项,和/或关闭所述HT功能的选项。
- 如权利要求47所述的方法,其特征在于,在激活所述耳机的ANC功能之后,在所述第一界面中新增所述ANC功能的强度选项。
- 如权利要求48所述的方法,其特征在于,所述ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,所述第一稳态ANC强度选项和所述第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的ANC功能强度;所述自适应ANC强度选项对应的ANC功能强度与所述终端设备或所述耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。
- 如权利要求49所述的方法,其特征在于,所述方法还包括,当所述ANC强度自适 应选项被使能时,获取所述终端或所述耳机所处当前环境的场景类型;根据所述场景类型匹配出ANC功能的目标强度;根据所述目标强度控制所述耳机的ANC功能。
- 如权利要求50所述的方法,其特征在于,所述当前环境不同的场景类型包括第一场景和第二场景。
- 如权利要求43-51任一项所述的方法,其特征在于,所述事件声音包括人声或者符合预设频谱特性的其它声音。
- 如权利要求43-52任一项所述的方法,其特征在于,在激活所述耳机的HT功能之后,所述方法还包括:获取HT功能的目标强度;根据所述HT功能的目标强度控制所述耳机的HT功能。
- 如权利要求43-53任一项所述的方法,其特征在于,所述事件声音增强功能的选项使能、所述HT功能的选项使能或所述ANC功能的选项使能包括:响应用户针对相应功能选项的点选、相应功能的自适应切换,或者相应功能的快捷方式触发。
- 如权利要求43-54任一项所述的方法,其特征在于,所述ANC功能和HT功能共同开启用于对所述耳机采集到的外部环境的声音信号进行处理,所述处理包括针对采集的信号中所述事件声音的信噪比进行提升;其中,所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大。
- 一种降噪方法,其特征在于,所述方法应用于耳机,所述耳机至少支持主动降噪ANC功能和环境声透传HT功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述方法包括:通过所述第一麦克风采集第一信号,所述第一信号用于表征当前外部环境的声音;通过所述第二麦克风采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;控制ANC功能和HT功能均处于开启状态;至少利用所述HT功能、所述ANC功能对所述第一信号和所述第二信号进行目标处理得到目标信号;所述目标信号中的事件声音的信噪比大于所述第一信号中的事件声音的信噪比;通过所述扬声器播放所述目标信号。
- 如权利要求56所述的方法,其特征在于,所述至少利用所述HT功能、所述ANC功能对所述第一信号和所述第二信号进行目标处理得到目标信号包括:利用HT功能透传所述第一信号得到还原信号;针对所述还原信号中的事件声音信号进行增强,并对于所述还原信号中的非事件声音信号进行减弱,得到事件声音增强信号;利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号。
- 一种降噪方法,其特征在于,所述方法应用于耳机,所述耳机至少支持主动降噪ANC功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述方法包括:通过所述第一麦克风用于采集第一信号,所述第一信号用于表征当前外部环境的声音;通过所述第二麦克风用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;开启所述ANC功能;对所述第一信号中的事件声音信号进行增强,并对所述第一信号中的非事件声音信号进行减弱,得到事件声音增强信号;利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号;所述目标信号中的所述事件声音的信噪比大于所述第一信号中的所述事件声音的信噪比;其中,信号中所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大;通过所述扬声器播放所述目标信号。
- 一种信号处理方法,其特征在于,所述方法应用于耳机,所述耳机至少支持主动降噪ANC功能、环境声透传HT功能和AH功能,所述耳机包括HT滤波器组、反馈滤波器组和前馈滤波器组;所述方法包括:获取耳机的工作模式;当所述工作模式为ANC功能时,调用所述反馈滤波器组和前馈滤波器组,实现所述ANC功能;当所述工作模式为HT功能时,调用所述HT滤波器组和所述反馈滤波器组,实现所述HT功能;当所述工作模式为AH功能时,调用所述HT滤波器组、所述前馈滤波器组和所述反馈滤波器组,实现所述AH功能。
- 一种主动降噪ANC的强度调节方法,其特征在于,所述方法应用于耳机,所述耳机至少支持ANC功能;所述耳机包括第一麦克风、第三麦克风;所述第一麦克风更侧重于采集当前外部环境的声音,所述第三麦克风更侧重于拾音;当所述耳机的所述ANC功能处于开启状态时,通过所述第一麦克风针对当前环境采集第一信号;通过所述第三麦克风针对所述当前环境采集第二信号;根据所述第一信号和所述第二信号判断所述当前场景的噪声等级;其中,不同的噪声等级对应于不同的ANC强度;根据当前的噪声等级控制所述ANC功能。
- 如权利要求60所述的方法,其特征在于,所述根据所述第一信号和所述第二信号判断所述当前场景的噪声等级包括:利用第一信号和第二信号的相关性特征作语音活动检测;对非语音信号进行噪声跟踪;若所述噪声能量小于第一阈值,则将当前场景确定为安静场景;或者,若所述噪声的频谱重心位于低频段,且噪声能量大于第二阈值,则将当前场景确定为重噪场景;或者,若所述当前场景既不属于所述安静场景也不属于所述重噪场景时,则将当前场景确定为普通场景;其中,所述第二阈值大于所述第一阈值。
- 如权利要求61所述的方法,其特征在于,所述安静场景、所述普通场景、所述重噪 场景对应的ANC强度依次变大。
- 如权利要求60-62中任一项所述的方法,其特征在于,所述根据当前的噪声等级控制所述ANC功能包括:若检测到当前场景为新的噪声等级且持续预设时长;获取所述新的噪声等级对应的ANC强度;根据所述新的噪声等级对应的ANC强度控制所述ANC功能。
- 一种耳机控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:所述终端设备与耳机建立通信连接;所述耳机至少支持主动降噪ANC功能;显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机ANC功能的选项;当所述ANC功能的选项使能时,激活所述耳机的ANC功能;在所述第一界面中新增所述ANC功能的强度选项;根据所述ANC功能的强度选项的使能结果进行ANC降噪;其中,所述ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,所述第一稳态ANC强度选项和所述第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的且稳定的ANC功能强度;所述自适应ANC强度选项对应的ANC功能强度与所述终端设备或所述耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。
- 如权利要求64所述的方法,其特征在于,所述根据所述ANC功能的强度选项的使能结果进行ANC降噪包括:当第一稳态ANC强度选项使能时,获取第一稳态的ANC强度选项对应的第一ANC功能强度;根据所述第一ANC功能强度控制所述ANC功能;或者,当第二稳态ANC强度选项使能时,获取第二稳态的ANC强度选项对应的第二ANC功能强度;根据所述第二ANC功能强度控制所述ANC功能。
- 如权利要求64或65所述的方法,其特征在于,所述根据所述ANC功能的强度选项的使能结果进行ANC降噪包括:当所述自适应ANC强度选项使能时,获取所述终端设备或所述耳机所处当前环境的场景类型;根据当前环境的场景类型确定ANC强度;根据确定出来的ANC强度控制所述ANC功能。
- 如权利要求64-66中任一项所述的方法,其特征在于,所述当前环境不同的场景类型包括第一场景和第二场景。
- 一种耳机控制方法,其特征在于,所述方法应用于终端设备,所述方法包括:所述终端设备与耳机建立通信连接;所述耳机至少支持环境声透传HT功能,显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机HT功能的选项;当所述HT功能的选项使能时,激活所述耳机的HT功能;在所述第一界面中新增所述事件声音增强的选项;所述事件声音为外部环境中符合预设事件条件的声音;当所述事件声音增强功能的选项使能时,控制所述耳机针对所述耳机采集的信号中所述事件声音的信噪比进行提升;其中,所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大。
- 如权利要求68所述的方法,其特征在于,所述事件声音包括人声或者符合预设频谱特性的其它声音。
- 如权利要求68或69所述的方法,其特征在于,所述第一界面中包括用于控制所述耳机ANC功能的选项、用于关闭所述ANC功能的选项,和/或关闭所述HT功能的选项。
- 如权利要求70所述的方法,其特征在于,所述方法还包括:获取ANC功能的第一强度,并根据所述第一强度控制所述耳机的ANC功能;或者,获取HT功能的第二强度,并根据所述第二强度控制所述耳机的HT功能;或者,获取事件声音增强的第三强度,并根据所述第三强度控制所述耳机的事件声音增强功能。
- 一种耳机控制装置,其特征在于,所述装置应用于终端设备,所述终端设备与耳机建立通信连接;所述耳机支持主动降噪ANC功能和环境声透传HT功能;所述装置包括:显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括事件声音增强的功能选项;所述事件声音为外部环境中符合预设事件条件的声音;处理模块,用于当所述事件声音增强功能的选项使能时,控制所述耳机的ANC功能和HT功能均处于开启状态。
- 如权利要求72所述的装置,其特征在于,所述第一界面中包括用于控制所述耳机HT功能的选项;当所述HT功能的选项使能时,所述处理模块用于激活所述耳机的HT功能;所述显示模块还用于:在所述第一界面中新增所述事件声音增强的选项。
- 如权利要求73所述的装置,其特征在于,所述处理模块还用于控制所述耳机保持所述HT功能处于开启状态;激活所述耳机的ANC功能。
- 如权利要求72-74中任一项所述的装置,其特征在于,所述第一界面中包括用于控制所述耳机ANC功能的选项;所述处理模块还用于:当所述ANC功能的选项使能时,激活所述耳机的ANC功能。
- 如权利要求72-75中任一项所述的装置,其特征在于,所述第一界面中还包括用于关闭所述ANC功能的选项,和/或关闭所述HT功能的选项。
- 如权利要求72-76中任一项所述的装置,其特征在于,所述事件声音包括人声或者符合预设频谱特性的其它声音。
- 一种降噪装置,其特征在于,所述装置应用于耳机,所述耳机至少支持主动降噪ANC功能和环境声透传HT功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述装置包括:采集模块,用于通过所述第一麦克风采集第一信号,所述第一信号用于表征当前外部环境的声音;还用于通过所述第二麦克风采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;接收模块,用于接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;处理模块,用于控制ANC功能和HT功能均处于开启状态,并至少利用所述HT功能、 所述ANC功能对所述第一信号和所述第二信号进行目标处理得到目标信号;所述目标信号中的事件声音的信噪比大于所述第一信号中的事件声音的信噪比;播放模块,通过所述扬声器播放所述目标信号。
- 如权利要求78所述的装置,其特征在于,所述处理模块具体用于:利用HT功能透传所述第一信号得到还原信号;针对所述还原信号中的事件声音信号进行增强,并对于所述还原信号中的非事件声音信号进行减弱,得到事件声音增强信号;利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号。
- 一种降噪装置,其特征在于,所述装置应用于耳机,所述耳机至少支持主动降噪ANC功能,所述耳机包括第一麦克风、第二麦克风和扬声器;所述装置包括:采集模块,用于通过所述第一麦克风采集第一信号,所述第一信号用于表征当前外部环境的声音;还用于通过所述第二麦克风采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;接收模块,用于接收事件声音增强的指令;所述事件声音为外部环境中符合预设事件条件的声音;处理模块,用于开启所述ANC功能;对所述第一信号中的事件声音信号进行增强,并对所述第一信号中的非事件声音信号进行减弱,得到事件声音增强信号;利用ANC功能对所述第一信号、所述第二信号和所述事件声音增强信号进行处理得到所述目标信号;所述目标信号中的所述事件声音的信噪比大于所述第一信号中的所述事件声音的信噪比;其中,信号中所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大;播放模块,通过所述扬声器播放所述目标信号。
- 一种信号处理装置,其特征在于,所述装置应用于耳机,所述耳机至少支持主动降噪ANC功能、环境声透传HT功能和AH功能,所述耳机包括HT滤波器组、反馈滤波器组和前馈滤波器组;所述装置包括:获取模块,用于获取耳机的工作模式;调用模块,用于:当所述工作模式为ANC功能时,调用所述反馈滤波器组和前馈滤波器组,实现所述ANC功能;当所述工作模式为HT功能时,调用所述HT滤波器组和所述反馈滤波器组,实现所述HT功能;当所述工作模式为AH功能时,调用所述HT滤波器组、所述前馈滤波器组和所述反馈滤波器组,实现所述AH功能。
- 一种主动降噪ANC的强度调节装置,其特征在于,所述装置应用于耳机,所述耳机至少支持ANC功能;所述耳机包括第一麦克风、第三麦克风;所述第一麦克风更侧重于采集当前外部环境的声音,所述第三麦克风更侧重于拾音;采集模块,用于当所述耳机的所述ANC功能处于开启状态时,通过所述第一麦克风针对当前环境采集第一信号;通过所述第三麦克风针对所述当前环境采集第二信号;识别模块,用于根据所述第一信号和所述第二信号判断所述当前场景的噪声等级;其中,不同的噪声等级对应于不同的ANC强度;处理模块,用于根据当前的噪声等级控制所述ANC功能。
- 如权利要求82所述的装置,其特征在于,所述识别模块具体用于:利用第一信号和第二信号的相关性特征作语音活动检测;对非语音信号进行噪声跟踪;若所述噪声能量小于第一阈值,则将当前场景确定为安静场景;或者,若所述噪声的频谱重心位于低频段,且噪声能量大于第二阈值,则将当前场景确定为重噪场景;或者,若所述当前场景既不属于所述安静场景也不属于所述重噪场景时,则将当前场景确定为普通场景;其中,所述第二阈值大于所述第一阈值。
- 如权利要求83所述的装置,其特征在于,所述安静场景、所述普通场景、所述重噪场景对应的ANC强度依次变大。
- 如权利要求82-84中任一项所述的装置,其特征在于,所述处理模块具体用于:若检测到当前场景为新的噪声等级且持续预设时长;获取所述新的噪声等级对应的ANC强度;根据所述新的噪声等级对应的ANC强度控制所述ANC功能。
- 一种耳机控制装置,其特征在于,所述装置应用于终端设备,所述终端设备与耳机建立通信连接;所述耳机至少支持主动降噪ANC功能;所述装置包括:显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机ANC功能的选项;处理模块,用于当所述ANC功能的选项使能时,激活所述耳机的ANC功能;所述显示模块还用于在所述ANC功能的选项使能之后在所述第一界面中新增所述ANC功能的强度选项;所述处理模块还用于根据所述ANC功能的强度选项的使能结果进行ANC降噪;其中,所述ANC功能的强度选项至少包括第一稳态ANC强度选项、第二稳态ANC强度选项和自适应ANC强度选项;其中,所述第一稳态ANC强度选项和所述第二稳态ANC强度选项分别对应于第一场景和第二场景,并分别对应于不同的且稳定的ANC功能强度;所述自适应ANC强度选项对应的ANC功能强度与所述终端设备或所述耳机所处的当前环境的场景类型有关;当前环境不同的场景类型对应不同的ANC强度。
- 如权利要求86所述的装置,其特征在于,所述处理模块具体用于:当第一稳态ANC强度选项使能时,获取第一稳态的ANC强度选项对应的第一ANC功能强度;根据所述第一ANC功能强度控制所述ANC功能;或者,当第二稳态ANC强度选项使能时,获取第二稳态的ANC强度选项对应的第二ANC功能强度;根据所述第二ANC功能强度控制所述ANC功能;或者,当所述自适应ANC强度选项使能时,获取所述终端设备或所述耳机所处当前环境的场景类型;根据当前环境的场景类型确定ANC强度;根据确定出来的ANC强度控制所述ANC功能。
- 一种耳机控制装置,其特征在于,所述装置应用于终端设备,所述终端设备与耳机建立通信连接;所述耳机至少支持环境声透传HT功能,所述装置包括:显示模块,用于显示第一界面;所述第一界面用于设置耳机的功能;所述第一界面中包括用于控制所述耳机HT功能的选项;处理模块,用于当所述HT功能的选项使能时,激活所述耳机的HT功能;显示模块还用于在所述HT功能的选项使能后,在所述第一界面中新增所述事件声音增强的选项;所述事件声音为外部环境中符合预设事件条件的声音;所述处理模块还用于当所述事件声音增强功能的选项使能时,控制所述耳机针对所述耳机采集的信号中所述事件声音的信噪比进行提升;其中,所述事件声音的信噪比越高,表明信号中所述事件声音的能量占比越大。
- 如权利要求88所述的装置,其特征在于,所述事件声音包括人声或者符合预设频谱特性的其它声音。
- 如权利要求88或89所述的装置,其特征在于,所述第一界面中包括用于控制所述耳机ANC功能的选项、用于关闭所述ANC功能的选项,和/或关闭所述HT功能的选项。
- 如权利要求90所述的装置,其特征在于,所述处理模块还具体用于:获取ANC功能的第一强度,并根据所述第一强度控制所述耳机的ANC功能;或者,获取HT功能的第二强度,并根据所述第二强度控制所述耳机的HT功能;或者,获取事件声音增强的第三强度,并根据所述第三强度控制所述耳机的事件声音增强功能。
- 一种终端设备,其特征在于,所述终端设备包括存储器、处理器、触摸屏、总线;所述存储器、所述触摸屏以及所述处理器通过所述总线相连;所述存储器用于存储计算机程序和指令;所述触摸屏用于显示界面;所述处理器用于调用所述存储器中存储的所述计算机程序、指令,可执行如权利要求43-55,64-67,68-71中任一项所述方法。
- 一种耳机,所述耳机包括第一麦克风、第二麦克风、处理器、存储器以及扬声器;所述第一麦克风,用于采集第一信号,所述第一信号用于表征当前外部环境的声音;所述第二麦克风,用于采集第二信号,所述第二信号用于表征佩戴所述耳机的用户耳道内部环境的声音;所述扬声器用于播放音频,所述存储器,用于存储程序或指令;所述处理器,用于调用所述程序或指令,以使得所述耳机执行如权利要求56-57,58,59,60-63中任一项所述方法。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP21834164.2A EP4171060A4 (en) | 2020-06-30 | 2021-06-30 | MODE CONTROL METHOD AND APPARATUS, AND TERMINAL DEVICE |
BR112022026923A BR112022026923A2 (pt) | 2020-06-30 | 2021-06-30 | Método e aparelho de controle de modo e dispositivo terminal |
KR1020237002989A KR20230027296A (ko) | 2020-06-30 | 2021-06-30 | 모드 제어 방법 및 장치, 그리고 단말 디바이스 |
US18/148,080 US20230164475A1 (en) | 2020-06-30 | 2022-12-29 | Mode Control Method and Apparatus, and Terminal Device |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010616084 | 2020-06-30 | ||
CN202010616084.7 | 2020-06-30 | ||
CN202010949885.5A CN113873379B (zh) | 2020-06-30 | 2020-09-10 | 一种模式控制方法、装置及终端设备 |
CN202010949885.5 | 2020-09-10 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/148,080 Continuation US20230164475A1 (en) | 2020-06-30 | 2022-12-29 | Mode Control Method and Apparatus, and Terminal Device |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022002110A1 true WO2022002110A1 (zh) | 2022-01-06 |
Family
ID=78982086
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/103435 WO2022002110A1 (zh) | 2020-06-30 | 2021-06-30 | 一种模式控制方法、装置及终端设备 |
Country Status (6)
Country | Link |
---|---|
US (1) | US20230164475A1 (zh) |
EP (1) | EP4171060A4 (zh) |
KR (1) | KR20230027296A (zh) |
CN (1) | CN113873379B (zh) |
BR (1) | BR112022026923A2 (zh) |
WO (1) | WO2022002110A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114640938A (zh) * | 2022-05-18 | 2022-06-17 | 深圳市听多多科技有限公司 | 一种基于蓝牙耳机芯片的助听功能实现方法及蓝牙耳机 |
CN114640937A (zh) * | 2022-05-18 | 2022-06-17 | 深圳市听多多科技有限公司 | 一种基于穿戴设备系统的助听功能实现方法及穿戴设备 |
EP4290885A1 (en) * | 2022-06-07 | 2023-12-13 | Starkey Laboratories, Inc. | Context-based situational awareness for hearing instruments |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
USD921669S1 (en) * | 2019-09-09 | 2021-06-08 | Apple Inc. | Display screen or portion thereof with animated graphical user interface |
TWI802108B (zh) * | 2021-05-08 | 2023-05-11 | 英屬開曼群島商意騰科技股份有限公司 | 降低聲學回音之語音處理裝置及其方法 |
CN113938787B (zh) * | 2021-12-16 | 2022-03-15 | 深圳市鑫正宇科技有限公司 | 基于数字均衡技术的骨传导耳机 |
CN114466278B (zh) * | 2022-04-11 | 2022-08-16 | 北京荣耀终端有限公司 | 一种耳机模式对应的参数确定方法、耳机、终端和系统 |
CN117751585A (zh) * | 2022-06-20 | 2024-03-22 | 北京小米移动软件有限公司 | 智能耳机的控制方法、装置、电子设备和存储介质 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140126734A1 (en) * | 2012-11-02 | 2014-05-08 | Bose Corporation | Providing Ambient Naturalness in ANR Headphones |
CN106303839A (zh) * | 2016-09-29 | 2017-01-04 | 中山市天键电声有限公司 | 基于手机app的anc降噪控制方法 |
US20180077482A1 (en) * | 2015-05-15 | 2018-03-15 | Huawei Technologies Co., Ltd. | Noise Reduction Headset Setting Method, Terminal, and Noise Reduction Headset |
US10284955B2 (en) * | 2013-05-23 | 2019-05-07 | Comhear, Inc. | Headphone audio enhancement system |
CN110825446A (zh) * | 2019-10-28 | 2020-02-21 | Oppo广东移动通信有限公司 | 参数配置方法、装置、存储介质及电子设备 |
CN111107461A (zh) * | 2019-12-13 | 2020-05-05 | 恒玄科技(北京)有限公司 | 一种降噪耳机的配置方法、装置及智能终端、降噪耳机 |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9716939B2 (en) * | 2014-01-06 | 2017-07-25 | Harman International Industries, Inc. | System and method for user controllable auditory environment customization |
CN104618829A (zh) * | 2014-12-29 | 2015-05-13 | 歌尔声学股份有限公司 | 耳机环境声音的调节方法和耳机 |
JP6588098B2 (ja) * | 2015-01-26 | 2019-10-09 | 深▲せん▼市冠旭電子股▲ふん▼有限公司Shenzhen Grandsun Electronic Co., Ltd. | ヘッドセットの騒音低減を制御する方法及び装置 |
WO2017101067A1 (zh) * | 2015-12-17 | 2017-06-22 | 华为技术有限公司 | 一种对周围环境音的处理方法及设备 |
US10809972B2 (en) * | 2016-09-27 | 2020-10-20 | Sony Corporation | Information processing device, information processing method, and program |
CN110049403A (zh) * | 2018-01-17 | 2019-07-23 | 北京小鸟听听科技有限公司 | 一种基于场景识别的自适应音频控制装置和方法 |
CN108429963A (zh) * | 2018-05-08 | 2018-08-21 | 歌尔股份有限公司 | 一种耳机及降噪方法 |
-
2020
- 2020-09-10 CN CN202010949885.5A patent/CN113873379B/zh active Active
-
2021
- 2021-06-30 KR KR1020237002989A patent/KR20230027296A/ko not_active Application Discontinuation
- 2021-06-30 EP EP21834164.2A patent/EP4171060A4/en active Pending
- 2021-06-30 BR BR112022026923A patent/BR112022026923A2/pt unknown
- 2021-06-30 WO PCT/CN2021/103435 patent/WO2022002110A1/zh unknown
-
2022
- 2022-12-29 US US18/148,080 patent/US20230164475A1/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140126734A1 (en) * | 2012-11-02 | 2014-05-08 | Bose Corporation | Providing Ambient Naturalness in ANR Headphones |
US10284955B2 (en) * | 2013-05-23 | 2019-05-07 | Comhear, Inc. | Headphone audio enhancement system |
US20180077482A1 (en) * | 2015-05-15 | 2018-03-15 | Huawei Technologies Co., Ltd. | Noise Reduction Headset Setting Method, Terminal, and Noise Reduction Headset |
CN106303839A (zh) * | 2016-09-29 | 2017-01-04 | 中山市天键电声有限公司 | 基于手机app的anc降噪控制方法 |
CN110825446A (zh) * | 2019-10-28 | 2020-02-21 | Oppo广东移动通信有限公司 | 参数配置方法、装置、存储介质及电子设备 |
CN111107461A (zh) * | 2019-12-13 | 2020-05-05 | 恒玄科技(北京)有限公司 | 一种降噪耳机的配置方法、装置及智能终端、降噪耳机 |
Non-Patent Citations (1)
Title |
---|
See also references of EP4171060A4 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114640938A (zh) * | 2022-05-18 | 2022-06-17 | 深圳市听多多科技有限公司 | 一种基于蓝牙耳机芯片的助听功能实现方法及蓝牙耳机 |
CN114640937A (zh) * | 2022-05-18 | 2022-06-17 | 深圳市听多多科技有限公司 | 一种基于穿戴设备系统的助听功能实现方法及穿戴设备 |
CN114640938B (zh) * | 2022-05-18 | 2022-08-23 | 深圳市听多多科技有限公司 | 一种基于蓝牙耳机芯片的助听功能实现方法及蓝牙耳机 |
CN114640937B (zh) * | 2022-05-18 | 2022-09-02 | 深圳市听多多科技有限公司 | 一种基于穿戴设备系统的助听功能实现方法及穿戴设备 |
EP4290885A1 (en) * | 2022-06-07 | 2023-12-13 | Starkey Laboratories, Inc. | Context-based situational awareness for hearing instruments |
Also Published As
Publication number | Publication date |
---|---|
KR20230027296A (ko) | 2023-02-27 |
CN113873379B (zh) | 2023-05-02 |
US20230164475A1 (en) | 2023-05-25 |
BR112022026923A2 (pt) | 2023-03-14 |
EP4171060A4 (en) | 2024-03-13 |
EP4171060A1 (en) | 2023-04-26 |
CN113873379A (zh) | 2021-12-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022002166A1 (zh) | 一种耳机噪声处理方法、装置及耳机 | |
WO2022002110A1 (zh) | 一种模式控制方法、装置及终端设备 | |
CN113169760B (zh) | 无线短距离音频共享方法及电子设备 | |
WO2021083128A1 (zh) | 一种声音处理方法及其装置 | |
WO2021000817A1 (zh) | 环境音处理方法及相关装置 | |
CN113571035B (zh) | 降噪方法及降噪装置 | |
CN115729511A (zh) | 一种播放音频的方法及电子设备 | |
CN114466107A (zh) | 音效控制方法、装置、电子设备及计算机可读存储介质 | |
JP2023156321A (ja) | 通話方法および装置 | |
CN113141483B (zh) | 基于视频通话的共享屏幕方法及移动设备 | |
CN111065020B (zh) | 音频数据处理的方法和装置 | |
WO2022257563A1 (zh) | 一种音量调节的方法,电子设备和系统 | |
CN114339429A (zh) | 音视频播放控制方法、电子设备和存储介质 | |
WO2022089563A1 (zh) | 一种声音增强方法、耳机控制方法、装置及耳机 | |
WO2024046416A1 (zh) | 一种音量调节方法、电子设备及系统 | |
CN113938556B (zh) | 来电提示方法、装置和电子设备 | |
WO2024066933A9 (zh) | 扬声器控制方法及设备 | |
WO2023197999A1 (zh) | 显示方法及电子设备 | |
WO2024046182A1 (zh) | 一种音频播放方法、系统及相关装置 | |
WO2024032035A1 (zh) | 一种语音信号的输出方法和电子设备 | |
CN117528370A (zh) | 信号处理方法及装置、设备控制方法及装置 | |
CN118098261A (zh) | 一种音频处理方法、设备及系统 | |
CN115706752A (zh) | 来电音量控制方法及电子设备 | |
CN117692845A (zh) | 声场校准方法、电子设备及系统 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21834164 Country of ref document: EP Kind code of ref document: A1 |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112022026923 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 2021834164 Country of ref document: EP Effective date: 20230120 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 112022026923 Country of ref document: BR Kind code of ref document: A2 Effective date: 20221228 |