WO2022218027A1 - 音频播放方法、装置、计算机可读存储介质及电子设备 - Google Patents
音频播放方法、装置、计算机可读存储介质及电子设备 Download PDFInfo
- Publication number
- WO2022218027A1 WO2022218027A1 PCT/CN2022/076239 CN2022076239W WO2022218027A1 WO 2022218027 A1 WO2022218027 A1 WO 2022218027A1 CN 2022076239 W CN2022076239 W CN 2022076239W WO 2022218027 A1 WO2022218027 A1 WO 2022218027A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user
- information
- audio
- target
- emotion
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 96
- 230000008451 emotion Effects 0.000 claims description 135
- 230000008909 emotion recognition Effects 0.000 claims description 68
- 230000001815 facial effect Effects 0.000 claims description 33
- 230000005236 sound signal Effects 0.000 claims description 20
- 230000002996 emotional effect Effects 0.000 claims description 16
- 230000004044 response Effects 0.000 claims description 16
- 239000000284 extract Substances 0.000 claims description 15
- 238000000926 separation method Methods 0.000 claims description 12
- 238000004590 computer program Methods 0.000 claims description 11
- 230000001755 vocal effect Effects 0.000 claims description 7
- 238000004519 manufacturing process Methods 0.000 abstract 2
- 230000008685 targeting Effects 0.000 abstract 1
- 238000012549 training Methods 0.000 description 20
- 238000010586 diagram Methods 0.000 description 9
- 238000000605 extraction Methods 0.000 description 8
- 238000004891 communication Methods 0.000 description 7
- 239000013598 vector Substances 0.000 description 7
- 238000012545 processing Methods 0.000 description 5
- 239000011435 rock Substances 0.000 description 5
- 238000001514 detection method Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000004422 calculation algorithm Methods 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000036651 mood Effects 0.000 description 2
- 238000003062 neural network model Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000033764 rhythmic process Effects 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- 241000282412 Homo Species 0.000 description 1
- 208000006930 Pseudomyxoma Peritonei Diseases 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 229920000306 polymethylpentene Polymers 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 230000006798 recombination Effects 0.000 description 1
- 238000005215 recombination Methods 0.000 description 1
- 238000013077 scoring method Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/54—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for retrieval
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/165—Management of the audio stream, e.g. setting of volume, audio stream path
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/166—Detection; Localisation; Normalisation using acquisition arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
- G06V40/176—Dynamic expression
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0008—Associated control or indicating means
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/02—Feature extraction for speech recognition; Selection of recognition unit
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/1815—Semantic context, e.g. disambiguation of the recognition hypotheses based on word meaning
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0272—Voice signal separating
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0272—Voice signal separating
- G10L21/028—Voice signal separating using properties of sound source
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/031—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
- G10H2210/056—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction or identification of individual instrumental parts, e.g. melody, chords, bass; Identification or separation of instrumental parts by their characteristic voices or timbres
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/031—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
- G10H2210/066—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for pitch analysis as part of wider processing for musical purposes, e.g. transcription, musical performance evaluation; Pitch recognition, e.g. in polyphonic sounds; Estimation or use of missing fundamental
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/075—Musical metadata derived from musical analysis or for use in electrophonic musical instruments
- G10H2240/085—Mood, i.e. generation, detection or selection of a particular emotional content or atmosphere in a musical piece
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/121—Musical libraries, i.e. musical databases indexed by musical parameters, wavetables, indexing schemes using musical parameters, musical rule bases or knowledge bases, e.g. for automatic composing methods
- G10H2240/131—Library retrieval, i.e. searching a database or selecting a specific musical piece, segment, pattern, rule or parameter set
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
Definitions
- the present disclosure relates to the field of computer technology, and in particular, to an audio playback method, an apparatus, a computer-readable storage medium, and an electronic device.
- Embodiments of the present disclosure provide an audio playback method, apparatus, computer-readable storage medium, and electronic device.
- An embodiment of the present disclosure provides an audio playback method, the method includes: acquiring intent judgment data collected for at least one user in a target space; determining, based on the intent judgment data, a target vocalization intent of the at least one user; For the target vocalization intention, determine feature information representing the current feature of the at least one user; extract and play audio corresponding to the feature information from a preset audio library.
- an audio playback device the device includes: an acquisition module for acquiring intent judgment data collected for at least one user in a target space; a first determination module for based on The intent judgment data determines the target vocalization intent of the at least one user; the second determination module determines feature information representing the current feature of the at least one user based on the target vocalization intent; the first playback module uses It is used to extract and play the audio corresponding to the feature information from the preset audio library.
- a computer-readable storage medium stores a computer program, and the computer program is used to execute the above-mentioned audio playback method.
- an electronic device includes: a processor; a memory for storing instructions executable by the processor; a processor for reading the executable instructions from the memory, and Execute the instruction to implement the above audio playback method.
- the target utterance of at least one user is determined Intent, then determine the feature information according to the target vocalization intent, and finally extract the audio corresponding to the feature information from the preset audio library and play it, so that the electronic device can automatically determine the user's target vocalization intent, and when it is determined that the user has vocalization intent
- the audio playback is automatically performed by the electronic device, and the user does not need to actively trigger the audio playback operation, which reduces the steps for the user to perform the audio playback operation and improves the convenience of the audio playback operation.
- the played audio is adapted to the characteristics of the user, so that the audio that the user wants to listen to is more accurately played, and the pertinence of the automatically played audio is improved.
- FIG. 1 is a system diagram to which the present disclosure is applied.
- FIG. 2 is a schematic flowchart of an audio playback method provided by an exemplary embodiment of the present disclosure.
- FIG. 3 is a schematic flowchart of an audio playback method provided by another exemplary embodiment of the present disclosure.
- FIG. 4 is a schematic flowchart of an audio playback method provided by another exemplary embodiment of the present disclosure.
- FIG. 5 is a schematic flowchart of an audio playback method provided by another exemplary embodiment of the present disclosure.
- FIG. 6 is a schematic flowchart of an audio playback method provided by another exemplary embodiment of the present disclosure.
- FIG. 7 is a schematic flowchart of an audio playback method provided by another exemplary embodiment of the present disclosure.
- FIG. 8 is a schematic flowchart of an audio playback method provided by another exemplary embodiment of the present disclosure.
- FIG. 9 is a schematic structural diagram of an audio playback device provided by an exemplary embodiment of the present disclosure.
- FIG. 10 is a schematic structural diagram of an audio playback apparatus provided by another exemplary embodiment of the present disclosure.
- FIG. 11 is a structural diagram of an electronic device provided by an exemplary embodiment of the present disclosure.
- a plurality may refer to two or more, and “at least one” may refer to one, two or more.
- the term "and/or" in the present disclosure is only an association relationship to describe associated objects, indicating that there can be three kinds of relationships, for example, A and/or B, it can mean that A exists alone, and A and B exist at the same time , there are three cases of B alone.
- the character "/" in the present disclosure generally indicates that the related objects are an "or" relationship.
- Embodiments of the present disclosure can be applied to electronic devices such as terminal devices, computer systems, servers, etc., which can operate with numerous other general-purpose or special-purpose computing system environments or configurations.
- Examples of well-known terminal equipment, computing systems, environments and/or configurations suitable for use with terminal equipment, computer systems, servers, etc. electronic equipment include, but are not limited to: personal computer systems, server computer systems, thin clients, thick clients computer, handheld or laptop devices, microprocessor-based systems, set-top boxes, programmable consumer electronics, network personal computers, minicomputer systems, mainframe computer systems, and distributed cloud computing technology environments including any of the foregoing, among others.
- Electronic devices such as terminal devices, computer systems, servers, etc., may be described in the general context of computer system-executable instructions, such as program modules, being executed by the computer system.
- program modules may include routines, programs, object programs, components, logic, data structures, etc. that perform particular tasks or implement particular abstract data types.
- Computer systems/servers may be implemented in distributed cloud computing environments where tasks are performed by remote processing devices that are linked through a communications network.
- program modules may be located on local or remote computing system storage media including storage devices.
- the user In the current audio playback system, the user usually needs to manually select the audio to be played, or trigger the audio playback by means of voice recognition, gesture recognition, or the like. These methods often require the user to actively interact with the audio playback system, and cannot automatically determine the user's vocal intention. The convenience of audio playback is insufficient, and it is impossible to automatically play the corresponding audio according to the user's characteristics. The pertinence of audio playback Not enough either.
- FIG. 1 illustrates an exemplary system architecture 100 of an audio playback method or audio playback apparatus to which embodiments of the present disclosure may be applied.
- the system architecture 100 may include a terminal device 101 , a network 102 , a server 103 and an information collection device 104 .
- the network 102 is a medium for providing a communication link between the terminal device 101 and the server 103 .
- the network 102 may include various connection types, such as wired, wireless communication links, or fiber optic cables, among others.
- the user can use the terminal device 101 to interact with the server 103 through the network 102 to receive or send messages and the like.
- Various communication client applications such as audio players, video players, web browser applications, instant communication tools, etc., may be installed on the terminal device 101 .
- the terminal device 101 may be various electronic devices capable of audio playback, including but not limited to, such as vehicle-mounted terminals, mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablets), PMPs (portables) mobile terminals such as multimedia players), etc., as well as stationary terminals such as digital TVs, desktop computers, smart home appliances, and the like.
- the information collection device 104 may be various devices for collecting user-related information (including intention decision data), including but not limited to at least one of the following: a camera, a microphone, and the like.
- the terminal device 101 is set in a space 105 with a limited range, and the information collection device 104 is associated with the space 105 .
- the information collection device 104 may be installed in the space 105 to collect various information such as images and sounds of the user, or may be installed outside the space 105 to collect various information such as images and sounds around the space 105 .
- the space 105 may be a variety of confined spaces, such as the interior of a vehicle, the interior of a room, and the like.
- the server 103 may be a server that provides various services, such as a background audio server that provides support for the audio played on the terminal device 101 .
- the background audio server can process the received intent judgment data to obtain information such as the user's target vocalization intent, the user's feature information, and the audio to be played).
- the audio playback method provided by the embodiments of the present disclosure may be executed by the server 103 or by the terminal device 101 , and correspondingly, the audio playback device may be set in the server 103 or in the terminal device 101 middle.
- the audio playback method provided by the embodiments of the present disclosure may also be performed jointly by the terminal device 101 and the server 103.
- the steps of acquiring intent judgment data and determining the target vocalization intent are performed by the terminal device 101, and the steps of determining feature information and extracting audio Executed by the server 103, correspondingly, each module included in the audio playback apparatus may be set in the terminal device 101 and the server 103, respectively.
- terminal devices, networks and servers in FIG. 1 are merely illustrative. According to implementation needs, there can be any number of terminal devices, networks, servers and information collection devices.
- the above-mentioned system architecture may not include the network and the server, but only the terminal device and the information collection device.
- FIG. 2 is a schematic flowchart of an audio playback method provided by an exemplary embodiment of the present disclosure. This embodiment can be applied to an electronic device (the terminal device 101 or the server 103 shown in FIG. 1 ). As shown in FIG. 2 , the method includes the following steps:
- Step 201 Obtain intent judgment data collected for at least one user in the target space.
- the electronic device may acquire intent judgment data collected for at least one user in the target space.
- the target space eg, the space 105 in FIG. 1
- the intention determination data may be various data used to determine the user's intention, for example, including but not limited to at least one of the following: the user's face image data, the user's voice, and the like.
- Step 202 Determine, based on the intent judgment data, a target vocalization intent of at least one user.
- the electronic device may determine the target vocalization intention of at least one user based on the intention judgment data.
- the utterance type represented by the target utterance intention may be preset.
- the target vocalization intent may include, but is not limited to, at least one of the following: singing intent, recitation intent, and the like.
- the electronic device may select a corresponding method to determine the target vocalization intention according to the type of intention judgment data.
- the intent judgment data includes the user's face image data
- emotion recognition can be performed on the facial image to obtain the emotion type, and if the emotion type is joy, it can be determined that the above-mentioned at least one user has a target vocalization intention (eg, singing intention) ).
- the intent determination data includes a voice signal emitted by the user
- the voice signal can be identified, and if the identification result indicates that the user is humming, it can be determined that there is a target voice intent.
- Step 203 Based on the target vocalization intention, determine feature information representing the current feature of the at least one user.
- the electronic device may determine feature information that characterizes the current feature of at least one user.
- the current characteristics of the user may include, but are not limited to, at least one of the following: the user's emotion, the number of users, the user's listening habits, and the like.
- the electronic device may determine the feature information in a manner corresponding to the above-mentioned various features respectively. For example, a facial image captured by a camera of the user may be obtained, and emotion recognition may be performed on the facial image to obtain characteristic information representing the current emotion of the user.
- the user's historical playing records may be acquired, and the type of audio that the user is accustomed to listen to may be determined according to the historical playing records as the feature information.
- Step 204 Extract and play audio corresponding to the feature information from the preset audio library.
- the electronic device may extract and play audio corresponding to the feature information from a preset audio library.
- the preset audio library may be set in the above electronic device, or may be set in other electronic devices communicatively connected with the above electronic device.
- the above-mentioned characteristic information corresponds to the type of audio, and the electronic device can determine the type of audio to be played according to the characteristic information, and from the audio of this type, select (for example, select by playback volume, randomly select, etc.) audio to play.
- audio playback marked as joy type may be extracted from a preset audio library.
- the characteristic information indicates that the user is accustomed to listening to rock music
- the audio playback of the rock type may be extracted from the preset audio library.
- the intent judgment data is collected for at least one user in the target space
- the target vocalization intention of the user is determined according to the intent determination data
- the feature information is determined according to the target vocalization intent
- the preset vocalization intent is determined.
- the audio corresponding to the feature information is extracted from the audio library and played, so that the electronic device can actively determine the user's target vocalization intention without the need for the user to trigger the audio playback operation. Playing the audio reduces the steps for the user to perform the audio playing operation, and improves the convenience of the audio playing operation.
- the played audio is adapted to the characteristics of the user, so that the audio that the user wants to listen to is more accurately played, and the pertinence of the automatically played audio is improved.
- the target vocalization intention of the at least one user may be determined based on any of the following methods:
- Mode 1 in response to determining that the intent judgment data includes at least one user's face image, input the face image into a pre-trained third emotion recognition model to obtain emotion category information; if the emotion category information is preset emotion type information, determine at least one. A user has a target vocal intent.
- the third emotion recognition model may be obtained by training a preset initial model for training the third emotion recognition model using a preset training sample set in advance.
- the training samples in the training sample set may include sample face images and corresponding emotion category information.
- the electronic device can use the sample face image as the input of the initial model (for example, including a convolutional neural network, a classifier, etc.), and use the emotion category information corresponding to the input sample face image as the expected output of the initial model to train the initial model. , to obtain the above-mentioned third emotion recognition model.
- the preset emotions represented by the above-mentioned preset emotion type information may be various emotions such as excitement, joy, sadness, etc.
- the emotion type information output by the third emotion recognition model representing the user's emotion is the above-mentioned preset emotion, it is determined that at least one user There is a target vocal intent. For example, when the emotion type information indicates that the user's emotion is excited, it means that the user may want to sing to express his mood, and it is determined that the user has the intention to sing.
- Mode 2 In response to determining that the intent judgment data includes at least one user's voice information, voice recognition is performed on the voice information to obtain a voice recognition result; if the voice recognition result indicates that at least one user instructs to play audio, it is determined that at least one user has a target vocalization intention.
- the method for performing speech recognition on sound information is the prior art, and details are not described here.
- a voice of "this song is good, I want to sing" is recognized by a user, it is determined that the above-mentioned at least one user has a target vocalization intention (ie, singing intention).
- Manner 3 In response to determining that the intent judgment data includes voice information of at least one user, perform melody recognition on the voice information to obtain a melody recognition result; if the melody recognition result indicates that at least one user is uttering in a target form, it is determined that at least one user has a target. Voice intent.
- the vocalization of the above-mentioned target form corresponds to the target vocalization intention.
- vocalizations in the target form may include singing, reciting, humming, and the like.
- the method of performing melody recognition on sound information is in the prior art, and is usually performed according to the following steps: performing melody extraction on the human voice input to the melody recognition model through note segmentation and pitch extraction, and obtaining a note sequence through melody extraction.
- the electronic device further matches the note sequence output by the melody recognition model with the audio note sequence in the audio library.
- the similarity between the output note sequence and a certain audio note sequence is greater than a preset similarity threshold, it means that the user Singing (ie, vocalization in the form of a target) is being performed, and it is determined that the at least one user has an intention to vocalize the target.
- This implementation provides a variety of ways to determine the user's target vocalization intention, thereby realizing the comprehensive detection of the user's target vocalization intention through multi-modal methods such as emotion recognition, speech recognition, and melody recognition, and the detection accuracy is improved. High, the audio can be played to the user based on the target vocalization intention in the follow-up without the user's manual operation, thereby improving the convenience of the audio playback operation.
- the feature information may be determined in at least one of the following manners:
- Manner 1 Obtain historical audio playback records for at least one user; determine listening habit information of at least one user based on the historical audio playback records; and determine feature information based on the listening habit information.
- the electronic device may acquire historical audio playback records locally or remotely, and the listening habit information is used to characterize features such as the type of audio that the user often listens to, listening time, and the like. For example, the audio type with the most listening times may be determined as the listening habit information according to the historical audio playback records. Generally, the listening habit information may be included as the characteristic information.
- Method 2 Obtain a face image of at least one user, input the face image into a pre-trained fourth emotion recognition model, and obtain emotion category information representing the current emotion of at least one user; and determine feature information based on the emotion category information.
- the fourth emotion recognition model may be a neural network model used for emotion classification of facial images, which may be the same as or different from the third emotion recognition model described in the above-mentioned optional implementation manner, but the training method is the same as The method for training the third emotion recognition model is basically the same, and will not be repeated here.
- the emotion category information can be included as the information of the feature information.
- Manner 3 Obtain at least one environment image of the environment where the user is located, input the environment image into a pre-trained environment recognition model, and obtain environment type information; and determine feature information based on the environment type information.
- the environment image may be obtained by photographing an environment other than the target space by a camera.
- the environment recognition model may be a neural network model for classifying environmental images, and the electronic device may use a preset training sample set in advance to obtain the environment recognition model by training a preset initial model for training the environment recognition model.
- the training samples in the training sample set may include sample environment images and corresponding environment type information.
- the electronic device can use the sample environment image as the input of the initial model (for example, including a convolutional neural network, a classifier, etc.), and use the environment type information corresponding to the input sample environment image as the expected output of the initial model, train the initial model, and obtain The above environment recognition model.
- the environment type information is used to represent the type of the environment where the at least one user is located.
- the type of the environment is a location type such as a suburb, a highway, and a village, and it can also be a weather type such as a sunny day, a rainy day, and a snowy day.
- the environment type information can be included as the information of the characteristic information.
- an in-space image is obtained by photographing the target space; based on the in-space image, the number of people in the target space is determined; and feature information is determined based on the number of people.
- the in-space image may be an image captured by a camera set in the target space, the number of in-space images may be one or more, and the electronic device may determine which of the in-space images is based on an existing target detection method. People and counting people. Generally, the number of people can be included as the information of the characteristic information.
- this implementation can comprehensively detect the current state of the user, and obtain more comprehensive feature information, which can help to more targetedly extract the user's interest based on the feature information. to improve the accuracy of playing audio for users.
- step 204 may be performed as follows:
- the characteristic information includes listening habit information
- audio corresponding to the listening habit is extracted and played.
- audio corresponding to the emotion category information is extracted and played.
- audio corresponding to the environment type information is extracted and played.
- audio corresponding to the number of people is extracted and played.
- rock-type audio may be extracted and played. If the emotion category information indicates that the user's current emotion is happy, fast-paced audio can be extracted and played. If the environment type information indicates that the current environment of the user is in the wild, the audio of the soothing rhythm type can be extracted and played. If the determined number of users is 2 or more, chorus-type audio can be extracted and played.
- the feature information includes at least two of listening habit information, emotion category information, environment type information, and number of people
- the intersection of the audio contained in the audio types corresponding to the various information can be taken as the audio to be played.
- the extracted audio can be more attractive to the user, thereby improving the accuracy of playing the audio for the user.
- FIG. 3 a schematic flowchart of still another embodiment of an audio playing method is shown. As shown in FIG. 3 , on the basis of the above-mentioned embodiment shown in FIG. 2 , after step 204, the following steps may be further included:
- Step 205 extracting user audio information from the current mixed sound signal.
- the above-mentioned mixed sound signal may be a signal collected by the information collection device 104 (ie, a microphone) as shown in FIG. 1 , which is arranged in the above-mentioned target space.
- User audio information is the sound made by a user.
- the sound signal collected by the microphone includes a noise signal, or includes a sound signal sent by at least two users at the same time, and the sound signal collected at this time is a mixed sound signal. That is, the mixed sound signal may include the noise signal, or may include the sound information uttered by the user, or may include both the noise signal and the sound signal uttered by the user.
- existing speech separation methods for example, Blind Source Separation (BSS, Blind Source Separation) method, Auditory Scene Analysis (ASA, Auditory Scene Analysis) method, etc.
- BSS Blind Source Separation
- ASA Auditory Scene Analysis
- User audio information corresponding to the user respectively.
- Step 206 in the case that the user audio information meets the preset condition, play the user audio information.
- the electronic device may analyze the extracted user audio information, and if the user audio information satisfies a preset condition, play the user audio information. As an example, if the electronic device recognizes that the user audio information represents that the user is singing, the electronic device plays the user audio information whose volume is amplified through the speaker. Or, if the electronic device recognizes that the melody of the user's audio information representing the sound made by the user matches the currently playing audio, the electronic device plays the user's audio information.
- steps 205-206 are performed while the audio described in step 204 is playing.
- the played audio can be music.
- user audio information is extracted in real time from the mixed sound signal currently sent by at least one user. If the user audio information matches the played music, the user audio information is played, thereby realizing A scene where the user sings to the music.
- an existing feedback sound elimination method may also be used to filter out the sound signal collected by the microphone and played from the speaker, thereby reducing the interference of the feedback sound on the playback of user audio information.
- Fig. 3 corresponds to the method provided by the embodiment.
- the user audio information can be played simultaneously with the audio extracted from the preset audio library, and there is no need for the user to separately provide a special purpose for playing the user's audio information.
- the microphone for sound just use the microphone for collecting the mixed sound of each user in the target space to extract the sound made by the user from the mixed sound signal and play it simultaneously with the currently playing audio, which simplifies the process of playing user audio information.
- the required hardware improves the convenience for the user to achieve the target vocal intention.
- playing the user audio information that meets the preset conditions can avoid the interference on the playing of the user audio information caused by playing out the user chat and other content.
- step 205 further includes the following steps:
- Step 2051 Acquire initial audio information collected by an audio collection device set in the target space.
- the initial audio information may include a mixed sound signal.
- the audio collection device is a device included in the information collection device 104 shown in FIG. 1 .
- the number of audio collection devices may be one or more, and the number of channels of initial audio information is consistent with the number of audio collection devices, that is, each audio collection device collects a channel of initial audio information.
- the number of audio collection devices may match the number of seats in the vehicle. That is, an audio capture device is installed near each seat.
- Step 2052 Perform vocal separation on the initial audio information to obtain at least one channel of user audio information.
- At least one channel of user audio information corresponds to one user respectively.
- the electronic device can use the existing voice separation method to extract the user audio information corresponding to each user from the initial audio information.
- a blind source separation algorithm may be used to separate at least one channel of user audio information from the initial audio information.
- at least one channel of user audio information can be separated from the initial audio information collected by each audio collection device by using an existing microphone array-based speech separation algorithm.
- FIG. 4 corresponds to the method provided by the embodiment.
- the user audio information of multiple users can be collected in real time, and each channel of user audio information can be collected in real time.
- the audio information eliminates the sound interference of other users, so that the user audio information played subsequently can clearly reflect the voices of each user, and the quality of playing the voices of multiple users is improved.
- step 206 in the foregoing embodiment corresponding to FIG. 3 may be performed as follows:
- the volume of at least one channel of user audio information is adjusted to the target volume respectively, the user audio information after the volume adjustment is synthesized, and the synthesized user audio information is played.
- the target volume corresponding to each channel of user audio information may be the same or different.
- the volume of the user audio information with the highest volume can be used as the target volume, and the volumes of other user audio information can be adjusted to the target volume; or a fixed volume can be set as the target volume, and the user audio information of each channel can be set to the same volume target volume.
- each channel of user audio information can be combined into stereo playback, or combined into the same channel for playback.
- the volume of each user audio information being played can be made consistent or reach the respective set volume, so as to prevent the volume from being too low during playback due to the low volume emitted by the user.
- the above-mentioned step 206 may play user audio information based on at least one of the following methods:
- Manner 1 Perform melody recognition on the user audio information to obtain the user melody information; match the user melody information with the melody information of the currently playing audio, and play the user audio information based on the obtained first matching result.
- the method of performing melody recognition on user audio information is the prior art, which is usually carried out according to the following steps: performing melody extraction on the user audio information input to the melody recognition model through note segmentation and fundamental tone extraction, and obtaining a sequence of notes through melody extraction as the melody information.
- the electronic device further calculates the similarity between the melody information output by the melody recognition model and the melody information of the currently playing audio, if the similarity (that is, the first matching result) is greater than or equal to the preset first similarity threshold, it can be determined that the first If the matching result meets the preset conditions, the user audio information can be played.
- Method 2 Perform speech recognition on the user audio information to obtain a speech recognition result; match the speech recognition result with the corresponding text information of the currently playing audio, and play the user audio information based on the obtained second matching result.
- the speech recognition result may be text information. It should be noted that the method for performing speech recognition on user audio information is in the prior art, and details are not described herein again.
- the corresponding text information of the currently playing audio is the text information that has established a corresponding relationship with the audio in advance. For example, if the currently playing audio is a song, the corresponding text information may be lyrics; if the currently playing audio is poetry reading, the corresponding text The information is the original text of the poem read aloud.
- the electronic device may perform similarity calculation on the speech recognition result and the above-mentioned corresponding text information, and if the similarity (ie, the second matching result) is greater than or equal to a preset second similarity threshold, it may be determined that the second matching result meets the preset condition, User audio information can be played.
- the electronic device may execute any one of the above-mentioned manners 1 and 2 to play user audio information.
- the first and second manners above may also be performed simultaneously, and if it is determined based on the first matching result and the second matching result that the user audio information can be played in both manners, the user audio information is played.
- the first mode and/or the second mode may be performed for each channel of user audio information.
- the user audio information can be played when certain conditions are met, so as to avoid playing the user audio information irrelevant to the currently playing audio, and make the user audio information played.
- the information matches the currently playing audio to a higher degree, thereby improving the quality of playing the user's audio information.
- the above step 206 further includes:
- the pitch of the user's audio information is determined.
- the method for determining the pitch of the user audio information is in the prior art, and details are not described herein again.
- Step 1 Adjust the pitch of the currently playing audio to a target pitch that matches the pitch of the user audio information.
- the pitch of the currently playing audio may be compared with the pitch of the user's audio information, and if the difference between the two is outside the preset difference range, the pitch of the currently playing audio is adjusted to match the user's audio pitch.
- the pitch difference of the audio information is within a preset difference range.
- the user audio information is the audio information of the user singing
- the currently playing audio is song music
- the pitch of the user audio information is higher or lower than the pitch of the currently playing music
- the pitch of the music is adapted to the pitch of the user's singing, that is, the difficulty of singing along with the played music is adjusted, so that the user can better adapt to the played music.
- Step 2 output recommendation information for recommending audio corresponding to the pitch of the user audio information.
- the audio corresponding to the pitch of the user audio information may be the audio whose difference value from the pitch of the user audio information is within a preset difference value range.
- the recommended information can be output in the form of prompt sounds, displayed text, images, etc. After the recommended information is output, the user can choose whether to play the recommended audio, so that the pitch of the replayed audio matches the user's pitch.
- this implementation makes the pitch of the played audio automatically adapt to the user's pitch, so that the playback effect of the user's audio information is better. There is no need to adjust the pitch of the played audio through active means such as manual or voice control, which improves the convenience of adjusting the audio.
- FIG. 5 a schematic flowchart of still another embodiment of the audio playing method is shown. As shown in FIG. 5 , on the basis of the above-mentioned embodiment shown in FIG. 3 , after step 206 , the following steps may be further included:
- Step 207 Determine a target user corresponding to the user audio information from at least one user, and acquire a face image of the target user.
- the face image may be an image captured by a camera provided in the target space and included in the information acquisition device 104 in FIG. 1 .
- the electronic device when it extracts the user audio information from the mixed sound signal, it can determine the position of the sound source corresponding to the user audio information based on the existing voice separation method (for example, using an existing microphone array-based multi-sound area voice)
- the separation method determines which position in the target space the user audio information corresponds to), the position of the sound source is the user's position, and the user's position can be determined by the image captured by the user, and then the user's face image corresponding to the user's audio information can be obtained. .
- Step 208 Input the respective face images of the at least one user into the pre-trained first emotion recognition model to obtain emotion category information corresponding to the at least one user. That is to say, in this step, the facial image of the target user corresponding to the user audio information is input into the pre-trained first emotion recognition model, and correspondingly, the emotion category information corresponding to the target user is obtained.
- the first emotion recognition model may be the same as at least one of the third emotion recognition model and the fourth emotion recognition model described in the above-mentioned optional implementation manner, or may be different, but the training method is the same as that of the third emotion recognition model and the fourth emotion recognition model.
- the training method of at least one of the four emotion recognition models is basically the same, and will not be repeated here.
- Step 209 based on the emotion category information, determine a first score representing the degree of matching between the emotion of at least one user and the type of the currently playing audio. If the emotion category information in this step is the emotion category information corresponding to the target user, the determined first score is used to represent the matching degree between the emotion of the target user and the type of the currently playing audio.
- the first score may be obtained based on a probability value corresponding to the output emotion category information calculated by the first emotion recognition model.
- the first emotion recognition model can classify the input facial image, and obtain a plurality of emotion category information and probability values corresponding to each emotion category information respectively, and the emotion category information corresponding to the maximum probability value can be determined as the one identified this time. Emotion category information for face images.
- the first score may be determined according to the probability corresponding to this type of emotion category information. If the emotion category information of the face image recognized this time includes multiple types of emotion category information, the emotion category information that matches the type of the currently playing audio can be determined from the multiple emotion category information as the target emotion category information, and then based on the target emotion category information The corresponding probability determines the first score. The larger the value of the first score, the greater the degree of matching with the currently playing audio.
- the corresponding relationship between the type of the currently playing audio and the emotion category information may be preset. For example, if the type of the currently playing audio is marked as "cheerful", the first score may be obtained based on the probability corresponding to the emotion category information representing the cheerful emotion output by the model.
- Step 210 based on the first score, determine and output the score of the user audio information.
- the score of the user's audio information can be output in various ways, such as displaying on a display screen, outputting the sound of the score through a speaker, and the like.
- the first score may be determined as the score of the user audio information.
- step 209 may be performed as follows: based on the user audio information, determine a second score representing the degree of matching between the user audio information and the currently played audio, that is, in this step, the second score is determined based on the user audio information, The second score is used to represent the degree of matching between the user audio information and the currently playing audio.
- Step 210 may be performed as follows: based on the second score, determine and output the score of the user audio information.
- the second score may be determined by using an existing method for scoring user audio information. For example, when the user audio information indicates that the user is singing, the second score may be determined based on an existing singing scoring method. Further, the second score may be determined as the score of the user's audio information.
- step 210 may also be performed as follows: based on the first score and the second score, determine and output the score of the user audio information.
- the first score and the second score may be weighted and summed based on the preset weights corresponding to the first score and the second score respectively, to obtain the score of the user audio information.
- the method provided in the corresponding embodiment of FIG. 5 determines the score of the user audio information based on facial image recognition and/or audio scoring, so that the score can fully reflect the matching degree between the user audio information and the played audio, and improves the performance of the user audio information. Scoring accuracy.
- step 208 may be performed as follows:
- the emotion category information in the first emotion category information sequence respectively corresponds to a face image subsequence.
- the number of the user's face images is at least two, that is, the user's face image sequence is input to the first emotion recognition model.
- the face image sequence of a certain user may be the user's face image sequence.
- the sequence of emotion category information can be represented in the form of a vector, where each value in the vector corresponds to a subsequence of face images and represents a certain emotion category.
- Each facial image subsequence may include at least one facial image.
- the duration of the currently playing audio is 3 minutes, and the user's face is photographed for 3 minutes during the playback.
- the 3-minute face image sequence can be divided into 100 face image subsequences, and each subsequence is input into the first subsequence in turn.
- An emotion recognition model obtaining a vector including 100 values as an emotion category information sequence.
- Step 2091 Acquire the video corresponding to the currently playing audio, and extract the facial image sequence of the target person from the video.
- the target person may be a person related to the currently playing audio.
- the corresponding video may be a video including an image of the singer of the song
- the target person may be the singer of the song or a person performing with the song.
- the target person can be manually set in advance, or can be obtained by recognizing the video by electronic equipment. For example, based on the existing mouth motion recognition method, the person whose mouth motion frequency matches the rhythm of the song is identified as the target person.
- the electronic device can use an existing facial image detection method to extract the facial image sequence of the target person from the image frames included in the video according to the preset or recognized target person.
- Step 2092 Input the facial image sequence into the first emotion recognition model to obtain the second emotion category information sequence.
- This step is basically the same as the above step of determining the first emotion category information sequence, and will not be repeated here.
- Step 2093 Determine the similarity between the first emotion category information sequence and the second emotion category information sequence.
- the first emotion category information sequence and the second emotion category information sequence may both be in the form of vectors, and the electronic device may determine the distance between the vectors, and determine the similarity based on the distance (for example, the inverse of the distance is the similarity).
- Step 2094 based on the similarity, determine a first score.
- the similarity may be determined as the first score, or the similarity may be scaled according to a preset ratio to obtain the first score.
- this implementation can accurately determine the degree of consistency between the user's emotions and the emotions of the original video, and the obtained first score is more accurate It reflects the degree of consistency between the user's emotion and the currently playing audio, thereby improving the accuracy of scoring the user's audio information.
- FIG. 7 a schematic flowchart of still another embodiment of an audio playing method is shown. As shown in FIG. 7 , on the basis of the above-mentioned embodiment shown in FIG. 3 , after step 206 , the following steps may be further included:
- Step 211 Determine a target user corresponding to the user audio information from at least one user, and acquire a face image of the target user.
- This step is basically the same as the above-mentioned step 207, and will not be repeated here.
- Step 212 input the facial image of the target user and the user audio information corresponding to the user audio information into the pre-trained second emotion recognition model to obtain emotion category information.
- the second emotion recognition model in this step is different from the above-mentioned first emotion recognition model, third emotion recognition model, and fourth emotion recognition model.
- the second emotion recognition model can receive images and audios as input at the same time. Audio is jointly analyzed to output emotion category information.
- the second emotion recognition model can be obtained by training a preset initial model for training the second emotion recognition model using a preset training sample set in advance.
- the training samples in the training sample set may include sample face images, sample audio information, and corresponding emotion category information.
- the electronic device can use the sample face image and sample audio information as the input of the initial model (for example, including neural networks, classifiers, etc.), and use the emotion category information corresponding to the input sample face image and sample audio information as the expected output of the initial model.
- the initial model is trained to obtain the above-mentioned third emotion recognition model.
- the neural network included in the initial model can determine the feature information of the input sample face image and sample audio information
- the classifier can classify the feature information, compare the actual output information with the expected output, and adjust the parameters of the initial model to make The gap between the actual output and the expected output gradually decreases until convergence, so that the above-mentioned second emotion recognition model is obtained by training.
- Step 213 based on the emotion category information, determine and output a score representing the matching degree between the emotion of the target user corresponding to the user audio information and the type of the currently playing audio.
- the score may be obtained based on the probability value corresponding to the output emotion category information calculated by the second emotion recognition model.
- the method for determining the score based on the probability value is basically the same as the method for determining the first score in the foregoing step 209, and details are not repeated here.
- FIG. 7 corresponds to the method provided by the embodiment.
- step 212 may be performed as follows:
- the emotion category information in the third emotion category information sequence respectively corresponds to a face image subsequence.
- the definition of the third emotion category information sequence is basically the same as that of the above-mentioned first emotion category information, and will not be repeated here.
- step 213 can be performed as follows:
- Step 2131 Acquire the video corresponding to the currently playing audio, and extract the facial image sequence of the target person from the video.
- This step is basically the same as the above-mentioned step 2091, and will not be repeated here.
- Step 2132 Input the facial image sequence and the currently playing audio into the second emotion recognition model to obtain a fourth emotion category information sequence.
- This step is basically the same as the above step of determining the third emotion category information sequence, and will not be repeated here.
- Step 2133 Determine the similarity between the third emotional category information sequence and the fourth emotional category information sequence.
- the third emotional category information sequence and the fourth emotional category information sequence may both be in the form of vectors, and the electronic device may determine the distance between the vectors, and determine the similarity based on the distance (for example, the inverse of the distance is the similarity).
- Step 2134 based on the similarity, determine a score representing the degree of matching between the user's emotion corresponding to the user's audio information and the type of the currently playing audio.
- the similarity may be determined as a score, or the similarity may be scaled according to a preset ratio to obtain a score.
- the third emotional category information sequence and the fourth emotional category information sequence in this implementation are obtained based on the user's face image and user audio information, the images and audio are integrated during the emotional classification. Therefore, the two emotional categories The accuracy of the information sequence in representing emotions is higher. Therefore, the score determined by the similarity between the two emotion category information sequences can more accurately represent the degree of agreement between the user's emotion and the emotion of the original video, which further improves the user's audio quality. The accuracy of the information to be scored.
- FIG. 9 is a schematic structural diagram of an audio playback device provided by an exemplary embodiment of the present disclosure. This embodiment can be applied to electronic equipment.
- the audio playback apparatus includes: an acquisition module 901 for acquiring intent judgment data collected for at least one user in the target space; a first determination module 902 for Based on the intention judgment data, determine the target vocalization intention of at least one user; the second determination module 903 is used for determining the feature information representing the current feature of the at least one user based on the target vocalization intention; the first playing module 904 is used for Extract and play the audio corresponding to the feature information from the preset audio library.
- the obtaining module 901 may obtain the intent judgment data collected for at least one user in the target space.
- the target space eg, the space 105 in FIG. 1
- the intent determination data may be various pieces of information used to determine the user's intent, for example, including but not limited to at least one of the following: the user's face image, the user's voice, and the like.
- the first determination module 902 may determine the target vocalization intention of at least one user based on the intention judgment data.
- the utterance type represented by the target utterance intention may be preset.
- the target vocalization intent may include, but is not limited to, at least one of the following: singing intent, recitation intent, and the like.
- the first determining module 902 may select a corresponding manner to determine the target vocalization intent according to the type of intent determination data.
- the intent judgment data when the intent judgment data includes a face image of the user, emotion recognition can be performed on the facial image to obtain the emotion type, and if the emotion type is joy, it can be determined that the above-mentioned at least one user has a target vocalization intention (eg, singing intention) .
- the intent determination data includes a voice signal emitted by the user, the voice signal can be identified, and if the identification result indicates that the user is humming, it can be determined that there is a target voice intent.
- the second determining module 903 may determine feature information representing the current feature of at least one user.
- the current characteristics of the user may include, but are not limited to, at least one of the following, the user's mood, the number of users, the user's listening habits, and the like.
- the second determining module 903 may determine the feature information in a manner corresponding to the above-mentioned various features respectively. For example, a facial image captured by a camera of the user may be acquired, and emotion recognition may be performed on the facial image to obtain characteristic information representing the current emotion of the user. For another example, the user's historical playing records may be acquired, and the type of audio that the user is accustomed to listen to may be determined according to the historical playing records as the feature information.
- the first playing module 904 may extract and play audio corresponding to the feature information from a preset audio library.
- the preset audio library may be set in the above electronic device, or may be set in other electronic devices communicatively connected with the above electronic device.
- the above-mentioned feature information corresponds to the type of audio, and the first playback module 904 can determine the type of audio to be played according to the feature information, and select (for example, select by playback volume, randomly select, etc.) audio from the audio of this type. play.
- audio playback marked as joy type may be extracted from a preset audio library.
- the characteristic information indicates that the user is accustomed to listening to rock music
- the audio playback of the rock type may be extracted from the preset audio library.
- FIG. 10 is a schematic structural diagram of an audio playback apparatus provided by another exemplary embodiment of the present disclosure.
- the apparatus further includes: an extraction module 905, configured to extract user audio information from the current mixed sound signal; a second playback module 906, configured to extract user audio information when the user audio information meets preset conditions In this case, the user audio information is played.
- the apparatus further includes: a third determination module 907, configured to determine a target user corresponding to the user audio information from at least one user and acquire a face image of the target user; a first emotion recognition module 908, is used to input the facial image of the target user corresponding to the user audio information into the pre-trained first emotion recognition model to obtain the emotion category information corresponding to the target user respectively; the fourth determination module 909 is used to determine the representative user based on the emotion category information The first score of the matching degree between the emotion of the target user corresponding to the audio information and the type of the currently playing audio; and/or, the fifth determination module 910 is configured to determine, based on the user audio information, characterizing the user audio information and the currently playing audio The second score of the matching degree of ; the sixth determination module 911 is configured to determine and output the score of the user audio information based on the first score and/or the second score.
- a third determination module 907 configured to determine a target user corresponding to the user audio information from at least one user and acquire a face image of the target
- the first emotion recognition module 908 includes: a first emotion recognition unit 9081, configured to input the face image of at least one user into the first emotion recognition model to obtain the first emotion recognition model corresponding to the at least one user.
- an emotion category information sequence wherein the emotion category information in the first emotion category information sequence corresponds to a face image subsequence respectively
- the first determination unit 9082 is configured to determine, based on the emotion category information, the emotion representing at least one user and the
- the first score of the degree of matching of the type of the currently playing audio includes: a first acquiring unit 9083 for acquiring the video corresponding to the currently playing audio, and extracting the face image sequence of the target person from the video;
- the second emotion recognition The unit 9084 is used to input the facial image sequence into the first emotion recognition model to obtain the second emotion category information sequence;
- the second determination unit 9085 is used to determine the relationship between the first emotion category information sequence and the second emotion category information sequence. similarity;
- the third determining unit 9086 is configured to determine the first score based on
- the apparatus further includes: a seventh determination module 912, configured to determine a target user corresponding to the user audio information from at least one user and acquire a face image of the target user; a second emotion recognition module 913, For inputting the face image of the target user corresponding to the user audio information and the user audio information into the pre-trained second emotion recognition model to obtain emotion category information; the eighth determination module 914 is used to determine the representative user audio based on the emotion category information. The information corresponding to the target user's emotion and the currently playing audio type match the score and output.
- a seventh determination module 912 configured to determine a target user corresponding to the user audio information from at least one user and acquire a face image of the target user
- a second emotion recognition module 913 For inputting the face image of the target user corresponding to the user audio information and the user audio information into the pre-trained second emotion recognition model to obtain emotion category information
- the eighth determination module 914 is used to determine the representative user audio based on the emotion category information.
- the second emotion recognition module 913 is further configured to: input the user's face image and user audio information corresponding to the user audio information into the second emotion recognition model to obtain a third emotion category information sequence, wherein , the emotional category information in the third emotional category information sequence corresponds to a sub-sequence of facial images respectively;
- the eighth determination module 914 includes: a second acquisition unit 9141 for acquiring the video corresponding to the currently playing audio, and from the video Extract the facial image sequence of the target person;
- the third emotion recognition unit 9142 is used to input the facial image sequence and the currently played audio into the second emotion recognition model to obtain the fourth emotion category information sequence;
- the fourth determination unit 9143 is used for is used to determine the similarity between the third emotional category information sequence and the fourth emotional category information sequence;
- the fifth determination unit 9144 is used to determine, based on the similarity, the user's emotion corresponding to the user's audio information and the type of the currently playing audio match score.
- the extraction module 905 includes: a third acquisition unit 9051, configured to acquire initial audio information collected by an audio acquisition device set in the target space, where the initial audio information includes a mixed sound signal; a separation unit 9052, It is used to separate the human voice from the initial audio information to obtain at least one channel of user audio information, wherein at least one channel of user audio information corresponds to one user respectively.
- a third acquisition unit 9051 configured to acquire initial audio information collected by an audio acquisition device set in the target space, where the initial audio information includes a mixed sound signal
- a separation unit 9052 It is used to separate the human voice from the initial audio information to obtain at least one channel of user audio information, wherein at least one channel of user audio information corresponds to one user respectively.
- the second playing module 906 is further configured to: adjust the volume of at least one channel of user audio information to the target volume respectively, synthesize the adjusted user audio information, and play the synthesized user audio information.
- the second playing module 906 includes: a first melody identification unit 9061, configured to perform melody identification on the user audio information to obtain the user melody information; compare the user melody information with the melody information of the currently playing audio Carry out matching, and play user audio information based on the obtained first matching result; and/or, the first voice recognition unit 9062 is used to perform voice recognition on user audio information to obtain a voice recognition result; the voice recognition result and the currently played audio The corresponding text information is matched, and the user audio information is played based on the obtained second matching result.
- a first melody identification unit 9061 configured to perform melody identification on the user audio information to obtain the user melody information
- compare the user melody information with the melody information of the currently playing audio Carry out matching and play user audio information based on the obtained first matching result
- the first voice recognition unit 9062 is used to perform voice recognition on user audio information to obtain a voice recognition result; the voice recognition result and the currently played audio The corresponding text information is matched, and the user audio information is played based on the obtained second matching result.
- the second playing module 906 includes: a sixth determining unit 9063, configured to determine the pitch of the user audio information; and an adjusting unit 9064, configured to adjust the pitch of the currently playing audio to be consistent with the user's a target pitch that matches the pitch of the audio information; and/or an output unit 9065, configured to output recommendation information for recommending audio corresponding to the pitch of the user's audio information.
- the first determination module 902 includes: a fourth emotion recognition unit 9021, configured to input the face image into a pre-trained third user in response to determining that the intent judgment data includes a face image of at least one user The emotion recognition model, to obtain emotion category information; if the emotion category information is preset emotion type information, it is determined that at least one user has a target vocalization intention; or, the second speech recognition unit 9022 is used to determine that the data includes at least one user in response to the determination of the intention.
- a fourth emotion recognition unit 9021 configured to input the face image into a pre-trained third user in response to determining that the intent judgment data includes a face image of at least one user The emotion recognition model, to obtain emotion category information; if the emotion category information is preset emotion type information, it is determined that at least one user has a target vocalization intention; or, the second speech recognition unit 9022 is used to determine that the data includes at least one user in response to the determination of the intention.
- the second melody recognition unit 9023 is used to respond to the Determine that the intent judgment data includes at least one user's voice information, perform melody recognition on the voice information, and obtain a melody recognition result; if the melody recognition result indicates that at least one user is vocalizing in a target form, it is determined that at least one user has a target vocalization intention.
- the second determination module 903 includes: a seventh determination unit 9031, configured to acquire historical audio playback records for at least one user; and determine listening habit information of at least one user based on the historical audio playback records; Based on the listening habit information, determine characteristic information; and/or, a fifth emotion recognition unit 9032, configured to acquire a facial image of at least one user, input the facial image into a pre-trained fourth emotion recognition model, and obtain a representation of the at least one user
- the emotion category information of the current emotion; based on the emotion category information, determine the feature information; and/or, the environment recognition unit 9033 is used to obtain the environment image of the environment where at least one user is located, and input the environment image into the pre-trained environment recognition model , obtain environment type information; based on the environment type information, determine the feature information; and/or, the eighth determination unit 9034 is used to obtain the in-space image obtained by photographing the target space; based on the in-space image, determine the number of people in the target space; The number of people to
- the first playing module 904 includes: a first playing unit 9041, configured to extract and play audio corresponding to the listening habit in response to determining that the feature information includes listening habit information; a second playing unit 9042, In response to determining that the feature information includes emotion category information, extract and play the audio corresponding to the emotion category information; the third playing unit 9043 is used to extract and play the audio corresponding to the environment type information in response to determining that the feature information includes environment type information. Audio; the fourth playing unit 9044 is configured to extract and play audio corresponding to the number of people in response to determining that the characteristic information includes the number of people.
- the audio playback device collects intent judgment data from at least one user in the target space, determines the target vocalization intention of the user according to the intent determination data, determines feature information according to the target vocalization intent, and finally selects the desired vocalization intent from the preset
- the audio corresponding to the feature information is extracted from the audio library and played, so that the electronic device can automatically determine the user's target vocalization intention, and when it is determined that the user has the vocalization intent, the electronic device can automatically play the audio without the need for the user.
- the operation of actively triggering the audio playback reduces the steps for the user to perform the audio playback operation and improves the convenience of the audio playback operation.
- the played audio is adapted to the characteristics of the user, so that the audio that the user wants to listen to is more accurately played, and the pertinence of the automatically played audio is improved.
- the electronic device may be any one or both of the terminal device 101 and the server 103 as shown in FIG. 1 , or a stand-alone device independent of them, the stand-alone device may communicate with the terminal device 101 and the server 103 to obtain data from them Receive the collected input signal.
- FIG. 11 illustrates a block diagram of an electronic device according to an embodiment of the present disclosure.
- the electronic device 1100 includes one or more processors 1101 and a memory 1102 .
- the processor 1101 may be a central processing unit (Central Processing Unit, CPU) or other forms of processing units with data processing capabilities and/or instruction execution capabilities, and may control other components in the electronic device 1100 to perform desired functions.
- CPU Central Processing Unit
- Memory 1102 may include one or more computer program products, which may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory.
- Volatile memory may include, for example, random access memory (Random Access Memory, RAM) and/or cache memory (cache).
- the non-volatile memory may include, for example, a read-only memory (Read-Only Memory, ROM), a hard disk, a flash memory, and the like.
- One or more computer program instructions may be stored on the computer-readable storage medium, and the processor 1101 may execute the program instructions to implement the audio playback method and/or other desired functions of the various embodiments of the present disclosure above.
- Various contents such as intent decision data, feature information, audio, and the like may also be stored in the computer-readable storage medium.
- the electronic device 1100 may also include an input device 1103 and an output device 1104 interconnected by a bus system and/or other form of connection mechanism (not shown).
- the input device 1103 may be a device such as a camera, a microphone, and the like, for inputting intention judgment data.
- the input device 1103 may be a communication network connector for receiving the input intention judgment data from the terminal device 101 and the server 103 .
- the output device 1104 can output various information, including extracted audio, to the outside.
- the output devices 1104 may include, for example, displays, speakers, and communication networks and their connected remote output devices, among others.
- the electronic device 1100 may also include any other appropriate components according to the specific application.
- embodiments of the present disclosure may also be computer program products comprising computer program instructions that, when executed by a processor, cause the processor to perform the "exemplary method" described above in this specification
- the steps in the audio playback method according to various embodiments of the present disclosure are described in the section.
- the computer program product may write program code for performing operations of embodiments of the present disclosure in any combination of one or more programming languages, including object-oriented programming languages, such as Java, C++, etc. , also includes conventional procedural programming languages, such as "C" language or similar programming languages.
- the program code may execute entirely on the user computing device, partly on the user device, as a stand-alone software package, partly on the user computing device and partly on a remote computing device, or entirely on the remote computing device or server execute on.
- embodiments of the present disclosure may also be computer-readable storage media having computer program instructions stored thereon that, when executed by a processor, cause the processor to perform the above-described "Example Method" section of this specification Steps in an audio playback method according to various embodiments of the present disclosure described in .
- the computer-readable storage medium may employ any combination of one or more readable media.
- the readable medium may be a readable signal medium or a readable storage medium.
- the readable storage medium may include, for example, but not limited to, electrical, magnetic, optical, electromagnetic, infrared, or semiconductor systems, apparatuses or devices, or a combination of any of the above.
- readable storage media include: electrical connections with one or more wires, portable disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read-only memory (Erasable Programmable Read-Only Memory, EPROM) or flash memory), optical fiber, portable compact disk read-only memory (Compact Disc Read-Only Memory, CD-ROM), optical storage devices, magnetic storage devices, Or any suitable combination of the above.
- the methods and apparatus of the present disclosure may be implemented in many ways.
- the methods and apparatus of the present disclosure may be implemented in software, hardware, firmware, or any combination of software, hardware, and firmware.
- the above-described order of steps for the method is for illustration only, and the steps of the method of the present disclosure are not limited to the order specifically described above unless specifically stated otherwise.
- the present disclosure can also be implemented as programs recorded in a recording medium, the programs including machine-readable instructions for implementing methods according to the present disclosure.
- the present disclosure also covers a recording medium storing a program for executing the method according to the present disclosure.
- each component or each step may be decomposed and/or recombined. These disaggregations and/or recombinations should be considered equivalents of the present disclosure.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Theoretical Computer Science (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Acoustics & Sound (AREA)
- General Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Signal Processing (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Quality & Reliability (AREA)
- Artificial Intelligence (AREA)
- Psychiatry (AREA)
- Hospice & Palliative Care (AREA)
- Child & Adolescent Psychology (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- User Interface Of Digital Computer (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (12)
- 一种音频播放方法,包括:获取针对目标空间内的至少一个用户采集的意图判决数据;基于所述意图判决数据,确定所述至少一个用户具有的目标发声意图;基于所述目标发声意图,确定表征所述至少一个用户的当前特征的特征信息;从预设音频库中提取并播放与所述特征信息对应的音频。
- 根据权利要求1所述的方法,其中,在所述提取并播放与所述特征信息对应的音频之后,所述方法还包括:从当前的混合声音信号中提取用户音频信息;在所述用户音频信息符合预设条件的情况下,播放所述用户音频信息。
- 根据权利要求2所述的方法,其中,在所述播放所述用户音频信息之后,所述方法还包括:从所述至少一个用户中确定所述用户音频信息对应的目标用户并获取所述目标用户的脸部图像;将所述用户音频信息对应的目标用户的脸部图像输入预先训练的第一情绪识别模型,得到所述目标用户对应的情绪类别信息;基于所述情绪类别信息,确定表征所述用户音频信息对应的目标用户的情绪与当前播放的音频的类型的匹配程度的第一评分;和/或,基于所述用户音频信息,确定表征所述用户音频信息与所述当前播放的音频的匹配程度的第二评分;基于所述第一评分和/或所述第二评分,确定所述用户音频信息的评分并输出。
- 根据权利要求2所述的方法,其中,在所述播放所述用户音频信息之后,所述方法还包括:从所述至少一个用户中确定所述用户音频信息对应的目标用户并获取所 述目标用户的脸部图像;将所述用户音频信息对应的目标用户的脸部图像和所述用户音频信息输入预先训练的第二情绪识别模型,得到情绪类别信息;基于所述情绪类别信息,确定表征所述用户音频信息对应的目标用户的情绪与当前播放的音频的类型的匹配程度的评分并输出。
- 根据权利要求4所述的方法,其中,所述将所述用户音频信息对应的目标用户的脸部图像和所述用户音频信息输入预先训练的第二情绪识别模型,得到情绪类别信息,包括:将所述用户音频信息对应的目标用户的脸部图像和所述用户音频信息输入所述第二情绪识别模型,得到第三情绪类别信息序列,其中,所述第三情绪类别信息序列中的情绪类别信息分别对应于一个脸部图像子序列;所述基于所述情绪类别信息,确定表征所述用户音频信息对应的目标用户的情绪与当前播放的音频的类型的匹配程度的评分,包括:获取所述当前播放的音频对应的视频,并从所述视频中提取目标人物的脸部图像序列;将所述脸部图像序列和所述当前播放的音频输入所述第二情绪识别模型,得到第四情绪类别信息序列;确定所述第三情绪类别信息序列和所述第四情绪类别信息序列之间的相似度;基于所述相似度,确定表征所述用户音频信息对应的用户的情绪与当前播放的音频的类型的匹配程度的评分。
- 根据权利要求2所述的方法,其中,所述从当前的混合声音信号中提取用户音频信息,包括:获取设置在所述目标空间的音频采集设备采集的初始音频信息,所述初始音频信息包括所述混合声音信号;对所述初始音频信息进行人声分离,得到至少一路用户音频信息,其中,所述至少一路用户音频信息分别对应于一个用户。
- 根据权利要求2所述的方法,其中,所述基于所述用户音频信息,播 放所述用户音频信息,包括:对所述用户音频信息进行旋律识别,得到用户旋律信息;将所述用户旋律信息与当前播放的音频的旋律信息进行匹配,基于得到的第一匹配结果播放所述用户音频信息;和/或,对所述用户音频信息进行语音识别,得到语音识别结果;将所述语音识别结果与当前播放的音频的对应文本信息进行匹配,基于得到的第二匹配结果播放所述用户音频信息。
- 根据权利要求1所述的方法,其中,所述基于所述意图判决数据,确定所述至少一个用户具有的目标发声意图,包括:响应于确定所述意图判决数据包括所述至少一个用户的脸部图像,将所述脸部图像输入预先训练的第三情绪识别模型,得到情绪类别信息;如果所述情绪类别信息为预设情绪类型信息,确定所述至少一个用户有目标发声意图;或者,响应于确定所述意图判决数据包括所述至少一个用户的声音信息,对所述声音信息进行语音识别,得到语音识别结果;如果所述语音识别结果表征所述至少一个用户指示播放音频,确定所述至少一个用户有目标发声意图;或者响应于确定所述意图判决数据包括所述至少一个用户的声音信息,对所述声音信息进行旋律识别,得到旋律识别结果;如果所述旋律识别结果表征所述至少一个用户正在进行目标形式的发声,确定所述至少一个用户有目标发声意图。
- 根据权利要求1所述的方法,其中,所述确定表征所述至少一个用户的当前特征的特征信息,包括:获取针对所述至少一个用户的历史音频播放记录;基于所述历史音频播放记录,确定所述至少一个用户的收听习惯信息;基于所述收听习惯信息,确定所述特征信息;和/或,获取所述至少一个用户的脸部图像,将所述脸部图像输入预先训练的第四情绪识别模型,得到表征所述至少一个用户当前的情绪的情绪类别信息;基于所述情绪类别信息,确定所述特征信息;和/或,获取所述至少一个用户所处的环境的环境图像,将所述环境图像输入预先训练的环境识别模型,得到环境类型信息;基于所述环境类型信息,确定所述特征信息;和/或,获取对所述目标空间拍摄得到空间内图像;基于所述空间内图像,确定所述目标空间内的人数;基于所述人数,确定所述特征信息。
- 一种音频播放装置,包括:获取模块,用于获取针对目标空间内的至少一个用户采集的意图判决数据;第一确定模块,用于基于所述意图判决数据,确定所述至少一个用户具有的目标发声意图;第二确定模块,用于确定表征所述至少一个用户的当前特征的特征信息;第一播放模块,用于从预设音频库中提取并播放与所述特征信息对应的音频。
- 一种计算机可读存储介质,所述存储介质存储有计算机程序,所述计算机程序用于执行上述权利要求1-9任一所述的方法。
- 一种电子设备,所述电子设备包括:处理器;用于存储所述处理器可执行指令的存储器;所述处理器,用于从所述存储器中读取所述可执行指令,并执行所述指令以实现上述权利要求1-9任一所述的方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2022573581A JP7453712B2 (ja) | 2021-04-16 | 2022-02-14 | オーディオ再生方法、装置、コンピュータ可読記憶媒体及び電子機器 |
US18/247,754 US20240004606A1 (en) | 2021-04-16 | 2022-02-14 | Audio playback method and apparatus, computer readable storage medium, and electronic device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110410353.9 | 2021-04-16 | ||
CN202110410353.9A CN113126951B (zh) | 2021-04-16 | 2021-04-16 | 音频播放方法、装置、计算机可读存储介质及电子设备 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022218027A1 true WO2022218027A1 (zh) | 2022-10-20 |
Family
ID=76777173
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2022/076239 WO2022218027A1 (zh) | 2021-04-16 | 2022-02-14 | 音频播放方法、装置、计算机可读存储介质及电子设备 |
Country Status (4)
Country | Link |
---|---|
US (1) | US20240004606A1 (zh) |
JP (1) | JP7453712B2 (zh) |
CN (1) | CN113126951B (zh) |
WO (1) | WO2022218027A1 (zh) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113126951B (zh) * | 2021-04-16 | 2024-05-17 | 深圳地平线机器人科技有限公司 | 音频播放方法、装置、计算机可读存储介质及电子设备 |
CN114120939B (zh) * | 2021-11-26 | 2024-10-11 | 合肥若叶无间网络科技有限公司 | 一种古琴调音器的实现方法 |
CN114999534A (zh) * | 2022-06-10 | 2022-09-02 | 中国第一汽车股份有限公司 | 一种车载音乐的播放控制方法、装置、设备和存储介质 |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107632814A (zh) * | 2017-09-25 | 2018-01-26 | 珠海格力电器股份有限公司 | 音频信息的播放方法、装置和系统、存储介质、处理器 |
US10102851B1 (en) * | 2013-08-28 | 2018-10-16 | Amazon Technologies, Inc. | Incremental utterance processing and semantic stability determination |
CN110111795A (zh) * | 2019-04-23 | 2019-08-09 | 维沃移动通信有限公司 | 一种语音处理方法及终端设备 |
CN110413250A (zh) * | 2019-06-14 | 2019-11-05 | 华为技术有限公司 | 一种语音交互方法、装置及系统 |
CN111199732A (zh) * | 2018-11-16 | 2020-05-26 | 深圳Tcl新技术有限公司 | 一种基于情感的语音交互方法、存储介质及终端设备 |
CN111523981A (zh) * | 2020-04-29 | 2020-08-11 | 深圳追一科技有限公司 | 虚拟试用方法、装置、电子设备及存储介质 |
CN112397065A (zh) * | 2020-11-04 | 2021-02-23 | 深圳地平线机器人科技有限公司 | 语音交互方法、装置、计算机可读存储介质及电子设备 |
CN113126951A (zh) * | 2021-04-16 | 2021-07-16 | 深圳地平线机器人科技有限公司 | 音频播放方法、装置、计算机可读存储介质及电子设备 |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000099014A (ja) * | 1998-09-22 | 2000-04-07 | Ntt Data Corp | 音楽自動採点装置、音楽自動採点モデル作成装置、音楽自動採点モデル作成方法、音楽自動採点方法、及び、記録媒体 |
JP2000330576A (ja) * | 1999-05-19 | 2000-11-30 | Taito Corp | カラオケの歌唱評価方法と装置 |
JP4037081B2 (ja) | 2001-10-19 | 2008-01-23 | パイオニア株式会社 | 情報選択装置及び方法、情報選択再生装置並びに情報選択のためのコンピュータプログラム |
JP2004163590A (ja) | 2002-11-12 | 2004-06-10 | Denso Corp | 再生装置及びプログラム |
JP4916005B2 (ja) | 2007-02-28 | 2012-04-11 | 株式会社第一興商 | カラオケシステム |
US8583615B2 (en) * | 2007-08-31 | 2013-11-12 | Yahoo! Inc. | System and method for generating a playlist from a mood gradient |
CN102970427A (zh) * | 2012-11-16 | 2013-03-13 | 广东欧珀移动通信有限公司 | 一种手机播放歌曲的方法 |
US10373611B2 (en) * | 2014-01-03 | 2019-08-06 | Gracenote, Inc. | Modification of electronic system operation based on acoustic ambience classification |
JP6409652B2 (ja) | 2015-03-30 | 2018-10-24 | ブラザー工業株式会社 | カラオケ装置、プログラム |
JP6866715B2 (ja) * | 2017-03-22 | 2021-04-28 | カシオ計算機株式会社 | 情報処理装置、感情認識方法、及び、プログラム |
CN107609034A (zh) * | 2017-08-09 | 2018-01-19 | 深圳市汉普电子技术开发有限公司 | 一种智能音箱的音频播放方法、音频播放装置及存储介质 |
WO2019114426A1 (zh) * | 2017-12-15 | 2019-06-20 | 蔚来汽车有限公司 | 车载音乐的匹配方法、装置及车载智能控制器 |
JP6944391B2 (ja) | 2018-01-31 | 2021-10-06 | 株式会社第一興商 | カラオケ装置 |
CN108848416A (zh) * | 2018-06-21 | 2018-11-20 | 北京密境和风科技有限公司 | 音视频内容的评价方法和装置 |
CN109299318A (zh) * | 2018-11-13 | 2019-02-01 | 百度在线网络技术(北京)有限公司 | 音乐推荐的方法、装置、存储介质和终端设备 |
CN111754965B (zh) * | 2019-03-29 | 2023-11-14 | 比亚迪股份有限公司 | 车载k歌装置、方法和车辆 |
CN110096611A (zh) * | 2019-04-24 | 2019-08-06 | 努比亚技术有限公司 | 一种歌曲推荐方法、移动终端及计算机可读存储介质 |
CN110197677A (zh) * | 2019-05-16 | 2019-09-03 | 北京小米移动软件有限公司 | 一种播放控制方法、装置及播放设备 |
CN111968611B (zh) * | 2020-08-12 | 2024-04-23 | 上海仙塔智能科技有限公司 | K歌方法、车载终端及计算机可读存储介质 |
-
2021
- 2021-04-16 CN CN202110410353.9A patent/CN113126951B/zh active Active
-
2022
- 2022-02-14 WO PCT/CN2022/076239 patent/WO2022218027A1/zh active Application Filing
- 2022-02-14 JP JP2022573581A patent/JP7453712B2/ja active Active
- 2022-02-14 US US18/247,754 patent/US20240004606A1/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10102851B1 (en) * | 2013-08-28 | 2018-10-16 | Amazon Technologies, Inc. | Incremental utterance processing and semantic stability determination |
CN107632814A (zh) * | 2017-09-25 | 2018-01-26 | 珠海格力电器股份有限公司 | 音频信息的播放方法、装置和系统、存储介质、处理器 |
CN111199732A (zh) * | 2018-11-16 | 2020-05-26 | 深圳Tcl新技术有限公司 | 一种基于情感的语音交互方法、存储介质及终端设备 |
CN110111795A (zh) * | 2019-04-23 | 2019-08-09 | 维沃移动通信有限公司 | 一种语音处理方法及终端设备 |
CN110413250A (zh) * | 2019-06-14 | 2019-11-05 | 华为技术有限公司 | 一种语音交互方法、装置及系统 |
CN111523981A (zh) * | 2020-04-29 | 2020-08-11 | 深圳追一科技有限公司 | 虚拟试用方法、装置、电子设备及存储介质 |
CN112397065A (zh) * | 2020-11-04 | 2021-02-23 | 深圳地平线机器人科技有限公司 | 语音交互方法、装置、计算机可读存储介质及电子设备 |
CN113126951A (zh) * | 2021-04-16 | 2021-07-16 | 深圳地平线机器人科技有限公司 | 音频播放方法、装置、计算机可读存储介质及电子设备 |
Also Published As
Publication number | Publication date |
---|---|
US20240004606A1 (en) | 2024-01-04 |
CN113126951B (zh) | 2024-05-17 |
CN113126951A (zh) | 2021-07-16 |
JP2023527473A (ja) | 2023-06-28 |
JP7453712B2 (ja) | 2024-03-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11842730B2 (en) | Modification of electronic system operation based on acoustic ambience classification | |
WO2022218027A1 (zh) | 音频播放方法、装置、计算机可读存储介质及电子设备 | |
Czyzewski et al. | An audio-visual corpus for multimodal automatic speech recognition | |
US10490195B1 (en) | Using system command utterances to generate a speaker profile | |
US11854550B2 (en) | Determining input for speech processing engine | |
CN109074806B (zh) | 控制分布式音频输出以实现语音输出 | |
CN106373569B (zh) | 语音交互装置和方法 | |
JP7126613B2 (ja) | ドメイン分類器を使用したニューラルネットワークにおけるドメイン適応のためのシステム及び方法 | |
Eronen et al. | Audio-based context recognition | |
JP7108144B2 (ja) | クロスドメインバッチ正規化を使用したニューラルネットワークにおけるドメイン適応のためのシステム及び方法 | |
JP2020034895A (ja) | 応答方法及び装置 | |
US12119022B2 (en) | Cognitive assistant for real-time emotion detection from human speech | |
KR102495888B1 (ko) | 사운드를 출력하기 위한 전자 장치 및 그의 동작 방법 | |
CN114121006A (zh) | 虚拟角色的形象输出方法、装置、设备以及存储介质 | |
US20240321264A1 (en) | Automatic speech recognition | |
Geiger et al. | Learning new acoustic events in an hmm-based system using map adaptation | |
CN110232911B (zh) | 跟唱识别方法、装置、存储介质及电子设备 | |
CN111429882A (zh) | 播放语音的方法、装置及电子设备 | |
CN111627417B (zh) | 播放语音的方法、装置及电子设备 | |
US9412395B1 (en) | Narrator selection by comparison to preferred recording features | |
US20210082427A1 (en) | Information processing apparatus and information processing method | |
US20240203446A1 (en) | Method of operating sound recognition device identifying speaker and electronic device having the same | |
CN114360523A (zh) | 关键词数据集获取、模型训练方法、装置、设备及介质 | |
JP2020086011A (ja) | 抽出装置、学習装置、抽出方法、抽出プログラム、学習方法および学習プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2022573581 Country of ref document: JP Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22787257 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 18247754 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 22.02.2024) |