CN112185344A - Voice interaction method and device, computer readable storage medium and processor - Google Patents

Voice interaction method and device, computer readable storage medium and processor Download PDF

Info

Publication number
CN112185344A
CN112185344A CN202011034411.4A CN202011034411A CN112185344A CN 112185344 A CN112185344 A CN 112185344A CN 202011034411 A CN202011034411 A CN 202011034411A CN 112185344 A CN112185344 A CN 112185344A
Authority
CN
China
Prior art keywords
voice
voiceprint
determining
tone
preset
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011034411.4A
Other languages
Chinese (zh)
Inventor
焦金珂
李健
武卫东
陈明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Sinovoice Technology Co Ltd
Original Assignee
Beijing Sinovoice Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Sinovoice Technology Co Ltd filed Critical Beijing Sinovoice Technology Co Ltd
Priority to CN202011034411.4A priority Critical patent/CN112185344A/en
Publication of CN112185344A publication Critical patent/CN112185344A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • G10L13/033Voice editing, e.g. manipulating the voice of the synthesiser
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • G10L13/04Details of speech synthesis systems, e.g. synthesiser structure or memory management
    • G10L13/047Architecture of speech synthesisers
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • G10L17/22Interactive procedures; Man-machine interfaces

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The application provides a voice interaction method, a device, a computer readable storage medium and a processor, wherein the voice interaction method comprises the steps of obtaining voice data of a speaker; determining a preset voice tone to be broadcasted according to the voice data; and broadcasting by adopting a preset voice tone. The method determines the preset voice tone of the broadcast according to the acquired voice data and broadcasts by adopting the preset voice tone, can realize intelligent recommendation and switching of the broadcast tone when voice interaction is carried out with different users, realizes broadcast by adopting different voice tones for different speakers, does not depend on division and definition of voice styles such as active type, steady type, humorous type, lovely type or more true type, can accurately predict and judge the preset voice tone corresponding to the speaker, meets the requirement of recommending different tones for different people, and remarkably improves the interestingness and personalized experience of the users in the using process.

Description

Voice interaction method and device, computer readable storage medium and processor
Technical Field
The present application relates to the field of voice interaction, and in particular, to a voice interaction method, apparatus, computer-readable storage medium, processor, and voice interaction system.
Background
The current voice interaction technology is widely applied to the fields of intelligent robots, intelligent sound boxes, intelligent vehicle-mounted equipment, intelligent homes and the like, and people can control equipment or a system to execute commands or complete question and answer conversations through voice conversations. However, when the device performs voice interaction, it is usually monotonous to perform synthesized broadcast by using a tone preset by a system.
In order to increase interest and individuation of interaction, a plurality of tone libraries are set by individual equipment, and a user needs to manually switch configuration in system setting, but the problems that tone is automatically switched and different tone libraries are recommended to different people during real-time voice interaction cannot be met.
The current voice interaction intelligent recommendation generally focuses on content recommendation, i.e., personalized content is recommended for different people, such as music, stories or questions and answers. In the current voice interaction system, a user and equipment have a conversation to obtain a fixed tone broadcast preset by the system, namely, in the face of different users, a machine can be broadcasted by synthesizing the same tone, different broadcast tones cannot be intelligently recommended for different users, and more personalized services cannot be provided.
The above information disclosed in this background section is only for enhancement of understanding of the background of the technology described herein and, therefore, certain information may be included in the background that does not form the prior art that is already known in this country to a person of ordinary skill in the art.
Disclosure of Invention
The present application mainly aims to provide a voice interaction method, a voice interaction device, a computer-readable storage medium, a processor, and a voice interaction system, so as to solve the problem that it is difficult to perform voice broadcast of different timbres for different speakers in the prior art.
According to an aspect of an embodiment of the present invention, there is provided a voice interaction method, including: acquiring voice data of a speaker; determining a preset voice tone to be broadcasted according to the voice data; and broadcasting by adopting the preset voice timbre.
Optionally, according to the voice data, determining a preset voice tone of broadcast, including extracting voiceprint features of the voice data; and determining the preset voice tone according to the voiceprint characteristics.
Optionally, determining the predetermined voice timbre according to the voiceprint feature, including determining a voiceprint feature in a voiceprint database, which is matched with the voiceprint feature of the voice data, as a target voiceprint feature; determining the person corresponding to the target voiceprint feature as a target person; and determining the preset voice tone corresponding to the target person as the preset voice tone.
Optionally, determining a voiceprint feature in a voiceprint database, which is matched with the voiceprint feature of the voice data, as a target voiceprint feature, including obtaining a voiceprint similarity between the voiceprint feature of the voice data and each voiceprint feature in the voiceprint database; determining whether the voiceprint similarity is greater than a voiceprint similarity threshold; and under the condition that the voiceprint similarity is greater than the voiceprint similarity threshold, determining the voiceprint feature in the voiceprint database corresponding to the maximum voiceprint similarity as the target voiceprint feature.
Optionally, determining a preset voice tone corresponding to the target person as the preset voice tone, including searching the preset voice tone corresponding to the target person in a tone library; and determining the preset voice tone as the preset voice tone.
Optionally, according to the voice data, determining a preset voice tone of broadcast, including extracting voice features of the voice data; and determining the preset voice tone according to the voice characteristics of the voice data.
Optionally, determining the predetermined voice tone according to the voice feature of the voice data, including obtaining voice similarity between the voice feature of the voice data and each voice feature in a voice feature library; determining the voice features in the voice feature library corresponding to the maximum voice similarity as target voice features; acquiring target identity attribute information corresponding to the target voice feature, wherein the target identity attribute information comprises at least one of the following information: gender, age group, language; and determining the preset voice tone corresponding to the target identity attribute information as the preset voice tone.
According to another aspect of the embodiments of the present invention, a voice interaction apparatus is provided, which includes an obtaining unit, a determining unit, and a broadcasting unit, where the obtaining unit is configured to obtain voice data of a speaker; the determining unit is used for determining the preset voice tone of the broadcast according to the voice data; the broadcasting unit is used for broadcasting by adopting the preset voice timbre.
According to still another aspect of embodiments of the present invention, there is provided a computer-readable storage medium including a stored program, wherein the program performs any one of the methods described above.
According to a further aspect of the embodiments of the present invention, there is provided a processor for executing a program, where the program executes to perform any one of the methods described above.
According to another aspect of embodiments of the present invention, there is also provided a voice interaction system, comprising one or more processors, memory, and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors, the one or more programs including instructions for performing the method of any of the above.
In the embodiment of the invention, the voice interaction method determines the broadcasted preset voice timbre according to the acquired voice data and broadcasts by adopting the preset voice timbre, so that the broadcasted timbre can be intelligently recommended and switched when voice interaction is carried out with different users, and broadcasting of different voice timbres by adopting different voice timbres can be realized.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this application, illustrate embodiments of the application and, together with the description, serve to explain the application and are not intended to limit the application. In the drawings:
FIG. 1 shows a schematic flow chart of a voice interaction method generation according to an embodiment of the present application;
fig. 2 is a schematic diagram illustrating components of a voice interaction apparatus according to an embodiment of the present application.
Wherein the figures include the following reference numerals:
10. an acquisition unit; 20. a determination unit; 30. and a broadcasting unit.
Detailed Description
It should be noted that the embodiments and features of the embodiments in the present application may be combined with each other without conflict. The present application will be described in detail below with reference to the embodiments with reference to the attached drawings.
In order to make the technical solutions better understood by those skilled in the art, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only partial embodiments of the present application, but not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
It should be noted that the terms "first," "second," and the like in the description and claims of this application and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It should be understood that the data so used may be interchanged under appropriate circumstances such that embodiments of the application described herein may be used. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
For convenience of description, some terms or expressions referred to in the embodiments of the present application are explained below:
broadcasting timbre: in voice interaction, a user interacts with a machine, and the machine answers, which is generally broadcasted by using preset timbre of a speaker by using a voice synthesis technology. Different timbres (such as the sound of a man and a woman, the sound of a low lying man, the sound of a sweet woman, English, cantonese and the like) can be set for synthesized broadcasting.
And (3) voice classification: the audio is classified in the specified category by extracting audio features. Such as gender identification, age group identification, language (dialect) identification, etc., all belong to speech classification. Namely, the speaker is subjected to feature extraction and group division through voice data of the speaker.
As mentioned in the background, it is difficult to perform voice broadcasting with different timbres for different speakers in the prior art, and in order to solve the above problems, in an exemplary embodiment of the present application, a voice interaction method, an apparatus, a computer-readable storage medium, a processor, and a voice interaction system are provided.
According to an embodiment of the present application, a voice interaction method is provided.
Fig. 1 is a flow chart generated by a voice interaction method according to an embodiment of the present application. As shown in fig. 1, the method comprises the steps of:
step S101, acquiring voice data of a speaker;
step S102, determining a preset voice tone to be broadcasted according to the voice data;
and step S103, broadcasting by adopting the preset voice timbre.
According to the voice interaction method, the broadcasted preset voice timbre is determined according to the acquired voice data and is broadcasted by adopting the preset voice timbre, so that the broadcasted timbre can be intelligently recommended and switched when voice interaction is carried out with different users, broadcasting of different voice timbres by adopting different speakers is realized, meanwhile, the method does not need to rely on division and definition of voice styles such as active type, stable type, humorous type, lovely type or more true type, the preset voice timbre corresponding to the speaker can be accurately predicted and judged, the requirement of recommending different timbres to different crowds is met, and interestingness and personalized experience of the users in the using process are remarkably improved.
In a specific embodiment of the present application, determining the broadcasted predetermined voice timbre according to the voice data includes: extracting the voiceprint characteristics of the voice data; and determining the preset voice tone according to the voiceprint characteristics. By extracting the voiceprint characteristics of the voice data and determining the preset voice timbre to broadcast according to the voiceprint characteristics, the problem that synthesis and broadcast are monotonous by using the timbre preset by a system is further avoided, and the interestingness and individuation of the interactive process are increased.
In another specific embodiment of the present application, determining the predetermined voice timbre according to the voiceprint feature includes: determining the voiceprint characteristics matched with the voiceprint characteristics of the voice data in the voiceprint database as target voiceprint characteristics; determining the person corresponding to the target voiceprint feature as a target person; and determining the preset voice tone corresponding to the target person as the preset voice tone. The target voiceprint characteristics are found by acquiring the voiceprint characteristics of the speaker and matching the voiceprint characteristics with the voiceprint characteristics in the voiceprint database, the target voiceprint characteristics are confirmed as the target personnel, the preset voice timbre corresponding to the target personnel is reported as the preset voice timbre, the reported timbre is further intelligently recommended, interestingness of a user in a voice interaction process is met, and satisfaction is improved.
According to an embodiment of the present application, determining a voiceprint feature in a voiceprint database that matches a voiceprint feature of the voice data as a target voiceprint feature includes: acquiring the voiceprint similarity between the voiceprint features of the voice data and each voiceprint feature in the voiceprint database; determining whether the voiceprint similarity is greater than a voiceprint similarity threshold; and under the condition that the voiceprint similarity is greater than the voiceprint similarity threshold, determining the voiceprint feature in the voiceprint database corresponding to the maximum voiceprint similarity as the target voiceprint feature. According to the method, the voiceprint similarity is obtained and compared with the voiceprint similarity threshold value to determine the target voiceprint characteristics, so that the intelligently recommended broadcast tone is matched with the voice data of the user, and the experience and satisfaction of the user in the voice interaction process are further guaranteed.
Specifically, when the obtained voiceprint features are matched with each voiceprint feature in the voiceprint database to obtain the voiceprint similarity, only when the voiceprint similarity reaches the set voiceprint similarity threshold, determining a target voiceprint feature and broadcasting the corresponding preset voice timbre; when a plurality of voiceprint similarities reaching the voiceprint similarity threshold are obtained through matching, determining the voiceprint feature corresponding to the maximum voiceprint similarity as the target voiceprint feature and broadcasting the preset voice timbre corresponding to the target voiceprint feature; and when the voiceprint similarity obtained through matching does not reach the voiceprint similarity threshold value, broadcasting by using default synthesized timbre.
In order to further ensure that the broadcasted predetermined voice tone has a high matching degree with the voice data of the speaker, and increase the interest and personalization of the interaction process, according to another embodiment of the present application, determining the preset voice tone corresponding to the target person as the predetermined voice tone includes: searching the preset voice timbre corresponding to the target person in a timbre library; and determining the preset voice tone as the preset voice tone.
In another exemplary embodiment of the present application, determining a predetermined voice tone of a broadcast according to the voice data includes: extracting voice characteristics of the voice data; and determining the preset voice tone according to the voice characteristics of the voice data. The preset voice timbre is determined by extracting the voice features, the voice styles such as active type, stable type, model, lovely type or more true type do not need to be divided, the implementability is high, the voice features of the speaker can be predicted and judged accurately, and the requirement of intelligently recommending corresponding timbres for different people is further met.
Specifically, the speech features include pitch frequency and/or formant bandwidth, and may also include features such as MFCC (mel frequency cepstral coefficient), LPC (linear prediction coefficient), LPCC (linear prediction cepstral coefficient), and/or LSF (line spectral frequency).
In another specific embodiment of the present application, the determining the predetermined voice tone according to the voice feature of the voice data includes: acquiring voice similarity between the voice features of the voice data and each voice feature in a voice feature library; determining the voice features in the voice feature library corresponding to the maximum voice similarity as target voice features; acquiring target identity attribute information corresponding to the target voice feature, wherein the target identity attribute information comprises at least one of the following information: gender, age group, language; and determining the preset voice tone corresponding to the target identity attribute information as the preset voice tone. The preset voice tone is determined by determining the voice feature corresponding to the maximum voice similarity as a target voice feature and acquiring the identity attribute information such as gender, age group, language and the like corresponding to the target voice feature, so that the voice tone intelligently recommended and switched in the voice interaction process is more fit with the identity attribute information of a speaker, and the interestingness and satisfaction of a user in the using process are further improved. Of course, the target identity attribute information may include at least one of gender, age group, and language, and other attribute information such as a speech rate and a speaking rhythm.
It should be noted that the steps illustrated in the flowcharts of the figures may be performed in a computer system such as a set of computer-executable instructions and that, although a logical order is illustrated in the flowcharts, in some cases, the steps illustrated or described may be performed in an order different than presented herein.
The embodiment of the present application further provides a voice interaction apparatus, and it should be noted that the voice interaction apparatus according to the embodiment of the present application may be used to execute the voice interaction method according to the embodiment of the present application. The following describes a voice interaction apparatus provided in an embodiment of the present application.
Fig. 2 is a schematic composition diagram of a voice interaction apparatus according to an embodiment of the present application. As shown in fig. 2, the apparatus includes an obtaining unit 10, a determining unit 20, and a broadcasting unit 30, wherein the obtaining unit is configured to obtain voice data of a speaker; the determining unit is used for determining the preset voice tone of the broadcast according to the voice data; the broadcasting unit is used for broadcasting by adopting the preset voice timbre.
The voice interaction device determines the broadcasted preset voice tone by the determining unit according to the voice data acquired by the acquiring unit, and broadcasts by the broadcasting unit by using the preset voice tone, so that intelligent recommendation of the voice tone is realized, identity attribute information of a speaker can be accurately predicted and judged, matching of the intelligently recommended voice tone with the identity attribute information of the speaker is guaranteed, the requirement of recommending different tones for different crowds is met, the degree of individuation is high, and the use feeling and interestingness of a user in the interaction process are remarkably improved.
According to an exemplary embodiment of the present application, the determining unit includes a first extracting module and a first determining module, wherein the first extracting module is configured to extract a voiceprint feature of the voice data; the first determining module is configured to determine the predetermined voice timbre according to the voiceprint feature. Through extracting the voiceprint characteristics of the voice data and determining the preset voice timbre to broadcast according to the voiceprint characteristics, the problem that synthesis and broadcast are monotonous by using the timbre preset by a system is avoided, and the interestingness and individuation of the interactive process are increased.
According to another embodiment of the present application, the first determining module includes a first determining submodule, a second determining submodule, and a third determining submodule, wherein the first determining submodule is configured to determine a voiceprint feature in a voiceprint database, which matches a voiceprint feature of the voice data, as a target voiceprint feature; the second determining submodule is used for determining the person corresponding to the target voiceprint characteristic as a target person; the third determining submodule is configured to determine that the preset voice tone corresponding to the target person is the predetermined voice tone. The target voiceprint characteristics are found by acquiring the voiceprint characteristics of the speaker and matching the voiceprint characteristics with the voiceprint characteristics in the voiceprint database, the target voiceprint characteristics are confirmed as the target personnel, the preset voice timbre corresponding to the target personnel is reported as the preset voice timbre, the reported timbre is further intelligently recommended, interestingness of a user in a voice interaction process is met, and satisfaction is improved.
In another specific embodiment of the present application, the first determining sub-module is further configured to obtain voiceprint similarities between voiceprint features of the voice data and each voiceprint feature in the voiceprint database; determining whether the voiceprint similarity is greater than a voiceprint similarity threshold; and under the condition that the voiceprint similarity is greater than the voiceprint similarity threshold, determining the voiceprint feature in the voiceprint database corresponding to the maximum voiceprint similarity as the target voiceprint feature. The device determines the target voiceprint characteristics by acquiring the voiceprint similarity and comparing the voiceprint similarity with the voiceprint similarity threshold, so that the intelligently recommended broadcast tone is matched with the voice data of the user, and the experience and satisfaction of the user in the voice interaction process are further ensured.
Specifically, when the obtained voiceprint features are matched with each voiceprint feature in the voiceprint database to obtain the voiceprint similarity, only when the voiceprint similarity reaches the set voiceprint similarity threshold, determining a target voiceprint feature and broadcasting the corresponding preset voice timbre; when a plurality of voiceprint similarities reaching the voiceprint similarity threshold are obtained through matching, determining the voiceprint feature corresponding to the maximum voiceprint similarity as the target voiceprint feature and broadcasting the preset voice timbre corresponding to the target voiceprint feature; and when the voiceprint similarity obtained through matching does not reach the voiceprint similarity threshold value, broadcasting by using default synthesized timbre.
In order to further ensure that the broadcasted predetermined voice tone has a high matching degree with the voice data of the speaker, and increase the interest and personalization of the interaction process, according to an embodiment of the present application, the third determining sub-module is further configured to search the preset voice tone corresponding to the target person in a tone library; and determining the preset voice tone as the preset voice tone.
According to another specific embodiment of the present application, the determining unit further includes a second extracting module and a second determining module, wherein the second extracting module is configured to extract a voice feature of the voice data; the second determining module is configured to determine the predetermined voice tone according to a voice feature of the voice data. The preset voice timbre is determined by extracting the voice features, the voice styles such as active type, stable type, model, lovely type or more true type do not need to be divided, the implementability is high, the voice features of the speaker can be predicted and judged accurately, and the requirement of intelligently recommending corresponding timbres for different crowds is further met.
In another embodiment of the present application, the second determining module includes a first obtaining sub-module, a fourth determining sub-module, a second obtaining sub-module, and a fifth determining sub-module, where the first obtaining sub-module is configured to obtain a voice similarity between a voice feature of the voice data and each voice feature in a voice feature library; the fourth determining submodule is used for determining the voice feature in the voice feature library corresponding to the maximum voice similarity as a target voice feature; the second obtaining sub-module is configured to obtain target identity attribute information corresponding to the target voice feature, where the target identity attribute information includes at least one of: gender, age group, language; the fifth determining submodule is configured to determine that the preset voice tone corresponding to the target identity attribute information is the predetermined voice tone. The preset voice tone is determined by obtaining the identity attribute information such as gender, age group, language and the like, so that the voice tone intelligently recommended and switched in the voice interaction process is more fit with the identity attribute information of the speaker, and the interestingness and satisfaction of a user in the using process are further improved. Of course, the target identity attribute information may include at least one of gender, age group, and language, and other attribute information such as a speech rate and a speaking rhythm.
The voice interaction device comprises a processor and a memory, the acquiring unit 10, the determining unit 20, the broadcasting unit 30 and the like are stored in the memory as program units, and the processor executes the program units stored in the memory to realize corresponding functions.
The processor comprises a kernel, and the kernel calls the corresponding program unit from the memory. The kernel can be set to be one or more than one, and the problem that voice broadcasting with different timbres is difficult to carry out in the prior art aiming at different speakers is solved by adjusting kernel parameters.
The memory may include volatile memory in a computer readable medium, Random Access Memory (RAM) and/or nonvolatile memory such as Read Only Memory (ROM) or flash memory (flash RAM), and the memory includes at least one memory chip.
An embodiment of the present invention provides a storage medium, on which a program is stored, and the program, when executed by a processor, implements the above-mentioned voice interaction method.
The embodiment of the invention provides a processor, which is used for running a program, wherein the voice interaction method is executed when the program runs.
The embodiment of the invention provides equipment, which comprises a processor, a memory and a program which is stored on the memory and can run on the processor, wherein when the processor executes the program, at least the following steps are realized:
step S101, acquiring voice data of a speaker;
step S102, determining a preset voice tone to be broadcasted according to the voice data;
and step S103, broadcasting by adopting the preset voice timbre.
The device herein may be a server, a PC, a PAD, a mobile phone, etc.
The present application further provides a computer program product adapted to perform a program of initializing at least the following method steps when executed on a data processing device:
step S101, acquiring voice data of a speaker;
step S102, determining a preset voice tone to be broadcasted according to the voice data;
and step S103, broadcasting by adopting the preset voice timbre.
In the above embodiments of the present invention, the descriptions of the respective embodiments have respective emphasis, and for parts that are not described in detail in a certain embodiment, reference may be made to related descriptions of other embodiments.
In the embodiments provided in the present application, it should be understood that the disclosed technology can be implemented in other ways. The above-described embodiments of the apparatus are merely illustrative, and for example, the above-described division of the units may be a logical division, and in actual implementation, there may be another division, for example, multiple units or components may be combined or may be integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, units or modules, and may be in an electrical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated unit may be stored in a computer-readable storage medium if it is implemented in the form of a software functional unit and sold or used as a separate product. Based on such understanding, the technical solution of the present invention may be embodied in the form of a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the above methods according to the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a removable hard disk, a magnetic or optical disk, and other various media capable of storing program codes.
In order to make the technical solutions of the present application more clearly understood by those skilled in the art, the following description will be given with reference to specific embodiments.
Example 1
In private occasions such as home furnishing and vehicle-mounted, a user can conveniently register voiceprint characteristics on own intelligent equipment operating the voice interaction method, and the voice interaction method can conveniently identify which member a speaker is according to the voiceprint characteristics. By configuring a preset voice tone matched with the voiceprint features as a preset voice tone, for example, a male owner is set as a preset voice tone for broadcasting, a female owner is set as a preset voice tone for broadcasting, a boy is set as a preset voice tone for broadcasting, and a default is D preset voice tone for broadcasting, so that after the voice data are collected by using the voice interaction method, target personnel can be determined through the voiceprint features, and then the preset voice tone is automatically switched to and synthesized into a broadcast. Particularly, if strangers outside the family, i.e., persons who have not previously registered voiceprint features and are not configured with personalized timbres, do voice interaction, the voice interaction method cannot identify the user identity, and at this time, the default timbre D is used for broadcasting.
Example 2
In public occasions such as a hall and the like, the user does not register the voiceprint features in advance, and intelligent equipment operating the voice interaction method cannot accurately identify user identity information. At this time, the corresponding relationship between the target identity attribute information and the predetermined voice timbre is preset through the voice interaction method, for example, a middle-aged man + cantonese is set as a timbre broadcast, and a girl + mandarin is set as a B timbre broadcast, so that the intelligent device extracts the voice features of the voice data after acquiring the voice data, obtains the target identity attribute information through the voice features, for example, a man, a middle-aged person, a cantonese, a woman, a child, and a mandarin, determines the predetermined voice timbre, and automatically switches to the predetermined voice timbre and synthesizes the predetermined voice timbre into a broadcast. The attributes here include the following categories: age group, children, young, middle-aged, elderly; sex, male, female; the language, mandarin, english, cantonese, Sichuan and Shanghai languages, of course, the target identity attribute information may further include other information, the age group and the language may further include other categories, which user attributes can be identified through the voice features, depending on the development of the current voice classification technology.
From the above description, it can be seen that the above-described embodiments of the present application achieve the following technical effects:
1) the preset voice timbre determined and broadcasted according to the acquired voice data can be broadcasted, the broadcasted timbres can be intelligently recommended and switched when voice interaction is carried out with different users, the fact that different voice timbres are broadcasted is achieved, meanwhile, the method does not need to rely on division and definition of voice styles such as active type, stable type, humorous type, lovely type or more genuine type, the preset voice timbre corresponding to the speaker can be accurately predicted and judged, the requirement of recommending different timbres to different crowds is met, and interestingness and personalized experience of the users in the using process are remarkably improved.
2) The voice interaction device comprises an acquisition unit, a determination unit, a broadcast unit and a broadcast unit, wherein the determination unit determines the preset voice tone color of the broadcast according to the voice data acquired by the acquisition unit, and the broadcast unit broadcasts the preset voice tone color.
The above description is only a preferred embodiment of the present application and is not intended to limit the present application, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present application shall be included in the protection scope of the present application.

Claims (11)

1. A method of voice interaction, comprising:
acquiring voice data of a speaker;
determining a preset voice tone to be broadcasted according to the voice data;
and broadcasting by adopting the preset voice timbre.
2. The method of claim 1, wherein determining a predetermined voice tone for the broadcast based on the voice data comprises:
extracting voiceprint features of the voice data;
and determining the preset voice tone according to the voiceprint characteristics.
3. The method of claim 2, wherein determining the predetermined voice timbre based on the voiceprint feature comprises:
determining a voiceprint feature matched with the voiceprint feature of the voice data in a voiceprint database as a target voiceprint feature;
determining the person corresponding to the target voiceprint feature as a target person;
and determining the preset voice tone corresponding to the target person as the preset voice tone.
4. The method of claim 3, wherein determining the voiceprint features in the voiceprint database that match the voiceprint features of the speech data as target voiceprint features comprises:
acquiring voiceprint similarity between the voiceprint features of the voice data and each voiceprint feature in the voiceprint database;
determining whether the voiceprint similarity is greater than a voiceprint similarity threshold;
and under the condition that the voiceprint similarity is greater than the voiceprint similarity threshold, determining the voiceprint feature in the voiceprint database corresponding to the maximum voiceprint similarity as the target voiceprint feature.
5. The method according to claim 3 or 4, wherein determining the preset voice timbre corresponding to the target person as the predetermined voice timbre comprises:
searching the preset voice tone corresponding to the target person in a tone library;
and determining the preset voice tone as the preset voice tone.
6. The method of claim 1, wherein determining a predetermined voice tone for the broadcast based on the voice data comprises:
extracting voice features of the voice data;
and determining the preset voice tone according to the voice characteristics of the voice data.
7. The method of claim 6, wherein determining the predetermined voice timbre based on the voice characteristic of the voice data comprises:
acquiring voice similarity between the voice features of the voice data and each voice feature in a voice feature library;
determining the voice features in the voice feature library corresponding to the maximum voice similarity as target voice features;
acquiring target identity attribute information corresponding to the target voice feature, wherein the target identity attribute information comprises at least one of the following information: gender, age group, language;
and determining the preset voice tone corresponding to the target identity attribute information as the preset voice tone.
8. A voice interaction apparatus, comprising:
the acquisition unit is used for acquiring voice data of a speaker;
the determining unit is used for determining the preset voice tone of the broadcast according to the voice data;
and the broadcasting unit is used for broadcasting the preset voice timbre.
9. A computer-readable storage medium, characterized in that the storage medium comprises a stored program, wherein the program performs the method of any one of claims 1 to 7.
10. A processor, characterized in that the processor is configured to run a program, wherein the program when running performs the method of any of claims 1 to 7.
11. A voice interaction system, comprising: one or more processors, memory, and one or more programs stored in the memory and configured to be executed by the one or more processors, the one or more programs including instructions for performing the method of any of claims 1-7.
CN202011034411.4A 2020-09-27 2020-09-27 Voice interaction method and device, computer readable storage medium and processor Pending CN112185344A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011034411.4A CN112185344A (en) 2020-09-27 2020-09-27 Voice interaction method and device, computer readable storage medium and processor

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011034411.4A CN112185344A (en) 2020-09-27 2020-09-27 Voice interaction method and device, computer readable storage medium and processor

Publications (1)

Publication Number Publication Date
CN112185344A true CN112185344A (en) 2021-01-05

Family

ID=73943762

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011034411.4A Pending CN112185344A (en) 2020-09-27 2020-09-27 Voice interaction method and device, computer readable storage medium and processor

Country Status (1)

Country Link
CN (1) CN112185344A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114999438A (en) * 2021-05-08 2022-09-02 中移互联网有限公司 Audio playing method and device
WO2023185004A1 (en) * 2022-03-29 2023-10-05 青岛海尔空调器有限总公司 Tone switching method and apparatus
WO2023207472A1 (en) * 2022-04-28 2023-11-02 腾讯音乐娱乐科技(深圳)有限公司 Audio synthesis method, electronic device and readable storage medium

Citations (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1156871A (en) * 1995-11-17 1997-08-13 雅马哈株式会社 Personal information database system
JP2006263348A (en) * 2005-03-25 2006-10-05 Toshiba Corp Device, method, and program for identifying user
CN101467204A (en) * 2005-05-27 2009-06-24 普提克斯科技股份有限公司 Method and system for bio-metric voice print authentication
CN103236259A (en) * 2013-03-22 2013-08-07 乐金电子研发中心(上海)有限公司 Voice recognition processing and feedback system, voice response method
CN105425953A (en) * 2015-11-02 2016-03-23 小天才科技有限公司 Man-machine interaction method and system
CN106328139A (en) * 2016-09-14 2017-01-11 努比亚技术有限公司 Voice interaction method and voice interaction system
US20170287489A1 (en) * 2016-04-01 2017-10-05 Intel Corporation Synthetic oversampling to enhance speaker identification or verification
CN107357875A (en) * 2017-07-04 2017-11-17 北京奇艺世纪科技有限公司 A kind of voice search method, device and electronic equipment
CN107507620A (en) * 2017-09-25 2017-12-22 广东小天才科技有限公司 Voice broadcast sound setting method and device, mobile terminal and storage medium
CN108737872A (en) * 2018-06-08 2018-11-02 百度在线网络技术(北京)有限公司 Method and apparatus for output information
CN109065035A (en) * 2018-09-06 2018-12-21 珠海格力电器股份有限公司 information interaction method and device
CN109101801A (en) * 2018-07-12 2018-12-28 北京百度网讯科技有限公司 Method for identity verification, device, equipment and computer readable storage medium
CN109273002A (en) * 2018-10-26 2019-01-25 蔚来汽车有限公司 Vehicle configuration method, system, vehicle device and vehicle
CN109272984A (en) * 2018-10-17 2019-01-25 百度在线网络技术(北京)有限公司 Method and apparatus for interactive voice
CN109873907A (en) * 2019-03-29 2019-06-11 彭舒婷 Call processing method, device, computer equipment and storage medium
CN110047490A (en) * 2019-03-12 2019-07-23 平安科技(深圳)有限公司 Method for recognizing sound-groove, device, equipment and computer readable storage medium
CN110085225A (en) * 2019-04-24 2019-08-02 北京百度网讯科技有限公司 Voice interactive method, device, intelligent robot and computer readable storage medium
CN110248021A (en) * 2019-05-10 2019-09-17 百度在线网络技术(北京)有限公司 A kind of smart machine method for controlling volume and system
CN110379432A (en) * 2019-07-22 2019-10-25 嘉兴沐栗服饰有限公司 A kind of speech recognition system accurately judging talker according to wavelength and word speed
CN110797032A (en) * 2020-01-06 2020-02-14 深圳中创华安科技有限公司 Voiceprint database establishing method and voiceprint identification method
CN110838294A (en) * 2019-11-11 2020-02-25 效生软件科技(上海)有限公司 Voice verification method and device, computer equipment and storage medium
CN111210829A (en) * 2020-02-19 2020-05-29 腾讯科技(深圳)有限公司 Speech recognition method, apparatus, system, device and computer readable storage medium
CN111292734A (en) * 2018-12-06 2020-06-16 阿里巴巴集团控股有限公司 Voice interaction method and device
US20200227049A1 (en) * 2019-01-11 2020-07-16 Baidu Online Network Technology (Beijing) Co., Ltd. Method, apparatus and device for waking up voice interaction device, and storage medium
CN111554302A (en) * 2020-03-31 2020-08-18 深圳壹账通智能科技有限公司 Strategy adjusting method, device, terminal and storage medium based on voiceprint recognition

Patent Citations (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1156871A (en) * 1995-11-17 1997-08-13 雅马哈株式会社 Personal information database system
JP2006263348A (en) * 2005-03-25 2006-10-05 Toshiba Corp Device, method, and program for identifying user
CN101467204A (en) * 2005-05-27 2009-06-24 普提克斯科技股份有限公司 Method and system for bio-metric voice print authentication
CN103236259A (en) * 2013-03-22 2013-08-07 乐金电子研发中心(上海)有限公司 Voice recognition processing and feedback system, voice response method
CN105425953A (en) * 2015-11-02 2016-03-23 小天才科技有限公司 Man-machine interaction method and system
US20170287489A1 (en) * 2016-04-01 2017-10-05 Intel Corporation Synthetic oversampling to enhance speaker identification or verification
CN106328139A (en) * 2016-09-14 2017-01-11 努比亚技术有限公司 Voice interaction method and voice interaction system
CN107357875A (en) * 2017-07-04 2017-11-17 北京奇艺世纪科技有限公司 A kind of voice search method, device and electronic equipment
CN107507620A (en) * 2017-09-25 2017-12-22 广东小天才科技有限公司 Voice broadcast sound setting method and device, mobile terminal and storage medium
CN108737872A (en) * 2018-06-08 2018-11-02 百度在线网络技术(北京)有限公司 Method and apparatus for output information
CN109101801A (en) * 2018-07-12 2018-12-28 北京百度网讯科技有限公司 Method for identity verification, device, equipment and computer readable storage medium
CN109065035A (en) * 2018-09-06 2018-12-21 珠海格力电器股份有限公司 information interaction method and device
CN109272984A (en) * 2018-10-17 2019-01-25 百度在线网络技术(北京)有限公司 Method and apparatus for interactive voice
CN109273002A (en) * 2018-10-26 2019-01-25 蔚来汽车有限公司 Vehicle configuration method, system, vehicle device and vehicle
CN111292734A (en) * 2018-12-06 2020-06-16 阿里巴巴集团控股有限公司 Voice interaction method and device
US20200227049A1 (en) * 2019-01-11 2020-07-16 Baidu Online Network Technology (Beijing) Co., Ltd. Method, apparatus and device for waking up voice interaction device, and storage medium
CN110047490A (en) * 2019-03-12 2019-07-23 平安科技(深圳)有限公司 Method for recognizing sound-groove, device, equipment and computer readable storage medium
CN109873907A (en) * 2019-03-29 2019-06-11 彭舒婷 Call processing method, device, computer equipment and storage medium
CN110085225A (en) * 2019-04-24 2019-08-02 北京百度网讯科技有限公司 Voice interactive method, device, intelligent robot and computer readable storage medium
CN110248021A (en) * 2019-05-10 2019-09-17 百度在线网络技术(北京)有限公司 A kind of smart machine method for controlling volume and system
CN110379432A (en) * 2019-07-22 2019-10-25 嘉兴沐栗服饰有限公司 A kind of speech recognition system accurately judging talker according to wavelength and word speed
CN110838294A (en) * 2019-11-11 2020-02-25 效生软件科技(上海)有限公司 Voice verification method and device, computer equipment and storage medium
CN110797032A (en) * 2020-01-06 2020-02-14 深圳中创华安科技有限公司 Voiceprint database establishing method and voiceprint identification method
CN111210829A (en) * 2020-02-19 2020-05-29 腾讯科技(深圳)有限公司 Speech recognition method, apparatus, system, device and computer readable storage medium
CN111554302A (en) * 2020-03-31 2020-08-18 深圳壹账通智能科技有限公司 Strategy adjusting method, device, terminal and storage medium based on voiceprint recognition

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114999438A (en) * 2021-05-08 2022-09-02 中移互联网有限公司 Audio playing method and device
CN114999438B (en) * 2021-05-08 2023-08-15 中移互联网有限公司 Audio playing method and device
WO2023185004A1 (en) * 2022-03-29 2023-10-05 青岛海尔空调器有限总公司 Tone switching method and apparatus
WO2023207472A1 (en) * 2022-04-28 2023-11-02 腾讯音乐娱乐科技(深圳)有限公司 Audio synthesis method, electronic device and readable storage medium

Similar Documents

Publication Publication Date Title
CN109658916B (en) Speech synthesis method, speech synthesis device, storage medium and computer equipment
CN112185344A (en) Voice interaction method and device, computer readable storage medium and processor
Aloufi et al. Emotionless: Privacy-preserving speech analysis for voice assistants
CN110335595A (en) Slotting based on speech recognition asks dialogue method, device and storage medium
US11646026B2 (en) Information processing system, and information processing method
JP2020034895A (en) Responding method and device
CN108737872A (en) Method and apparatus for output information
CN109065052B (en) Voice robot
CN108735200A (en) A kind of speaker's automatic marking method
CN112562681B (en) Speech recognition method and apparatus, and storage medium
EP1280137B1 (en) Method for speaker identification
CN113314119A (en) Voice recognition intelligent household control method and device
CN115424606A (en) Voice interaction method, voice interaction device and computer readable storage medium
CN111081249A (en) Mode selection method, device and computer readable storage medium
CN113643684B (en) Speech synthesis method, device, electronic equipment and storage medium
CN114817514A (en) Method and device for determining reply audio, storage medium and electronic device
CN110767238A (en) Blacklist identification method, apparatus, device and storage medium based on address information
CN112820270A (en) Voice broadcasting method and device and intelligent equipment
CN113506565B (en) Speech recognition method, device, computer readable storage medium and processor
CN114514576A (en) Data processing method, device and storage medium
CN117746834B (en) Voice generation method and device based on large model, storage medium and electronic device
JP7545934B2 (en) Program, device and method for associating person's speech with user's comment
CN113220913B (en) Multimedia file processing method, device and interactive system
CN114299914A (en) Intelligent voice service system and operation method thereof
CN113257224A (en) TTS (text to speech) optimization method and system for multi-turn conversation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination