CN105654950B - Adaptive voice feedback method and device - Google Patents
Adaptive voice feedback method and device Download PDFInfo
- Publication number
- CN105654950B CN105654950B CN201610060206.2A CN201610060206A CN105654950B CN 105654950 B CN105654950 B CN 105654950B CN 201610060206 A CN201610060206 A CN 201610060206A CN 105654950 B CN105654950 B CN 105654950B
- Authority
- CN
- China
- Prior art keywords
- information
- voice
- voice feedback
- mentioned
- word speed
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 45
- 230000003044 adaptive effect Effects 0.000 title claims abstract description 31
- 230000008451 emotion Effects 0.000 claims abstract description 38
- 230000036651 mood Effects 0.000 claims description 54
- 238000001228 spectrum Methods 0.000 claims description 6
- 230000006870 function Effects 0.000 description 16
- 230000006854 communication Effects 0.000 description 10
- 238000004891 communication Methods 0.000 description 9
- 238000005516 engineering process Methods 0.000 description 9
- 238000010586 diagram Methods 0.000 description 7
- 238000012545 processing Methods 0.000 description 7
- 230000002996 emotional effect Effects 0.000 description 6
- 230000002123 temporal effect Effects 0.000 description 6
- 241001672694 Citrus reticulata Species 0.000 description 5
- 238000004590 computer program Methods 0.000 description 5
- 235000013399 edible fruits Nutrition 0.000 description 3
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 235000012907 honey Nutrition 0.000 description 2
- 230000007935 neutral effect Effects 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000005611 electricity Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 210000003127 knee Anatomy 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 235000012054 meals Nutrition 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/225—Feedback of the input speech
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
This application discloses adaptive voice feedback method and devices.One specific embodiment of the method includes: to obtain input information;Identify the scene information of the input information;Parse the input information obtain user emotion information, exchange way information, in subject content information at least one of, wherein the exchange way information includes language category information;According in the user emotion information, the language category information, the subject content information at least one of and the scene information generate user property label;The user property label is matched with the applicable label of voice feedback pattern trained in advance, and obtains matching degree;Voice feedback is carried out using with the highest voice feedback pattern of the user property matching degree.The embodiment realizes adaptive feedback voice, improves the specific aim and validity of voice feedback.
Description
Technical field
This application involves field of computer technology, and in particular to Internet technical field more particularly to adaptive voice are anti-
Present method and apparatus.
Background technique
With the development of computer technology especially Internet technology, the function of the client application of terminal device is increasingly
Diversification.Voice assistant is a kind of can to realize or substitute partially our inquiries and operation on mobile phone by interactive voice
Application, certainly, realize voice assistant function further include have voice assistant function application or have voice assistant function
The website of energy.By such application or website, the convenience of operating terminal equipment can be greatly improved.However, existing above-mentioned
Using or the interaction of website and people also only reside within how correctly to understand the voice input of people, and give answer as early as possible and ask
Topic executes certain operations (such as inquire, show, application operating), and function is more single, and specific aim is not strong.
Summary of the invention
The purpose of the application is to propose a kind of improved adaptive voice feedback method and device, to solve background above
The technical issues of technology segment is mentioned.
In a first aspect, this application provides a kind of adaptive voice feedback methods, which comprises obtain input letter
Breath;Identify the scene information of the input information;Parse the input information obtain user emotion information, exchange way information,
At least one of in subject content information, wherein the exchange way information includes language category information;According to user's feelings
Thread information, the language category information, in the subject content information at least one of and the scene information generate user and belong to
Property label;The user property label is matched with the applicable label of voice feedback pattern trained in advance, and is obtained
With degree;Voice feedback is carried out using with the highest voice feedback pattern of the user property matching degree.
In some embodiments, the exchange way information further include: word speed information, group sentence mode information or close language classification
Information;And it includes: basis that the use, which carries out voice feedback with the highest voice feedback pattern of the user property matching degree,
The word speed information, described group of sentence mode information or the close language classification information adjust the voice feedback pattern;Use process
The voice feedback pattern of adjustment is fed back.
In some embodiments, described to be believed according to the word speed information, described group of sentence mode information or the close language classification
Breath includes: that the word speed of the voice feedback pattern is adjusted to corresponding with the word speed information adjusting the voice feedback pattern
Word speed;The group sentence mode of the voice feedback pattern is adjusted to consistent with described group of sentence mode information;Acquisition pre-establishes
With the close consistent close language set of language classification information, wherein the close language set include positive common-use words and with it is described normal
The corresponding close language of term;Words and phrases in the voice feedback pattern are compared with the positive common-use words in the close language set;
If the words and phrases are identical as the positive common-use words, the words and phrases are replaced with close language corresponding with the positive common-use words.
In some embodiments, the input information includes: voice messaging and/or video information;And parsing is described defeated
Entering information to obtain user emotion information includes: that the parsing voice messaging obtains in word speed information, prosody information or spectrum information
At least one of;The word speed information, prosody information are compared with word speed threshold value, intonation threshold value, obtain voice mood knot
Fruit;It parses the video information and obtains video mood result;It is obtained based on the voice mood result and the video mood result
User emotion information out.
In some embodiments, the method also includes: according to from the recommendation set preestablished choose and institute
State scene information, the user emotion information, the associated recommendation of subject content information;Recommendation is generated to execute
Request;The recommendation is sent to client and executes request, so that client chooses whether to permit the execution recommendation
Content;If the client executes request to the recommendation and issues license, the recommendation is executed.
In some embodiments, the scene information includes at least one of the following: that temporal information, location information or terminal are answered
Use classification information.
Second aspect, this application provides a kind of adaptive voice feedback device, described device includes: acquisition module, is matched
It sets for obtaining input information;Identification module is configured to identify the scene information of the input information;Parsing module, configuration
For parse the input information obtain user emotion information, exchange way information, in subject content information at least one of,
In, the exchange way information includes language category information;Generation module is configured to according to the user emotion information, institute
At least one in predicate speech classification information, the subject content information and scene information generation user property label;?
With module, it is configured to match the user property label with the applicable label of voice feedback pattern trained in advance,
And obtain matching degree;Feedback module is configured to carry out using with the highest voice feedback pattern of the user property matching degree
Voice feedback.
In some embodiments, the exchange way information further include: word speed information, group sentence mode information or close language classification
Information;And the feedback module includes adjusting submodule, the adjusting submodule is configured to: according to the word speed information,
Described group of sentence mode information or the close language classification information adjust the voice feedback pattern;Use the voice feedback being adjusted
Pattern is fed back.
In some embodiments, described to be believed according to the word speed information, described group of sentence mode information or the close language classification
Breath includes: that the word speed of the voice feedback pattern is adjusted to corresponding with the word speed information adjusting the voice feedback pattern
Word speed;The group sentence mode of the voice feedback pattern is adjusted to consistent with described group of sentence mode information;Acquisition pre-establishes
With the close consistent close language set of language classification information, wherein the close language set include positive common-use words and with it is described normal
The corresponding close language of term;Words and phrases in the voice feedback pattern are compared with the positive common-use words in the close language set;
If the words and phrases are identical as the positive common-use words, the words and phrases are replaced with close language corresponding with the positive common-use words.
In some embodiments, the input information includes: voice messaging and/or video information;And parsing is described defeated
Entering information to obtain user emotion information includes: that the parsing voice messaging obtains in word speed information, prosody information or spectrum information
At least one of;The word speed information, prosody information are compared with word speed threshold value, intonation threshold value, obtain voice mood knot
Fruit;It parses the video information and obtains video mood result;It is obtained based on the voice mood result and the video mood result
User emotion information out.
In some embodiments, described device further includes recommending module, and the recommending module is configured to: according to from preparatory
It is chosen in the recommendation set set up related to the scene information, the user emotion information, the subject content information
The recommendation of connection;It generates recommendation and executes request;The recommendation is sent to client and executes request, for client
It chooses whether to permit the execution recommendation;Permitted if the client executes request and issue to the recommendation
Can, then execute the recommendation.
In some embodiments, the scene information includes at least one of the following: that temporal information, location information or terminal are answered
Use classification information.
Adaptive voice feedback method and device provided by the present application input information by obtaining;Identification input information
Scene information, parsing input information obtain user emotion information, exchange way information, in subject content information at least one of,
Further according in user emotion information, the language category information, subject content information at least one of and scene information generate use
Then family attribute tags match user property label with the applicable label of voice feedback pattern trained in advance, and
To matching degree, finally uses and carry out voice feedback with the highest voice feedback pattern of the user property matching degree, realize certainly
Feedback voice is adapted to, the specific aim and validity of voice feedback are improved.
Detailed description of the invention
By reading a detailed description of non-restrictive embodiments in the light of the attached drawings below, the application's is other
Feature, objects and advantages will become more apparent upon:
Fig. 1 is can be using the adaptive voice feedback method of the application or the embodiment of adaptive voice feedback device
Exemplary system architecture;
Fig. 2 is the flow chart according to one embodiment of the adaptive voice feedback method of the application;
Fig. 3 is the flow chart according to another embodiment of the adaptive voice feedback method of the application;
Fig. 4 is the data flow schematic diagram of an application scenarios of embodiment illustrated in fig. 3;
Fig. 5 is the structural schematic diagram according to one embodiment of the adaptive voice feedback device of the application;
Fig. 6 is adapted for the structural representation of the computer system for the terminal device or server of realizing the embodiment of the present application
Figure.
Specific embodiment
The application is described in further detail with reference to the accompanying drawings and examples.It is understood that this place is retouched
The specific embodiment stated is used only for explaining related invention, rather than the restriction to the invention.It also should be noted that in order to
Convenient for description, part relevant to related invention is illustrated only in attached drawing.
It should be noted that in the absence of conflict, the features in the embodiments and the embodiments of the present application can phase
Mutually combination.The application is described in detail below with reference to the accompanying drawings and in conjunction with the embodiments.
Fig. 1 is shown can be using the implementation of the adaptive voice feedback method or adaptive voice feedback device of the application
The exemplary system architecture 100 of example.
As shown in Figure 1, system architecture 100 may include terminal device 101,102,103, network 104 and server 105.
Network 104 between terminal device 101,102,103 and server 105 to provide the medium of communication link.Network 104 can be with
Including various connection types, such as wired, wireless communication link or fiber optic cables etc..
User can be used terminal device 101,102,103 and be fed back by network 104 and server 105, to receive or send out
Send message (such as voice messaging) etc..Various telecommunication customer end applications can be installed on terminal device 101,102,103, such as
The application of voice assistant class, the application of document management class, searching class application, mailbox client, social platform software etc..
Terminal device 101,102,103 can be various electronic equipments, including but not limited to smart phone, tablet computer,
E-book reader, MP3 player (Moving Picture Experts Group Audio Layer III, dynamic image
Expert's compression standard audio level 3), MP4 (Moving Picture Experts Group Audio Layer IV, dynamic shadow
As expert's compression standard audio level 4) player, pocket computer on knee and desktop computer etc..
Server 105 can be to provide the server of various services, such as to the voice on terminal device 101,102,103
The backstage Speech processing services device that the offers such as assistant's class application are supported.Above-mentioned backstage Speech processing services device can be set to from terminal
The standby voice received such as is stored, is analyzed at the processing, and processing result is fed back to above-mentioned terminal device and is executed corresponding
Operation.
As shown in Figure 1, by installing corresponding voice assistant class application, Huo Zhetong on terminal device 101,102,103
Cross on terminal device 101,102,103 with voice assistant function communication class application, or by terminal device 101,
102, browsing has the communication class website of voice assistant function on 103, these terminal devices can with the form of speech message to
Server 105 issues object dispatching request, can execute above-mentioned adaptive voice feedback method by server 105 later.Phase
Ying Di, adaptive voice feedback device can be set in server 105.
It should be understood that the number of terminal device, network and server in Fig. 1 is only schematical.According to realization need
It wants, can have any number of terminal device, network and server.
With continued reference to Fig. 2, the process of one embodiment of the adaptive voice feedback method according to the application is shown
200.Above-mentioned adaptive voice feedback method, comprising the following steps:
Step 201, input information is obtained.
In the present embodiment, electronic equipment (such as the service shown in FIG. 1 of adaptive voice feedback method operation thereon
Device) it can be from the input information for locally or remotely obtaining user.When in the memory that input information has been stored in electronic equipment
When, above-mentioned electronic equipment directly can obtain above-mentioned input information from local memory.Alternatively, when above-mentioned electronic equipment is pair
When the background server that the voice assistant class application on terminal device is supported, wired connection mode or nothing can be passed through
Line connection type obtains input information from above-mentioned terminal device.Above-mentioned radio connection include but is not limited to 3G/4G connection,
WiFi connection, bluetooth connection, WiMAX connection, Zigbee connection, UWB (ultra wideband) connection and other now
The radio connection known or developed in the future.
In the present embodiment, input information includes but is not limited to voice messaging, video information, image information or text letter
Breath.
Step 202, the scene information of identification input information.
In the present embodiment, the scene information for inputting information refers to scene locating when user sends this input information, on
Stating scene information can include but is not limited to temporal information, location information or terminal applies classification information.Herein, temporal information
Refer to that user sends the time of above-mentioned input information.Location information refers to that user sends the place of above-mentioned input information, above-mentionedly
Point can be specific address, such as the city x, the x province area the x street x;Address above mentioned is also possible to locating place, such as family, company, doctor
Institute.Herein, terminal applies classification information can be user and indicate to operate the information of which kind of terminal applies, for example, if user is logical
It crosses voice instruction and opens map class application, then terminal applies classification information is exactly this map class application;Optionally, terminal applies
Classification information can also be used in above-mentioned input information that user enters and sends which kind of terminal applies information, such as
User uses communication class application to send one of text information, video information, image information or a variety of, then terminal applies class
Other information is exactly this communication class application.
In some optional implementations of the present embodiment, location-based service (LBS, Location Based can be passed through
Services, also known as positioning service) identification location information.
In some optional implementations of the present embodiment, several scenes can be established according to the statistical data of history
Information model, such as " evening-family-take-away ", " morning-company's meeting-degree is secret ", " Sunday-open air-map ", " default ".It will obtain
The scene information matching degree of the scene information and scene information Model Matching of the input information taken, judgement and current input information is most
High scene information model makees this scene information model if this scene information model meets predetermined matching degree threshold value
For scene information.If the matching degree of scene information model and the scene information of identification cannot all reach scheduled matching degree threshold
Value, then using the scene information of identification as the scene information of input information;Alternatively, being made with " default " model of scene information model
For the scene information for inputting information.
Step 203, parsing input information obtain user emotion information, exchange way information, in subject content information extremely
One item missing.
In the present embodiment, the emotional information of user refers to emotional state information when user inputs above-mentioned input information,
Such as it is pleasant, sad, tire out, have much energy.
In some optional implementations of the present embodiment, parsing above-mentioned voice messaging and obtaining user emotion information can lead to
It crosses following manner realization: parsing above-mentioned voice messaging and obtain at least one in pace value, intonation value or spectrum information;It will be above-mentioned
Pace value, intonation value are compared with word speed threshold value, intonation threshold value, obtain voice mood result;Above-mentioned video information is parsed to obtain
To video mood result;User emotion information is obtained based on above-mentioned voice mood result and above-mentioned video mood result.
In the present embodiment, above-mentioned exchange way information includes but is not limited to language category information, above-mentioned language category letter
Breath can be languages classification information, such as Chinese, English, Japanese;It is different that above-mentioned language category information is also possible to language of the same race
Dialect information, such as Guangdong language, Ningbo words, mandarin, then such as Americanese, British English.
In the present embodiment, above-mentioned subject content information refers to the semantic information that above-mentioned input information is transmitted, for example, defeated
Entering information is " I will order take-away ", in some optional implementations, handles by word cutting and semantics recognition is handled, obtain master
Topic content information can be " ordering " " take-away ".
In some optional implementations of the present embodiment, the main theme using speech recognition technology identification voice messaging
Content information.Such as the voice messaging of user is identified as shopping, inquiry, the square XX, excellent clothing library, then by subject content information
It is expressed as a series of subject key words { shopping, inquiry, the square XX, excellent clothing library }.
In some optional implementations of the present embodiment, corresponding operation is executed according to subject content information, for example, defeated
Entering information is " I will order take-away ", has the operation for ordering the terminal applies for taking out function then executing and opening, right rear line pushes away
Recommend take-away ordering information.
Step 204, according in user emotion information, language category information, subject content information at least one of and scene
Information generates user property label.
In the present embodiment, user property label is for characterize and the mood of user, used language, being presently in field
The situation of the correlation such as scape, for example, being taken out surely by dialect, mood is dejected, then using when at 9 points in a female user evening is in
Family attribute tags can be { at night, family, female voice is dejected, Shanghai native language }.
Step 205, above-mentioned user property label is matched with the applicable label of voice feedback pattern trained in advance,
And obtain matching degree.
In the present embodiment, several voice feedback patterns of voice are trained in advance, and above-mentioned voice feedback pattern includes feature
Label and applicable label, wherein features described above label is used to characterize the feature of this voice feedback pattern, for example, voice feedback sample
The feature tag of formula is { female voice, Lin Zhiling, mandarin are after 80s };Above-mentioned applicable label is suitable for characterizing this voice feedback pattern
Which share in situation, such as it is { scene 1, scene 2, mood 1, mood that feature tag, which is the applicable label of voice feedback pattern,
2, dialect 1, close language 1, close language 2, theme 1, theme 2 }.
Step 206, voice feedback is carried out using with the highest voice feedback pattern of above-mentioned user property matching degree.
In the present embodiment, above-mentioned steps 205 are based on, by above-mentioned user property label and above-mentioned applicable tag match,
It obtains the highest voice feedback pattern of matching degree and is fed back, such as mandarin input of the user when driving in a good humor, be
System will acquire { female voice, Lin Zhiling, mandarin are after 80s } this voice feedback pattern and carry out voice feedback.
The method provided by the above embodiment of the application is by obtaining input information;The scene information of identification input information,
Parsing input information obtain user emotion information, exchange way information, in subject content information at least one of, further according to user
Emotional information, the language category information, in subject content information at least one of and scene information generate user property label,
Then user property label is matched with the applicable label of voice feedback pattern trained in advance, and obtains matching degree, most
After use and carry out voice feedback with the highest voice feedback pattern of the user property matching degree, realize adaptive backchannel
Sound improves the specific aim and validity of voice feedback.
With further reference to Fig. 3, it illustrates the processes 300 of another embodiment of adaptive voice feedback method.It should be certainly
Adapt to the process 300 of voice feedback method, comprising the following steps:
Step 301, input information is obtained.
In the present embodiment, electronic equipment (such as the service shown in FIG. 1 of adaptive voice feedback method operation thereon
Device) it can be from the input information for locally or remotely obtaining user.
In the present embodiment, input information can include but is not limited to voice messaging, video information, image information or text
Information.
Step 302, the scene information of identification input information.
In the present embodiment, the scene information for inputting information refers to scene locating when user sends this input information, on
Stating scene information can include but is not limited to temporal information, location information or terminal applies classification information.
In some optional implementations of the present embodiment, several scenes can be established according to the statistical data of history
Information model, such as " evening-family-take-away ", " morning-company's meeting-degree is secret ", " Sunday-open air-map ", " default ".It will obtain
The scene information matching degree of the scene information and scene information Model Matching of the input information taken, judgement and current input information is most
High scene information model makees this scene information model if this scene information model meets predetermined matching degree threshold value
For scene information.If the matching degree of scene information model and the scene information of identification cannot all reach scheduled matching degree threshold
Value, then using the scene information of identification as the scene information of input information;Alternatively, being made with " default " model of scene information model
For the scene information for inputting information.
Step 303, parsing input information obtains user emotion information, exchange way information, subject content information.
In the present embodiment, the emotional information of user refers to emotional state information when user inputs above-mentioned input information,
Such as it is pleasant, sad, tire out, have much energy.
In the present embodiment, parsing above-mentioned voice messaging and obtaining user emotion information can be accomplished by the following way: solution
It analyses above-mentioned voice messaging and obtains pace value, intonation value;Above-mentioned pace value, intonation value and word speed threshold value, intonation threshold value are compared
Compared with obtaining voice mood result;It parses above-mentioned video information and obtains video mood result;Based on above-mentioned voice mood result and upper
It states video mood result and obtains user emotion information.It is understood that above-mentioned word speed threshold value, intonation threshold value according to user not
It is same and different.
In some optional implementations of the present embodiment, if above-mentioned pace value is lower than word speed threshold value, and intonation value is low
In intonation threshold value, then determining that voice mood result is negative.If above-mentioned pace value is higher than word speed threshold value, and intonation value is higher than
Intonation threshold value, then determining that voice mood result is front.If only one is higher than its threshold value for above-mentioned pace value, intonation value,
So determine that voice mood result is neutrality.
In some optional implementations of the present embodiment, parsing above-mentioned video information and obtaining video mood result can lead to
It crosses video Dynamic Recognition or sampled images knowledge is realized otherwise.
In some optional implementations of the present embodiment, when voice mood result and video mood result be all front or
When negative, mood result information is just positive or negative;If when voice mood result and inconsistent video mood result, determined
Mood result information is neutrality.
It can be above-mentioned voice mood result and above-mentioned video mood knot in some optional implementations of the present embodiment
Fruit sets up different weights, such as when voice input quality is preferable and video input is second-rate, sets for voice mood result
Vertical higher weights.It is 1, -1,0 by front, negative, the neutral numerical value respectively that is arranged, by voice mood result and video mood result
Weight and front, negative, neutral corresponding numerical operation, obtain mood end value, judge which kind of feelings is mood end value be in
The numerical intervals of thread obtain mood result information.
In some optional implementations of the present embodiment, above-mentioned exchange way information includes: language category information, word speed
Value, group sentence mode information or close language classification information.
Above-mentioned subject content information refers to the semantic information that above-mentioned input information is transmitted, for example, input information is that " I wants
Order take-away ", in some optional implementations, is handled by word cutting and semantics recognition is handled, show that subject content information can
To be " ordering " " take-away ".
In some optional implementations of the present embodiment, the main theme using speech recognition technology identification voice messaging
Content information.Such as the voice messaging of user is identified as shopping, inquiry, the square XX, excellent clothing library, then by subject content information
It is expressed as a series of subject key words { shopping, inquiry, the square XX, excellent clothing library }.
In some optional implementations of the present embodiment, corresponding operation is executed according to subject content information, for example, defeated
Entering information is " I will order take-away ", has the operation for ordering the terminal applies for taking out function then executing and opening, right rear line pushes away
Recommend take-away ordering information.
Step 304, according to above-mentioned user emotion information, above-mentioned language category information, above-mentioned subject content information and above-mentioned
Scene information generates user property label.
In the present embodiment, user property label for characterize and the mood of user, used language, expressed by user
The subject content situation related to scene etc. is presently in pass through dialect for example, 9 points when being at night of a female user
Fixed to take out, mood is dejected, then user property label can be { at night, family, female voice is dejected, Shanghai native language, orders take-away }.
Step 305, above-mentioned user property label is matched with the applicable label of voice feedback pattern trained in advance,
And obtain matching degree.
In the present embodiment, several voice feedback patterns of voice are trained in advance, and above-mentioned voice feedback pattern includes feature
Label and applicable label, wherein features described above label is used to characterize the feature of this voice feedback pattern, for example, voice feedback sample
The feature tag of formula is { female voice, Lin Zhiling, mandarin are after 80s };Above-mentioned applicable label is suitable for characterizing this voice feedback pattern
Which share in situation, such as it is { scene 1, scene 2, mood 1, mood that feature tag, which is the applicable label of voice feedback pattern,
2, dialect 1, close language 1, close language 2, theme 1, theme 2 }.
Step 306, above-mentioned voice feedback pattern is adjusted according to word speed information, group sentence mode information or close language classification information.
In some optional implementations of the present embodiment, above-mentioned voice feedback pattern is adjusted according to above-mentioned word speed information
It can be accomplished by the following way: the word speed of above-mentioned voice feedback pattern is adjusted to word speed corresponding with above-mentioned word speed information,
Herein, word speed information is the value range inputted near pace value numerical value in user.For example, user is releived using slower
Voice, then the word speed of voice feedback pattern is also also adjusted to the slower voice releived, i.e., by voice feedback pattern
Word speed is transferred in this value range.For another example the word speed of above-mentioned voice feedback pattern is adjusted to different from above-mentioned word speed information
In range, such as user uses faster voice, and possible user is in an irritable state, then by voice feedback pattern
Word speed is the slower voice releived.
In some optional implementations of the present embodiment, can by adjusting voice in voice feedback pattern word with
The word speed of interval adjustment voice feedback pattern between word, between sentence and sentence.
In some optional implementations of the present embodiment, above-mentioned voice feedback is adjusted according to above-mentioned group of sentence mode information
Pattern can be accomplished by the following way: the group sentence mode of above-mentioned voice feedback pattern is adjusted to and above-mentioned group of sentence mode information
Unanimously;Above-mentioned group of sentence mode information is obtained by speech recognition technology, such as identifies that the group sentence mode for obtaining user is accustomed to as { meaning
Language, subject, object }, then the group sentence mode of adjustment voice feedback pattern is { predicate, subject, object }.For another example if with
The group sentence mode at family is accustomed to as { have a meal, first }, if Normal Feedback may be " getting ahead ", according to herein may according to be adjusted to " walk,
First ".
In some optional implementations of the present embodiment, above-mentioned voice feedback pattern is adjusted according to close language classification information
It can be accomplished by the following way: obtain the above-mentioned close consistent close language set of language classification information, wherein above-mentioned close language set includes
Positive common-use words and close language corresponding with above-mentioned positive common-use words;It will be in the words and phrases and above-mentioned close language set in above-mentioned voice feedback pattern
Positive common-use words be compared;If above-mentioned words and phrases are identical as above-mentioned positive common-use words, with corresponding close with above-mentioned positive common-use words
Language replaces above-mentioned words and phrases.
Step 307, it is fed back using the voice feedback pattern being adjusted.
In the present embodiment, above-mentioned steps 306 are based on, carry out voice feedback using the voice feedback pattern being adjusted.
In some optional implementations of the present embodiment, the process 300 of above-mentioned adaptive voice feedback method may be used also
To include step 308, after feeding back voice, recommendation is executed;Above-mentioned execution recommendation can be realized by following steps:
According to from being chosen in the recommendation set preestablished with above-mentioned scene information, above-mentioned user emotion information, in above-mentioned theme
Hold the associated recommendation of information;It generates recommendation and executes request;Above-mentioned recommendation, which is sent, to client executes request,
So that client chooses whether to permit the above-mentioned recommendation of above-mentioned execution;If above-mentioned client asks the execution of above-mentioned recommendation
It asks sending to permit, then executes above-mentioned recommendation.For example, recognition result is (scene information) happy (user outdoors in the morning
Emotional information) exercise (subject content information), then can recommend play a small pure and fresh song, if user license,
Execute the operation for playing and recommending song.For another example being taken out surely by dialect when at 9 points in a female user evening is in, passing through use
Family carries out voice and video identification when voice input, judges that user is very tired very hungry very dejected, then takes out surely in rapid execution
Meanwhile the same sex dialect sound releived can be used, it inquires whether the user needs to play the music liked, or dial the electricity of common boudoir honey
Words, if permitted, play the music or the phone for dialing common boudoir honey that user likes.
In the present embodiment, step 301, step 302, step 303, step 304 and the step 305 in above-mentioned implementation process
It is essentially identical with step 201, step 202, step 203, step 204 and the step 205 in previous embodiment respectively, herein no longer
It repeats.
From figure 3, it can be seen that the main difference of embodiment corresponding with Fig. 2 is the adaptive language in the present embodiment
The process 300 of sound feedback method, which has had more, adjusts above-mentioned voice according to word speed information, group sentence mode information or close language classification information
It feeds back the step 306 of pattern and executes the step 308 of recommendation after feeding back voice.By the step 306 of increase and 308,
The scheme of the present embodiment description significantly more efficient can feed back preparatory information, improve the specific aim of feedback.
With further reference to Fig. 5, as the realization to method shown in above-mentioned each figure, this application provides a kind of adaptive voices
One embodiment of feedback device, the Installation practice is corresponding with embodiment of the method shown in Fig. 2, which can specifically answer
For in various electronic equipments.
As shown in figure 5, the above-mentioned adaptive voice feedback device 500 of the present embodiment includes: to obtain module 501, configuration is used
Information is inputted in obtaining;Identification module 502 is configured to identify the scene information of above-mentioned input information;Parsing module 503, matches
Set for parse above-mentioned input information obtain user emotion information, exchange way information, in subject content information at least one of,
Wherein, above-mentioned exchange way information includes language category information;Generation module 504 is configured to be believed according to above-mentioned user emotion
At least one in breath, above-mentioned language category information, above-mentioned subject content information and above-mentioned scene information generation user property mark
Label;Matching module 505, be configured to by the applicable label of above-mentioned user property label and voice feedback pattern trained in advance into
Row matching, and obtain matching degree;Feedback module 506 is configured to using anti-with the above-mentioned highest voice of user property matching degree
It presents pattern and carries out voice feedback.
In an optional embodiment of the present embodiment, above-mentioned exchange way information further include: word speed information, group sentence side
Formula information or close language classification information;And above-mentioned feedback module includes adjusting submodule, above-mentioned adjusting submodule is configured to:
Above-mentioned voice feedback pattern is adjusted according to above-mentioned word speed information, above-mentioned group of sentence mode information or above-mentioned close language classification information;It uses
The voice feedback pattern being adjusted is fed back.
It is above-mentioned according to above-mentioned word speed information, above-mentioned group of sentence mode information in an optional embodiment of the present embodiment
Or above-mentioned close language classification information adjust above-mentioned voice feedback pattern include: by the word speed of above-mentioned voice feedback pattern be adjusted to
The corresponding word speed of above-mentioned word speed information;The group sentence mode of above-mentioned voice feedback pattern is adjusted to and above-mentioned group of sentence mode information one
It causes;It obtains pre-establishing with the above-mentioned close consistent close language set of language classification information, wherein above-mentioned close language set includes just common
Language and close language corresponding with above-mentioned positive common-use words;By in above-mentioned voice feedback pattern words and phrases with it is normal in above-mentioned close language set
Term is compared;If above-mentioned words and phrases are identical as above-mentioned positive common-use words, replaced with close language corresponding with above-mentioned positive common-use words
Above-mentioned words and phrases.
In an optional embodiment of the present embodiment, above-mentioned input information includes: voice messaging and/or video letter
Breath;And parsing above-mentioned input information to obtain user emotion information includes: that the above-mentioned voice messaging of parsing obtains word speed information, language
At least one of in tune information or spectrum information;Above-mentioned word speed information, prosody information and word speed threshold value, intonation threshold value are compared
Compared with obtaining voice mood result;It parses above-mentioned video information and obtains video mood result;Based on above-mentioned voice mood result and upper
It states video mood result and obtains user emotion information.
In an optional embodiment of the present embodiment, above-mentioned apparatus 500 further includes recommending module 507, above-mentioned recommendation
Module is configured to: being believed according to choosing from the recommendation set preestablished with above-mentioned scene information, above-mentioned user emotion
Breath, the above-mentioned associated recommendation of subject content information;It generates recommendation and executes request;Above-mentioned recommendation is sent to client
Content executes request, so that client chooses whether to permit the above-mentioned recommendation of above-mentioned execution;If above-mentioned client is to above-mentioned
Recommendation executes request and issues license, then executes above-mentioned recommendation.
In an optional embodiment of the present embodiment, above-mentioned scene information include at least one of the following: temporal information,
Location information or terminal applies classification information.
It will be understood by those skilled in the art that above-mentioned adaptive voice feedback device 500 further includes some other known knots
Structure, such as processor, memory etc., in order to unnecessarily obscure embodiment of the disclosure, these well known structures in Fig. 5 not
It shows.
Below with reference to Fig. 6, it illustrates the calculating of the terminal device or server that are suitable for being used to realize the embodiment of the present application
The structural schematic diagram of machine system 600.
As shown in fig. 6, computer system 600 includes central processing unit (CPU) 601, it can be read-only according to being stored in
Program in memory (ROM) 602 or be loaded into the program in random access storage device (RAM) 603 from storage section 608 and
Execute various movements appropriate and processing.In RAM 603, also it is stored with system 600 and operates required various programs and data.
CPU 601, ROM 602 and RAM 603 are connected with each other by bus 604.Input/output (I/O) interface 605 is also connected to always
Line 604.
I/O interface 605 is connected to lower component: the importation 606 including keyboard, mouse etc.;It is penetrated including such as cathode
The output par, c 607 of spool (CRT), liquid crystal display (LCD) etc. and loudspeaker etc.;Storage section 608 including hard disk etc.;
And the communications portion 609 of the network interface card including LAN card, modem etc..Communications portion 609 via such as because
The network of spy's net executes communication process.Driver 610 is also connected to I/O interface 605 as needed.Detachable media 611, such as
Disk, CD, magneto-optic disk, semiconductor memory etc. are mounted on as needed on driver 610, in order to read from thereon
Computer program be mounted into storage section 608 as needed.
Particularly, in accordance with an embodiment of the present disclosure, it may be implemented as computer above with reference to the process of flow chart description
Software program.For example, embodiment of the disclosure includes a kind of computer program product comprising be tangibly embodied in machine readable
Computer program on medium, above-mentioned computer program include the program code for method shown in execution flow chart.At this
In the embodiment of sample, which can be downloaded and installed from network by communications portion 609, and/or from removable
Medium 611 is unloaded to be mounted.When the computer program is executed by central processing unit (CPU) 601, execute in the present processes
The above-mentioned function of limiting.
Flow chart and block diagram in attached drawing are illustrated according to the system of the various embodiments of the application, method and computer journey
The architecture, function and operation in the cards of sequence product.In this regard, each box in flowchart or block diagram can generation
A part of one module, program segment or code of table, a part of above-mentioned module, program segment or code include one or more
Executable instruction for implementing the specified logical function.It should also be noted that in some implementations as replacements, institute in box
The function of mark can also occur in a different order than that indicated in the drawings.For example, two boxes succeedingly indicated are practical
On can be basically executed in parallel, they can also be executed in the opposite order sometimes, and this depends on the function involved.Also it wants
It is noted that the combination of each box in block diagram and or flow chart and the box in block diagram and or flow chart, Ke Yiyong
The dedicated hardware based system of defined functions or operations is executed to realize, or can be referred to specialized hardware and computer
The combination of order is realized.
Being described in unit involved in the embodiment of the present application can be realized by way of software, can also be by hard
The mode of part is realized.Described unit also can be set in the processor, for example, can be described as: a kind of processor packet
Include acquisition module.Wherein, the title of these units does not constitute the restriction to the unit itself under certain conditions, for example, instead
Compilation unit is also described as " for obtaining the unit of input information ".
As on the other hand, present invention also provides a kind of nonvolatile computer storage media, the non-volatile calculating
Machine storage medium can be nonvolatile computer storage media included in above-mentioned apparatus in above-described embodiment;It is also possible to
Individualism, without the nonvolatile computer storage media in supplying terminal.Above-mentioned nonvolatile computer storage media is deposited
One or more program is contained, when said one or multiple programs are executed by an equipment, so that above equipment: will be upper
The word speed for stating voice feedback pattern is adjusted to word speed corresponding with above-mentioned word speed information;By the group sentence side of above-mentioned voice feedback pattern
Formula is adjusted to consistent with above-mentioned group of sentence mode information;It obtains pre-establishing with the above-mentioned close consistent close language collection of language classification information
It closes, wherein above-mentioned close language set includes positive common-use words and close language corresponding with above-mentioned positive common-use words;By above-mentioned voice feedback pattern
In words and phrases be compared with the positive common-use words in above-mentioned close language set;If above-mentioned words and phrases are identical as above-mentioned positive common-use words,
Above-mentioned words and phrases are replaced with close language corresponding with above-mentioned positive common-use words.
Above description is only the preferred embodiment of the application and the explanation to institute's application technology principle.Those skilled in the art
Member is it should be appreciated that invention scope involved in the application, however it is not limited to technology made of the specific combination of above-mentioned technical characteristic
Scheme, while should also cover in the case where not departing from foregoing invention design, it is carried out by above-mentioned technical characteristic or its equivalent feature
Any combination and the other technical solutions formed.Such as features described above has similar function with (but being not limited to) disclosed herein
Can technical characteristic replaced mutually and the technical solution that is formed.
Claims (12)
1. a kind of adaptive voice feedback method, which is characterized in that the described method includes:
Obtain input information;
Identify the scene information of the input information;
Parse the input information obtain user emotion information, exchange way information, in subject content information at least one of,
In, the exchange way information includes language category information;
According in the user emotion information, the language category information, the subject content information at least one of and it is described
Scene information generates user property label;
The user property label is matched with the applicable label of voice feedback pattern trained in advance, and is matched
Degree;
Voice feedback is carried out using with the highest voice feedback pattern of the user property tag match degree.
2. the method according to claim 1, wherein the exchange way information further include: word speed information, group sentence
Mode information or close language classification information;And
The use carries out voice feedback with the highest voice feedback pattern of the user property tag match degree
The voice feedback pattern is adjusted according to the word speed information, described group of sentence mode information or the close language classification information;
It is fed back using the voice feedback pattern being adjusted.
3. according to the method described in claim 2, it is characterized in that, described believe according to the word speed information, described group of sentence mode
Breath or the close language classification information include: adjusting the voice feedback pattern
The word speed of the voice feedback pattern is adjusted to word speed corresponding with the word speed information;
The group sentence mode of the voice feedback pattern is adjusted to consistent with described group of sentence mode information;
It obtains pre-establishing with the close consistent close language set of language classification information, wherein the close language set includes normal
Term and close language corresponding with the positive common-use words;By in the voice feedback pattern words and phrases and the close language set in just
Common-use words are compared;If the words and phrases are identical as the positive common-use words, replaced with close language corresponding with the positive common-use words
Change the words and phrases.
4. method according to claim 1-3, which is characterized in that the input information include: voice messaging and/
Or video information;And
It parses the input information and obtains user emotion information and include:
It parses the voice messaging and obtains at least one in word speed information, prosody information or spectrum information;
The word speed information, prosody information are compared with word speed threshold value, intonation threshold value, obtain voice mood result;
It parses the video information and obtains video mood result;
User emotion information is obtained based on the voice mood result and the video mood result.
5. according to the method described in claim 4, it is characterized in that, the method also includes:
It is chosen and the scene information, the user emotion information, the master according to from the recommendation set preestablished
Inscribe the associated recommendation of content information;
It generates recommendation and executes request;
The recommendation is sent to client and executes request, so that client chooses whether to permit described execute in the recommendation
Hold;
If the client executes request to the recommendation and issues license, the recommendation is executed.
6. according to the method described in claim 5, it is characterized in that, the scene information includes at least one of the following: that the time believes
Breath, location information or terminal applies classification information.
7. a kind of adaptive voice feedback device, which is characterized in that described device includes:
Module is obtained, is configured to obtain input information;
Identification module is configured to identify the scene information of the input information;
Parsing module is configured to parse the input information and obtains user emotion information, exchange way information, subject content letter
At least one of in breath, wherein the exchange way information includes language category information;
Generation module is configured to according in the user emotion information, the language category information, the subject content information
At least one of and the scene information generate user property label;
Matching module is configured to carry out on the applicable label of the user property label and voice feedback pattern trained in advance
Matching, and obtain matching degree;
Feedback module is configured to anti-using voice is carried out with the highest voice feedback pattern of the user property tag match degree
Feedback.
8. device according to claim 7, which is characterized in that the exchange way information further include: word speed information, group sentence
Mode information or close language classification information;And
The feedback module includes adjusting submodule, and the adjusting submodule is configured to:
The voice feedback pattern is adjusted according to the word speed information, described group of sentence mode information or the close language classification information;
It is fed back using the voice feedback pattern being adjusted.
9. device according to claim 8, which is characterized in that described to be believed according to the word speed information, described group of sentence mode
Breath or the close language classification information include: adjusting the voice feedback pattern
The word speed of the voice feedback pattern is adjusted to word speed corresponding with the word speed information;
The group sentence mode of the voice feedback pattern is adjusted to consistent with described group of sentence mode information;
It obtains pre-establishing with the close consistent close language set of language classification information, wherein the close language set includes normal
Term and close language corresponding with the positive common-use words;By in the voice feedback pattern words and phrases and the close language set in just
Common-use words are compared;If the words and phrases are identical as the positive common-use words, replaced with close language corresponding with the positive common-use words
Change the words and phrases.
10. according to the described in any item devices of claim 7-9, which is characterized in that the input information includes: voice messaging
And/or video information;And
It parses the input information and obtains user emotion information and include:
It parses the voice messaging and obtains at least one in word speed information, prosody information or spectrum information;
The word speed information, prosody information are compared with word speed threshold value, intonation threshold value, obtain voice mood result;
It parses the video information and obtains video mood result;
User emotion information is obtained based on the voice mood result and the video mood result.
11. device according to claim 10, which is characterized in that described device further includes recommending module, the recommendation mould
Block is configured to:
It is chosen and the scene information, the user emotion information, the master according to from the recommendation set preestablished
Inscribe the associated recommendation of content information;
It generates recommendation and executes request;
The recommendation is sent to client and executes request, so that client chooses whether to permit described execute in the recommendation
Hold;
If the client executes request to the recommendation and issues license, the recommendation is executed.
12. device according to claim 11, which is characterized in that the scene information includes at least one of the following: the time
Information, location information or terminal applies classification information.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610060206.2A CN105654950B (en) | 2016-01-28 | 2016-01-28 | Adaptive voice feedback method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610060206.2A CN105654950B (en) | 2016-01-28 | 2016-01-28 | Adaptive voice feedback method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105654950A CN105654950A (en) | 2016-06-08 |
CN105654950B true CN105654950B (en) | 2019-07-16 |
Family
ID=56488908
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610060206.2A Active CN105654950B (en) | 2016-01-28 | 2016-01-28 | Adaptive voice feedback method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105654950B (en) |
Families Citing this family (59)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10630751B2 (en) | 2016-12-30 | 2020-04-21 | Google Llc | Sequence dependent data message consolidation in a voice activated computer network environment |
US10956485B2 (en) | 2011-08-31 | 2021-03-23 | Google Llc | Retargeting in a search environment |
US9703757B2 (en) | 2013-09-30 | 2017-07-11 | Google Inc. | Automatically determining a size for a content item for a web page |
US10431209B2 (en) | 2016-12-30 | 2019-10-01 | Google Llc | Feedback controller for data transmissions |
US10614153B2 (en) | 2013-09-30 | 2020-04-07 | Google Llc | Resource size-based content item selection |
CN106297789B (en) * | 2016-08-19 | 2020-01-14 | 北京光年无限科技有限公司 | Personalized interaction method and system for intelligent robot |
CN106486111B (en) * | 2016-10-14 | 2020-02-07 | 北京光年无限科技有限公司 | Multi-TTS engine output speech speed adjusting method and system based on intelligent robot |
US10157607B2 (en) | 2016-10-20 | 2018-12-18 | International Business Machines Corporation | Real time speech output speed adjustment |
CN106531162A (en) * | 2016-10-28 | 2017-03-22 | 北京光年无限科技有限公司 | Man-machine interaction method and device used for intelligent robot |
CN106657543B (en) * | 2016-10-31 | 2020-02-07 | 北京小米移动软件有限公司 | Voice information processing method and device |
CN106504743B (en) * | 2016-11-14 | 2020-01-14 | 北京光年无限科技有限公司 | Voice interaction output method for intelligent robot and robot |
CN108121721A (en) * | 2016-11-28 | 2018-06-05 | 渡鸦科技(北京)有限责任公司 | Intension recognizing method and device |
JP6795387B2 (en) * | 2016-12-14 | 2020-12-02 | パナソニック株式会社 | Voice dialogue device, voice dialogue method, voice dialogue program and robot |
CN108205526A (en) * | 2016-12-20 | 2018-06-26 | 百度在线网络技术(北京)有限公司 | A kind of method and apparatus of determining Technique Using Both Text information |
CN106815321A (en) * | 2016-12-27 | 2017-06-09 | 深圳前海勇艺达机器人有限公司 | Chat method and device based on intelligent chat robots |
CN106782540B (en) * | 2017-01-17 | 2021-04-13 | 联想(北京)有限公司 | Voice equipment and voice interaction system comprising same |
CN106782521A (en) * | 2017-03-22 | 2017-05-31 | 海南职业技术学院 | A kind of speech recognition system |
CN106992012A (en) * | 2017-03-24 | 2017-07-28 | 联想(北京)有限公司 | Method of speech processing and electronic equipment |
CN106782544A (en) * | 2017-03-29 | 2017-05-31 | 联想(北京)有限公司 | Interactive voice equipment and its output intent |
CN107146610B (en) * | 2017-04-10 | 2021-06-15 | 易视星空科技无锡有限公司 | Method and device for determining user intention |
CN107331388A (en) * | 2017-06-15 | 2017-11-07 | 重庆柚瓣科技有限公司 | A kind of dialect collection system based on endowment robot |
CN107291900B (en) * | 2017-06-22 | 2020-06-05 | 美味不用等(上海)信息科技股份有限公司 | Information feedback and tracking system |
CN107316641B (en) * | 2017-06-30 | 2021-06-15 | 联想(北京)有限公司 | Voice control method and electronic equipment |
CN107545029A (en) * | 2017-07-17 | 2018-01-05 | 百度在线网络技术(北京)有限公司 | Voice feedback method, equipment and the computer-readable recording medium of smart machine |
CN107393530B (en) * | 2017-07-18 | 2020-08-25 | 国网山东省电力公司青岛市黄岛区供电公司 | Service guiding method and device |
CN107274900B (en) * | 2017-08-10 | 2020-09-18 | 北京京东尚科信息技术有限公司 | Information processing method for control terminal and system thereof |
CN107657017B (en) * | 2017-09-26 | 2020-11-13 | 百度在线网络技术(北京)有限公司 | Method and apparatus for providing voice service |
CN107767869B (en) * | 2017-09-26 | 2021-03-12 | 百度在线网络技术(北京)有限公司 | Method and apparatus for providing voice service |
CN107818787B (en) * | 2017-10-31 | 2021-02-05 | 努比亚技术有限公司 | Voice information processing method, terminal and computer readable storage medium |
CN107871500B (en) * | 2017-11-16 | 2021-07-20 | 百度在线网络技术(北京)有限公司 | Method and device for playing multimedia |
CN108257596B (en) * | 2017-12-22 | 2021-07-23 | 北京小蓦机器人技术有限公司 | Method and equipment for providing target presentation information |
CN108091324B (en) | 2017-12-22 | 2021-08-17 | 北京百度网讯科技有限公司 | Tone recognition method and device, electronic equipment and computer-readable storage medium |
CN108281141B (en) * | 2017-12-22 | 2019-10-18 | 北京小蓦机器人技术有限公司 | It is a kind of for providing the method and apparatus of voice-response information |
CN108257037A (en) * | 2018-01-18 | 2018-07-06 | 封玉涛 | It is a kind of with social scene turn to point of penetration apply fusion method and device |
CN108319485A (en) * | 2018-01-29 | 2018-07-24 | 出门问问信息科技有限公司 | Information interacting method, device, equipment and storage medium |
CN108335700B (en) * | 2018-01-30 | 2021-07-06 | 重庆与展微电子有限公司 | Voice adjusting method and device, voice interaction equipment and storage medium |
CN109036405A (en) * | 2018-07-27 | 2018-12-18 | 百度在线网络技术(北京)有限公司 | Voice interactive method, device, equipment and storage medium |
CN109240488A (en) * | 2018-07-27 | 2019-01-18 | 重庆柚瓣家科技有限公司 | A kind of implementation method of AI scene engine of positioning |
CN110930998A (en) * | 2018-09-19 | 2020-03-27 | 上海博泰悦臻电子设备制造有限公司 | Voice interaction method and device and vehicle |
CN110930999A (en) * | 2018-09-19 | 2020-03-27 | 上海博泰悦臻电子设备制造有限公司 | Voice interaction method and device and vehicle |
CN109346076A (en) * | 2018-10-25 | 2019-02-15 | 三星电子(中国)研发中心 | Interactive voice, method of speech processing, device and system |
CN109672724B (en) * | 2018-11-01 | 2022-07-29 | 百度在线网络技术(北京)有限公司 | Information pushing method, device and equipment |
CN109377979B (en) * | 2018-12-07 | 2021-09-24 | 思必驰科技股份有限公司 | Method and system for updating welcome language |
CN111292737A (en) * | 2018-12-07 | 2020-06-16 | 阿里巴巴集团控股有限公司 | Voice interaction and voice awakening detection method, device, equipment and storage medium |
CN109697290B (en) * | 2018-12-29 | 2023-07-25 | 咪咕数字传媒有限公司 | Information processing method, equipment and computer storage medium |
CN109686362B (en) * | 2019-01-02 | 2021-04-02 | 百度在线网络技术(北京)有限公司 | Voice broadcasting method and device and computer readable storage medium |
CN109741744B (en) * | 2019-01-14 | 2021-03-09 | 博拉网络股份有限公司 | AI robot conversation control method and system based on big data search |
CN111724774B (en) * | 2019-03-22 | 2024-05-17 | 斑马智行网络(香港)有限公司 | Voice interaction and vehicle-mounted voice interaction method, device, equipment and storage medium |
CN110347817B (en) * | 2019-07-15 | 2022-03-18 | 网易(杭州)网络有限公司 | Intelligent response method and device, storage medium and electronic equipment |
CN110827797B (en) * | 2019-11-06 | 2022-04-12 | 北京沃东天骏信息技术有限公司 | Voice response event classification processing method and device |
CN111354350B (en) * | 2019-12-26 | 2024-04-05 | 阿里巴巴集团控股有限公司 | Voice processing method and device, voice processing equipment and electronic equipment |
CN111179903A (en) * | 2019-12-30 | 2020-05-19 | 珠海格力电器股份有限公司 | Voice recognition method and device, storage medium and electric appliance |
CN111241822A (en) * | 2020-01-03 | 2020-06-05 | 北京搜狗科技发展有限公司 | Emotion discovery and dispersion method and device under input scene |
CN111310009A (en) * | 2020-01-16 | 2020-06-19 | 珠海格力电器股份有限公司 | User classification method and device, storage medium and computer equipment |
CN111724173A (en) * | 2020-06-18 | 2020-09-29 | 中国银行股份有限公司 | Robot self-adjusting method, device, equipment and computer storage medium |
CN112017646A (en) * | 2020-08-21 | 2020-12-01 | 博泰车联网(南京)有限公司 | Voice processing method and device and computer storage medium |
CN115312079A (en) * | 2021-04-20 | 2022-11-08 | 北京沃东天骏信息技术有限公司 | Information display method and device, electronic equipment and computer readable medium |
CN113435962A (en) * | 2021-06-07 | 2021-09-24 | 布瑞克农业大数据科技集团有限公司 | Agricultural product online collecting and purchasing method, system and storage medium thereof |
CN114356276A (en) * | 2021-12-22 | 2022-04-15 | 科大讯飞股份有限公司 | Voice interaction method and related device |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103577544A (en) * | 2013-10-11 | 2014-02-12 | 北京百度网讯科技有限公司 | Method and device for providing information to be sent |
CN104038836A (en) * | 2014-06-03 | 2014-09-10 | 四川长虹电器股份有限公司 | Television program intelligent pushing method |
CN104836720A (en) * | 2014-02-12 | 2015-08-12 | 北京三星通信技术研究有限公司 | Method for performing information recommendation in interactive communication, and device |
CN104881108A (en) * | 2014-02-27 | 2015-09-02 | 青岛海尔机器人有限公司 | Intelligent man-machine interaction method and device |
CN105206269A (en) * | 2015-08-14 | 2015-12-30 | 百度在线网络技术(北京)有限公司 | Voice processing method and device |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090234655A1 (en) * | 2008-03-13 | 2009-09-17 | Jason Kwon | Mobile electronic device with active speech recognition |
-
2016
- 2016-01-28 CN CN201610060206.2A patent/CN105654950B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103577544A (en) * | 2013-10-11 | 2014-02-12 | 北京百度网讯科技有限公司 | Method and device for providing information to be sent |
CN104836720A (en) * | 2014-02-12 | 2015-08-12 | 北京三星通信技术研究有限公司 | Method for performing information recommendation in interactive communication, and device |
CN104881108A (en) * | 2014-02-27 | 2015-09-02 | 青岛海尔机器人有限公司 | Intelligent man-machine interaction method and device |
CN104038836A (en) * | 2014-06-03 | 2014-09-10 | 四川长虹电器股份有限公司 | Television program intelligent pushing method |
CN105206269A (en) * | 2015-08-14 | 2015-12-30 | 百度在线网络技术(北京)有限公司 | Voice processing method and device |
Also Published As
Publication number | Publication date |
---|---|
CN105654950A (en) | 2016-06-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105654950B (en) | Adaptive voice feedback method and device | |
US20210232761A1 (en) | Methods and systems for improving machine learning performance | |
US11315546B2 (en) | Computerized system and method for formatted transcription of multimedia content | |
CN105701088B (en) | The method and apparatus manually talked with are switched to from machine dialogue | |
CN107491534A (en) | Information processing method and device | |
CN104735468B (en) | A kind of method and system that image is synthesized to new video based on semantic analysis | |
CN105786793B (en) | Parse the semantic method and apparatus of spoken language text information | |
CN106383875B (en) | Man-machine interaction method and device based on artificial intelligence | |
CN108986805B (en) | Method and apparatus for sending information | |
KR20160055930A (en) | Systems and methods for actively composing content for use in continuous social communication | |
CN106227792B (en) | Method and apparatus for pushed information | |
CN112883731B (en) | Content classification method and device | |
CN104462064A (en) | Method and system for prompting content input in information communication of mobile terminals | |
CN108268450B (en) | Method and apparatus for generating information | |
CN106407361A (en) | Method and device for pushing information based on artificial intelligence | |
CN111506794A (en) | Rumor management method and device based on machine learning | |
CN110554782A (en) | Expression input image synthesis method and system | |
CN108900612A (en) | Method and apparatus for pushed information | |
CN110084658A (en) | The matched method and apparatus of article | |
US11361759B2 (en) | Methods and systems for automatic generation and convergence of keywords and/or keyphrases from a media | |
CN106201010B (en) | Method for adding word bank and device | |
CN111354350B (en) | Voice processing method and device, voice processing equipment and electronic equipment | |
CN110990632A (en) | Video processing method and device | |
CN112632962B (en) | Method and device for realizing natural language understanding in man-machine interaction system | |
CN110399494A (en) | Method and apparatus for generating information |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |