CN106653006A - Search method and device based on voice interaction - Google Patents

Search method and device based on voice interaction Download PDF

Info

Publication number
CN106653006A
CN106653006A CN201611019821.5A CN201611019821A CN106653006A CN 106653006 A CN106653006 A CN 106653006A CN 201611019821 A CN201611019821 A CN 201611019821A CN 106653006 A CN106653006 A CN 106653006A
Authority
CN
China
Prior art keywords
participle
term
search
combination
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201611019821.5A
Other languages
Chinese (zh)
Other versions
CN106653006B (en
Inventor
郎文静
李裕东
朱群燕
石远
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN201611019821.5A priority Critical patent/CN106653006B/en
Publication of CN106653006A publication Critical patent/CN106653006A/en
Application granted granted Critical
Publication of CN106653006B publication Critical patent/CN106653006B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L2015/081Search algorithms, e.g. Baum-Welch or Viterbi

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Machine Translation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention provides a search method and device based on voice interaction. The search method based on voice interaction comprises the steps of acquiring a first search word according to voice data provided by a user and used for searching, and acquiring a plurality of second search words for context information of the first search word; generating multiple segmentation combinations according to the first segmentation result of the first search word and the second segmentation result of each second search word; according to the correlation score of each segmentation combination, selecting a target segmentation combination from the multiple segmentation combinations as a third search word; and searching according to the third search word. The method and the device for searching in combination with the context information of voice search improve the search accuracy and the user search experience.

Description

Searching method and device based on interactive voice
Technical field
The present invention relates to Internet technical field, more particularly to a kind of searching method and device based on interactive voice.
Background technology
With the continuous development of Internet technology, the input cost of keyboard is also constantly lifted therewith.Phonetic search, based on strong Big speech recognition capabilities, support quickly to initiate search by voice command, make search faster, more direct, more intelligent.It is related In technology, traditional phonetic search to receive instruction, feedback command pattern, in search every time, with the last time search by this search Rope and search for independent next time, i.e. in same retrieval sequence, equal onrelevant between search.
Under this mode, if user is carrying out when questioning closely or supplementing search of correlation to current term, need to repeat Requirement of main body simultaneously provides complete Search Requirement expression, it is impossible to enough scan for reference to the contextual information of phonetic search, in language It is not high that sound search expression searches for precision in the case of simplifying.
The content of the invention
It is contemplated that at least solving one of technical problem in correlation technique to a certain extent.
For this purpose, it is an object of the present invention to propose a kind of searching method based on interactive voice, can be with reference to voice The contextual information of search is scanned for, and lifts search precision, lifts user's search experience degree.
Further object is that proposing a kind of searcher based on interactive voice.
A further object of the present invention is to propose a kind of searcher based on interactive voice.
Further object is that proposing a kind of non-transitorycomputer readable storage medium.
Further object is that proposing a kind of computer program.
To reach above-mentioned purpose, the searching method based on interactive voice that first aspect present invention embodiment is proposed, including: First term is obtained according to the speech data for search that user provides, and is retrieved as the letter above of first term Multiple second terms of breath;According to the first participle result and the second participle of each the second term of first term As a result various participle combinations are generated;According to the relevance score of every kind of participle combination, from various participle combinations mesh is selected Mark participle group is incorporated as the 3rd term;Scanned for according to the 3rd term.
The searching method based on interactive voice that first aspect present invention embodiment is proposed, by according to the first term Second word segmentation result of first participle result and each the second term generates various participle combinations;According to the combination of every kind of participle Relevance score, from selection target participle group in the combination of various participles the 3rd term is incorporated as;Entered according to the 3rd term Line search, can scan for reference to the contextual information of phonetic search, lift search precision, lift user's search experience Degree.
To reach above-mentioned purpose, the searcher based on interactive voice that second aspect present invention embodiment is proposed, including: First acquisition module, the speech data for search for being provided according to user obtains the first term;Second acquisition module, For being retrieved as multiple second terms of the information above of first term;Generation module, for according to described first The first participle result of term and the second word segmentation result of each the second term generate various participle combinations;Selecting module, For the relevance score combined according to every kind of participle, from selection target participle group in various participles combinations the is incorporated as Three terms;Search module, for being scanned for according to the 3rd term.
The searcher based on interactive voice that second aspect present invention embodiment is proposed, by according to the first term Second word segmentation result of first participle result and each the second term generates various participle combinations;According to the combination of every kind of participle Relevance score, from selection target participle group in the combination of various participles the 3rd term is incorporated as;Entered according to the 3rd term Line search, can scan for reference to the contextual information of phonetic search, lift search precision, lift user's search experience Degree.
To reach above-mentioned purpose, the searcher based on interactive voice that third aspect present invention embodiment is proposed, its spy Levy and be, including:
Processor;
For storing the memory of processor executable;
Wherein, the processor is configured to:
First term is obtained according to the speech data for search that user provides, and is retrieved as first term Information above multiple second terms;
Generated according to the first participle result of first term and the second word segmentation result of each the second term many Plant participle combination;
According to the relevance score of every kind of participle combination, combine and make from selection target participle in various participle combinations For the 3rd term;
Scanned for according to the 3rd term.
The searcher based on interactive voice that third aspect present invention embodiment is proposed, by according to the first term Second word segmentation result of first participle result and each the second term generates various participle combinations;According to the combination of every kind of participle Relevance score, from selection target participle group in the combination of various participles the 3rd term is incorporated as;Entered according to the 3rd term Line search, can scan for reference to the contextual information of phonetic search, lift search precision, lift user's search experience Degree.
To reach above-mentioned purpose, the non-transitorycomputer readable storage medium that fourth aspect present invention embodiment is proposed, When the instruction in the storage medium is performed by the processor of mobile terminal so that mobile terminal is able to carry out one kind and is based on The searching method of interactive voice, methods described includes:
First term is obtained according to the speech data for search that user provides, and is retrieved as first term Information above multiple second terms;
Generated according to the first participle result of first term and the second word segmentation result of each the second term many Plant participle combination;
According to the relevance score of every kind of participle combination, combine and make from selection target participle in various participle combinations For the 3rd term;
Scanned for according to the 3rd term.
The non-transitorycomputer readable storage medium that fourth aspect present invention embodiment is proposed, by according to the first retrieval The first participle result of word and the second word segmentation result of each the second term generate various participle combinations;According to every kind of participle group The relevance score of conjunction, from selection target participle group in the combination of various participles the 3rd term is incorporated as;According to the 3rd retrieval Word is scanned for, and can be scanned for reference to the contextual information of phonetic search, lifts search precision, lifts user's search body Degree of testing.
To reach above-mentioned purpose, the computer program that fifth aspect present invention embodiment is proposed, when the computer When instruction processing unit in program product is performed, a kind of searching method based on interactive voice is performed, methods described includes:
First term is obtained according to the speech data for search that user provides, and is retrieved as first term Information above multiple second terms;
Generated according to the first participle result of first term and the second word segmentation result of each the second term many Plant participle combination;
According to the relevance score of every kind of participle combination, combine and make from selection target participle in various participle combinations For the 3rd term;
Scanned for according to the 3rd term.
The computer program that fifth aspect present invention embodiment is proposed, by the first participle according to the first term As a result various participle combinations are generated with the second word segmentation result of each the second term;Commented according to the correlation of every kind of participle combination Point, it is incorporated as the 3rd term from selection target participle group in the combination of various participles;Scanned for according to the 3rd term, energy Enough scan for reference to the contextual information of phonetic search, lift search precision, lift user's search experience degree.
The additional aspect of the present invention and advantage will be set forth in part in the description, and partly will become from the following description Obtain substantially, or recognized by the practice of the present invention.
Description of the drawings
The above-mentioned and/or additional aspect of the present invention and advantage will become from the following description of the accompanying drawings of embodiments It is substantially and easy to understand, wherein:
Fig. 1 is the schematic flow sheet of the searching method based on interactive voice that one embodiment of the invention is proposed;
Fig. 2 is the schematic flow sheet of the searching method based on interactive voice that another embodiment of the present invention is proposed;
Fig. 3 is the schematic flow sheet of the searching method based on interactive voice that another embodiment of the present invention is proposed;
Fig. 4 is phonetic search interface schematic diagram in the embodiment of the present invention;
Fig. 5 is the structural representation of the searcher based on interactive voice that one embodiment of the invention is proposed;
Fig. 6 is the structural representation of the searcher based on interactive voice that another embodiment of the present invention is proposed.
Specific embodiment
Embodiments of the invention are described below in detail, the example of the embodiment is shown in the drawings, wherein from start to finish Same or similar label represents same or similar element or the element with same or like function.Below with reference to attached The embodiment of figure description is exemplary, is only used for explaining the present invention, and is not considered as limiting the invention.Conversely, this Inventive embodiment includes all changes fallen in the range of the spirit and intension of attached claims, modification and is equal to Thing.
Fig. 1 is the schematic flow sheet of the searching method based on interactive voice that one embodiment of the invention is proposed.Language should be based on The searching method of sound interaction is configured in the searcher based on interactive voice illustrate.Searching for interactive voice should be based on Suo Fangfa can be applied in the search engine of electronic equipment, and search engine is the search information for referring to receiving user's input, And collect from internet and search information related information, after information is organized and processed, provide the user retrieval clothes Business, the system that user will be showed with search information related information.
Wherein, electronic equipment is, for example, PC (Personal Computer, PC), and cloud device or movement set It is standby, mobile device such as smart mobile phone, or panel computer etc..
Referring to Fig. 1, should be based on the searching method of interactive voice includes:
S11:First term is obtained according to the speech data for search that user provides, and is retrieved as the first term Information above multiple second terms.
Traditional phonetic search to receive instruction, feedback command pattern, in search every time, with the last time search by this search Rope and search for independent next time, i.e. in same retrieval sequence, equal onrelevant between search, if user is to current retrieval Word carries out when questioning closely or supplementing search of correlation, need to repeat requirement of main body and provide complete Search Requirement expression, it is impossible to be enough Scan for reference to the contextual information of phonetic search, search for precision in the case where phonetic search expression formula simplifies not high.
And in embodiments of the invention, understood with the deep layer demand of current search above by search, if historical search word In the initial participle of the second word segmentation result of first participle result and the second term comprising the first term when combining, then root Scan for according to the combination of initial participle, can effectively lift search efficiency;If combining without initial participle in historical search word, root The first participle and the second participle are combined into process according to the type information of the first term, various participle combinations are obtained, according to Related algorithm calculates the relevance score of above-mentioned every kind of participle combination, and the maximum participle of score value is combined as final retrieval Word is scanned for, and can be scanned for reference to the contextual information of phonetic search, lifts search precision, lifts user's search body Degree of testing.
In an embodiment of the present invention, the speech data for search that can be provided with receive user, obtains and voice number According to corresponding first text, and using corresponding first text as the first term.
Alternatively, user can click on the microphone button in search engine search box, be input into speech data, search engine In voice acquisition module collect the speech data, obtain the first text corresponding with speech data, and by corresponding first Text is used as the first term.
It is understood that user can relatively be questioned closely search to current search result or supplement search, when The information above of front first term can be multiple.
For example, the speech data that user is input in a search engine is " U.S. ", and Search Results displayed page provides the U.S. Relevant information, when user wants to inquire about the demographic data in the U.S., is input into speech data " population is how many ", with reference to information above, can Know, the first term is " population is how many ", the second term is " U.S. ", or, user can continue to question closely " where capital is In ", understanding with reference to information above, the first term is " where capital is ", and the second term is:" U.S. " and " have a large population It is few ".
S12:Generated according to the first participle result of the first term and the second word segmentation result of each the second term many Plant participle combination.
In an embodiment of the present invention, the first term and the second term can be carried out at cutting word using correlation technique Reason, obtains the morphological information of the first term and the second term, i.e. obtain first participle result and the second word segmentation result.
It should be noted that the first term not occurred for history, based on context feature, at cutting word Reason, can filter out the first identical participle, and can filter out phonetic identical fuzzy phoneme.
In an embodiment of the present invention, it can be determined that with the presence or absence of comprising first participle result and second in historical search word The initial participle combination of word segmentation result;If there is the combination of the initial participle comprising first participle result and the second word segmentation result, Initial participle is combined as the participle combination for generating;If not existing initial comprising first participle result and the second word segmentation result Participle is combined, then extract the fisrt feature information of each first participle in first participle result, and is extracted in the second word segmentation result The second feature information of each the second participle;According to fisrt feature information and second feature information, the class of the first term is determined Type information;The first participle and the second participle are combined by process according to the type information of the first term, various participles are obtained Combination.
In some embodiments, referring to Fig. 2, step S12 is specifically included:
S21:Judge in historical search word with the presence or absence of the initial participle comprising first participle result and the second word segmentation result Combination, if so, performs S22, otherwise, performs S23.
Alternatively, judge in historical search word with the presence or absence of initial point comprising first participle result and the second word segmentation result Word combination, when there is the combination of the initial participle comprising first participle result and the second word segmentation result in historical search word, directly Initial participle is combined as the participle combination for generating, scanned for according to the participle combination for generating, can effectively lift search Efficiency;When there is no the combination of the initial participle comprising first participle result and the second word segmentation result in historical search word, perform S23。
S22:Initial participle is combined as the participle combination for generating.
Alternatively, there is the combination of the initial participle comprising first participle result and the second word segmentation result in historical search word When, directly initial participle is combined as the participle combination for generating, scanned for according to the participle combination for generating, can effectively carry Rise search efficiency.
S23:The fisrt feature information of each first participle in first participle result is extracted, and is extracted in the second word segmentation result The second feature information of each the second participle.
In an embodiment of the present invention, fisrt feature information/second feature information includes at least one of:First point The click frequency, general colloquial style after entity type, weight, the user's history search frequency, the user's history search of word/the second participle Punishment weight, and editing distance.
Wherein, the entity type of the first participle/the second participle for the first participle/the second participle entity information, such as people Name, place name etc..
The weights size of the weight of the first participle/the second participle describes its weight in the first term/the second term The property wanted.
The user's history search frequency is used to determine the relevance score of the every kind of participle combination in subsequent step.
General colloquial punishment weight represents and the extensive normalization of colloquial style is carried out to the first participle/the second participle.By There are a large amount of colloquial style statements, such as " I wants to search ... " " what meaning is consulted ... is " in speech data, it is therefore desirable to Carry out the extensive normalization of colloquial style.Specifically, the first participle/the second participle is entered according to artificial constructed colloquial style vocabulary The extensive normalization of row colloquial style, one punishment weight of each participle correspondence in colloquial style vocabulary, punishment weight is bigger, shows The spoken words degree of participle is higher.
Editing distance is a kind of common similarity calculating method.
It is alternatively possible to extract the fisrt feature information of each first participle in first participle result, and extract second point The second feature information of each the second participle in word result, to determine the first inspection according to fisrt feature information and second feature information The type information of rope word.
S24:According to fisrt feature information and second feature information, the type information of the first term is determined.
Wherein, type information includes:First term be the second term supplement search first kind information, first Term is the Second Type information for questioning closely search of the second term.
In an embodiment of the present invention, first kind information is:Main body semanteme is constant, demand strong correlation, for example, " everyday Extremely run down load ", pronouns, general term for nouns, numerals and measure words " extremely running " everyday based on, " download " is demand word;Second Type information is:The semantic change of main body, demand It is semantic constant.
It is alternatively possible to by entity attribute knowledge base recognize the entity attribute of the first term and the second term into Point.It should be noted that in natural language the meaning of the same race can with different form of presentations, be also in entity attribute description as This, i.e., same attribute requirements have different form of presentations, and for example, " population how many ", " how many population ", " population is many It is few " be all " population " attribute requirements different expression, it is therefore desirable to taken out from different form of presentations by trunk extraction technique Take identical demand trunk.
Specifically, by judging the entity attribute information and multiple second terms of information above of current first term Entity attribute correlation, i.e., according to fisrt feature information and second feature information, and then judge that the first term is attribute Change or Entity Change.
For example, when the first term is:" that body weight", the second term is:When " height of Liu Dehua ", by reality Body attributes match, it is " Liu Dehua " to obtain entity, and " height " " body weight " is the attribute of " Liu Dehua ", therefore, it is known that first The type information of term is:Attribute is changed, and main body semanteme is constant, i.e. first kind information.
Or, when the first term is:" that Yao Ming", the second term is:When " height of Liu Dehua ", by reality Body attributes match, it is " Liu Dehua " " Yao Ming " to obtain entity, and their common attributes are " height ", therefore, it is known that the first inspection The type information of rope word is:The semantic change of main body, demand semanteme is constant, i.e. Second Type information.
S25:The first participle and the second participle are combined by process according to the type information of the first term, obtain various Participle is combined.
As a kind of example, when the first term be the second term question closely search Second Type information when, for example, When the first term is:" that France", the second term is:When " where the capital in the U.S. is ", the various participles for obtaining Combination can be as shown in table 1.
Table 1
It is initial comprising first participle result and the second word segmentation result by existing in historical search word in the present embodiment When participle is combined, initial participle is combined as the participle combination for generating, search efficiency can be effectively lifted, by searching in history When there is no initial participle comprising first participle result and the second word segmentation result in rope word and combining, according to fisrt feature information and Second feature information, determines the type information of the first term, according to the type information of the first term by the first participle and Two participles are combined process, obtain various participle combinations, carry out in order to subsequently maximum to relevance score participle combination Search, can make user express Search Requirement freely so that phonetic search interactive experience is more intelligent, lift search accurate Degree, lifts user's search experience degree.
In some embodiments, after step S12, also include:
S31:In determining various participle combinations, frequency of occurrence of every kind of participle combination in user's search history, Yi Jiyu Participle combines the corresponding first user historical search frequency of the first participle and the second user history of corresponding second participle is searched The rope frequency.
It is alternatively possible to assume that frequency of occurrence of the participle combination in user's search history is Nxyi, combine right with participle The first user historical search frequency of the first participle answered is Nxi, the second user of the second participle corresponding with participle combination goes through The history search frequency is Nyi, wherein, i=1,2 ..., M, M are the number of participle combination, and the total search frequency of user's history is N.
S32:Phase is determined according to frequency of occurrence, the first user historical search frequency, and the second user historical search frequency The scoring of closing property.
In an embodiment of the present invention, can be according to frequency of occurrence Nxyi, first user historical search frequency Nxi, second use Family historical search frequency Nyi, and preset formula determines relevance score Corr (i), wherein, preset formula is:
Corr (i)=Math.log10 (N/Nxi)*Math.log10(N/Nyi)*Nxyi/(Nxi+Nyi-Nxyi);
By determining phase according to frequency of occurrence, the first user historical search frequency, and the second user historical search frequency The scoring of closing property, scans in order to subsequently maximum to relevance score participle combination, and user can be made to express inspection freely Rope demand so that phonetic search interactive experience is more intelligent, lifts search precision, lifts user's search experience degree.
S13:According to the relevance score of every kind of participle combination, combine and make from selection target participle in the combination of various participles For the 3rd term.
In an embodiment of the present invention, the combination of relevance score highest participle can be selected from the combination of various participles to make Combine for target participle, and target participle is combined as the 3rd term.
For example, when the Second Type information for questioning closely search that the first term is the second term, and the first term are: " that France", the second term is:When " where the capital in the U.S. is ", every kind of participle in table 1 is obtained according to preset formula As shown in table 2 (the first term is the place of the first kind information for supplementing search of the second term to the relevance score of combination Reason process is similar to, and here is omitted).
Table 2
The combination of relevance score highest participle can be selected from the combination of various participles:The capital of France is where, to make Combine for target participle, and target participle is combined as the 3rd term, to be scanned for according to the 3rd term.
S14:Scanned for according to the 3rd term.
Used as a kind of example, referring to Fig. 4, Fig. 4 is phonetic search interface schematic diagram in the embodiment of the present invention, works as user input Speech data be " U.S. " (the second term) when, Search Results displayed page provide U.S.'s relevant information, when user wants to look into When asking the demographic data in the U.S., only speech data " population is how many " (the first term) need to be directly inputted, with reference to information above, this Embodiment can the term of automatic identification first be the second term supplement search first kind information, obtain the 3rd retrieval Word is " U.S. population is how many ", and triggers Search Results, meets user's request.
In the same manner, user can continue to question closely " where capital is " (the first term), understand with reference to information above, the second inspection Rope word is:" U.S. " and " population is how many ", the present embodiment can the term of automatic identification first be that the supplement of the second term is searched The first kind information of rope, obtains the 3rd term for " where U.S. capital is ", displaying Search Results.
Further, in order to meet the colloquial style speech expression of user, user can also naturally question closely " that France ", with reference to information above, it is that the second term questions closely the second of search that the present embodiment can automatically identify the first term Type information, obtains the 3rd term for " where the capital of France is ", displaying Search Results.
In the present embodiment, by the first participle result according to the first term and the second participle of each the second term As a result various participle combinations are generated;According to the relevance score of every kind of participle combination, from selection target point in the combination of various participles Phrase is incorporated as the 3rd term;Scanned for according to the 3rd term, can be entered with reference to the contextual information of phonetic search Line search, lifts search precision, lifts user's search experience degree.
Fig. 5 is the structural representation of the searcher based on interactive voice that one embodiment of the invention is proposed.Language should be based on The searcher 500 of sound interaction can pass through being implemented in combination in for software, hardware or both.
Referring to Fig. 5, should be based on the searcher 500 of interactive voice can include:First acquisition module 510, second is obtained Module 520, generation module 530, selecting module 540, and search module 550.Wherein,
First acquisition module 510, the speech data for search for being provided according to user obtains the first term.
Alternatively, the first acquisition module 510 specifically for:The speech data for search that receive user is provided, obtains The first text corresponding with speech data, and using corresponding first text as the first term.
Second acquisition module 520, for being retrieved as multiple second terms of the information above of the first term.
Generation module 530, for second point according to the first participle result of the first term and each the second term Word result generates various participle combinations.
Selecting module 540, for the relevance score combined according to every kind of participle, from selection target in the combination of various participles Participle group is incorporated as the 3rd term.
Alternatively, selecting module 540 specifically for:Relevance score highest participle group is selected from the combination of various participles Cooperate to be combined for target participle, and target participle is combined as the 3rd term.
Search module 550, for being scanned for according to the 3rd term.
In some embodiments, referring to Fig. 6, should be based on the searcher 500 of interactive voice can also include:
Alternatively, generation module 530 includes:
Judging submodule 531, for judging historical search word in the presence or absence of comprising first participle result and the second participle As a result initial participle combination.
First processes submodule 532, for there is the initial participle comprising first participle result and the second word segmentation result During combination, initial participle is combined as the participle combination for generating.
Extracting sub-module 533, for there is no the initial participle group comprising first participle result and the second word segmentation result During conjunction, extract the fisrt feature information of each first participle in first participle result, and extract in the second word segmentation result each the The second feature information of two participles.
Alternatively, fisrt feature information/second feature information includes at least one of:
Point after entity type, weight, the user's history search frequency, the user's history search of the first participle/the second participle Hit the frequency, general colloquial punishment weight, and editing distance.
Determination sub-module 534, for according to fisrt feature information and second feature information, determining the type of the first term Information, wherein, type information includes:First term is the first kind information for supplementing search of the second term, the first inspection Rope word is the Second Type information for questioning closely search of the second term.
Second processing submodule 535, for being entered the first participle and the second participle according to the type information of the first term Row combined treatment, obtains various participle combinations.
First determining module 560, in determining that various participles are combined, every kind of participle is combined in user's search history Frequency of occurrence, and the first user historical search frequency with the participle corresponding first participle of combination and corresponding second participle The second user historical search frequency.
Second determining module 570, for being gone through according to frequency of occurrence, the first user historical search frequency, and second user The history search frequency determines relevance score.
It should be noted that to the explanation of the searching method embodiment based on interactive voice in earlier figures 1- Fig. 4 embodiments The searcher 500 based on interactive voice for being also applied for the embodiment is illustrated, it realizes that principle is similar to, and here is omitted.
In the present embodiment, by the first participle result according to the first term and the second participle of each the second term As a result various participle combinations are generated;According to the relevance score of every kind of participle combination, from selection target point in the combination of various participles Phrase is incorporated as the 3rd term;Scanned for according to the 3rd term, can be entered with reference to the contextual information of phonetic search Line search, lifts search precision, lifts user's search experience degree.
It should be noted that in describing the invention, term " first ", " second " etc. are not only used for describing purpose, and not It is understood that to indicate or implying relative importance.Additionally, in describing the invention, unless otherwise stated, the implication of " multiple " It is two or more.
In flow chart or here any process described otherwise above or method description are construed as, expression includes It is one or more for realizing specific logical function or process the step of the module of code of executable instruction, fragment or portion Point, and the scope of the preferred embodiment of the present invention includes other realization, wherein can not press shown or discussion suitable Sequence, including according to involved function by it is basic simultaneously in the way of or in the opposite order, carry out perform function, this should be of the invention Embodiment person of ordinary skill in the field understood.
It should be appreciated that each several part of the present invention can be realized with hardware, software, firmware or combinations thereof.Above-mentioned In embodiment, the software that multiple steps or method can in memory and by suitable instruction execution system be performed with storage Or firmware is realizing.For example, if realized with hardware, and in another embodiment, can be with well known in the art Any one of row technology or their combination are realizing:With for realizing the logic gates of logic function to data-signal Discrete logic, the special IC with suitable combinational logic gate circuit, programmable gate array (PGA), scene Programmable gate array (FPGA) etc..
Those skilled in the art are appreciated that to realize all or part of step that above-described embodiment method is carried Suddenly the hardware that can be by program to instruct correlation is completed, and described program can be stored in a kind of computer-readable storage medium In matter, the program upon execution, including one or a combination set of the step of embodiment of the method.
Additionally, each functional unit in each embodiment of the invention can be integrated in a processing module, it is also possible to It is that unit is individually physically present, it is also possible to which two or more units are integrated in a module.Above-mentioned integrated mould Block both can be realized in the form of hardware, it would however also be possible to employ the form of software function module is realized.The integrated module is such as Fruit is realized and as independent production marketing or when using using in the form of software function module, it is also possible to be stored in a computer In read/write memory medium.
Storage medium mentioned above can be read-only storage, disk or CD etc..
In the description of this specification, reference term " one embodiment ", " some embodiments ", " example ", " specifically show The description of example " or " some examples " etc. means to combine specific features, structure, material or spy that the embodiment or example are described Point is contained at least one embodiment of the present invention or example.In this manual, to the schematic representation of above-mentioned term not Necessarily refer to identical embodiment or example.And, the specific features of description, structure, material or feature can be any One or more embodiments or example in combine in an appropriate manner.
Although embodiments of the invention have been shown and described above, it is to be understood that above-described embodiment is example Property, it is impossible to limitation of the present invention is interpreted as, one of ordinary skill in the art within the scope of the invention can be to above-mentioned Embodiment is changed, changes, replacing and modification.

Claims (12)

1. a kind of searching method based on interactive voice, it is characterised in that comprise the following steps:
First term is obtained according to the speech data for search that user provides, and is retrieved as the upper of first term Multiple second terms of literary information;
Various points are generated according to the first participle result of first term and the second word segmentation result of each the second term Word combination;
According to the relevance score of every kind of participle combination, from selection target participle group in the various participles combination the is incorporated as Three terms;
Scanned for according to the 3rd term.
2. the searching method of interactive voice is based on as claimed in claim 1, it is characterised in that described according to the described first retrieval The first participle result of word and the second word segmentation result of each the second term generate various participle combinations, including:
Judge in historical search word with the presence or absence of the initial participle comprising the first participle result and second word segmentation result Combination;
If there is the combination of the initial participle comprising the first participle result and second word segmentation result, will be described first Beginning participle is combined as the participle combination for generating;
If there is no the combination of the initial participle comprising the first participle result and second word segmentation result, institute is extracted The fisrt feature information of each first participle in first participle result is stated, and is extracted each second point in second word segmentation result The second feature information of word;
According to the fisrt feature information and the second feature information, the type information of first term is determined, wherein, The type information includes:First term is the first kind information, described for supplementing search of second term First term is the Second Type information for questioning closely search of second term;
The first participle and second participle are combined by process according to the type information of first term, are obtained Various participle combinations.
3. the searching method of interactive voice is based on as claimed in claim 2, it is characterised in that the fisrt feature information/institute Second feature information is stated including at least one of:
After entity type, weight, the user's history search frequency, the user's history search of the first participle/second participle The click frequency, general colloquial punishment weight, and editing distance.
4. the searching method of interactive voice is based on as claimed in claim 1, it is characterised in that described according to every kind of participle group Before the relevance score of conjunction, also include:
In determining the various participles combination, frequency of occurrence of every kind of participle combination in user's search history, and with it is described Participle combines the corresponding first user historical search frequency of the first participle and the second user history of corresponding second participle is searched The rope frequency;
It is true according to the frequency of occurrence, the first user historical search frequency, and the second user historical search frequency The fixed relevance score.
5. the searching method of interactive voice is based on as claimed in claim 1, it is characterised in that described from various participle groups Selection target participle group is incorporated as the 3rd term in conjunction, including:
The relevance score highest participle combination is selected from various participle combinations as target participle combination, And the target participle is combined as the 3rd term.
6. the searching method of interactive voice is based on as claimed in claim 1, it is characterised in that the use provided according to user The first term is obtained in the speech data of search, including:
The speech data for search that receive user is provided, obtains corresponding with the speech data the first text, and by institute Corresponding first text is stated as first term.
7. a kind of searcher based on interactive voice, it is characterised in that include:
First acquisition module, the speech data for search for being provided according to user obtains the first term;
Second acquisition module, for being retrieved as multiple second terms of the information above of first term;
Generation module, for being tied according to the second participle of the first participle result of first term and each the second term Fruit generates various participle combinations;
Selecting module, for the relevance score combined according to every kind of participle, from selection target point in various participle combinations Phrase is incorporated as the 3rd term;
Search module, for being scanned for according to the 3rd term.
8. the searcher of interactive voice is based on as claimed in claim 7, it is characterised in that the generation module includes:
Judging submodule, for judging historical search word in the presence or absence of comprising the first participle result and second participle As a result initial participle combination;
First processes submodule, for exist it is described initial comprising the first participle result and second word segmentation result When participle is combined, the initial participle is combined as the participle combination for generating;
Extracting sub-module, for there is no include the first participle result and second word segmentation result initial point During word combination, the fisrt feature information of each first participle in the first participle result is extracted, and extract second participle As a result the second feature information of each the second participle in;
Determination sub-module, for according to the fisrt feature information and the second feature information, determining first term Type information, wherein, the type information includes:First term is that second term supplements the of search One type information, first term are the Second Type information for questioning closely search of second term;
Second processing submodule, for according to the type information of first term by the first participle and described second point Word is combined process, obtains various participle combinations.
9. the searcher of interactive voice is based on as claimed in claim 8, it is characterised in that the fisrt feature information/institute Second feature information is stated including at least one of:
After entity type, weight, the user's history search frequency, the user's history search of the first participle/second participle The click frequency, general colloquial punishment weight, and editing distance.
10. the searcher of interactive voice is based on as claimed in claim 7, it is characterised in that also included:
First determining module, in determining that various participles are combined, every kind of participle combines going out in user's search history The existing frequency, and the first user historical search frequency and corresponding second participle with the participle corresponding first participle of combination The second user historical search frequency;
Second determining module, for according to the frequency of occurrence, the first user historical search frequency, and described second uses The family historical search frequency determines the relevance score.
11. searchers based on interactive voice as claimed in claim 7, it is characterised in that the selecting module is specifically used In:
The relevance score highest participle combination is selected from various participle combinations as target participle combination, And the target participle is combined as the 3rd term.
12. searchers based on interactive voice as claimed in claim 7, it is characterised in that the first acquisition module tool Body is used for:
The speech data for search that receive user is provided, obtains corresponding with the speech data the first text, and by institute Corresponding first text is stated as first term.
CN201611019821.5A 2016-11-17 2016-11-17 Searching method and device based on interactive voice Active CN106653006B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611019821.5A CN106653006B (en) 2016-11-17 2016-11-17 Searching method and device based on interactive voice

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611019821.5A CN106653006B (en) 2016-11-17 2016-11-17 Searching method and device based on interactive voice

Publications (2)

Publication Number Publication Date
CN106653006A true CN106653006A (en) 2017-05-10
CN106653006B CN106653006B (en) 2019-11-08

Family

ID=58807746

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611019821.5A Active CN106653006B (en) 2016-11-17 2016-11-17 Searching method and device based on interactive voice

Country Status (1)

Country Link
CN (1) CN106653006B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107133345A (en) * 2017-05-22 2017-09-05 北京百度网讯科技有限公司 Exchange method and device based on artificial intelligence
CN107608957A (en) * 2017-09-06 2018-01-19 百度在线网络技术(北京)有限公司 Text modification method, apparatus and its equipment based on voice messaging
CN108538291A (en) * 2018-04-11 2018-09-14 百度在线网络技术(北京)有限公司 Sound control method, terminal device, cloud server and system
CN112259096A (en) * 2020-10-23 2021-01-22 海信视像科技股份有限公司 Voice data processing method and device

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101075435A (en) * 2007-04-19 2007-11-21 深圳先进技术研究院 Intelligent chatting system and its realizing method
CN101140587A (en) * 2007-10-15 2008-03-12 深圳市迅雷网络技术有限公司 Searching method and apparatus
CN101281745A (en) * 2008-05-23 2008-10-08 深圳市北科瑞声科技有限公司 Interactive system for vehicle-mounted voice
US20110144995A1 (en) * 2009-12-15 2011-06-16 At&T Intellectual Property I, L.P. System and method for tightly coupling automatic speech recognition and search
CN103369398A (en) * 2013-07-01 2013-10-23 安徽广电信息网络股份有限公司 Voice searching method and voice searching system based on television EPG (electronic program guide) information
CN103995870A (en) * 2014-05-21 2014-08-20 百度在线网络技术(北京)有限公司 Interactive searching method and device
CN103995880A (en) * 2014-05-27 2014-08-20 百度在线网络技术(北京)有限公司 Interactive searching method and device
CN104102723A (en) * 2014-07-21 2014-10-15 百度在线网络技术(北京)有限公司 Search content providing method and search engine
CN104239459A (en) * 2014-09-02 2014-12-24 百度在线网络技术(北京)有限公司 Voice search method, voice search device and voice search system
CN104714954A (en) * 2013-12-13 2015-06-17 中国电信股份有限公司 Information searching method and system based on context understanding
CN105279227A (en) * 2015-09-11 2016-01-27 百度在线网络技术(北京)有限公司 Voice search processing method and device of homonym

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101075435A (en) * 2007-04-19 2007-11-21 深圳先进技术研究院 Intelligent chatting system and its realizing method
CN101140587A (en) * 2007-10-15 2008-03-12 深圳市迅雷网络技术有限公司 Searching method and apparatus
CN101281745A (en) * 2008-05-23 2008-10-08 深圳市北科瑞声科技有限公司 Interactive system for vehicle-mounted voice
US20110144995A1 (en) * 2009-12-15 2011-06-16 At&T Intellectual Property I, L.P. System and method for tightly coupling automatic speech recognition and search
CN103369398A (en) * 2013-07-01 2013-10-23 安徽广电信息网络股份有限公司 Voice searching method and voice searching system based on television EPG (electronic program guide) information
CN104714954A (en) * 2013-12-13 2015-06-17 中国电信股份有限公司 Information searching method and system based on context understanding
CN103995870A (en) * 2014-05-21 2014-08-20 百度在线网络技术(北京)有限公司 Interactive searching method and device
CN103995880A (en) * 2014-05-27 2014-08-20 百度在线网络技术(北京)有限公司 Interactive searching method and device
CN104102723A (en) * 2014-07-21 2014-10-15 百度在线网络技术(北京)有限公司 Search content providing method and search engine
CN104239459A (en) * 2014-09-02 2014-12-24 百度在线网络技术(北京)有限公司 Voice search method, voice search device and voice search system
CN105279227A (en) * 2015-09-11 2016-01-27 百度在线网络技术(北京)有限公司 Voice search processing method and device of homonym

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107133345A (en) * 2017-05-22 2017-09-05 北京百度网讯科技有限公司 Exchange method and device based on artificial intelligence
US10664504B2 (en) 2017-05-22 2020-05-26 Beijing Baidu Netcom Science And Technology Co., Ltd. Interaction method and apparatus based on artificial intelligence
CN107133345B (en) * 2017-05-22 2020-11-06 北京百度网讯科技有限公司 Interaction method and device based on artificial intelligence
CN107608957A (en) * 2017-09-06 2018-01-19 百度在线网络技术(北京)有限公司 Text modification method, apparatus and its equipment based on voice messaging
CN108538291A (en) * 2018-04-11 2018-09-14 百度在线网络技术(北京)有限公司 Sound control method, terminal device, cloud server and system
US11127398B2 (en) 2018-04-11 2021-09-21 Baidu Online Network Technology (Beijing) Co., Ltd. Method for voice controlling, terminal device, cloud server and system
CN112259096A (en) * 2020-10-23 2021-01-22 海信视像科技股份有限公司 Voice data processing method and device

Also Published As

Publication number Publication date
CN106653006B (en) 2019-11-08

Similar Documents

Publication Publication Date Title
KR101581816B1 (en) Voice recognition method using machine learning
CN111191022B (en) Commodity short header generation method and device
EP2717176A2 (en) Method for searching for information using the web and method for voice conversation using same
CN102867512A (en) Method and device for recognizing natural speech
CN106653006B (en) Searching method and device based on interactive voice
US8583417B2 (en) Translation device and computer program product
KR20130082835A (en) Method and appartus for providing contents about conversation
CN102867511A (en) Method and device for recognizing natural speech
CN111414763A (en) Semantic disambiguation method, device, equipment and storage device for sign language calculation
Akhtiamov et al. Speech and Text Analysis for Multimodal Addressee Detection in Human-Human-Computer Interaction.
US20130173251A1 (en) Electronic device and natural language analysis method thereof
CN107748784A (en) A kind of method that structured data searching is realized by natural language
KR101410601B1 (en) Spoken dialogue system using humor utterance and method thereof
CN110808032A (en) Voice recognition method and device, computer equipment and storage medium
US20180039632A1 (en) Information processing apparatus, information processing method, and computer program
KR101695014B1 (en) Method for building emotional lexical information and apparatus for the same
CN108345612A (en) A kind of question processing method and device, a kind of device for issue handling
CN104182381A (en) character input method and system
CN108710653B (en) On-demand method, device and system for reading book
Henderson et al. PolyResponse: A rank-based approach to task-oriented dialogue with application in restaurant search and booking
CN102970618A (en) Video on demand method based on syllable identification
CN109063182B (en) Content recommendation method based on voice search questions and electronic equipment
CN113361252A (en) Text depression tendency detection system based on multi-modal features and emotion dictionary
CN112765977A (en) Word segmentation method and device based on cross-language data enhancement
CN109684357B (en) Information processing method and device, storage medium and terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant