CN103201707A - System and method for inputting text into electronic devices - Google Patents

System and method for inputting text into electronic devices Download PDF

Info

Publication number
CN103201707A
CN103201707A CN2011800532559A CN201180053255A CN103201707A CN 103201707 A CN103201707 A CN 103201707A CN 2011800532559 A CN2011800532559 A CN 2011800532559A CN 201180053255 A CN201180053255 A CN 201180053255A CN 103201707 A CN103201707 A CN 103201707A
Authority
CN
China
Prior art keywords
sequence
model
prediction
probability
text
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2011800532559A
Other languages
Chinese (zh)
Other versions
CN103201707B (en
Inventor
本杰明·麦德洛克
道格拉斯·亚历山大·哈珀·欧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Microsoft Technology Licensing LLC
Original Assignee
Touchtype Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Touchtype Ltd filed Critical Touchtype Ltd
Priority to CN201710761127.9A priority Critical patent/CN107506047B/en
Publication of CN103201707A publication Critical patent/CN103201707A/en
Application granted granted Critical
Publication of CN103201707B publication Critical patent/CN103201707B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • G06F3/023Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
    • G06F3/0233Character input methods
    • G06F3/0237Character input methods using prediction or retrieval techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/274Converting codes to words; Guess-ahead of partial word inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • G06F3/023Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
    • G06F3/0233Character input methods
    • G06F3/0236Character input methods using selection techniques to select from displayed items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Artificial Intelligence (AREA)
  • Machine Translation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

A text prediction engine, a system comprising a text prediction engine, and a method for generating sequence predictions. The text prediction engine, system and method generate a final set of sequence predictions, each with an associated probability value.

Description

Be used for text prediction engine, system and method to the electronic equipment input text
Technical field
The present invention relates generally to a kind of for text prediction engine, system and method to the electronic equipment input text.
Background technology
Some existing inventions utilize multiple different technologies that the improvement method of electronic equipment user version input is provided.Yet well-known, disclosed related system at first is faced with the problem of using stable and fully integrated probability model predictive user expection to write text.
Summary of the invention
In a first aspect of the present invention, a kind of text prediction engine is provided, comprising: at least one model, it is used for having first group of sequence that dependent probability is estimated from the generation of evidence source; The probability maker, it is used for receiving the described first group of sequence that has the dependent probability appraisal and generates one group of sequence prediction that has the dependent probability value; Wherein, received under the situation of all possible sequence the described probable value of normalization on all possible sequence prediction that is generated by described probability maker by described probability maker given.
Preferably, described text prediction engine comprises priority model, and it is used for generating the second group of sequence that has the dependent probability appraisal.
Preferably, described model generates first group of sequence according to the uncertainty in described evidence source and the described evidence source.Preferably, described probability maker is used for receiving described first, second group sequence that has the dependent probability appraisal.
Described probability maker preferably by n most probable sequence prediction represented the constant addition with the probable value that remains possible sequence prediction, is estimated the normalization factor of described probable value.Described constant is represented the probable value by the remaining possibility sequence prediction of described model and the generation of described priority model.
Described model comprises a plurality of models that have a plurality of first group of sequence of dependent probability appraisal for generation.In one embodiment, described a plurality of model generates a plurality of first group of sequence according to a plurality of evidences source.
Preferably, described text prediction engine is the part of a certain system, and described user input text by one or more users select, character input or speech recognition be input in this system.
Described text prediction engine comprises given linguistic context sequence according to corresponding model probability is weighted the probable value of described sequence prediction.In one embodiment, described a plurality of model comprises a plurality of language models corresponding with multiple different language; And the probable value of the described text prediction engine pair sequence prediction corresponding with the language model of the most probable language that relates to user input text is carried out five-star weighting.
Each evidence source is moulded by the corresponding model that is used for generating the sequence that has the dependent probability appraisal.Under the situation of given described sequence prediction, described probability maker preferably with each evidence source as other on evidence the separate component of having ready conditions in source handle.
In a preferred embodiment of described text prediction engine, described model comprises context model and input model, and described context model is used for receiving the text of user's input and generates one group of sequence with described input model to be estimated with relevant probability; And described priority model comprises for generating the target priority model that one group of sequence and dependent probability are estimated.Described input model preferably includes candidate's model and language model.Described context model preferably includes candidate's model and prefix matching model.Described target priority model preferably includes character model and a meta-model.
In a second aspect of the present invention, a kind of system is provided, comprising: user interface, it is used for receiving the text by user's input; Text prediction engine, it be used for to receive from the described text of described user interface input and generates one group of sequence prediction that has the dependent probability value, wherein, all possible sequence prediction on the described probable value of normalization; Wherein, described text prediction engine also is used for providing described sequence prediction to described user interface.
Preferably, described input model comprises candidate's model and language model.Preferably, described context model comprises candidate's model and prefix matching model.Preferably, described target priority model comprises character model and a meta-model.
In a third aspect of the present invention, a kind of method of process user text input is provided, comprising: receive the text that inputs to user interface; Use text prediction engine to generate one group of sequence prediction and relevant probable value, wherein, the described probable value of normalization on all possible sequence prediction; Described sequence prediction is offered described user interface.
The step that generates the normalization probable value preferably includes: by the probable value of n most probable sequence prediction is represented the constant addition with remaining probable value that may sequence prediction, estimate the normalization factor of described probable value.
This method also comprises: described sequence prediction is presented on the described user interface selects for the user.Preferably, by described text prediction engine described sequence prediction is sorted, show in order for described user interface.Only when the corresponding probable value of described sequence prediction during more than or equal to first threshold values, described sequence prediction is offered described user interface.Similarly, only when the corresponding probable value of described sequence prediction during more than or equal to first threshold values, said system offers described user interface with described sequence prediction.
Preferably, at least one in the described sequence prediction is equivalent to be inputed to by the user adjustment or the invulnerable release of the text of described user interface.
Described method also comprises: input has the sequence prediction greater than second threshold values or the probable value on second threshold values automatically.Similarly, in one embodiment, said system is imported the sequence prediction that has greater than second threshold values or the probable value on second threshold values automatically.
The probability maker that uses in this method preferably includes for generating a plurality of models that the corresponding model of one group of sequence prediction and dependent probability value and basis comprises the described probable value of probability weight of given linguistic context sequence.
The present invention also provides a kind of computer program, comprising: computer-readable medium wherein stores be used to making processor carry out the computer program of said method.
The invention still further relates to a kind of text prediction engine for the formation sequence prediction and be used for the formation sequence prediction for demonstration and user selected system and method.In one embodiment, the present invention relates to correct mistakes automatically list entries system and realize the method for this correction.In a preferred embodiment, the invention provides a text prediction engine and generate the one group of ultimate sequence prediction that has the dependent probability value by estimating in conjunction with the different probability of an arbitrarily sequence expection.Text prediction engine of the present invention, system and method can provide the prediction based on any corroboration source thus.Can be by distributing correct probability to each expected sequence but not be the sequence rank, realization this purpose.By distributing correct probable value, but analysis of allocated is given the evolution of the probability of different entries, and can compare given entry on two different time points or the probability of one group of entry.This means in particular prediction under given pre-set threshold value " trust " situation (confidence), can use pre-set threshold value to come the behavior of regulating system.For instance, only show the sequence predict, if perhaps the accuracy of system estimation probability surpasses 0.75, or the sequence that in other words dopes have at least 75% may be accurately, revise this moment automatically.If use certain special evaluation to come to be the element rank, such as those values that can't reliably compare between the sequence on the time point, then this reasoning is irrealizable.
In order to generate correct probable value, the present invention preferably provide the normalization of a kind of effective estimation all sequences to add and method.
Below with reference to following accompanying drawing, introduce the present invention in detail.
Description of drawings
Fig. 1 is the synoptic diagram of the senior predict of the present invention;
Fig. 2 is the synoptic diagram of the preferred predict example of the present invention.
Embodiment
Definition:
● the specification unit symbol that character-expression is basic;
● the finite aggregate of character set-character;
● sequence-sorted finite length character string;
● if the initial character of prefix-originate in each sequence is identical, and has continuous mapping one by one and length (s)≤length (s'), and then sequence s is the prefix of another sequence s ';
● proper prefix-as infructescence s is the prefix of sequence s ', and length (s)<length (s'), and then sequence s is the proper prefix of another sequence s ';
● language-a group (being generally limited) is specifically write or the order character of oral statement;
● text-from one or more language, extract write data;
● system-theme of the present invention;
● the user-with the mutual personnel of said system.
Generally but not unique, can implement system of the present invention referring to Fig. 1.Fig. 1 is the calcspar of advanced text predict of the present invention.Native system comprises one group of most probable sequence prediction S that expects input by the user of generation FText prediction engine.Each sequence prediction has a relative probable value.
As shown in Figure 1, text prediction engine preferably includes being used for from a plurality of evidences source e of a plurality of training 1, e 2, e 3, e 4Deng in make probability inference model M 1, M 2, M 3And probability maker (PG, probability generator).Yet, in other embodiments, a model of training and an evidence source can be arranged.
The potential evidence source e of more existing any types 1, e 2Deng.The example in these evidence sources comprises:
● the sequence that the user has imported;
● the current entry/phrase imported of user;
● the historical series by user's input of storage;
● user's mother tongue;
● the specific type language of input;
● positive input has the application of current sequence;
● the target message recipient in message transmission environment;
● time/date;
● as the position of the equipment of native system main frame.
Universal model
The target of native system is, expects that according to the user possibility of a certain sequence of input is the sequence rank in the given language subset.In Probability, this is equivalent to grade for the sequence in the S set, and this set is arranged by following expression formula:
P(s∈S|e,M) (1)
Wherein, e is the evidence that observes, and the trained model set of M for being used for making probability inference.In other words, native system will be in the set of all sequences that can extract prediction the evaluation condition probability.Target sequence is expressed as s.
In order to simplify the anabolic process of the prediction that comes from the different pieces of information source, in a preferred embodiment, target sequence s is defined as the prediction that comes from particular source.
Each model among the M is trained on particular source.Therefore, can be by the model representation particular source among the M, and the S set in the expression formula (1) relates to all different entries (or sequence) that generated by the model among the M.Provide the prediction entry by interrogation model.This entry is relevant with its source model.Because this entry is relevant with its source model, so it is different from and comes from other models but the entry of morphology unanimity.This relevant can being implied in the above-mentioned data.Yet this entry can be marked with the identifier relevant with this entry source model.
Predict preferably that at this two other the identical prediction that will come from the different pieces of information source is considered as difference in conjunction with in handling.Come from the sequence of different models for combination to obtain predicting list, by removing the prediction of repetition, this sequence is simply sorted.In this preferred operations, given morphology entry/sequence is carried out the most probable assessment, and abandon (least possibility) entry that any morphology repeats.
Be example with a limiting examples, if M comprises two linguistic context language models, French (LM French) and English (LM English), entry " pain " may appear in these two models, and occurs twice in S, once be related with the French model, and another time is related with the English model.Like this under the situation of given one group of special evidence (wherein, in this case, this evidence is prediction entry " pain " linguistic context before), entry " pain " carried out the assessment of opening in twice minute.
These assessments relate to two different sequences (is come from the French model, and is come from the English model), but because its morphology is identical, need not all to present to the user.Therefore, according to this preferred embodiment, given morphology sequence is carried out the most probable assessment, and discard any morphology dittograph bar.
In order to expect that according to the user possibility of a certain sequence of input is the sequence rank in the given language subset, need in the calculation expression (1) conditional probability P (s ∈ S|e, M).In order to determine this probability, it is as follows to use Bayesian formula (Bayes ' rule) to rearrange this expression formula:
P ( e | s , M ) P ( s | M ) P ( e | M ) - - - ( 2 )
And the target sequence in this expression formula denominator of marginalisation, obtain thus:
P ( e | s , M ) P ( s | M ) Σ j = 1 | S | P ( e | s j , M ) P ( s j | M ) - - - ( 3 )
In a preferred embodiment, (e|s M), supposes: under the situation of given above-mentioned target sequence, evidence can be split into according to correlation model [M in order to calculate P 1M N] under the non-overlapped set [e that obtains respectively of a certain distribution 1E N].This independence assumption can be write as:
P ( e | s , M ) = Π i = 1 N [ P ( e i | s , M i ) ] - - - ( 4 )
And be expressed as follows:
Suppose 1: under the situation of given target sequence, evidence is split as different set, thereby makes the evidence in each set separate conditionally to each other.
Wherein, each e iHas relative model M iLike this, can make up a framework, in this framework, can any evidence source be combined in the high mode of counting yield.In a preferred embodiment, model R ∈ M is preferentially relevant with target sequence.Consider this hypothesis, it is as follows that we can explain expression formula (3) again:
P ( s | R ) Π i = 1 N P ( e i | s , M i ) Σ j = 1 | S | P ( s j R ) Π i = 1 N P ( e i | s j , M i ) - - - ( 5 )
Therefore, in a preferred embodiment, can be by calculating the preferential P of target sequence (s|R) and each evidential probability P (e i| s, M i), calculate the conditional probability of expression formula (1).
Therefore denominator in the expression formula (5) is steady state value with respect to s, can not have influence on rank, and it is the normalization factor of the probable value that calculates or rather.In a preferred embodiment, this steady state value is estimated as subclass and a constant sum of most probable sequence, has to calculate the expression formula 13-15 of S(in vide infra to overcome) in the problem of conditional probability of all sequences.Because the Zipfian(Qi Pufu of some elephants of speaking naturally distributes) characteristic, this method is rationalized, the minority probability event has most probability mass in this characteristic.It is the example that power law distributes that Qi Pufu distributes, and wherein, the frequency of given event and its rank roughly are inversely proportional to.
Expression formula (5) provides and will import the principle method that the different evidences source of purpose combines about text, and at the given e of source on evidence 1, e 2... situation under by one group of model R that trained, M 1, M 2... generate one group of sequence S R, S 1, S 2... the conditional probability value P relevant with one group R, P 1, P 2... implement optimum decision system of the present invention.Model R is used for calculating priority target sequence probability P (s|R), each model M simultaneously 1, M 2... calculate each evidential probability P (e i| s, M i).Each model is exported one group of sequence S iWith one group of relevant conditional probability P iEach model M 1, M 2... comprise one or more submodels.Probability maker PG with sequence and relevant conditional probability as input, and export one group with probable value P FRelevant ultimate sequence S FProbability maker PG can combine prediction as above-mentioned preferred process, that is to say, will predict according to probability order rank, and only delete and repeat prediction arbitrarily.With final probable value P FRelevant sequence S FCan be presented on the user interface of native system with the form such as tabulation, browse and select for the user.The user can select by making prediction or operation be equipped with the device of native system by other means, carry out alternately with native system, so fresh evidence more.When text inputs to native system, renewable each model R, M 1M N
What the invention provides passes through two kinds of method for optimizing that marginalisation is presented on the evidential probability in the evidence candidate lexical or textual analysis calculating probability framework in the graphics frame, although can also use additive method.Below, will introduce this two kinds of method for optimizing.
Candidate's model 1
The probability that comes from the evidence in a certain evidence source in formation is estimated P (e i| s, M i) time, help usually to come representation model according to the candidate as the intergrade between ' user's expection ' sequence and the evidence that observes.If come representation model according to the candidate, probability P (e then i| s, M i) can be expressed as again:
P ( e | s , M ) = Σ j = 1 K P ( e | c j , s , M candidate ) P ( c j | s , M sequence ) - - - ( 6 )
Wherein, c jBe a certain candidate, and have two submodels of the M in given evidence source: candidate's model M CandidateWith series model M SequenceCritical assumptions herein are as follows:
Suppose 2: under given candidate's situation, the probability under the above-mentioned model can be expressed as candidate's marginalisation, and wherein, evidence is independent of target sequence conditionally.
Use this hypothesis, can from the evidence entry, delete the dependence for s:
P ( e | s , M ) = Σ j = 1 K P ( e | c j , s , M candidate ) P ( c j | s , M sequence ) - - - ( 7 )
The attribute of candidate's model can also be encoded into the form of graphical model.This graphical model has been described the relation as follows between variable and the model:
Figure BDA00003141303200091
Candidate's model 2
Another variant of candidate's model at first uses Bayesian formula conversion evidential probability:
P ( e | s , M ) = P ( s | e , M ) P ( e | M ) P ( s | M ) - - - ( 8 )
In one embodiment, evidence condition sequence probability can be expressed as again:
P ( s | e , M ) = Σ j = 1 K P ( s | c j , e , M sequence ) P ( c j | e , M candidate ) - - - ( 9 )
Wherein, c jBe a certain candidate, and still have two submodels of the M in given evidence source: candidate's model M CandidateWith series model M SequenceLike this, critical assumptions are as follows:
Suppose 3: under given candidate's situation, the probability under the above-mentioned model can be expressed as candidate's marginalisation, and wherein, target sequence is independent of evidence conditionally.
Use this hypothesis, can from the evidence entry, delete the dependence for s:
P ( s | e , M ) = Σ j = 1 K P ( s | c j , M candidate ) P ( c j | e , M sequence ) - - - ( 10 )
The graphical model of candidate's model of this version is as follows:
Figure BDA00003141303200095
And complete evidential probability is:
P ( e | s , M ) = Σ j = 1 K P ( s | c j , M sequence ) P ( c j | e , M candidate ) P ( e | M ) P ( s | M ) - - - ( 11 )
Particular module
With reference to showing prediction engine from two homologies not: Fig. 2 of the native system preferred embodiment of extraction evidence linguistic context (context) and the input (input), we utilize general candidate's model, have proposed a particular example of native system.Yet as mentioned above, native system is not limited to linguistic context and imports as evidence.If used other or extra evidence source, native system will be correspondingly according to this type of evidence source generation forecast.
Generally, linguistic context is represented the evidence that observes about user's input text, and the clear proof certificate that observes of input text in the ban about the user is then represented in input.For instance, if the user has imported English sequence " My name is B ", then we can think that the linguistic context evidence is sequence " My name is ", and the input evidence is sequence " B ".Yet, only just for example, it should be noted that in prevailing form this model can not made any concrete expression to the special shape of observed evidence.For example, in fact the input evidence may be the touch coordinate that some row come from dummy keyboard.
As shown in Figure 2, evidence (input and linguistic context) is used as the input of prediction engine, in this prediction engine, preferably exist three model R, M Context, M Input, each model preferably includes at least two submodels (character model, a meta-model; Candidate's model, language model; Candidate's model, prefix matching model).As shown in Figure 2, this prediction engine preferably includes target sequence priority model R.Although this feature is preferred, native system is not limited to comprise the embodiment of target sequence priority model R.
Target sequence priority model R comprises:
● character model-not fixedly the sequence in the language of vocabulary concept implement to distribute.General using Markov model (Markov model) is implemented this distribution in character string.
Character model is a kind of series model of being set up by the alternative entry of character.For example, if the set in the training is " explaining ", a meta-model may be following appearance:
P(e)=0.1
P(x)=0.1
P(p)=0.1
P(l)=0.1
P(a)=0.1
P(i)=0.2
P(n)=0.2
P(g)=0.1。
The ternary character model may be following appearance:
P(e)=0.1
P(x|e)=1.0
P(p|ex)=1.0
P(l|xp)=1.0
P(a|pl)=1.0
P(i|la)=1.0
P(n|ai)=1.0
P(i|in)=1.0
P(n|ni)=1.0
P(g|in)=1.0。
● the sequence of a meta-model-under the prerequisite of not considering linguistic context in language is implemented to distribute, and each sequence is carried out inter-process as primary entity.
For instance, if the set in the training is " the dog chased the cat ", then a Dui Ying gram language model may be:
P(the)->0.4
P(dog)->0.2
P(chased)->0.2
P(cat)->0.2。
Linguistic context evidence model M ContextComprise:
● candidate's model-under the given situation that special candidate's lexical or textual analysis arranged, implementation condition distributes on the linguistic context observed value.
● series model-under the situation of given special context, implementation condition distributes on the sequence in language or language set.In Fig. 2, series model is illustrated as language model.This language model comprises one group in a preferred embodiment corresponding to the language model of different language, for example, and LM French, LM German, LM EnglishDeng.
Input evidence model M InputComprise:
Candidate's model-under the situation of given special candidate's lexical or textual analysis, implementation condition distributes on the input observed value.
Series model-under the situation of given intended target sequence, implementation condition distributes on the candidate in language or language set.This model is illustrated as " prefix matching model " in Fig. 2.
Comprising each model of target sequence priority model R can be according to circumstances upgrade with the text of user's input.By using the dynamic language model, can calculate to a nicety the more text sequence of designated user expection of native system.
Each model has been exported one group of sequence S R, S Context, S InputAnd the dependent probability that input is used as probability maker PG is estimated P R, P Context, P InputProbability maker PG estimates P with the probability of above-mentioned model output R, P Context, P InputCombine, generate one group of ultimate sequence prediction S FProbable value P F
Ultimate sequence prediction S FCan will be shown to the user by user interface, and browse and select for the user, or can use this ultimate sequence prediction S by native system FWith the input text of righting the wrong automatically.In case automatically or by the user selected prediction, this input text preferably is added in order to generate the linguistic context evidence of next prediction.On the contrary, the multiword symbol adds next input relevant with current entry if the user is by importing more, and then this input text preferably is added into the input evidence is distributed to prediction with modification the prior probability of working as.
The special system of introducing in detail below in the present embodiment is how to be generated by Fundamentals of Mathematics.
The instantiation expression formula (5) of bringing two evidence sources into generates:
P ( s | R ) P ( context | s , M context ) P ( input | s , M input ) Z - - - ( 12 )
Wherein, the Z=normaliztion constant approximates:
Σ j = 1 | s | P ( s j | R ) P ( context | s j , M context ) P ( input | s j , M input ) - - - ( 13 )
As described below, this approximate value is applied to native system.The function z of an arrangement set T of let us imagination, for example:
z ( T ) = Σ j = 1 | T | P ( s j | R ) P ( context | s j , M context ) P ( input | s j , M input ) - - - ( 14 )
It is as follows to obtain z:
Z=z(T)+z({u})*k (15)
Wherein u represents one " the unknown " sequence, and k is | the appraisal of S|-|T|, wherein | S| is the sequence quantity in the set of all possible target sequence, and | T| is the quantity of the sequence of the type estimated as " known " that at least one bottom evidence model has.Each independent evidence condition model M can return P (e|u, appraisal M), that is: under the situation of given " the unknown " sequence, the distribution on the evidence observed value.This means that substantially each evidence condition model is smoothly responsible to the distribution of himself, but must be with relevant with the proportional k of " the unknown " sequence overall estimate quantity.In fact, each model all can be understood arrangement set S ', wherein
Figure BDA00003141303200136
, and P (e|s, appraisal M) will keep constant and equal P (e|u, M), although
Figure BDA00003141303200135
This characteristic smoothly be that native system consider to change the employed a kind of means of level of trust in the model relevant with each evidence source.
According to expression formula (12) and (14), (s ∈ S|e M), calculates following appraisal: the preferential P of target sequence (s|R) for the conditional probability P in definite above-mentioned special system example; Linguistic context probability P (context|s, M Context) and input probability P (input|s, M Input).These appraisals will be discussed below and how calculate these appraisals.
Target sequence is preferential
Preferred calculating target sequence is preferentially as follows:
P ( s | R ) = P ( s | R unigram ) if ( s ∈ V ) P ( s | R character ) otherwise
Wherein V is included in R UnigramIn arrangement set, the enforcement of model then realizes according to the known technology that makes up a level and smooth gram language model and level and smooth Markov chain character model.Some application technologies that are used for these models of enforcement are listed hereinafter.But the technology that other are fit to is also unlisted.
● level and smooth n unit's entry or character model (known in this field);
● as the patent documentation of<GB Patent Application No. 0917753.6〉in the multilingual model of self-adaptation put down in writing;
● as<document: Scheffler2008〉the middle PPM(prediction by partial matching that puts down in writing, partial match estimation) language model;
● under certain probability by the morphological analysis engine that constitutes the morphological component formation sequence.
By comprising target sequence priority model R, native system has improved the expected sequence accuracy of predicting.In addition, target sequence priority model R can realize the reasoning based on character of invisible (unseen) target sequence, that is to say, native system can infer better that the unknown object sequence is with all possible target sequence of general estimation.
The linguistic context probability
Preferably by second candidate's model evaluation linguistic context probability P (context|s, the M Context) to propose following expression formula (16).Although this mode is the preferred means of assessment probability, the present invention is not limited to make the probability of assessing in this way.
P ( context | s , M context ) = Σ j = 1 K P ( s | c j , M context - sequence ) P ( c j | context , M context - candidate ) P ( context | M context ) P ( s | M context ) - - - ( 16 )
Therefore, in order to determine the linguistic context probability, be calculated as follows every: the linguistic context sequence is estimated P (s|c j, M Context-sequence); Linguistic context candidate estimates P (c j| context, M Context-candidate); Linguistic context is preferentially estimated P (context|M Context); And target sequence is preferentially estimated P (s|M Context).These appraisals will be discussed below and how calculate these appraisals.
The linguistic context sequence is estimated
At given particular candidate sequence c jSituation under, the linguistic context sequence is estimated P (s|c j, M Context-sequence) be the probability of the target sequence s under the linguistic context series model.The linguistic context series model is a kind of function that returns the probability of target sequence under the situation of given linguistic context sequence, i.e. f in form s(t Target, t Context)=P (t Target| t Context, θ s), wherein, θ sIt is the parameter of model.Therefore, being calculated as of linguistic context sequence probability: P (s|c i, S)=f s(s, c i).Can use multiple different technologies to calculate this appraisal.For example, the level and smooth frequency analysis on the linguistic context training data, similar with equation (21) and to estimate relevant mode and explain with target sequence is preferential.Alternatively, can separately or be used in combination following any:
● n gram language model (known in this field);
● as the patent documentation of<GB Patent Application No. 0917753.6〉in the multilingual model of self-adaptation put down in writing;
● as<document: Scheffler2008〉the middle PPM(prediction by partial matching that puts down in writing, partial match estimation) language model;
● the HMM(Hidden Markov Model of generation, hidden Markov model) probability part-of-speech tagging device<reference: 008.LingPipe4.1.0.http: //alias-i.com/lingpipe (accessed September26,2011) or Thede, S.M., Harper, M.P., 1999 〉;
● be used for the natural language resolver of the probability of returning part sentence, as RASP<reference: Briscoe, E., J.Carroll and R.Watson2006 〉;
● admit to represent the input characteristics of linguistic context sequence and target sequence and the neural network of output probability (techniques well known).
Native system is not limited to above-mentioned technology, and any technology that can be used for calculating the linguistic context sequence probability all is applicable to native system.
As indicated above, M Context-sequenceCan comprise a plurality of language models corresponding to multiple different language.In order to determine the conditional probability of expression formula (16), can use the language model relevant with entry to determine this conditional probability.As a kind of explanation, mentioned in the above-mentioned example from English model (LM English) and French model (LM French) the middle prediction entry of extracting out " pain ".In this case, expression formula (16) is by P (context|pain, LM English) andP (context|pain, LM French) determine, wherein from French model (LM French) in " Pain " that extract out be different from from English model (LM English) middle " Pain " that extracts out, although should predict identical lexically.By entry is associated with its source model, native system has been simplified the processing mode of the identical entry of morphology, because native system has only kept the most probable entry in the identical entry of two or more morphology.In addition, native system has been simplified the conditional probability calculating of expression formula (16).This simplification is feasible, although because the morphology of entry is identical, entry has the different meaning of a word in different language, can treat this entry with a certain discrimination thus.
Transfer back to Fig. 2 like this, by model M ContextThe entry S set that generates ContextCan comprise M ContextIn the entry of any language model (or candidate's model).
Linguistic context candidate estimates
Linguistic context candidate estimates P (c j| context, M Context-candidate) be a kind of function, its form is f Context-candidate(t)=P (t| θ Context-candidate), wherein t is arbitrary sequence, θ Context-candidateBe model parameter.What like this, linguistic context candidate condition was estimated is calculated as follows: P (c j| context, M Context-candidate)=f Context-candidate(c j).
In an optimum decision system, linguistic context candidate is sequence, and the linguistic context candidate collection is expressed as directed acyclic graph (DAG, directed acyclic graph), and each node in the directed acyclic graph comprises the subsequence that contains one or more characters.Each border is assigned probability, and in a preferred embodiment, directed acyclic graph preferably also has the specific properties that each path is restricted to same length.In this article, the variant of directed acyclic graph is called as probabilistic limited sequence chart (PCSG, probabilistic constrained sequence graph).Each independent candidate sequence then is expressed as unique path by PCSG, and linguistic context candidate pattern function is calculated as the probability of the delegated path of linguistic context candidate model for given candidate's rreturn value.
In form, PCSG comprises the four-tuple that contains a group node N, root node r, one group of directed edge E and one group of parameter (probability) θ:
G=(N,r,E,θ) (17)
Boundary representation between two node n, the n ' is that (n → n'), the probability tables of shifting to n ' from n along the border is shown P (n'|n).Path through G originates in node r, and follows one from the outwardly directed border extension of each node of visiting, till arriving at the node that does not contain the border of going out (outgoing edge).The attribute of G is as follows:
1) G is directed acyclic graph (DAG);
2)
Figure BDA00003141303200168
That is: the node of all except root node must have at least one and enters border (incoming edge);
3) ∃ m , k ∈ N . ∀ n ∈ N . ( m → n ) ∈ E ⇒ ( n → k ) ∈ E , That is: add follow-up common path immediately again from the path that given node is told.This attribute has strictly limited the structure of this figure, and has hinted that all paths have identical length, has reduced the normalization needs that path probability is calculated.
Linguistic context candidate's pattern function calculates the probability following (being equal to linguistic context candidate estimation) of given path:
P(c j|context,M context-candidate)=f context-candidate(c j)=P(p j|G) (18)
Wherein, P (p j| G) be path probability, be calculated as the product on each border in the path:
P ( p j G ) = P ( n 1 | r ) Π k = 2 K P ( n k | n k = 1 ) - - - ( 19 )
Wherein, K is the boundary number in the path.It should be noted that this preferred formula is equivalent to internodal implicit independence assumption.This is that the sequence probability of candidate sequence is not modeled because in this case, and the variation probability among the candidate is modeled.Therefore, following Column Properties has kept borderline probability:
∀ n ∈ N . Σ ( n → m ) ∈ E P ( m | n ) = 1 - - - ( 20 )
That is to say that all borderline probability sums one of going out that come from given node n are decided to be 1.This means that also following expression formula is effective: That is: the probability sum in all paths equals 1 among the PCSG.
A certain example is illustrated these concepts with help.Consider following 12 linguistic context candidate sequences:
·“Sunday at3pm” ·“sunday at3pm” ·“Sun at3pm”
·“Sunday at3pm” ·“sunday at3pm” ·“Sun at3pm”
·“Sunday at3p.m.” ·“sunday at3p.m.” ·“Sun at3p.m.”
·“Sunday at3p.m.” ·“sunday at3p.m.” ·“Sun at3p.m.”
These linguistic context candidate sequences can be represented by ' | ' by clear and definite word circle of following PCSG(, and empty sequence is by " φ " expression) be expressed as:
Figure BDA00003141303200173
According to linguistic context candidate model and adopt expression formula (19), give the border with probability assignments, for example:
Figure BDA00003141303200174
Then, the candidate's probability that from PCSG, generates above-mentioned 12 sequences following (to be concise in expression clearly in order making, only to have listed three examples):
P("sunday at3pm"|"sunday at3pm",C)=0.6*1.0*1.0*0.6*1.0*0.7=0.252
P("Sunday at3pm"|"sunday at3pm",C)=0.3*1.0*1.0*0.4*1.0*0.7=0.084
P("sun at3p.m."|"sunday at3pm",C)=0.1*1.0*1.0*0.4*1.0*0.3=0.012
Be used for to make up DAG and change to the particular example that the model detail of node allocation probability will depend on native system.Above-mentioned graphic examples to three kinds of general variations are encoded:
● the borderline branch of word (the single meaning of possibility);
● the branch in capital and small letter (case) variation;
● the branch on morphology changes.
Be understood that the variation of any kind all can be encoded in this framework.Scheme before another example will turn to, for example, if native system dopes " on " and " in ", and the user has selected " in ", then it can be encoded into the branch that also has the small probability of distributing to " on " the probability right of distributing to " in " except having, with the possibility of the unexpected acceptance error suggestion of expression user.In these cases, enroll following principle:
● be lower than " Sun " of abbreviated form with the possibility of ' sunday ' of lowercase ' s ' beginning, and the possibility of " Sun " of abbreviated form is lower than full spelling distortion ' Sunday ';
● the possibility of the participle situation (tokenization case) that " pm " and numeral " 3 " are split slightly is lower than the not situation of participle;
● the possibility of fullstop distortion " p.m. " is a shade below no fullstop form " pm ".
In the following manner, preferably made up the particular example of linguistic context candidate PCSG at algorithm by homing sequence s:
1) by s being encapsulated among the node ns that is connected in root node, converts s to PCSG;
2) by introducing the branch node on the change point, iteration is torn open and is analysed ns.
For instance, consider PCSG construction algorithm that original series " sunday at3pm " is worked.At first, step 1:
Figure BDA00003141303200181
Native system is disposed probability participle device, and the result is as follows:
Figure BDA00003141303200191
It should be noted that, because the attribute 3 of above-mentioned PCSG, modification will always show as the structure infix form of branch-also-reclosing (branch-and-rejoin), for special circumstances, modification has a node branch, this is convenient to subsequent treatment very much, because it can not have influence on the probability in whole paths.To introduce in further detail hereinafter according to model and add boarder probability.Continue to introduce this algorithm, capital and small letter (case) deformation analysis device is disposed as follows:
At last, morphology deformation analysis device is disposed as follows:
Figure BDA00003141303200193
It should be noted that the attribute 3 because of PCSG, each branch must converge before the fork again.This means, in some cases, if occur two take-off points continuously, then must insert empty node.
Boarder probability preferably is assigned to PCSG.About the parameter of linguistic context candidate model, can preferably implement boarder probability and distribute.Visual interpretation for these probability has 2 points:
1) the probability appraisal of user's expected sequence of specific branch is distributed in their expressions.For example, if the user imports " Dear ben ", we perhaps can think and may they in fact want input " Dear Ben ";
2) their expressions are the compensation probability of effective spelling distortion of the sequence that observes for specific branch.For example, if user's input " See you on Thur ", then the selectable correct orthographic form of " Thur " can be " Thurs ".
Under the situation of given a certain background model information, the probability of distributing to specific border also can be subjected to correctly spelling the influence of the estimated probability of variant.For example, in fact can reuse the probability appraisal that linguistic context series model s tries to achieve different correct spelling variants.This probability is estimated and can be used in combination with other probabilistic quantities, generates branch's probability.The context of use series model means the actual linguistic context series model S that comprises of linguistic context candidate MODEL C by this way, obviously runs counter to the independence assumption (attribute 7 above) between candidate's model and the series model thus.Yet this hypothesis will not appear under the linguistic context situation, and is therefore comparatively safe.
Following example will help explanation.In a preferred embodiment, suppose that linguistic context candidate model uses following algorithm assigns probability:
1) sequence that observes obtains probability 0.8, and other sequences are evenly divided remainder equally;
2) estimate measurement numerical value by the linguistic context series model;
3) normalization numerical value makes it satisfy above-mentioned PCSG attribute 19.
According to above-mentioned PCSG example, following branching into:
Figure BDA00003141303200201
Because " sunday " is raw observation, at first the step 1 by above-mentioned algorithm is its allocation probability value 0.8, and other borders then respectively are assigned probable value 0.1.The appraisal of being returned by the linguistic context series model in this example is as follows:
P("sunday"|C S)=0.01
P("Sunday"|C S)=0.04
P("Sun"|C S)=0.02
Wherein, C SBe illustrated in linguistic context candidate model context of use series model in this case.Therefore, in this example, the probability (difference) of distributing to the not normalization on each border and normalization (through rounding up) is as follows:
Linguistic context is preferentially estimated
By the frequency of the normalization initiation sequence t relevant with linguistic context, can roughly estimate linguistic context and preferentially estimate P (context|M Context)
P ( context | M context ) ≅ freq ( t ) E t ′ freq ( t ′ ) - - - ( 21 )
Wherein freq (t) is the frequency of sequence t in the training data, and denominator is the frequency sum of all sequences in the training data.Sequence " t " in the expression formula (21) is the current linguistic context as the native system input.Preferentially according to the probable value of probability weight prediction, the corresponding model of wherein extracting this prediction out comprises the corresponding probability of given linguistic context sequence to linguistic context.In order to realize this process, linguistic context is preferentially according to the appraisal weight estimation value of expression formula (21).
In actual applications, for example can by settle in invisible sequence hypothesis (occurrence assumption) appears or be retracted into all sequences all limited (low level) in the sightless example estimate smoothly this appraisal.For instance, if linguistic context is ternary model, then prediction engine can be return and be constituted binary or monobasic appraisal.
Linguistic context preferentially provides dual function: help the normalization probability to estimate; And when can't providing useful information, context model provides simple " model detection ".Can't provide information (for example, when last entry is unknown) if the linguistic context sequence is estimated for the N meta-model, linguistic context is preferentially estimated can use most probable linguistic context this model of weighting in large quantities, and the prediction of this model is promoted on the prediction of other models." most probable linguistic context " is to estimate (21) maximal value on a plurality of model sets, for example language model set LM English, LM French, LM GermanFor instance, if linguistic context is " The dog chased ", then can expect that with appearing to compare in the French this linguistic context more likely appears in the English.Therefore, the conditional probability of expression formula (21) is for LM EnglishTo be maximum, and the probability maker thus in a large number weighting from LM EnglishPrediction probable value but not from LM FrenchThe probable value of prediction, therefore, LM EnglishMore be subjected to preferential " preference " estimated of linguistic context.
Therefore, consider linguistic context, linguistic context is preferentially estimated a large amount of weightings from the only language model in a plurality of language models relevant with multilingual.Like this, the preferential appraisal of linguistic context can detect the affiliated language of text that someone imports.
Target sequence is preferentially estimated
Can be according to preferentially estimating with linguistic context, the same mode of expression formula (21) is used through level and smooth training data frequency analysis estimation target sequence and is preferentially estimated P (s|M Context), that is: can be preferential by the target sequence frequence estimation target sequence on all sequences in the normalization linguistic context training data:
P ( s | M context ) ≅ freq ( s ) Σ s ′ freq ( s ′ )
Wherein freq (s) is the frequency of target sequence in the training data, and denominator is whole frequency sums of all target sequences in the training data.Denominator roughly is equivalent to entry (the comprising the dittograph bar) sum in the training data.
Input probability
Utilize first candidate's model assessment input probability P (input|s, the M Input):
P ( input | s , M input ) = Σ j = 1 K P ( input | c j , M input - candidate ) P ( c j | s , M input - sequence ) - - - ( 22 )
Therefore, in order to determine input probability, need to calculate following appraisal: the input candidate estimates P (input|c j, M Input-candidate) and list entries appraisal P (c j| s, M Input-sequence).Introduce this two kinds of appraisals below.
Input the candidate estimate
The input candidate estimates P (input|c j, M Input-candidate) incoming event and the function on the sequence: the f that are defined as observing Input-candidate(i, t)=P (i|t, θ Input-candidate), θ wherein Input-candidateIt is the parameter of model.Input observed reading i encodes in list entries expected structure (ISIS, input sequence intention structure) arbitrarily.This list entries expected structure is the ordered list with the arrangement set of probability mapping:
{(t 11→P(i 1|t 11),(t 12→P(i 1|t 12),...},{(t 21→P(i 2|t 21),(t 22→P(i 2|t 22),...},...
It should be noted that each appraisal has form P (i j| t Jk), if that is: the user has planned list entries t Jk, we should observe incoming event i jProbability what are.Consider following ISIS example:
{ ( H → 0.5 ) , ( h → 0.3 ) , ( g → 0.1 ) , ( j → 0.1 ) } { ( e → 0.8 ) , ( w → 0.1 ) , ( r → 0.1 ) }
This ISIS example is encoded for a kind of like this scheme, and whether native system estimation user plans to import the character ' H ' of for example following character ' e ' thereafter in this scheme, estimates that like this incoming event that observes has probability 0.5 and 0.8 respectively.
The method that generates these probability distribution is not theme of the present invention.Or rather, given prominence to a series of suitable technology, for instance:
-can distribute according to the character generating probability of given target keywords on the keyboard layout, for example qwerty keyboard if the user knocks the zone corresponding with " H " key, then may comprise the character " G " and " J " that have certain probability in ISIS.
-can distribute according to the distance between touch coordinate (touch-screen dummy keyboard) and the specified button coordinate (or certain distance function, for example square etc.) generating probability.
In a preferred systems, the input candidate is sequence, and the input candidate collection is expressed as the PCSG(EPCSG of expansion).EPCSG is a kind of PCSG, but has the additional structure of running counter to Standard PC SG attribute (definition hereinafter).As the linguistic context situation, represent each candidate sequence by unique path of passing EPCSG, and given candidate's input candidate pattern function rreturn value is calculated as the normalization probability of its delegated path.
Input candidate EPCSG generative process begins with the ordered list of native system by the sequence probability pair set that generates with user interactions, wherein the probability distribution in each subclass representative of consumer list entries expection.
The algorithm that is generated input candidate EPCSG by input ISIS comprised for two steps:
1) converts ISIS to PCSG;
2) insert additional generalized structure (generalizing structure), thereby generate EPCSG.
Step 1 is flat-footed.Root node with new PCSG begins, and this algorithm is each the structure branch that distributes among the ISIS.Above-mentioned ISIS result in the step 1 is as follows:
Figure BDA00003141303200241
Step 2 is used two existing PCSG of additional structure finishing.First structure is empty node subpath (its suitable PCSG framework), and second structure is ' asterisk wildcard ' structure (converting PCSG to EPCSG).The application example of step 2 is as follows:
Figure BDA00003141303200242
The symbol of asterisk wildcard (being expressed as ' * ') is actually the simple expression way of the branch that comprises/generate each symbol in the character set.The asterisk wildcard structure is a limited circulation, has therefore run counter to the acyclic attribute of Standard PC SG.The EPCSG expansion allows only to use the asterisk wildcard circulation at convergence point, and numerical value e, w are the probability constants of predesignating.It should be noted that in this case each take-off point has sky node additional (being two in this case), and each convergence point has asterisk wildcard additional (being in this case).These generalized structures can be omitted the one or more characters in the target sequence (having asterisk wildcard probability w) or insert one or more error characters (having hole node probability e) with respect to the user.Be understood that how structure that these are the extra details that adds PCSG can change according to the different instances along with native system such as computational resource, series model length.
Empty node subpath makes native system can abandon the character of user error input, and not so this error character can cause incorrect chain through PCGS.
Rely on these additional generalized structures (especially asterisk wildcard branch), the quantity through the path of PCSG is increased sharply.For example, suppose that size is 50 character set, 1020 different paths can be arranged through the PCSG of above-mentioned simplification.For the ISIS of reality, there are tens of very hundreds and thousands of different paths.This optimum decision system preferably uses this combination of following technical finesse to increase sharply in mode alone or in combination.
● use trie (word lookup tree, techniques well known) to neglect and predict that those are not the paths of sequence prefix in the vocabulary;
● the probability of use threshold values is deleted those unlikely relatively paths.Threshold values is set to the ratio of the differential of the lower sequence of current most probable sequence and possibility.Given threshold values t, the path L of current investigation.If the maintenance following relationship is then deleted path n 1N L:
P ( n 1 | r ) &Pi; j = 2 L P ( n j | n j - 1 ) arg max m &lsqb; P ( m 1 | r ) &Pi; j = 2 L P ( m j | m j - 1 ) &rsqb; < t - - - ( 23 )
● list entries model T is used to the probability threshold values equally.Given difference or limited threshold values t, and be L:{c by length 1..., c KThe arrangement set that forms of all paths.If maintenance following relationship, then deleted representation particular sequence c PGiven path p:
P ( c p | T ) arg max j &lsqb; P ( c j | T ) &rsqb; < t - - - ( 24 )
Can also dispose separately or with above-mentioned wherein one or all technical combinations dispose the technology that other are applicable to that treatment combination increases sharply.
List entries is estimated
Under the situation of given target sequence, list entries is estimated P (c j| s, M Input-sequence) be the distribution on the candidate sequence, and can be estimated as normalized target function:
P ( c j | s , M input - sequence ) = &delta; ( s , c j ) z - - - ( 25 )
Wherein, if t ' is the prefix of t, then δ (t, t')=1, otherwise δ (t, t')=0, and the z=∑ kδ (s, c k), i.e. all candidate's sums.
It should be noted that if present candidate's uniqueness, and allow candidate collection to comprise all possible sequence, then can recomputate normalization factor: Z=length (s).For instance, given target sequence " the " always just in time has three matching candidates: " t ", " the " and " the ".
Therefore, the invention provides a kind of text prediction engine and system of routine, and the particular example of text prediction engine or system, text prediction engine or system can generate one group and have dependent probability value P respectively FSequence prediction S F
The present invention also provides a kind of correlation method for the treatment of the user version input.Be back to Fig. 1 and said system, this method comprises that receiving the text that inputs to such as user interfaces such as electronic equipments imports; Use the prediction of text prediction engine formation sequence SF and relevant probable value PF; And this sequence prediction provided to described user interface.
As the introduction for native system, conventional method comprises: by comprising the prediction of text prediction engine formation sequence and the dependent probability value of one or more models.In a preferred embodiment, this method comprises: use at least one evidence source e by target priority model R and at least one 1, e 2... the model M of generation forecast 1, M 2... the formation sequence prediction.As mentioned above, about this system, and especially expression formula (12) is to (15), and this method comprises: the probable value normalization factor generation normalization probable value of being obtained by the constant sum of the possible sequence prediction of the probable value of n most probable sequence prediction and expression residue by estimation.
With reference to Fig. 2, in above preferred embodiment, final prediction sets S FAnd relevant probable value P FBy probability maker PG according to respectively from target priority model R, context model M ContextWith input model M InputThe prediction sets S that extracts out R, S Context, S InputGenerate.In this embodiment, the linguistic context of user input sequence is used as from context model M ContextIn extract the evidence of prediction out, and the relevant user input sequence of the current word of attempting importing with the user is used as from input model M InputThe middle evidence of extracting prediction out.
Other aspects and the said system of this method are similar, for example, and in a certain embodiment of this method, if the corresponding probable value of sequence prediction then only offers user interface with these sequence predictions respectively more than or equal to first threshold values.
As mentioned above, related to one and implement the system of generalized structure to determine that the linguistic context candidate estimates in PCSG, in a preferred embodiment of this method, at least one group of sequence prediction is equivalent to be inputed to by the user adjustment or the invulnerable release of the text input of user interface.
Description according to said system is analogized, and can determine other aspects of the inventive method at an easy rate.
Following is claims of the non-exhaustive nothing left of above-described embodiment of stating in this application:
1. system comprises:
Text prediction engine, it comprises:
At least one is used for having from one group of evidence source generation the model of the sequence of dependent probability appraisal;
Be used for generating one group of model that has the sequence of dependent probability appraisal; And
The probability maker, it is used for receiving the sequence of respectively organizing that has the dependent probability appraisal and also generates the sequence prediction that has the dependent probability value.
2. the system of embodiment 1 comprises a plurality of for generated a plurality of models that have the arrangement set of dependent probability appraisal by some evidence sources.
3. in the system of embodiment 1 or 2, the probability maker is used for generating one group of sequence prediction according to an any amount corroboration source.
4. in the system of embodiment 3, one of them evidence source comprises user input text.Modes such as this user input text can be selected by the user, character input, speech recognition are imported.
5. system comprises:
User interface is used for receiving user input text;
According to the text prediction engine of first aspect or other suitable text prediction engine arbitrarily, be used for receiving the text input that comes from user interface and generate the sequence prediction that has the dependent probability value.
6. in the system of embodiment 5, text prediction engine comprises:
Context model is used for receiving the text of being imported by the user and generates one group of sequence and relevant probability appraisal;
Input model is used for receiving the text of being imported by the user and generates one group of sequence and relevant probability appraisal;
One model is used for generating one group of sequence and estimates with relevant probability; And
The probability maker is used for receiving from above-mentioned a plurality of models and respectively organizes sequence and relevant probability appraisal and generate one group of sequence prediction and relevant probable value.
7. in the system of embodiment 6, described user interface is used for the above-mentioned sequence prediction of demonstration and selects for the user.
8. in the system of embodiment 7, this system is described sequence prediction ordering according to described probable value and described sequence prediction is shown as ordered set.
9. in embodiment 6 to 8 in the system of any one embodiment, wherein, this system preferably uses the automatic correction of described sequence prediction to input to the wrong input text of user interface.
10. in embodiment 6 to 8 in the system of any one embodiment, described text prediction engine is used for generating based on the sequence prediction in a corroboration source arbitrarily.
11. the method for a process user text input comprises:
Reception inputs to the text of user interface;
Utilize the prediction of prediction engine formation sequence and relevant probable value;
This sequence prediction is offered described user interface.
12. in the method for embodiment 11, comprising: described sequence prediction is presented at described user interface selects for the user.
13. in the method for embodiment 12, comprising: according to the dependent probability value of described sequence prediction, be described sequence prediction ordering.
14. in the method for embodiment 13, comprising: show that sorted sequence prediction selects for the user.
15. in embodiment 11 to 14, in the method for any one embodiment, comprise and utilize the automatic correction of described sequence prediction to input to the step of the wrong input text of user interface.
16. in the text prediction engine of in aforementioned any one embodiment, using, comprising: context model, input model, target priority model and probability maker.
17. in the text prediction engine of embodiment 16, described context model receives the text of being imported by the user with described input model user and generates one group of sequence and relevant probability appraisal.
18. in the text prediction engine of embodiment 16 or 17, described target priority model is used for generating one group of sequence to be estimated with relevant probability.
19. in the text prediction engine of an embodiment in embodiment 16 to 18, described probability maker is used for respectively organizing sequence and relevant probability appraisal from described model reception, and generates each to one group of sequence prediction of probable value should be arranged.
The above only is preferred embodiment of the present invention, and is within the spirit and principles in the present invention all, any modification of doing, is equal to replacement etc., all should be included within protection scope of the present invention.

Claims (36)

1. text prediction engine comprises:
At least one model, it is used for having first group of sequence that dependent probability is estimated from the generation of evidence source;
The probability maker, it is used for receiving the described first group of sequence that has the dependent probability appraisal and generates one group of sequence prediction that has the dependent probability value;
Wherein, received under the situation of all possible sequence the described probable value of normalization on all possible sequence prediction that is generated by described probability maker by described probability maker given.
2. text prediction engine according to claim 1 is characterized in that, also comprises priority model, and it is used for generating the second group of sequence that has the dependent probability appraisal.
3. text prediction engine according to claim 1 and 2 is characterized in that, described model generates first group of sequence according to the uncertainty in described evidence source and the described evidence source.
4. according to claim 2 or 3 described text prediction engine, it is characterized in that described probability maker receives and has described first, second group sequence that dependent probability is estimated.
5. according to any described text prediction engine in the claim 1 to 4, it is characterized in that, described probability maker is estimated the normalization factor of described probable value by n most probable sequence prediction is represented the constant addition with remaining probable value that may sequence prediction.
6. text prediction engine according to claim 5 is characterized in that, described constant is represented the probable value by the remaining possibility sequence prediction of described model and the generation of described priority model.
7. according to any described text prediction engine in the claim 1 to 6, it is characterized in that described model comprises a plurality of models that have a plurality of first group of sequence of dependent probability appraisal for generation.
8. text prediction engine according to claim 7 is characterized in that, described a plurality of models generate a plurality of first group of sequence according to a plurality of evidences source.
9. text prediction engine according to claim 8 is characterized in that, one in described a plurality of evidences source comprises user input text.
10. text prediction engine according to claim 9 is characterized in that, described text prediction engine is the part of a certain system, and described user input text by one or more users select, character input or speech recognition be input in this system.
11., it is characterized in that described text prediction engine comprises given linguistic context sequence according to corresponding model probability is weighted the probable value of described sequence prediction according to claim 7 or 8 described text prediction engine.
12. text prediction engine according to claim 11 is characterized in that, described a plurality of models comprise a plurality of language models corresponding with multiple different language; And the probable value of the described text prediction engine pair sequence prediction corresponding with the language model of the most probable language that relates to user input text is carried out five-star weighting.
13., it is characterized in that each evidence source is moulded by the corresponding model that is used for generating the sequence that has the dependent probability appraisal according to any described text prediction engine in the claim 10 to 12.
14. text prediction engine according to claim 13 is characterized in that, under the situation of given described sequence prediction, described probability maker with each evidence source as other on evidence the separate component of having ready conditions in source handle.
15. according to aforementioned any described text prediction engine of claim, it is characterized in that, described model comprises context model and input model, and described context model is used for receiving the text of user's input and generates one group of sequence with described input model to be estimated with relevant probability; And
Described priority model comprises for generating the target priority model that one group of sequence and dependent probability are estimated.
16. text prediction engine according to claim 15 is characterized in that, described input model comprises candidate's model and language model.
17., it is characterized in that described context model comprises candidate's model and prefix matching model according to claim 15 or 16 described text prediction engine.
18. according to any described text prediction engine in the claim 15 to 17, it is characterized in that described target priority model comprises character model and a meta-model.
19. the description of aforementioned texts prediction engine is with reference to accompanying drawing and as shown in drawings.
20. a system comprises:
User interface, it is used for receiving the text by user's input;
Text prediction engine, it be used for to receive from the described text of described user interface input and generates one group of sequence prediction that has the dependent probability value, wherein, all possible sequence prediction on the described probable value of normalization;
Wherein, described text prediction engine also is used for providing described sequence prediction to described user interface.
21. system according to claim 20 is characterized in that, described text prediction engine is as any described prediction engine in the claim 1 to 19.
22., it is characterized in that only having one during more than or equal to the probable value of the first probability threshold values when described sequence prediction according to claim 20 or 21 described systems, described text prediction engine provides sequence prediction to described user interface.
23., it is characterized in that if described text prediction engine has the corresponding probable value that is equal to or greater than the second probability threshold values, then described text prediction engine is from the list entries prediction of the described system of trend according to any described system in the claim 20 to 22.
24. according to any described system in the claim 20 to 23, it is characterized in that described system is presented at described sequence prediction on the described user interface and selects for the user.
25. system according to claim 24 is characterized in that, described probability maker sorts to described sequence prediction according to the dependent probability value of described sequence prediction, and described user interface shows described sequence prediction as ordered set.
26. the description of aforementioned system is with reference to accompanying drawing and as shown in drawings.
27. the method for a process user input text comprises:
Reception inputs to the text of user interface;
Use text prediction engine to generate one group of sequence prediction and relevant probable value, wherein, the described probable value of normalization on all possible sequence prediction;
Described sequence prediction is offered described user interface.
28. method according to claim 27, it is characterized in that, the step that generates the normalization probable value comprises: by the probable value of n most probable sequence prediction is represented the constant addition with remaining probable value that may sequence prediction, estimate the normalization factor of described probable value.
29. according to claim 27 or 28 described methods, it is characterized in that, also comprise: described sequence prediction is presented on the described user interface selects for the user.
30. method according to claim 29 is characterized in that, by described text prediction engine described sequence prediction is sorted, and shows in order for described user interface.
31. according to any described method in the claim 29 to 30, it is characterized in that, only when the corresponding probable value of described sequence prediction during more than or equal to first threshold values, described sequence prediction offered described user interface.
32. according to any described method in the claim 29 to 31, it is characterized in that at least one in the described sequence prediction is equivalent to be inputed to by the user adjustment or the invulnerable release of the text of described user interface.
33. according to any described method in the claim 29,30 and 32, it is characterized in that, also comprise: input has the sequence prediction greater than second threshold values or the probable value on second threshold values automatically.
34., it is characterized in that described probability maker comprises for a plurality of models that generate one group of sequence prediction and dependent probability value according to any described method in the claim 27 to 33; Described probable value comprises that according to corresponding model the probability of given linguistic context sequence is weighted.
35. the description of preceding method is with reference to accompanying drawing and as shown in drawings.
36. a computer program comprises: computer-readable medium wherein stores for making the processor enforcement of rights require the computer program of 27 to 35 any described methods.
CN201180053255.9A 2010-09-29 2011-09-29 Text prediction engine from text to electronic equipment, system and method for inputting Active CN103201707B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710761127.9A CN107506047B (en) 2010-09-29 2011-09-29 Text prediction engine, system and method for inputting text to electronic device

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
GB1016385.5 2010-09-29
GBGB1016385.5A GB201016385D0 (en) 2010-09-29 2010-09-29 System and method for inputting text into electronic devices
PCT/GB2011/001419 WO2012042217A1 (en) 2010-09-29 2011-09-29 System and method for inputting text into electronic devices

Related Child Applications (1)

Application Number Title Priority Date Filing Date
CN201710761127.9A Division CN107506047B (en) 2010-09-29 2011-09-29 Text prediction engine, system and method for inputting text to electronic device

Publications (2)

Publication Number Publication Date
CN103201707A true CN103201707A (en) 2013-07-10
CN103201707B CN103201707B (en) 2017-09-29

Family

ID=43128163

Family Applications (2)

Application Number Title Priority Date Filing Date
CN201710761127.9A Active CN107506047B (en) 2010-09-29 2011-09-29 Text prediction engine, system and method for inputting text to electronic device
CN201180053255.9A Active CN103201707B (en) 2010-09-29 2011-09-29 Text prediction engine from text to electronic equipment, system and method for inputting

Family Applications Before (1)

Application Number Title Priority Date Filing Date
CN201710761127.9A Active CN107506047B (en) 2010-09-29 2011-09-29 Text prediction engine, system and method for inputting text to electronic device

Country Status (5)

Country Link
US (2) US9384185B2 (en)
EP (1) EP2622437B1 (en)
CN (2) CN107506047B (en)
GB (1) GB201016385D0 (en)
WO (1) WO2012042217A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106164893A (en) * 2014-04-04 2016-11-23 触摸式有限公司 System and method for one or more inputs that input is associated with multi input target
CN107688398A (en) * 2016-08-03 2018-02-13 中国科学院计算技术研究所 Determine the method and apparatus and input reminding method and device of candidate's input
CN108073679A (en) * 2017-11-10 2018-05-25 中国科学院信息工程研究所 Stochastic model set of strings generation method, equipment and readable storage medium storing program for executing under a kind of String matching scene
CN110929518A (en) * 2019-12-09 2020-03-27 朱利 Text sequence labeling algorithm using overlapping splitting rule

Families Citing this family (160)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US10002189B2 (en) 2007-12-20 2018-06-19 Apple Inc. Method and apparatus for searching using an active ontology
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
US20100030549A1 (en) 2008-07-31 2010-02-04 Lee Michael M Mobile device having human language translation capability with positional feedback
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US9431006B2 (en) 2009-07-02 2016-08-30 Apple Inc. Methods and apparatuses for automatic speech recognition
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
GB201003628D0 (en) * 2010-03-04 2010-04-21 Touchtype Ltd System and method for inputting text into electronic devices
GB201200643D0 (en) 2012-01-16 2012-02-29 Touchtype Ltd System and method for inputting text
US10037319B2 (en) 2010-09-29 2018-07-31 Touchtype Limited User input prediction
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US10417037B2 (en) 2012-05-15 2019-09-17 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US20150199332A1 (en) * 2012-07-20 2015-07-16 Mu Li Browsing history language model for input method editor
GB201216640D0 (en) * 2012-09-18 2012-10-31 Touchtype Ltd Formatting module, system and method for formatting an electronic character sequence
US9547647B2 (en) 2012-09-19 2017-01-17 Apple Inc. Voice-based media searching
IN2013CH00469A (en) 2013-01-21 2015-07-31 Keypoint Technologies India Pvt Ltd
EP2946272A4 (en) 2013-01-21 2016-11-02 Keypoint Technologies India Pvt Ltd Text input system and method
KR20150104615A (en) 2013-02-07 2015-09-15 애플 인크. Voice trigger for a digital assistant
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US20140278349A1 (en) * 2013-03-14 2014-09-18 Microsoft Corporation Language Model Dictionaries for Text Predictions
US10748529B1 (en) 2013-03-15 2020-08-18 Apple Inc. Voice activated device for use with a voice-based digital assistant
WO2014197334A2 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
WO2014197335A1 (en) 2013-06-08 2014-12-11 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
EP3008641A1 (en) 2013-06-09 2016-04-20 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
CN105453026A (en) 2013-08-06 2016-03-30 苹果公司 Auto-activating smart responses based on activities from remote devices
US10296160B2 (en) 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
TWI566107B (en) 2014-05-30 2017-01-11 蘋果公司 Method for processing a multi-part voice command, non-transitory computer readable storage medium and electronic device
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9760559B2 (en) * 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US10152299B2 (en) 2015-03-06 2018-12-11 Apple Inc. Reducing response latency of intelligent automated assistants
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US10460227B2 (en) 2015-05-15 2019-10-29 Apple Inc. Virtual assistant in a communication session
US10200824B2 (en) 2015-05-27 2019-02-05 Apple Inc. Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US9578173B2 (en) 2015-06-05 2017-02-21 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US20160378747A1 (en) 2015-06-29 2016-12-29 Apple Inc. Virtual assistant for media playback
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10740384B2 (en) 2015-09-08 2020-08-11 Apple Inc. Intelligent automated assistant for media search and playback
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10331312B2 (en) 2015-09-08 2019-06-25 Apple Inc. Intelligent automated assistant in a media environment
US10786182B2 (en) 2015-09-09 2020-09-29 The Joan and Irwin Jacobs Technion-Cornell Institute System and method for passive remote monitoring of patients' fine motor behavior
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10956666B2 (en) 2015-11-09 2021-03-23 Apple Inc. Unconventional virtual assistant interactions
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
KR102462365B1 (en) 2016-02-29 2022-11-04 삼성전자주식회사 Method and apparatus for predicting text input based on user demographic information and context information
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US11227589B2 (en) 2016-06-06 2022-01-18 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
DK179588B1 (en) 2016-06-09 2019-02-22 Apple Inc. Intelligent automated assistant in a home environment
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
DK179343B1 (en) 2016-06-11 2018-05-14 Apple Inc Intelligent task discovery
DK179049B1 (en) 2016-06-11 2017-09-18 Apple Inc Data driven natural language event detection and classification
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US10311046B2 (en) * 2016-09-12 2019-06-04 Conduent Business Services, Llc System and method for pruning a set of symbol-based sequences by relaxing an independence assumption of the sequences
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US11550751B2 (en) * 2016-11-18 2023-01-10 Microsoft Technology Licensing, Llc Sequence expander for data entry/information retrieval
GB201619724D0 (en) * 2016-11-22 2017-01-04 Microsoft Technology Licensing Llc Trained data input system
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
DK201770383A1 (en) 2017-05-09 2018-12-14 Apple Inc. User interface for correcting recognition errors
US10417266B2 (en) 2017-05-09 2019-09-17 Apple Inc. Context-aware ranking of intelligent response suggestions
DK201770439A1 (en) 2017-05-11 2018-12-13 Apple Inc. Offline personal assistant
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
DK180048B1 (en) 2017-05-11 2020-02-04 Apple Inc. MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
DK201770429A1 (en) 2017-05-12 2018-12-14 Apple Inc. Low-latency intelligent automated assistant
DK201770431A1 (en) 2017-05-15 2018-12-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
DK201770432A1 (en) 2017-05-15 2018-12-21 Apple Inc. Hierarchical belief states for digital assistants
DK201770411A1 (en) 2017-05-15 2018-12-20 Apple Inc. Multi-modal interfaces
DK179560B1 (en) 2017-05-16 2019-02-18 Apple Inc. Far-field extension for digital assistant services
US20180336275A1 (en) 2017-05-16 2018-11-22 Apple Inc. Intelligent automated assistant for media exploration
US10403278B2 (en) 2017-05-16 2019-09-03 Apple Inc. Methods and systems for phonetic matching in digital assistant services
US20180336892A1 (en) 2017-05-16 2018-11-22 Apple Inc. Detecting a trigger of a digital assistant
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
US10657328B2 (en) 2017-06-02 2020-05-19 Apple Inc. Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling
CN109521888B (en) * 2017-09-19 2022-11-01 北京搜狗科技发展有限公司 Input method, device and medium
US10445429B2 (en) 2017-09-21 2019-10-15 Apple Inc. Natural language understanding using vocabularies with compressed serialized tries
US10755051B2 (en) 2017-09-29 2020-08-25 Apple Inc. Rule-based natural language processing
US10636424B2 (en) 2017-11-30 2020-04-28 Apple Inc. Multi-turn canned dialog
US10733982B2 (en) 2018-01-08 2020-08-04 Apple Inc. Multi-directional dialog
US10733375B2 (en) 2018-01-31 2020-08-04 Apple Inc. Knowledge-based framework for improving natural language understanding
US10789959B2 (en) 2018-03-02 2020-09-29 Apple Inc. Training speaker recognition models for digital assistants
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
US10909331B2 (en) 2018-03-30 2021-02-02 Apple Inc. Implicit identification of translation payload with neural machine translation
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US10984780B2 (en) 2018-05-21 2021-04-20 Apple Inc. Global semantic word embeddings using bi-directional recurrent neural networks
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
DK179822B1 (en) 2018-06-01 2019-07-12 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
DK180639B1 (en) 2018-06-01 2021-11-04 Apple Inc DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT
US11386266B2 (en) 2018-06-01 2022-07-12 Apple Inc. Text correction
DK201870355A1 (en) 2018-06-01 2019-12-16 Apple Inc. Virtual assistant operation in multi-device environments
US10496705B1 (en) 2018-06-03 2019-12-03 Apple Inc. Accelerated task performance
US11010561B2 (en) 2018-09-27 2021-05-18 Apple Inc. Sentiment prediction from textual data
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US11170166B2 (en) 2018-09-28 2021-11-09 Apple Inc. Neural typographical error modeling via generative adversarial networks
US10839159B2 (en) 2018-09-28 2020-11-17 Apple Inc. Named entity normalization in a spoken dialog system
US11475898B2 (en) 2018-10-26 2022-10-18 Apple Inc. Low-latency multi-speaker speech recognition
US11638059B2 (en) 2019-01-04 2023-04-25 Apple Inc. Content playback on multiple devices
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
US11423908B2 (en) 2019-05-06 2022-08-23 Apple Inc. Interpreting spoken requests
DK201970509A1 (en) 2019-05-06 2021-01-15 Apple Inc Spoken notifications
US11475884B2 (en) 2019-05-06 2022-10-18 Apple Inc. Reducing digital assistant latency when a language is incorrectly determined
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
DK180129B1 (en) 2019-05-31 2020-06-02 Apple Inc. User activity shortcut suggestions
US11289073B2 (en) 2019-05-31 2022-03-29 Apple Inc. Device text to speech
DK201970510A1 (en) 2019-05-31 2021-02-11 Apple Inc Voice identification in digital assistant systems
US11496600B2 (en) 2019-05-31 2022-11-08 Apple Inc. Remote execution of machine-learned models
US11360641B2 (en) 2019-06-01 2022-06-14 Apple Inc. Increasing the relevance of new available information
US11227599B2 (en) 2019-06-01 2022-01-18 Apple Inc. Methods and user interfaces for voice-based control of electronic devices
WO2021056255A1 (en) 2019-09-25 2021-04-01 Apple Inc. Text detection using global geometry estimators
US11061543B1 (en) 2020-05-11 2021-07-13 Apple Inc. Providing relevant data items based on context
US11043220B1 (en) 2020-05-11 2021-06-22 Apple Inc. Digital assistant hardware abstraction
US11755276B2 (en) 2020-05-12 2023-09-12 Apple Inc. Reducing description length based on confidence
US11545145B2 (en) 2020-05-29 2023-01-03 Samsung Electronics Co., Ltd. Machine action based on language-independent graph rewriting of an utterance
US11490204B2 (en) 2020-07-20 2022-11-01 Apple Inc. Multi-device audio adjustment coordination
US11438683B2 (en) 2020-07-21 2022-09-06 Apple Inc. User identification using headphones
US11181988B1 (en) 2020-08-31 2021-11-23 Apple Inc. Incorporating user feedback into text prediction models via joint reward planning
US20220318500A1 (en) * 2021-04-06 2022-10-06 Talent Unlimited Online Services Private Limited System and method for generating contextualized text using a character-based convolutional neural network architecture

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1387651A (en) * 1999-11-05 2002-12-25 微软公司 System and iterative method for lexicon, segmentation and language model joint optimization
US6865528B1 (en) * 2000-06-01 2005-03-08 Microsoft Corporation Use of a unified language model
EP1724692A2 (en) * 2005-05-18 2006-11-22 Ramin O. Assadollahi Device incorporating improved text input mechanism using the context of the input
CN1871597A (en) * 2003-08-21 2006-11-29 伊迪利亚公司 System and method for associating documents with contextual advertisements
CN1954315A (en) * 2004-03-16 2007-04-25 Google公司 Systems and methods for translating chinese pinyin to chinese characters
US20080195388A1 (en) * 2007-02-08 2008-08-14 Microsoft Corporation Context based word prediction
CN101286094A (en) * 2007-04-10 2008-10-15 谷歌股份有限公司 Multi-mode input method editor

Family Cites Families (131)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5027406A (en) * 1988-12-06 1991-06-25 Dragon Systems, Inc. Method for interactive speech recognition and training
CA2006163A1 (en) * 1988-12-21 1990-06-21 Alfred B. Freeman Keyboard express typing system
US5477451A (en) 1991-07-25 1995-12-19 International Business Machines Corp. Method and system for natural language translation
US5963671A (en) * 1991-11-27 1999-10-05 International Business Machines Corporation Enhancement of soft keyboard operations using trigram prediction
US5664059A (en) * 1993-04-29 1997-09-02 Panasonic Technologies, Inc. Self-learning speaker adaptation based on spectral variation source decomposition
US5612690A (en) 1993-06-03 1997-03-18 Levy; David Compact keypad system and method
US5671426A (en) 1993-06-22 1997-09-23 Kurzweil Applied Intelligence, Inc. Method for organizing incremental search dictionary
US6304841B1 (en) * 1993-10-28 2001-10-16 International Business Machines Corporation Automatic construction of conditional exponential models from elementary features
US5510981A (en) * 1993-10-28 1996-04-23 International Business Machines Corporation Language translation apparatus and method using context-based translation models
US5748512A (en) * 1995-02-28 1998-05-05 Microsoft Corporation Adjusting keyboard
US5680511A (en) 1995-06-07 1997-10-21 Dragon Systems, Inc. Systems and methods for word recognition
WO1997005541A1 (en) 1995-07-26 1997-02-13 King Martin T Reduced keyboard disambiguating system
US5953541A (en) 1997-01-24 1999-09-14 Tegic Communications, Inc. Disambiguating system for disambiguating ambiguous input sequences by displaying objects associated with the generated input sequences in the order of decreasing frequency of use
US6009444A (en) 1997-02-24 1999-12-28 Motorola, Inc. Text input device and method
US6054941A (en) 1997-05-27 2000-04-25 Motorola, Inc. Apparatus and method for inputting ideographic characters
DE69837979T2 (en) 1997-06-27 2008-03-06 International Business Machines Corp. System for extracting multilingual terminology
EP0998714A1 (en) * 1997-07-22 2000-05-10 Microsoft Corporation System for processing textual inputs using natural language processing techniques
US6052657A (en) 1997-09-09 2000-04-18 Dragon Systems, Inc. Text segmentation and identification of topic using language models
ATE221222T1 (en) 1997-09-25 2002-08-15 Tegic Communications Inc SYSTEM FOR SUPPRESSING AMBIGUITY IN A REDUCED KEYBOARD
US6125342A (en) 1997-11-18 2000-09-26 L & H Applications Usa, Inc. Pronoun semantic analysis system and method
US6219632B1 (en) 1997-11-20 2001-04-17 International Business Machines Corporation System for the facilitation of supporting multiple concurrent languages through the use of semantic knowledge representation
JP3272288B2 (en) 1997-12-24 2002-04-08 日本アイ・ビー・エム株式会社 Machine translation device and machine translation method
US6052443A (en) * 1998-05-14 2000-04-18 Motorola Alphanumeric message composing method using telephone keypad
US6253169B1 (en) 1998-05-28 2001-06-26 International Business Machines Corporation Method for improvement accuracy of decision tree based text categorization
US6104989A (en) 1998-07-29 2000-08-15 International Business Machines Corporation Real time detection of topical changes and topic identification via likelihood based methods
US6393399B1 (en) 1998-09-30 2002-05-21 Scansoft, Inc. Compound word recognition
US6321192B1 (en) 1998-10-22 2001-11-20 International Business Machines Corporation Adaptive learning method and system that matches keywords using a parsed keyword data structure having a hash index based on an unicode value
DE19849855C1 (en) 1998-10-29 2000-04-27 Ibm Method for using a computer system to generate a text expression automatically while retaining meaning determines a statistical model on a number of preset pairs of word meanings and associated expressions.
US7712053B2 (en) 1998-12-04 2010-05-04 Tegic Communications, Inc. Explicit character filtering of ambiguous text entry
US6885317B1 (en) * 1998-12-10 2005-04-26 Eatoni Ergonomics, Inc. Touch-typable devices based on ambiguous codes and methods to design such devices
US6460015B1 (en) 1998-12-15 2002-10-01 International Business Machines Corporation Method, system and computer program product for automatic character transliteration in a text string object
US6362752B1 (en) 1998-12-23 2002-03-26 Motorola, Inc. Keypad with strokes assigned to key for ideographic text input
DE60003177T2 (en) * 1999-03-18 2004-05-06 602531 British Columbia Ltd., Vancouver DATA ENTRY FOR PERSONNEL COMPUTER DEVICES
US6204848B1 (en) 1999-04-14 2001-03-20 Motorola, Inc. Data entry apparatus having a limited number of character keys and method
US6275792B1 (en) 1999-05-05 2001-08-14 International Business Machines Corp. Method and system for generating a minimal set of test phrases for testing a natural commands grammar
US7030863B2 (en) 2000-05-26 2006-04-18 America Online, Incorporated Virtual keyboard system with automatic correction
US7610194B2 (en) 2002-07-18 2009-10-27 Tegic Communications, Inc. Dynamic database reordering system
US7750891B2 (en) 2003-04-09 2010-07-06 Tegic Communications, Inc. Selective input system based on tracking of motion parameters of an input device
US6327561B1 (en) 1999-07-07 2001-12-04 International Business Machines Corp. Customized tokenization of domain specific text via rules corresponding to a speech recognition vocabulary
US6865258B1 (en) * 1999-08-13 2005-03-08 Intervoice Limited Partnership Method and system for enhanced transcription
US6993476B1 (en) 1999-08-26 2006-01-31 International Business Machines Corporation System and method for incorporating semantic characteristics into the format-driven syntactic document transcoding framework
US6484136B1 (en) 1999-10-21 2002-11-19 International Business Machines Corporation Language model adaptation via network of similar users
US6848080B1 (en) * 1999-11-05 2005-01-25 Microsoft Corporation Language input architecture for converting one text form to another text form with tolerance to spelling, typographical, and conversion errors
US7177795B1 (en) * 1999-11-10 2007-02-13 International Business Machines Corporation Methods and apparatus for semantic unit based automatic indexing and searching in data archive systems
US6490549B1 (en) 2000-03-30 2002-12-03 Scansoft, Inc. Automatic orthographic transformation of a text stream
US6519557B1 (en) 2000-06-06 2003-02-11 International Business Machines Corporation Software and method for recognizing similarity of documents written in different languages based on a quantitative measure of similarity
US6724936B1 (en) 2000-08-23 2004-04-20 Art-Advanced Recognition Technologies, Ltd. Handwriting input device and method using a single character set
US7092870B1 (en) 2000-09-15 2006-08-15 International Business Machines Corporation System and method for managing a textual archive using semantic units
US7277732B2 (en) * 2000-10-13 2007-10-02 Microsoft Corporation Language input system for mobile devices
CA2323856A1 (en) * 2000-10-18 2002-04-18 602531 British Columbia Ltd. Method, system and media for entering data in a personal computing device
US6963831B1 (en) 2000-10-25 2005-11-08 International Business Machines Corporation Including statistical NLU models within a statistical parser
GB0103053D0 (en) * 2001-02-07 2001-03-21 Nokia Mobile Phones Ltd A communication terminal having a predictive text editor application
US7395205B2 (en) * 2001-02-13 2008-07-01 International Business Machines Corporation Dynamic language model mixtures with history-based buckets
US7426505B2 (en) 2001-03-07 2008-09-16 International Business Machines Corporation Method for identifying word patterns in text
US6813616B2 (en) 2001-03-07 2004-11-02 International Business Machines Corporation System and method for building a semantic network capable of identifying word patterns in text
US7385591B2 (en) * 2001-03-31 2008-06-10 Microsoft Corporation Out-of-vocabulary word determination and user interface for text input via reduced keypad keys
US6625600B2 (en) 2001-04-12 2003-09-23 Telelogue, Inc. Method and apparatus for automatically processing a user's communication
US7269546B2 (en) 2001-05-09 2007-09-11 International Business Machines Corporation System and method of finding documents related to other documents and of finding related words in response to a query to refine a search
US6925433B2 (en) 2001-05-09 2005-08-02 International Business Machines Corporation System and method for context-dependent probabilistic modeling of words and documents
US6671670B2 (en) 2001-06-27 2003-12-30 Telelogue, Inc. System and method for pre-processing information used by an automated attendant
US20030007018A1 (en) * 2001-07-09 2003-01-09 Giovanni Seni Handwriting user interface for personal digital assistants and the like
US7610189B2 (en) 2001-10-18 2009-10-27 Nuance Communications, Inc. Method and apparatus for efficient segmentation of compound words using probabilistic breakpoint traversal
US7075520B2 (en) 2001-12-12 2006-07-11 Zi Technology Corporation Ltd Key press disambiguation using a keypad of multidirectional keys
GB0200352D0 (en) 2002-01-09 2002-02-20 Ibm Finite state dictionary and method of production thereof
US7111248B2 (en) * 2002-01-15 2006-09-19 Openwave Systems Inc. Alphanumeric information input method
US7949513B2 (en) * 2002-01-22 2011-05-24 Zi Corporation Of Canada, Inc. Language module and method for use with text processing devices
US7175438B2 (en) 2002-03-01 2007-02-13 Digit Wireless Fast typing system and method
ATE436083T1 (en) 2002-05-23 2009-07-15 Digit Wireless Llc ELECTRICAL KEY SWITCH
US7493253B1 (en) 2002-07-12 2009-02-17 Language And Computing, Inc. Conceptual world representation natural language understanding system and method
US7151530B2 (en) * 2002-08-20 2006-12-19 Canesta, Inc. System and method for determining an input selected by a user through a virtual interface
AU2003274592A1 (en) 2002-11-28 2004-06-18 Koninklijke Philips Electronics N.V. Method to assign word class information
US7251367B2 (en) 2002-12-20 2007-07-31 International Business Machines Corporation System and method for recognizing word patterns based on a virtual keyboard layout
US7098896B2 (en) 2003-01-16 2006-08-29 Forword Input Inc. System and method for continuous stroke word-based text input
US7453439B1 (en) 2003-01-16 2008-11-18 Forward Input Inc. System and method for continuous stroke word-based text input
US7129932B1 (en) * 2003-03-26 2006-10-31 At&T Corp. Keyboard for interacting on small devices
US7475010B2 (en) 2003-09-03 2009-01-06 Lingospot, Inc. Adaptive and scalable method for resolving natural language ambiguities
US7366666B2 (en) 2003-10-01 2008-04-29 International Business Machines Corporation Relative delta computations for determining the meaning of language inputs
WO2005050474A2 (en) 2003-11-21 2005-06-02 Philips Intellectual Property & Standards Gmbh Text segmentation and label assignment with user interaction by means of topic specific language models and topic-specific label statistics
US8136050B2 (en) 2003-11-21 2012-03-13 Nuance Communications, Inc. Electronic device and user interface and input method therefor
US7362305B2 (en) * 2004-02-10 2008-04-22 Senseboard Technologies Ab Data input device
US7706616B2 (en) 2004-02-27 2010-04-27 International Business Machines Corporation System and method for recognizing word patterns in a very large vocabulary based on a virtual keyboard layout
US7555732B2 (en) * 2004-03-12 2009-06-30 Steven Van der Hoeven Apparatus method and system for a data entry interface
US7187365B2 (en) 2004-03-31 2007-03-06 Motorola, Inc. Indic intermediate code and electronic device therefor
US7758264B2 (en) * 2004-08-13 2010-07-20 5 Examples, Inc. One-row keyboard
US20130304453A9 (en) * 2004-08-20 2013-11-14 Juergen Fritsch Automated Extraction of Semantic Content and Generation of a Structured Document from Speech
US7373248B2 (en) * 2004-09-10 2008-05-13 Atx Group, Inc. Systems and methods for off-board voice-automated vehicle navigation
US20060055669A1 (en) * 2004-09-13 2006-03-16 Mita Das Fluent user interface for text entry on touch-sensitive display
US7610191B2 (en) 2004-10-06 2009-10-27 Nuance Communications, Inc. Method for fast semi-automatic semantic annotation
US20060117307A1 (en) * 2004-11-24 2006-06-01 Ramot At Tel-Aviv University Ltd. XML parser
US7630980B2 (en) 2005-01-21 2009-12-08 Prashant Parikh Automatic dynamic contextual data entry completion system
US7734471B2 (en) * 2005-03-08 2010-06-08 Microsoft Corporation Online learning for dialog systems
US7487461B2 (en) 2005-05-04 2009-02-03 International Business Machines Corporation System and method for issuing commands based on pen motions on a graphical keyboard
US20090193334A1 (en) * 2005-05-18 2009-07-30 Exb Asset Management Gmbh Predictive text input system and method involving two concurrent ranking means
US8374846B2 (en) * 2005-05-18 2013-02-12 Neuer Wall Treuhand Gmbh Text input device and method
EP1727024A1 (en) * 2005-05-27 2006-11-29 Sony Ericsson Mobile Communications AB Automatic language selection for text input in messaging context
US7496513B2 (en) * 2005-06-28 2009-02-24 Microsoft Corporation Combined input processing for a computing device
WO2007022079A2 (en) * 2005-08-11 2007-02-22 Lane David M System and method for the anticipation and execution of icon selection in graphical user interfaces
US20070094024A1 (en) 2005-10-22 2007-04-26 International Business Machines Corporation System and method for improving text input in a shorthand-on-keyboard interface
US20070115343A1 (en) * 2005-11-22 2007-05-24 Sony Ericsson Mobile Communications Ab Electronic equipment and methods of generating text in electronic equipment
US8010343B2 (en) 2005-12-15 2011-08-30 Nuance Communications, Inc. Disambiguation systems and methods for use in generating grammars
US7574672B2 (en) * 2006-01-05 2009-08-11 Apple Inc. Text entry interface for a portable communication device
CN101034390A (en) * 2006-03-10 2007-09-12 日电(中国)有限公司 Apparatus and method for verbal model switching and self-adapting
US8462118B2 (en) 2006-06-19 2013-06-11 Nuance Communications, Inc. Data entry system and method of entering data
US7586423B2 (en) * 2006-06-30 2009-09-08 Research In Motion Limited Handheld electronic device and method for dual-mode disambiguation of text input
US7724957B2 (en) * 2006-07-31 2010-05-25 Microsoft Corporation Two tiered text recognition
US7856350B2 (en) * 2006-08-11 2010-12-21 Microsoft Corporation Reranking QA answers using language modeling
US7774197B1 (en) * 2006-09-27 2010-08-10 Raytheon Bbn Technologies Corp. Modular approach to building large language models
US7698326B2 (en) * 2006-11-27 2010-04-13 Sony Ericsson Mobile Communications Ab Word prediction
AU2007339735A1 (en) * 2007-01-04 2008-07-10 Thinking Solutions Pty Ltd Linguistic analysis
US8074172B2 (en) 2007-01-05 2011-12-06 Apple Inc. Method, system, and graphical user interface for providing word recommendations
US8225203B2 (en) 2007-02-01 2012-07-17 Nuance Communications, Inc. Spell-check for a keyboard system with automatic correction
US8201087B2 (en) 2007-02-01 2012-06-12 Tegic Communications, Inc. Spell-check for a keyboard system with automatic correction
US8768689B2 (en) 2007-02-14 2014-07-01 Nuance Communications, Inc. Method and system for translation management of source language text phrases
US7809575B2 (en) 2007-02-27 2010-10-05 Nuance Communications, Inc. Enabling global grammars for a particular multimodal application
US8065624B2 (en) * 2007-06-28 2011-11-22 Panasonic Corporation Virtual keypad systems and methods
US7953692B2 (en) * 2007-12-07 2011-05-31 Microsoft Corporation Predicting candidates using information sources
US8010465B2 (en) * 2008-02-26 2011-08-30 Microsoft Corporation Predicting candidates using input scopes
US8484582B2 (en) 2008-05-12 2013-07-09 Nuance Communications, Inc. Entry selection from long entry lists
ATE501478T1 (en) * 2008-06-11 2011-03-15 Exb Asset Man Gmbh APPARATUS AND METHOD WITH IMPROVED TEXT ENTRY MECHANISM
US20100121870A1 (en) 2008-07-03 2010-05-13 Erland Unruh Methods and systems for processing complex language text, such as japanese text, on a mobile device
CN101620469B (en) * 2008-07-04 2013-03-27 索尼(中国)有限公司 Character input device and method thereof
US8117144B2 (en) 2008-12-12 2012-02-14 Nuance Communications, Inc. Generating predilection cohorts
US8669941B2 (en) 2009-01-05 2014-03-11 Nuance Communications, Inc. Method and apparatus for text entry
GB0917753D0 (en) * 2009-10-09 2009-11-25 Touchtype Ltd System and method for inputting text into electronic devices
US20100315266A1 (en) * 2009-06-15 2010-12-16 Microsoft Corporation Predictive interfaces with usability constraints
US9110515B2 (en) 2009-08-19 2015-08-18 Nuance Communications, Inc. Method and apparatus for text input
US20110106792A1 (en) * 2009-11-05 2011-05-05 I2 Limited System and method for word matching and indexing
US8782556B2 (en) * 2010-02-12 2014-07-15 Microsoft Corporation User-centric soft keyboard predictive technologies
US9092425B2 (en) * 2010-12-08 2015-07-28 At&T Intellectual Property I, L.P. System and method for feature-rich continuous space language models
US20120167009A1 (en) 2010-12-22 2012-06-28 Apple Inc. Combining timing and geometry information for typing correction
US9223497B2 (en) * 2012-03-16 2015-12-29 Blackberry Limited In-context word prediction and word correction

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1387651A (en) * 1999-11-05 2002-12-25 微软公司 System and iterative method for lexicon, segmentation and language model joint optimization
US6865528B1 (en) * 2000-06-01 2005-03-08 Microsoft Corporation Use of a unified language model
CN1871597A (en) * 2003-08-21 2006-11-29 伊迪利亚公司 System and method for associating documents with contextual advertisements
CN1954315A (en) * 2004-03-16 2007-04-25 Google公司 Systems and methods for translating chinese pinyin to chinese characters
EP1724692A2 (en) * 2005-05-18 2006-11-22 Ramin O. Assadollahi Device incorporating improved text input mechanism using the context of the input
US20080195388A1 (en) * 2007-02-08 2008-08-14 Microsoft Corporation Context based word prediction
CN101286094A (en) * 2007-04-10 2008-10-15 谷歌股份有限公司 Multi-mode input method editor

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106164893A (en) * 2014-04-04 2016-11-23 触摸式有限公司 System and method for one or more inputs that input is associated with multi input target
CN106164893B (en) * 2014-04-04 2020-06-05 触摸式有限公司 System and method for inputting one or more inputs associated with a multi-input target
US10802710B2 (en) 2014-04-04 2020-10-13 Touchtype Ltd System and method for inputting one or more inputs associated with a multi-input target
CN107688398A (en) * 2016-08-03 2018-02-13 中国科学院计算技术研究所 Determine the method and apparatus and input reminding method and device of candidate's input
CN107688398B (en) * 2016-08-03 2019-09-17 中国科学院计算技术研究所 It determines the method and apparatus of candidate input and inputs reminding method and device
CN108073679A (en) * 2017-11-10 2018-05-25 中国科学院信息工程研究所 Stochastic model set of strings generation method, equipment and readable storage medium storing program for executing under a kind of String matching scene
CN108073679B (en) * 2017-11-10 2021-09-28 中国科学院信息工程研究所 Random pattern string set generation method and device in string matching scene and readable storage medium
CN110929518A (en) * 2019-12-09 2020-03-27 朱利 Text sequence labeling algorithm using overlapping splitting rule
CN110929518B (en) * 2019-12-09 2023-08-04 朱利 Text sequence labeling algorithm using overlapping splitting rule

Also Published As

Publication number Publication date
CN107506047B (en) 2021-03-26
CN107506047A (en) 2017-12-22
CN103201707B (en) 2017-09-29
US20130253912A1 (en) 2013-09-26
US10146765B2 (en) 2018-12-04
US20160283464A1 (en) 2016-09-29
GB201016385D0 (en) 2010-11-10
US9384185B2 (en) 2016-07-05
WO2012042217A9 (en) 2012-10-26
WO2012042217A1 (en) 2012-04-05
EP2622437A1 (en) 2013-08-07
EP2622437B1 (en) 2019-09-25

Similar Documents

Publication Publication Date Title
CN103201707A (en) System and method for inputting text into electronic devices
CN107836000B (en) Improved artificial neural network method and electronic device for language modeling and prediction
Sun et al. Sentiment analysis for Chinese microblog based on deep neural networks with convolutional extension features
US10037319B2 (en) User input prediction
US9052748B2 (en) System and method for inputting text into electronic devices
JP6492238B2 (en) User input prediction
Paulus et al. Global belief recursive neural networks
Tsuruoka et al. Learning with lookahead: Can history-based models rival globally optimized models?
CN111539197B (en) Text matching method and device, computer system and readable storage medium
CN110825721A (en) Hypertension knowledge base construction and system integration method under big data environment
KR20160097352A (en) System and method for inputting images or labels into electronic devices
CN107807968B (en) Question answering device and method based on Bayesian network and storage medium
CN104081320A (en) User data input prediction
CN111858944A (en) Entity aspect level emotion analysis method based on attention mechanism
EP3549031B1 (en) Language data prediction with neural networks and online learning
Liu et al. A recurrent neural network based recommendation system
KR102361616B1 (en) Method and apparatus for recognizing named entity considering context
CN105830058B (en) Dialog manager
US20220269939A1 (en) Graph-based labeling rule augmentation for weakly supervised training of machine-learning-based named entity recognition
Sadr et al. Unified topic-based semantic models: a study in computing the semantic relatedness of geographic terms
Ahanin et al. A multi-label emoji classification method using balanced pointwise mutual information-based feature selection
CN112270189A (en) Question type analysis node generation method, question type analysis node generation system and storage medium
JP2016197289A (en) Parameter learning device, similarity calculation device and method, and program
Neiswanger et al. Modeling citation networks using latent random offsets
Lindén Word sense discovery and disambiguation

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20200901

Address after: 98052 Microsoft Avenue, Redmond, Washington, USA

Patentee after: MICROSOFT TECHNOLOGY LICENSING, LLC

Address before: England Atsushi

Patentee before: TOUCHTYPE Ltd.