CN105718070A - Pinyin long sentence continuous type-in input method and Pinyin long sentence continuous type-in input system - Google Patents

Pinyin long sentence continuous type-in input method and Pinyin long sentence continuous type-in input system Download PDF

Info

Publication number
CN105718070A
CN105718070A CN201610029530.8A CN201610029530A CN105718070A CN 105718070 A CN105718070 A CN 105718070A CN 201610029530 A CN201610029530 A CN 201610029530A CN 105718070 A CN105718070 A CN 105718070A
Authority
CN
China
Prior art keywords
chinese character
long sentence
bhmm
model
phonetic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610029530.8A
Other languages
Chinese (zh)
Inventor
周诚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Gaoxin Computer Systems Co Ltd
Original Assignee
Shanghai Gaoxin Computer Systems Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Gaoxin Computer Systems Co Ltd filed Critical Shanghai Gaoxin Computer Systems Co Ltd
Priority to CN201610029530.8A priority Critical patent/CN105718070A/en
Publication of CN105718070A publication Critical patent/CN105718070A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • G06F3/023Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
    • G06F3/0233Character input methods

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Document Processing Apparatus (AREA)

Abstract

The invention relates to the field of an input method, and discloses a pinyin long sentence continuous type-in input method and a Pinyin long sentence continuous type-in input system. The pinyin long sentence continuous type-in input method comprises the following steps of pre-building a BHMM (Bidirectional Hidden Markov Model); continuously receiving Pinyin codes input by a user; according to the BHMM and the continuously received Pinyin codes, obtaining a long sentence formed by each Chinese character with the highest occurring probability in the BHMM; and outputting the long sentence formed by each Chinese character with the highest occurring probability in the BHMM. The long sentence continuous type-in input system comprises a client and a cloud server, wherein the cloud server comprises a model building module, a matching module and a return module; and the client comprises a receiving module, a sending module and an outputting module. The method and the system have the advantage that by building the BHMM, the precision of outputting the correct result during whole sentence or long sentence Pinyin-to-Chinese-character conversion in Pinyin long sentence continuous type-pin input is improved.

Description

The long sentence of a kind of phonetic even beats input method and system thereof
Technical field
The present invention relates to input method field, particularly to Pinyin Input.
Background technology
Development and progress along with computer technology, the technology of spelling input method there has also been progressive and improves, particularly the long sentence of phonetic even lose into, prior art is based on context to find an optimum sentence when given phonetic, a dynamic programming problems can be regarded as, find shortest path.Its core technology, for adopting HMM, utilizes the processing mode of statistics natural language to carry out computing and process, obtains the sentence of optimum.The shortcoming of Markov is in that its independence assumption, and it assumes to have ignored moment event and contacting between all events before the moment, exactly because but also having had such it is assumed that make algorithm become simple and clear.HMM adopts the joint ensemble of production to solve this conditional probability problem, and this method is not suitable for processing the situation of a lot of feature description observation sequence.Markov it is assumed that when causing the exponent number of model more high, the statistical result that will not bring to model on the contrary, even if exponent number improves again, also cannot cover all of language phenomenon.All these defects, the output result precision after causing whole sentence that phonetic long sentence even beats or long sentence company to beat is not high.
Summary of the invention
It is an object of the invention to provide the long sentence of a kind of phonetic and even beat input method and system thereof, by setting up two-way hidden Markov BHMM model, improve phonetic long sentence and even lose out the precision of correct result.
For solving above-mentioned technical problem, embodiments of the present invention provide the long sentence of a kind of phonetic and even beat input method, comprise the steps of
Pre-build two-way hidden Markov BHMM model;In described BHMM model, the number of times that when probability of occurrence of each Chinese character in long sentence is by forward-propagating, the top n Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously, the number of times that during back propagation, rear N number of Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously determines;Described N is the natural number more than 1;
Persistently receive the phonetic code of user's input;
According to described BHMM model and the described phonetic code continuing to receive, obtain the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model forms;
Export the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model of described acquisition forms.
Corresponding to this, it is a further object to provide the long sentence of phonetic and even beat input system, comprise: client and cloud server;
Cloud server comprises:
Model building module, is used for pre-building two-way hidden Markov BHMM model;In described BHMM model, the number of times that when probability of occurrence of each Chinese character in long sentence is by forward-propagating, the top n Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously, the number of times that during back propagation, rear N number of Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously determines;Described N is the natural number more than 1;
Matching module, for according to described BHMM model with from the phonetic code continuing to receive of described client, obtaining the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model forms;
Return module, for the long sentence that described matching module obtains is back to described client;
Described client comprises:
Receiver module, for persistently receiving the phonetic code of user's input;
Sending module, for sending the described phonetic code continuing to receive to described cloud server;
Output module, for exporting the described long sentence that described cloud server returns.
Embodiment of the present invention is in terms of existing technologies, provide long sentence input method and the long sentence input system of a kind of phonetic, by setting up two-way hidden Markov BHMM model, improve the precision of output correct result during phonetic conversion Chinese character in whole sentence or long sentence.
It addition, described BHMM model is based upon in cloud server;The described phonetic code continuing to receive is sent to described cloud server by client in real time, by described cloud server according to BHMM model and the described phonetic code continuing to receive, obtains described long sentence in real time, optimizes phonetic and convert the efficiency of Chinese character.
It addition, described according to BHMM model with the described phonetic code continuing to receive, obtain in the step of the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model forms, obtain described long sentence according to below equation:
w 1 , w 2 , ... , w L = A r g M a x Π i = 1 L ( P ( w i | w i - 1 , w i - 2 , ... w i - N ) + P ( w i ′ | w i + 1 ′ , w i + 2 ′ , ... w i + N ′ ) )
Wherein, w1,w2,...,wLRepresenting each Chinese character in the described long sentence obtained, L is the natural number more than 1;P (wi|wi-1,wi-2,...wi-N) represent in forward-propagating process, Chinese character wiTop n Chinese character wi-1,wi-2,...wi-NWhen occurring, Chinese character wiThe probability occurred;P (w'i|w'i+1,w'i+2,...w'i+N) represent in back-propagation process, Chinese character w'iRear N number of Chinese character w'i+1,w'i+2,...w'i+NWhen occurring, Chinese character w'iThe probability occurred.
It addition, define in this two-way hidden Markov BHMM model, N is 3, is the three two-way HMMs in rank (being alternatively referred to as the two-way HMM of quaternary).Specifically, it is simply that in communication process, current state is respectively by the impacts of before it and thereafter three states, and the result accuracy of output is higher than single order and second order, and speed is faster.
It addition, after the described phonetic code persistently receiving user's input, according to described BHMM model and the described phonetic code continuing to receive, obtain the long sentence that the big each Chinese character of probability of occurrence time forms in this BHMM model;Export the long sentence that the big each Chinese character of described probability of occurrence time forms, facilitate user to select.
Accompanying drawing explanation
The long sentence that Fig. 1 is the phonetic according to first embodiment of the invention connects the flow chart beating input method;
Fig. 2 is according to the two-way HMM ultimate principle figure of single order in first embodiment of the invention;
Fig. 3 a-3d is the output Comparative result schematic diagram of two kinds of input methods in first embodiment of the invention;
The long sentence that Fig. 4 is the phonetic according to second embodiment of the invention connects the structural representation beating input system.
Detailed description of the invention
For making the object, technical solutions and advantages of the present invention clearly, below in conjunction with accompanying drawing, the embodiments of the present invention are explained in detail.But, it will be understood by those skilled in the art that in each embodiment of the present invention, propose many ins and outs in order to make reader be more fully understood that the application.But, even without these ins and outs with based on the many variations of following embodiment and amendment, it is also possible to realize the application each claim technical scheme required for protection.
First embodiment of the invention relates to the long sentence input method of a kind of phonetic, implements flow process as it is shown in figure 1, details are as follows:
In step S101, cloud server pre-builds two-way hidden Markov BHMM model;This BHMM model specifies, the number of times that when probability of occurrence of each Chinese character in long sentence is by forward-propagating, the top n Chinese character of this Chinese character and this Chinese character occurs simultaneously in data base, the number of times that during back propagation, rear N number of Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously determines.Wherein it is desired to illustrate, data base of the present embodiment saves common long sentence and the phonetic code of corresponding long sentence thereof.
It should be noted that, one sentence has L word (or word) to form, phonetic is corresponding to the state of two-way hidden Markov BHMM model, candidate word is corresponding to the output symbol of BHMM, turn to another one word (or word) corresponding to State Transferring from a word (or word), from 26 letters, carry out the input of phonetic corresponding to from the observation symbol of a State-output.Next for the two-way hidden Markov model of single order (namely N takes 1), the principle of two-way HMM being illustrated, ultimate principle is as shown in Figure 2.
Wherein, py1,py2,...,pyLRepresent L phonetic;W1,w2,...,wLRepresent the Chinese character of each phonetic corresponding output respectively, i.e. candidate word.In forward transmitting procedure, as illustrated in solid line in figure 2, when moment t, corresponding phonetic is pyt, corresponding Chinese character is wt, phonetic pytPrevious phonetic pyt-1When occurring, input Pinyin pytProbability be P (pyt|pyt-1);Phonetic pytBe converted to Chinese character wtProbability P (wt|pyt), and the probability P (w of last exported word (or word) that is correct or that be referred to alternatively as the optimum sentence of compositiont|wt-1).In reverse transfer process, as shown by the dash line in figure 2, equally when moment t, phonetic py'tPrevious phonetic py't+1When occurring, input Pinyin py'tProbability be P (pyt'|py't+1);Phonetic py'tBe converted to Chinese character w'tProbability be P (w't|py't);The probability finally exporting correct word (or word) is: P (wt'|wt'+1).At this, in forward-propagating process, with # (pyt-1,pyt) represent the transfer number of phonetic when being transferred to moment t-1 by moment t, namely from pyt-1Transfer to pytNumber of times;With # (pyt,wt) represent pytBe converted to wtNumber of times, # (wt-1,wt) represent word wt-1And wtThe number of times simultaneously occurred, with # (pyt) represent phonetic pytThe number of times occurred, with # (pyt-1) represent phonetic pyt-1The number of times occurred, # (wt-1) represent candidate word (or word) # (wt-1) represent candidate word (or word) wt-1The number of times occurred;In back-propagation process, with # (py't+1,pyt') represent from py't+1Transfer to py'tNumber of times, with # (py't,w't) represent py'tBe converted to w'tNumber of times, with # (w't+1,wt') represent w't+1And w'tThe number of times simultaneously occurred, with # (py't) represent py'tThe number of times occurred, with # (py't+1) represent py't+1The number of times occurred, with # (wt') represent candidate word (or word) w'tThe number of times occurred, with # (w't+1) represent candidate word (or word) w't+1The number of times occurred.
In step s 102, client continues to receive the phonetic code of user's input, and according to the rule input that spelling input method is set on 26 buttons of a to z on keyboard, the phonetic code continuing to receive is to be whole phonetic codes of each individual character respectively.
In step s 103, the phonetic code continuing to receive is sent to cloud server by client.
In step S104, cloud server is according to BHMM model and the described phonetic code continuing to receive, the long sentence that maximum and suboptimum each Chinese character of probability of occurrence forms in this BHMM model in real time.
Concrete BHMM model algorithm step is as described below:
In model, the number of times that when probability of occurrence of each Chinese character in long sentence is by forward-propagating, the top n Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously, the number of times that during back propagation, rear N number of Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously determines.In the present embodiment, N takes 3, namely illustrates for the three two-way HMMs in rank (being alternatively referred to as the two-way HMM of quaternary).The three two-way HMMs in rank refer to, in forward-propagating process, the event of moment t is relevant with moment t-1, t-2 and the event of moment t-3;In back-propagation process, the event of moment t is relevant with moment t+1, t+2 and t+3 moment event.Specific algorithm step is as follows:
In forward-propagating process, calculate the transition probability at t phonetic
P ( py t | py t - 1 , py t - 2 , py t - 3 ) ≈ # ( py t - 1 py t - 2 py t - 3 , py t ) # ( py t - 1 py t - 2 py t - 3 )
Calculate the probability being converted to Chinese character at t phonetic, thus can calculate candidate word
P ( w t | py t , py t - 1 , py t - 2 ) = # ( py t py t - 1 py t - 2 , w t ) # ( py t py t - 1 py t - 2 )
Calculate the probability exporting correct word (or word)
P ( w t | w t - 1 , w t - 2 , w t - 3 ) = # ( w t - 1 w t - 2 w t - 3 , w t ) # ( w t - 1 w t - 2 w t - 3 )
In back-propagation process, calculate the probability of t input Pinyin
P ( py t ′ | py t + 1 ′ , py t + 2 ′ , py t + 3 ′ ) ≈ # ( py t + 1 ′ py t + 2 ′ py t + 3 ′ , py t ′ ) # ( py t + 1 ′ py t + 2 ′ py t + 3 ′ )
Calculating in t, phonetic is converted to the probability of Chinese character, thus can obtain candidate word
P ( w t ′ | py t ′ , py t + 1 ′ , py t + 2 ′ ) ≈ # ( py t ′ py t + 1 ′ py t + 2 ′ , w t ′ ) # ( py t ′ py t + 1 ′ py t + 2 ′ )
Calculate the probability of correct output word (or word)
P ( w t ′ | w t + 1 ′ , w t + 2 ′ , w t + 3 ′ ) ≈ # ( w t + 1 ′ w t + 2 ′ w t + 3 ′ , w t ′ ) # ( w t + 1 ′ w t + 2 ′ w t + 3 ′ )
Thus can calculate, the event of t institute, namely produce the final probability of candidate word (or word)
w t = M a x w ∈ W ( P ( w t | w t - 1 , w t - 2 , w t - 3 ) + P ( w t ′ | w t + 1 ′ , w t + 2 ′ , w t + 3 ′ ) )
Calculate the output of final optimal sentence
w 1 , w 2 , ... , w L = A r g M a x w ∈ W ( P ( py 1 , py 2 , ... , py N | w 1 , w 2 , ... , w N ) · P ( w 1 , w 2 , ... , w N ) ) ≈ A r g M a x Π i = 1 L ( P ( w i | w i - 1 , w i - 2 , w i - 3 ) + P ( w i ′ | w i + 1 ′ , w i + 2 ′ , w i + 3 ′ ) )
Wherein, in forward-propagating process, # (pyt-1pyt-2pyt-3,pyt) represent the number of times that phonetic shifts, # (pyt-1pyt-2pyt-3) represent the common number of times occurred, # (pytpyt-1pyt-2,wt) represent phonetic and be converted to Chinese character wtNumber of times, # (pytpyt-1pyt-2) represent the number of times that three phonetics occur simultaneously, # (wt-1wt-2wt-3,wt) represent wt-1wt-2wt-3Transfer to wtNumber of times, # (wt-1wt-2wt-3) represent the number of times that three words (or word) occur simultaneously;In reverse procedure, # (py't+1py't+2py't+3,py't) represent the transfer number of phonetic, # (py't+1py't+2py't+3) represent the number of times that phonetic occurs simultaneously, # (py'tpy't+1py't+2,w't) represent phonetic and be converted to Chinese character w'tNumber of times, # (py'tpy't+1py't+2) represent the number of times that three phonetics occur simultaneously, # (w't+1w't+2w't+3,wt') represent w't+1w't+2w't+3Transfer to wt' number of times, # (w't+1w't+2w't+3) represent the number of times that three words (or word) occur simultaneously.
In step S105, the long sentence that the maximum and secondary big each Chinese character of probability of occurrence in this BHMM model that output obtains forms.
Without the input method output result that two-way hidden Markov model processes, with the comparing result of the input method output result processed through two-way HMM as illustrated in figs. 3 a-3d, Fig. 3 a-3d contains 4 groups of examples, often in group example, being arranged above the input method output result processed without two-way hidden Markov model, lower section is the input method output result processed through two-way HMM.
It is seen that, in the present embodiment by pre-building two-way hidden Markov BHMM model, in solution prior art phonetic long sentence input method poor accuracy defect;Improve the precision of output correct result during phonetic conversion Chinese character in whole sentence or long sentence.
Second embodiment of the invention relates to the long sentence of a kind of phonetic and even beats input system, and this system includes client and cloud server, and Fig. 4 is the structural representation of present embodiment.
Cloud server comprises model building module, matching module and return module;Model building module, is used for pre-building two-way hidden Markov BHMM model;Matching module, for according to described BHMM model with from the phonetic code continuing to receive of described client, obtaining the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model forms;Return module, for the long sentence that described matching module obtains is back to described client.
Client comprises receiver module, sending module and output module;Receiver module, for persistently receiving the phonetic code of user's input;Sending module, for sending the described phonetic code continuing to receive to described cloud server;Output module, for exporting the described long sentence that described cloud server returns.
It is seen that, present embodiment is the system embodiment corresponding with the first or second embodiment, and present embodiment can be worked in coordination enforcement with the first or second embodiment.The relevant technical details mentioned in first embodiment is still effective in the present embodiment, in order to reduce repetition, repeats no more here.Correspondingly, the relevant technical details mentioned in present embodiment is also applicable in the first embodiment.
It will be understood by those skilled in the art that the respective embodiments described above are to realize specific embodiments of the invention, and in actual applications, it is possible in the form and details it is done various change, without departing from the spirit and scope of the present invention.

Claims (8)

1. the long sentence of a phonetic even beats input method, it is characterised in that comprise the steps of
Pre-build two-way hidden Markov BHMM model;In described BHMM model, the number of times that when probability of occurrence of each Chinese character in long sentence is by forward-propagating, the top n Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously, the number of times that during back propagation, rear N number of Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously determines;Described N is the natural number more than 1;
Persistently receive the phonetic code of user's input;
According to described BHMM model and the described phonetic code continuing to receive, obtain the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model forms;
Export the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model of described acquisition forms.
2. the long sentence of phonetic according to claim 1 even beats input method, it is characterised in that
Described BHMM model is based upon in cloud server;
Described according to BHMM model with the described phonetic code continuing to receive, obtain in the step of the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model forms, comprise following sub-step:
The described phonetic code continuing to receive is sent to described cloud server by client in real time, by described cloud server according to BHMM model and the described phonetic code continuing to receive, obtains described long sentence in real time.
3. the long sentence of phonetic according to claim 1 even beats input method, it is characterized in that, described according to BHMM model with the described phonetic code continuing to receive, obtain in the step of the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model forms, obtain described long sentence according to below equation:
w 1 , w 2 , ... , w L = A r g M a x Π i = 1 L ( P ( w i | w i - 1 , w i - 2 , ... w i - N ) + P ( w i ′ | w i + 1 ′ , w i + 2 ′ , ... w i + N ′ ) )
Wherein, w1,w2,...,wLRepresenting each Chinese character in the described long sentence obtained, L is the natural number more than 1;P (wi|wi-1,wi-2,...wi-N) represent in forward-propagating process, Chinese character wiTop n Chinese character wi-1,wi-2,...wi-NWhen occurring, Chinese character wiThe probability occurred;P (w 'i|w′i+1,w′i+2,...w′i+N) represent in back-propagation process, Chinese character w 'iRear N number of Chinese character w 'i+1,w′i+2,...w′i+NWhen occurring, Chinese character w 'iThe probability occurred.
4. the long sentence of phonetic according to claim 3 even beats input method, it is characterised in that described N is 3.
5. the long sentence of phonetic according to claim 1 even beats input method, it is characterised in that also comprise the steps of
After the described phonetic code persistently receiving user's input, according to described BHMM model and the described phonetic code continuing to receive, obtain the long sentence that the big each Chinese character of probability of occurrence time forms in this BHMM model;
Export the long sentence that the big each Chinese character of described probability of occurrence time forms, select for user.
6. the long sentence of a phonetic even beats input system, it is characterised in that comprise: client and cloud server;
Cloud server comprises:
Model building module, is used for pre-building two-way hidden Markov BHMM model;In described BHMM model, the number of times that when probability of occurrence of each Chinese character in long sentence is by forward-propagating, the top n Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously, the number of times that during back propagation, rear N number of Chinese character of this Chinese character and this Chinese character occurs in data base simultaneously determines;Described N is the natural number more than 1;
Matching module, for according to described BHMM model with from the phonetic code continuing to receive of described client, obtaining the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model forms;
Return module, for the long sentence that described matching module obtains is back to described client;
Described client comprises:
Receiver module, for persistently receiving the phonetic code of user's input;
Sending module, for sending the described phonetic code continuing to receive to described cloud server;
Output module, for exporting the described long sentence that described cloud server returns.
7. the long sentence of phonetic according to claim 6 even beats input system, it is characterized in that, described according to BHMM model with the described phonetic code continuing to receive, obtain in the step of the long sentence that each Chinese character that probability of occurrence is maximum in this BHMM model forms, obtain described long sentence according to below equation:
w 1 , w 2 , ... , w L = A r g M a x Π i = 1 L ( P ( w i | w i - 1 , w i - 2 , ... w i - N ) + P ( w i ′ | w i + 1 ′ , w i + 2 ′ , ... w i + N ′ ) )
Wherein, w1,w2,...,wLRepresenting each Chinese character in the described long sentence obtained, L is the natural number more than 1;P (wi|wi-1,wi-2,...wi-N) represent in forward-propagating process, Chinese character wiTop n Chinese character wi-1,wi-2,...wi-NWhen occurring, Chinese character wiThe probability occurred;P (w 'i|w′i+1,w′i+2,...w′i+N) represent in back-propagation process, Chinese character w 'iRear N number of Chinese character w 'i+1,w′i+2,...w′i+NWhen occurring, Chinese character w 'iThe probability occurred.
8. the long sentence of phonetic according to claim 7 even beats input system, it is characterised in that described N is 3.
CN201610029530.8A 2016-01-16 2016-01-16 Pinyin long sentence continuous type-in input method and Pinyin long sentence continuous type-in input system Pending CN105718070A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610029530.8A CN105718070A (en) 2016-01-16 2016-01-16 Pinyin long sentence continuous type-in input method and Pinyin long sentence continuous type-in input system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610029530.8A CN105718070A (en) 2016-01-16 2016-01-16 Pinyin long sentence continuous type-in input method and Pinyin long sentence continuous type-in input system

Publications (1)

Publication Number Publication Date
CN105718070A true CN105718070A (en) 2016-06-29

Family

ID=56147890

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610029530.8A Pending CN105718070A (en) 2016-01-16 2016-01-16 Pinyin long sentence continuous type-in input method and Pinyin long sentence continuous type-in input system

Country Status (1)

Country Link
CN (1) CN105718070A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110673748A (en) * 2019-09-27 2020-01-10 北京百度网讯科技有限公司 Method and device for providing candidate long sentences in input method
CN110850998A (en) * 2019-11-04 2020-02-28 北京华宇信息技术有限公司 Intelligent word-forming calculation optimization method and device for Chinese input method
CN111144096A (en) * 2019-12-11 2020-05-12 心医国际数字医疗系统(大连)有限公司 HMM-based pinyin completion training method, completion model, completion method and completion input method
CN112527127A (en) * 2020-12-23 2021-03-19 北京百度网讯科技有限公司 Training method and device for input method long sentence prediction model, electronic equipment and medium
CN113589946A (en) * 2020-04-30 2021-11-02 北京搜狗科技发展有限公司 Data processing method and device and electronic equipment

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1110802A (en) * 1994-04-18 1995-10-25 北京超凡电子科技有限公司 Intelligence phonetic letter input method
WO1997044758A1 (en) * 1996-05-23 1997-11-27 Apple Computer, Inc. Methods and apparatuses for handwriting recognition
CN1182906A (en) * 1997-03-27 1998-05-27 罗仁 Intelligent phoneme-shape code input method and application thereof
CN102103416A (en) * 2009-12-17 2011-06-22 新浪网技术(中国)有限公司 Chinese character input method and device
CN102508829A (en) * 2011-11-24 2012-06-20 百度在线网络技术(北京)有限公司 Pinyin-input cloud result prediction methods, system and device
CN102915122A (en) * 2012-07-19 2013-02-06 上海交通大学 Intelligent mobile platform Pinyin (phonetic transcriptions of Chinese characters) input method based on language models
CN103870001A (en) * 2012-12-11 2014-06-18 百度国际科技(深圳)有限公司 Input method candidate item generating method and electronic device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1110802A (en) * 1994-04-18 1995-10-25 北京超凡电子科技有限公司 Intelligence phonetic letter input method
WO1997044758A1 (en) * 1996-05-23 1997-11-27 Apple Computer, Inc. Methods and apparatuses for handwriting recognition
CN1182906A (en) * 1997-03-27 1998-05-27 罗仁 Intelligent phoneme-shape code input method and application thereof
CN102103416A (en) * 2009-12-17 2011-06-22 新浪网技术(中国)有限公司 Chinese character input method and device
CN102508829A (en) * 2011-11-24 2012-06-20 百度在线网络技术(北京)有限公司 Pinyin-input cloud result prediction methods, system and device
CN102915122A (en) * 2012-07-19 2013-02-06 上海交通大学 Intelligent mobile platform Pinyin (phonetic transcriptions of Chinese characters) input method based on language models
CN103870001A (en) * 2012-12-11 2014-06-18 百度国际科技(深圳)有限公司 Input method candidate item generating method and electronic device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
吴鹏: "基于Markov链的整句输入算法研究与实现", 《中国优秀硕士学位论文全文数据库》 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110673748A (en) * 2019-09-27 2020-01-10 北京百度网讯科技有限公司 Method and device for providing candidate long sentences in input method
CN110850998A (en) * 2019-11-04 2020-02-28 北京华宇信息技术有限公司 Intelligent word-forming calculation optimization method and device for Chinese input method
CN111144096A (en) * 2019-12-11 2020-05-12 心医国际数字医疗系统(大连)有限公司 HMM-based pinyin completion training method, completion model, completion method and completion input method
CN111144096B (en) * 2019-12-11 2023-09-29 心医国际数字医疗系统(大连)有限公司 Pinyin completion training method, completion model, completion method and completion input method based on HMM
CN113589946A (en) * 2020-04-30 2021-11-02 北京搜狗科技发展有限公司 Data processing method and device and electronic equipment
CN112527127A (en) * 2020-12-23 2021-03-19 北京百度网讯科技有限公司 Training method and device for input method long sentence prediction model, electronic equipment and medium
CN112527127B (en) * 2020-12-23 2022-01-28 北京百度网讯科技有限公司 Training method and device for input method long sentence prediction model, electronic equipment and medium

Similar Documents

Publication Publication Date Title
CN105718070A (en) Pinyin long sentence continuous type-in input method and Pinyin long sentence continuous type-in input system
CN116127045B (en) Training method for generating large language model and man-machine voice interaction method based on model
CN108647214B (en) Decoding method based on deep neural network translation model
CN1667699B (en) Generating large units of graphonemes with mutual information criterion for letter to sound conversion
WO2020182122A1 (en) Text matching model generation method and device
CN107506823B (en) Construction method of hybrid neural network model for dialog generation
CN100535890C (en) Representation of a deleted interpolation n-gram language model in ARPA standard format
CN108052512A (en) A kind of iamge description generation method based on depth attention mechanism
CN110275939B (en) Method and device for determining conversation generation model, storage medium and electronic equipment
CN112487173B (en) Man-machine conversation method, device and storage medium
CN107346340A (en) A kind of user view recognition methods and system
CN116244416A (en) Training method for generating large language model and man-machine voice interaction method based on model
CN110457661B (en) Natural language generation method, device, equipment and storage medium
KR20210154705A (en) Method, apparatus, device and storage medium for matching semantics
CN116932723B (en) Man-machine interaction system and method based on natural language processing
CN116127020A (en) Method for training generated large language model and searching method based on model
CN113901907A (en) Image-text matching model training method, image-text matching method and device
CN116127046A (en) Training method for generating large language model and man-machine voice interaction method based on model
CN111709244A (en) Deep learning method for identifying causal relationship of contradictory dispute events
KR20220034070A (en) Model training method and apparatus, font library establishment method and apparatus, and storage medium
CN113326367B (en) Task type dialogue method and system based on end-to-end text generation
CN112270181B (en) Sequence labeling method, system, computer readable storage medium and computer device
KR20220034080A (en) Training method for circulary generating network model, method and apparatus for establishing word library, electronic device, recording medium and computer program
CN109992785B (en) Content calculation method, device and equipment based on machine learning
CN110297895B (en) Dialogue method and system based on free text knowledge

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20160629

RJ01 Rejection of invention patent application after publication