CN110929514A - Text proofreading method and device, computer readable storage medium and electronic equipment - Google Patents

Text proofreading method and device, computer readable storage medium and electronic equipment Download PDF

Info

Publication number
CN110929514A
CN110929514A CN201911144534.0A CN201911144534A CN110929514A CN 110929514 A CN110929514 A CN 110929514A CN 201911144534 A CN201911144534 A CN 201911144534A CN 110929514 A CN110929514 A CN 110929514A
Authority
CN
China
Prior art keywords
error
word
words
error correction
sentence
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911144534.0A
Other languages
Chinese (zh)
Other versions
CN110929514B (en
Inventor
苏海波
苏萌
刘译璟
姚震
檀玉飞
黄伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baifendian Information Science & Technology Co Ltd
Original Assignee
Beijing Baifendian Information Science & Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baifendian Information Science & Technology Co Ltd filed Critical Beijing Baifendian Information Science & Technology Co Ltd
Priority to CN201911144534.0A priority Critical patent/CN110929514B/en
Publication of CN110929514A publication Critical patent/CN110929514A/en
Application granted granted Critical
Publication of CN110929514B publication Critical patent/CN110929514B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The disclosure relates to a text proofreading method and device, a computer-readable storage medium and an electronic device. The method comprises the following steps: determining error correction information of each sentence in the text to be corrected, wherein the error correction information comprises error words and at least one corresponding error word; aiming at each error word, respectively determining a first co-occurrence frequency and a second co-occurrence frequency of the error word and preceding and following words in a preset corpus; acquiring semantic features aiming at each error correction word corresponding to the error word; and judging whether the error-correcting words are correct or not according to at least the first co-occurrence frequency, the second co-occurrence frequency and the semantic features. The correctness of the error-correcting words is judged, and the text proofreading accuracy can be improved. When the correctness of the error-correcting words is judged, the collocation of the words before and after the words and the context semantic features are comprehensively considered, so that the accuracy of judging the correctness of the error-correcting words can be ensured, and the text proofreading accuracy is further improved. In addition, the work of proofreading is intelligent and automatic, the pressure of manual proofreading is reduced, the work efficiency is improved, and the labor cost is reduced.

Description

Text proofreading method and device, computer readable storage medium and electronic equipment
Technical Field
The present disclosure relates to the field of computer technologies, and in particular, to a text proofreading method and apparatus, a computer-readable storage medium, and an electronic device.
Background
In text processing, a quite mature computer application system is used for inputting, editing and typesetting, but the intermediate link of text proofreading mainly depends on a manual processing stage, and becomes a bottleneck that the whole industry is restricted from developing and the working efficiency is influenced in the fields of news, publishing, office printing and the like. The manual text proofreading is time-consuming and labor-consuming, and the accuracy of the proofreading is difficult to ensure.
Based on the problems, at present, an N-gram model is mainly adopted to detect errors in a text and give an error correction suggestion, but the method only considers the matching problem of front and rear words, and the accuracy rate of text correction is low.
Disclosure of Invention
In order to overcome the problems in the related art, the present disclosure provides a text proofreading method, a text proofreading apparatus, a computer-readable storage medium, and an electronic device.
In order to achieve the above object, according to a first aspect of embodiments of the present disclosure, there is provided a text proofing method, including:
determining error correction information of each sentence in a text to be corrected, wherein the error correction information comprises error words and at least one error correction word corresponding to the error words;
respectively determining a first frequency of occurrence of the error word and a preceding word of the error word in a preset corpus and a second frequency of occurrence of the error word and a following word of the error word in the preset corpus for each error word;
aiming at each error-correcting word corresponding to the error word, acquiring semantic features of the error word and the error-correcting word;
and judging whether the error-correcting words are correct or not at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features.
Optionally, the determining whether the error correction word is correct according to at least the first co-occurrence frequency, the second co-occurrence frequency, and the semantic feature includes:
and inputting at least the first co-occurrence frequency, the second co-occurrence frequency and the semantic features into a preset xgboost model to judge whether the error-correcting word is correct.
Optionally, the method further comprises:
marking words belonging to a single word in the error words and the error-correcting words as 1, and marking words belonging to multiple words in the error words and the error-correcting words as 0;
the determining whether the error-correcting word is correct at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features includes:
and judging whether the error-correcting words are correct or not according to the marks of the error-correcting words, the first co-occurrence frequency, the second co-occurrence frequency and the semantic features.
Optionally, the obtaining semantic features of the error word and the error correction word includes:
replacing the error words in the initial sentence to which the error words belong with the error correction words to obtain error correction sentences;
respectively acquiring a first vector A ═ A corresponding to the initial sentences through a Bert model1,A2,…,Am) And the second vector B corresponding to the error correction sentence is equal to (B)1,B2,…,Bn) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, AiA first score for characterizing the reasonableness of the ith character appearing in the initial sentence, i ═ 1,2, …, m, BjA second score for characterizing the reasonableness of the jth character of the error-corrected sentence appearing in the error-corrected sentence, j being 1,2, …, n;
and determining a first difference value between the average value of each second score in the second vector and the average value of each first score in the first vector as the semantic features of the error word and the error correction word.
Optionally, the obtaining semantic features of the error word and the error correction word includes:
replacing the error words in the initial sentence to which the error words belong with the error correction words to obtain error correction sentences;
respectively acquiring a first vector A ═ A corresponding to the initial sentences through a Bert model1,A2,…,Am) And the second vector B corresponding to the error correction sentence is equal to (B)1,B2,…,Bn) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, AiA first score for characterizing the reasonableness of the ith character appearing in the initial sentence, i ═ 1,2, …, m, BjA second score for characterizing the reasonableness of the jth character of the error-corrected sentence appearing in the error-corrected sentence, j being 1,2, …, n;
sequentially converting the first vector and the second vector through each preset conversion function in a plurality of preset conversion functions to obtain a plurality of third vectors corresponding to the first vector and a plurality of fourth vectors corresponding to the second vector;
respectively calculating a second difference value of the average value of each third score in the third vector and the average value of each fourth score in the fourth vector aiming at the third vector and the fourth vector obtained by converting each preset conversion function;
determining a plurality of the second differences as semantic features of the error word and the error correction word.
Optionally, the method further comprises:
marking a maximum value of the plurality of second difference values as 1, and marking second difference values except for the maximum value of the plurality of second difference values as 0;
the determining whether the error-correcting word is correct at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features includes:
and judging whether the error-correcting word is correct or not according to the mark of the second difference value, the first co-occurrence frequency, the second co-occurrence frequency and the semantic features.
Optionally, before the step of determining error correction information of each sentence in the text to be collated, the method further includes:
preprocessing the text to be corrected to obtain a new text to be corrected;
the determining of the error correction information of each sentence in the text to be corrected includes:
and determining the error correction information of each sentence in the new text to be corrected.
According to a second aspect of the embodiments of the present disclosure, there is provided a text proofing apparatus, the apparatus including:
the device comprises a first determining module, a second determining module and a correcting module, wherein the first determining module is used for determining error correction information of each sentence in a text to be corrected, and the error correction information comprises error words and at least one error correction word corresponding to the error words;
a second determining module, configured to determine, for each error word determined by the first determining module, a first frequency of co-occurrence of the error word and a preceding word of the error word in a preset corpus and a second frequency of co-occurrence of the error word and a following word of the error word in the preset corpus, respectively;
an obtaining module, configured to obtain, for each error correction word corresponding to the error word determined by the first determining module, a semantic feature of the error correction word in a corresponding sentence;
and the judging module is used for judging whether the error-correcting word is correct or not at least according to the first co-occurrence frequency and the second co-occurrence frequency determined by the second determining module and the semantic features acquired by the acquiring module.
According to a third aspect of embodiments of the present disclosure, there is provided a computer readable storage medium, on which a computer program is stored, which when executed by a processor, performs the steps of the method provided by the first aspect of the present disclosure.
According to a fourth aspect of the embodiments of the present disclosure, there is provided an electronic apparatus including:
a memory having a computer program stored thereon;
a processor for executing the computer program in the memory to implement the steps of the method provided by the first aspect of the present disclosure.
In the technical scheme, firstly, error words existing in each sentence in the text to be corrected and at least one error correction word corresponding to each error word are determined; then, respectively determining a first co-occurrence frequency and a second co-occurrence frequency of the error word and a preceding word and a following word thereof for each error word, and simultaneously acquiring corresponding semantic features for each error correction word corresponding to the error word; and finally, judging the correctness of the error-correcting words at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features. After the error words and the corresponding error correction words are obtained, the correctness of the error correction words is further judged, and therefore the accuracy of text proofreading can be improved. In addition, when the correctness of the error-correcting words is judged, the matching problem of the front words and the rear words is considered, and the context semantic features of the words are combined, so that the correctness judgment precision of the error-correcting words can be ensured, and the accuracy of text proofreading is further improved. In addition, the text proofreading method enables proofreading work to be intelligent and automatic, reduces pressure of manual proofreading, improves working efficiency and reduces labor cost.
Additional features and advantages of the disclosure will be set forth in the detailed description which follows.
Drawings
The accompanying drawings, which are included to provide a further understanding of the disclosure and are incorporated in and constitute a part of this specification, illustrate embodiments of the disclosure and together with the description serve to explain the disclosure without limiting the disclosure. In the drawings:
FIG. 1 illustrates a flow diagram of a method of text proofing according to an exemplary embodiment.
FIG. 2A is a flow diagram illustrating a method of obtaining semantic features in accordance with an exemplary embodiment.
FIG. 2B is a flow diagram illustrating a method of obtaining semantic features in accordance with another exemplary embodiment.
FIG. 3 illustrates a flow diagram of a method of text proofing according to another exemplary embodiment.
FIG. 4 illustrates a flow diagram of a method of text proofing according to another exemplary embodiment.
FIG. 5 illustrates a flow diagram of a method of text proofing according to another exemplary embodiment.
FIG. 6 illustrates a block diagram of a text proofing apparatus according to an exemplary embodiment.
FIG. 7 is a block diagram illustrating an electronic device in accordance with an example embodiment.
FIG. 8 is a block diagram illustrating an electronic device in accordance with an example embodiment.
Detailed Description
The following detailed description of specific embodiments of the present disclosure is provided in connection with the accompanying drawings. It should be understood that the detailed description and specific examples, while indicating the present disclosure, are given by way of illustration and explanation only, not limitation.
FIG. 1 illustrates a flow diagram of a method of text proofing according to an exemplary embodiment. As shown in fig. 1, the above method may include the following steps 101 to 104.
In step 101, error correction information of each sentence in the text to be corrected is determined, wherein the error correction information includes an error word and at least one error correction word corresponding to the error word.
In the present disclosure, the error correction information may include 0, 1, or a plurality of error words, and each error word may correspond to 1 or more error correction words (i.e., error correction suggestions). For example, the sentence "today is really good. "the error correction information is: the error word "weather" and its corresponding error word "weather".
Also, in the present disclosure, the error correction information of each sentence in the text to be collated may be determined in various ways. In one embodiment, the error correction information may be obtained by an N-gram model. Specifically, this can be achieved by: (1) after the text to be corrected is obtained, word segmentation processing, part-of-speech tagging and other operations can be performed on the text; (2) based on a huge corpus, an N-gram model is adopted to perform positioning operation of error words, and a possible error position is detected; (3) further detecting the position which is possibly wrong by using a part-of-speech N-gram method, and if the position is detected to be unreasonable, judging the position to be wrong, and defining the position to be a wrong word; (4) and carrying out error correction processing on the error words and providing corresponding error correction words. Therefore, the error correction information of each sentence in the text to be corrected can be acquired.
In another embodiment, the error correction information may be obtained by reverse error correction (i.e., character matching). Specifically, after the text to be corrected is obtained, word segmentation processing can be performed on the text to be corrected; and then, for each word segmentation, respectively matching the word segmentation with each word in a preset word bank to obtain a plurality of matching degrees, if the maximum value of the plurality of matching degrees is smaller than a preset matching degree threshold value, determining the word segmentation as an error word, and determining a word corresponding to one or more matching degrees which are ranked earlier in the preset word bank and the plurality of matching degrees as an error correction word corresponding to the error word, wherein the matching degrees are sorted from big to small. Therefore, the error correction information of each sentence in the text to be corrected can be acquired.
In addition, the preset lexicon may be a database composed of words, and the preset lexicon may be generated based on a huge preset lexicon or an existing lexicon, which is not specifically limited in the present disclosure. The preset matching degree threshold may be a value set by a user, or may be a default empirical value, which is not specifically limited in this disclosure.
In another embodiment, the error correction information may be obtained by an N-gram model and a reverse error correction, and then the obtained error correction information is de-overlapped, and the de-overlapped and combined error correction information is used as the error correction information of each sentence in the text to be corrected. Therefore, the comprehensiveness of error word detection can be improved, and the accuracy of text proofreading is improved.
In step 102, for each error word, a first frequency of occurrence of the error word and a preceding word of the error word in a predetermined corpus and a second frequency of occurrence of the error word and a following word of the error word in the predetermined corpus are respectively determined.
In the present disclosure, the predetermined corpus may be a database composed of text sentences. Further, in order to make the first co-occurrence frequency and the second co-occurrence frequency determined according to the technical solution of the embodiment of the present application more fit with the text to be collated, the embodiment of the present application preferably makes up the preset corpus by the text sentences in the field to which the text to be collated belongs. In the embodiment of the present application, the generation method of the preset corpus is not limited.
The preceding word of the error word may be a word located before and next to the error word in the initial sentence to which the error word belongs, and the following word of the error word may be a word located after and next to the error word in the initial sentence to which the error word belongs. For example, the sentence "today is really good. "the error correction information is: the wrong word "qi day" and the corresponding error-correcting word "weather", wherein "today" is the antecedent of the wrong word "qi day" and "true" is the consequent of the wrong word "qi day".
The first and second co-occurrence frequencies may be determined in various ways. In one embodiment, the determination may be by a mathematical statistical method. Specifically, the number of times that the error word and its preceding word appear in the preset corpus (i.e., the first co-occurrence frequency), and the number of times that the error word and its following word appear in the preset corpus (i.e., the second co-occurrence frequency) may be counted for each error word.
In another embodiment, the first co-occurrence frequency and the second co-occurrence frequency can be determined by an N-gram model. Since the specific manner of determining the first and second co-occurrence frequencies through the N-gram model is well known to those skilled in the art, it is not described in detail in this disclosure.
In step 103, for each error correction word corresponding to the error word, semantic features of the error word and the error correction word are obtained.
In this disclosure, after the error correction information is obtained in step 101, corresponding semantic features may be obtained for each error correction word corresponding to each error word in the error correction information. In particular, the semantic features described above may be obtained in a number of ways. In one embodiment, this may be achieved through steps 1031 to 1033 shown in fig. 2A.
In step 1031, the error words in the initial sentence to which the error words belong are replaced with error correction words to obtain an error-corrected sentence.
Illustratively, the above sentence "this day is really good. "the error correction information is: the wrong word "sky" and its corresponding wrong word "weather", wherein, the initial sentence that the wrong word "sky" belongs to is: "today is very good. ". Thus, replacing the error word "weather" in the initial sentence with the error correction word "weather", the error correction sentence "weather is really good today" can be obtained. ".
In step 1032, a first vector corresponding to the initial sentence and a second vector corresponding to the error-corrected sentence are obtained through the Bert model, respectively.
In the present disclosure, a first vector a ═ a (a)1,A2,…,Am) Second vector B ═ B (B)1,B2,…,Bn) Wherein m and n are the number of characters (including punctuation marks) contained in the initial sentence and the number of characters (including punctuation marks) contained in the error-corrected sentence, respectively, AiTo characterize the reasonableness of the ith character appearing in the initial sentence, i is 1,2, …, m, BjIn order to represent the plausibility of the jth character in the error-corrected sentence, j is 1,2, …, n. The number of characters included in the error word and the number of characters included in the corresponding error correction word may be the same or different, and m and n may be equal or different.
Bert(Bidirectional encoderrepresentation from transforms, i.e., bidirectionally encoded representations of transformers) is a method of pre-training language representation, a model that can be downloaded and used free of charge. Wherein the model can be used to extract high quality linguistic features from sentences in the text to be collated. In the present disclosure, the first vector a ═ corresponding to the initial sentence may be obtained through the Bert model respectively (a ═ a)1,A2,…,Am) And the second vector B corresponding to the error correction sentence is equal to (B)1,B2,…,Bn)。
In step 1033, a first difference between the average of each second score in the second vector and the average of each first score in the first vector is determined as the semantic feature of the error word and the error correction word.
In this disclosure, the size of the first difference may reflect the quality of the corresponding error-correcting word, where when the first difference is greater than 0, it indicates that the corresponding error-correcting word is relatively good, and if the first difference is less than or equal to 0, it indicates that the corresponding error-correcting word is relatively poor.
The first vector a corresponding to the initial sentence to which the error word belongs is obtained in step 1032 (a ═ a)1,A2,…,Am) And the corresponding second vector B ═ B of the error correction sentence1,B2,…,Bn) Then, the first vector a ═ (a) can be calculated, respectively1,A2,…,Am) Each first score AiThe average value of (i.e.,
Figure BDA0002281811340000091
) Second vector B ═ B1,B2,…,Bn) Of (2) each second score BjThe average value of (i.e.,
Figure BDA0002281811340000092
) (ii) a After that, the second vector B is (B)1,B2,…,Bn) Of (2) each second score BjIs equal to the first vector a (a)1,A2,…,Am) Each first score AiA first difference of the average values of (i.e.,
Figure BDA0002281811340000093
) Semantic features are determined for the erroneous words and the error-correcting words.
In another embodiment, the method can be implemented by steps 1031, 1032, 1034, 1035 and 1036 shown in fig. 2B.
In step 1031, the error words in the initial sentence to which the error words belong are replaced with error correction words to obtain an error-corrected sentence.
In step 1032, a first vector corresponding to the initial sentence and a second vector corresponding to the error-corrected sentence are obtained through the Bert model, respectively.
In step 1034, the first vector and the second vector are transformed sequentially through each of a plurality of preset transformation functions, so as to obtain a plurality of third vectors corresponding to the first vector and a plurality of fourth vectors corresponding to the second vector.
In this disclosure, the preset transformation function may be used to transform the first vector and the second vector to obtain a third vector and a fourth vector. The plurality of preset transformation functions may be preset by a user, or may be default (for example, y (X) is the ith preset transformation function of the plurality of preset transformation functions, X + CiX is a first vector or a second vector, Y (X) is a third vector corresponding to the first vector or a fourth vector corresponding to the second vector, CiConstant vectors corresponding to the ith preset transformation function and constant vectors C corresponding to the preset transformation functionsiDifferent), are not specifically limited in this disclosure.
After the first vectors of the initial sentences to which the error words belong and the second vectors corresponding to the corresponding error correction sentences are obtained in step 1032, each first vector can be transformed by using a plurality of preset transformation functions, so that a plurality of third vectors are obtained; meanwhile, each second vector may be transformed by using the plurality of preset transformation functions, so as to obtain a plurality of fourth vectors.
In step 1035, for each of the third and fourth vectors transformed by the predetermined transformation function, a second difference between the average of the third scores in the third vector and the average of the fourth scores in the fourth vector is calculated.
In this disclosure, after the plurality of third vectors corresponding to the initial sentence to which the error word belongs and the plurality of fourth vectors corresponding to the corresponding error-corrected sentence are obtained through the step 1034, a second difference value between an average value of each third score in the third vector and an average value of each fourth score in the fourth vector may be calculated for each of the third vectors and the fourth vectors obtained by transforming each preset transformation function. Thereby, a plurality of second difference values can be obtained.
It should be noted that the number of the preset transformation functions may be set by a user, or may be a default (for example, 6), and is not specifically limited in this disclosure.
In step 1036, a plurality of second differences are determined as semantic features of the error words and the error correction words.
After the second differences are obtained in step 1035, the second differences can be determined as semantic features of the error word and the error correction word.
Returning to fig. 1, in step 104, it is determined whether the error-correcting word is correct according to at least the first co-occurrence frequency, the second co-occurrence frequency, and the semantic features.
In this disclosure, for each error word, after the first frequency of occurrence of the error word and its preceding word in the preset corpus, the second frequency of occurrence of the error word and its following word in the preset corpus are obtained in the above step 102, and the corresponding semantic features are obtained in the above step 103, they may be input into a preset xgboost (extreme gradient boost) model to determine whether the error word corresponding to the error word is correct. The xgboost model is a binary model, the output of the model may be 0 or 1, 0 may be used to characterize the error-correcting word error, and 1 may be used to characterize the error-correcting word correct.
In addition, in the present disclosure, the construction of the xgboost model may be performed based on the manually collated text. Firstly, in the same manner as the step 102, for each sample error word obtained by manual proofreading, obtaining a third co-occurrence frequency of the sample error word and a previous word thereof in a preset corpus and a fourth co-occurrence frequency of the sample error word and a next word thereof in the preset corpus; meanwhile, corresponding reference semantic features are obtained in the same manner as in the step 103; and then, inputting at least the third co-occurrence frequency, the fourth co-occurrence frequency and the reference semantic features into an initial xgboost model for training to obtain the preset xgboost model. The specific construction manner of the xgboost model is well known to those skilled in the art, and therefore, will not be described in detail in the present disclosure.
In addition, the xgboost model may be optimized, for example, the model may be optimized by modifying the model parameters according to the training and testing effects.
In one embodiment, whether the error-correcting word is correct or not can be determined according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features. Specifically, the first co-occurrence frequency, the second co-occurrence frequency, and the semantic features may be input into the preset xgboost model, so as to determine whether the error-correcting word corresponding to the error word is correct according to the output of the preset xgboost model. That is, when the output of the preset xgboost model is 0, it indicates that the error-correcting word corresponding to the error word is incorrect; and when the output of the preset xgboost model is 1, indicating that the error-correcting word corresponding to the error word is correct.
In another embodiment, in order to further improve the accuracy of text proofreading, when determining whether the error-correcting word is correct, the number information of the characters included in the error-correcting word and the error-correcting word may be considered in addition to the first co-occurrence frequency, the second co-occurrence frequency, and the semantic features. Specifically, as shown in fig. 3, the method may further include the following step 105.
In step 105, a word belonging to a single word among the error word and the error-correcting word is labeled as 1, and a word belonging to a multi-word among the error word and the error-correcting word is labeled as 0.
After the error correction information is obtained in step 101, it may be determined whether each error word or error correction word is a single word, and if the error word or error correction word is a single word, the error word or error correction word is marked as 1, otherwise, the error word or error correction word is marked as 0. Thus, whether the error-correcting words are correct or not can be determined together according to the marks of the error-correcting words, the first co-occurrence frequency, the second co-occurrence frequency and the semantic features. Specifically, the flag of the error word, the first co-occurrence frequency, the second co-occurrence frequency, and the semantic features may be input into the preset xgboost model, so as to determine whether the error word corresponding to the error word is correct according to the output of the preset xgboost model.
It should be noted that, the step 105 may be executed before the step 102 or the step 103, may be executed after the step 102 or the step 103, may be executed simultaneously with the step 102 or the step 103, and is not particularly limited in this disclosure.
In another embodiment, in order to further improve the accuracy of text proofreading, when determining whether the error-correcting word is correct, the second difference information may be considered in addition to the first co-occurrence frequency, the second co-occurrence frequency, and the semantic features. Specifically, as shown in fig. 4, the method may further include the following step 106.
In step 106, the maximum value of the plurality of second difference values is marked as 1, and the second difference values except the maximum value of the plurality of second difference values are marked as 0.
After the plurality of second difference values are obtained in step 103 (i.e., step 1035), the maximum value of the plurality of second difference values may be marked as 1, and the second difference values except the maximum value of the plurality of second difference values may be marked as 0. In this way, whether the error-correcting word is correct or not can be determined jointly according to the mark of each second difference value, the first co-occurrence frequency, the second co-occurrence frequency and the semantic features. Specifically, the flag, the first co-occurrence frequency, the second co-occurrence frequency, and the semantic feature of each second difference may be input into the preset xgboost model, so as to determine whether the error-correcting word corresponding to the error word is correct according to the output of the preset xgboost model.
It should be noted that step 103 may be executed before step 102, may be executed after step 102, may be executed simultaneously with step 102, and is not particularly limited in this disclosure.
In the technical scheme, firstly, error words existing in each sentence in the text to be corrected and at least one error correction word corresponding to each error word are determined; then, respectively determining a first co-occurrence frequency and a second co-occurrence frequency of the error word and a preceding word and a following word thereof for each error word, and simultaneously acquiring corresponding semantic features for each error correction word corresponding to the error word; and finally, judging the correctness of the error-correcting words at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features. After the error words and the corresponding error correction words are obtained, the correctness of the error correction words is further judged, and therefore the accuracy of text proofreading can be improved. In addition, when the correctness of the error-correcting words is judged, the matching problem of the front words and the rear words is considered, and the context semantic features of the words are combined, so that the correctness judgment precision of the error-correcting words can be ensured, and the accuracy of text proofreading is further improved. In addition, the text proofreading method enables proofreading work to be intelligent and automatic, reduces pressure of manual proofreading, improves working efficiency and reduces labor cost.
In addition, in order to further improve the accuracy of text proofreading, before error correction information is obtained according to the text to be proofread, preprocessing can be performed on the text to be proofread. Specifically, as shown in fig. 5, before the step 101, the method may further include the following step 107.
In step 107, the text to be collated is preprocessed to obtain a new text to be collated.
In the present disclosure, this preprocessing may include filtering out of illegal characters (e.g., spaces, empty lines, etc.). After the text to be collated is preprocessed, a new text to be collated may be obtained, and then, error correction information may be obtained based on the new text to be collated, that is, the above step 101 is performed.
FIG. 6 is a block diagram illustrating a text proofing apparatus according to an example embodiment. Referring to fig. 6, the apparatus 600 may include: a first determining module 601, configured to determine error correction information of each sentence in a text to be corrected, where the error correction information includes an error word and at least one error correction word corresponding to the error word; a second determining module 602, configured to determine, for each error word determined by the first determining module 601, a first frequency of occurrence of the error word and a preceding word of the error word in a preset corpus, and a second frequency of occurrence of the error word and a following word of the error word in the preset corpus, respectively; an obtaining module 603, configured to obtain, for each error correction word corresponding to the error word determined by the first determining module 601, a semantic feature of the error correction word in a corresponding sentence; a determining module, configured to determine whether the error correcting word is correct at least according to the first co-occurrence frequency and the second co-occurrence frequency determined by the second determining module 602 and the semantic features acquired by the acquiring module 603.
Optionally, the determining module 604 is configured to input at least the first co-occurrence frequency, the second co-occurrence frequency, and the semantic features into a preset xgboost model to determine whether the error-correcting word is correct.
Optionally, the apparatus 600 may further include: the first marking module is used for marking words belonging to a single word in the error words and the error-correcting words as 1 and marking words belonging to multiple words in the error words and the error-correcting words as 0;
the decision module 604 is configured to: and judging whether the error-correcting words are correct or not according to the marks of the error words, the marks of the error-correcting words, the first co-occurrence frequency, the second co-occurrence frequency and the semantic features, which are obtained by the first marking module.
Optionally, the obtaining module 603 includes: the replacing submodule is used for replacing the error words in the initial sentence to which the error words belong with the error correction words to obtain error correction sentences; a vector obtaining submodule, configured to obtain, through a Bert model, first vectors a corresponding to the initial sentences respectively=(A1,A2,…,Am) And the second vector B corresponding to the error correction sentence is equal to (B)1,B2,…,Bn) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, AiA first score for characterizing the reasonableness of the ith character appearing in the initial sentence, i ═ 1,2, …, m, BjA second score for characterizing the reasonableness of the jth character of the error-corrected sentence appearing in the error-corrected sentence, j being 1,2, …, n; a first determining submodule, configured to determine a first difference between an average of each second score in the second vector and an average of each first score in the first vector as a semantic feature of the error word and the error correction word.
Optionally, the obtaining module 603 includes: the replacing submodule is used for replacing the error words in the initial sentence to which the error words belong with the error correction words to obtain error correction sentences; a vector obtaining sub-module, configured to obtain, through a Bert model, first vectors a ═ respectively corresponding to the initial sentences (a ═ a)1,A2,…,Am) And the second vector B corresponding to the error correction sentence is equal to (B)1,B2,…,Bn) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, AiA first score for characterizing the reasonableness of the ith character appearing in the initial sentence, i ═ 1,2, …, m, BjA second score for characterizing the reasonableness of the jth character of the error-corrected sentence appearing in the error-corrected sentence, j being 1,2, …, n; the transformation submodule is used for respectively transforming the first vector and the second vector through each preset transformation function in a plurality of preset transformation functions in sequence to obtain a plurality of third vectors corresponding to the first vector and a plurality of fourth vectors corresponding to the second vector; a calculation submodule, configured to calculate, for a third vector and a fourth vector obtained by transforming each of the preset transformation functions, an average value of each third score in the third vector and each fourth score in the fourth vector respectivelyA second difference in the averaged values; and the second determining submodule is used for determining a plurality of second difference values as semantic features of the error words and the error correction words.
Optionally, the apparatus 600 may further include: a second marking module, configured to mark a maximum value of the second difference values obtained by the calculation sub-module as 1, and mark a second difference value, excluding the maximum value, of the second difference values as 0;
the decision module 604 is configured to: and judging whether the error-correcting word is correct or not according to the mark of the second difference value, the first co-occurrence frequency, the second co-occurrence frequency and the semantic features, which are obtained by the second marking module.
Optionally, the apparatus 600 may further include: the preprocessing module is used for preprocessing the text to be corrected before the first determining module determines the error correction information of each sentence in the text to be corrected, so as to obtain a new text to be corrected;
and the first determining module is used for determining the error correction information of each sentence in the new text to be corrected.
With regard to the apparatus in the above-described embodiment, the specific manner in which each module performs the operation has been described in detail in the embodiment related to the method, and will not be elaborated here.
The present disclosure also provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements the steps of the above-mentioned text proofing method provided by the present disclosure.
Fig. 7 is a block diagram illustrating an electronic device 700 in accordance with an example embodiment. As shown in fig. 7, the electronic device 700 may include: a processor 701 and a memory 702. The electronic device 700 may also include one or more of a multimedia component 703, an input/output (I/O) interface 704, and a communication component 705.
The processor 701 is configured to control the overall operation of the electronic device 700, so as to complete all or part of the steps in the text proofreading method. The memory 702 is used to store various types of data to support operation at the electronic device 700, such as instructions for any application or method operating on the electronic device 700 and application-related data, such as contact data, transmitted and received messages, pictures, audio, video, and the like. The Memory 702 may be implemented by any type of volatile or non-volatile Memory device or combination thereof, such as Static Random Access Memory (SRAM), Electrically Erasable Programmable Read-Only Memory (EEPROM), Erasable Programmable Read-Only Memory (EPROM), Programmable Read-Only Memory (PROM), Read-Only Memory (ROM), magnetic Memory, flash Memory, magnetic disk, or optical disk. The multimedia components 703 may include screen and audio components. Wherein the screen may be, for example, a touch screen and the audio component is used for outputting and/or inputting audio signals. For example, the audio component may include a microphone for receiving external audio signals. The received audio signal may further be stored in the memory 702 or transmitted through the communication component 705. The audio assembly also includes at least one speaker for outputting audio signals. The I/O interface 704 provides an interface between the processor 701 and other interface modules, such as a keyboard, mouse, buttons, etc. These buttons may be virtual buttons or physical buttons. The communication component 705 is used for wired or wireless communication between the electronic device 700 and other devices. Wireless communication, such as Wi-Fi, bluetooth, Near Field Communication (NFC), 2G, 3G, 4G, NB-IOT, eMTC, or other 5G, etc., or a combination of one or more of them, which is not limited herein. The corresponding communication component 705 may thus include: Wi-Fi module, Bluetooth module, NFC module, etc.
In an exemplary embodiment, the electronic Device 700 may be implemented by one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), Field Programmable Gate Arrays (FPGAs), controllers, microcontrollers, microprocessors, or other electronic components for performing the text proofing method described above.
In another exemplary embodiment, a computer readable storage medium comprising program instructions which, when executed by a processor, implement the steps of the text proofing method described above is also provided. For example, the computer readable storage medium may be the memory 702 described above that includes program instructions executable by the processor 701 of the electronic device 700 to perform the text proofing method described above.
Fig. 8 is a block diagram illustrating an electronic device 800 in accordance with an example embodiment. For example, the electronic device 800 may be provided as a server. Referring to fig. 8, an electronic device 800 includes a processor 822, which may be one or more in number, and a memory 832 for storing computer programs executable by the processor 822. The computer programs stored in memory 832 may include one or more modules that each correspond to a set of instructions. Further, the processor 822 may be configured to execute the computer program to perform the text proofing method described above.
Additionally, the electronic device 800 may also include a power component 826 and a communication component 850, the power component 826 may be configured to perform power management of the electronic device 800, and the communication component 850 may be configured to enable communication, e.g., wired or wireless communication, of the electronic device 800. The electronic device 800 may also include input/output (I/O) interfaces 858. The electronic device 800 may operate based on an operating system stored in the memory 832, such as Windows Server, Mac OSXTM, UnixTM, LinuxTM, and the like.
In another exemplary embodiment, a computer readable storage medium comprising program instructions which, when executed by a processor, implement the steps of the text proofing method described above is also provided. For example, the computer readable storage medium may be the memory 832 including program instructions described above that are executable by the processor 822 of the electronic device 800 to perform the text proofing method described above.
In another exemplary embodiment, a computer program product is also provided, which comprises a computer program executable by a programmable apparatus, the computer program having code portions for performing the text proofing method described above when executed by the programmable apparatus.
The preferred embodiments of the present disclosure are described in detail with reference to the accompanying drawings, however, the present disclosure is not limited to the specific details of the above embodiments, and various simple modifications may be made to the technical solution of the present disclosure within the technical idea of the present disclosure, and these simple modifications all belong to the protection scope of the present disclosure.
It should be noted that the various features described in the above embodiments may be combined in any suitable manner without departing from the scope of the invention. In order to avoid unnecessary repetition, various possible combinations will not be separately described in this disclosure.
In addition, any combination of various embodiments of the present disclosure may be made, and the same should be considered as the disclosure of the present disclosure, as long as it does not depart from the spirit of the present disclosure.

Claims (10)

1. A method of text proofing, the method comprising:
determining error correction information of each sentence in a text to be corrected, wherein the error correction information comprises error words and at least one error correction word corresponding to the error words;
respectively determining a first frequency of occurrence of the error word and a preceding word of the error word in a preset corpus and a second frequency of occurrence of the error word and a following word of the error word in the preset corpus for each error word;
aiming at each error-correcting word corresponding to the error word, acquiring semantic features of the error word and the error-correcting word;
and judging whether the error-correcting words are correct or not at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features.
2. The method according to claim 1, wherein said determining whether the error correction word is correct according to at least the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature comprises:
and inputting at least the first co-occurrence frequency, the second co-occurrence frequency and the semantic features into a preset xgboost model to judge whether the error-correcting word is correct.
3. The method of claim 1, further comprising:
marking words belonging to a single word in the error words and the error-correcting words as 1, and marking words belonging to multiple words in the error words and the error-correcting words as 0;
the determining whether the error-correcting word is correct at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features includes:
and judging whether the error-correcting words are correct or not according to the marks of the error-correcting words, the first co-occurrence frequency, the second co-occurrence frequency and the semantic features.
4. The method according to any one of claims 1 to 3, wherein the obtaining semantic features of the error word and the error correction word comprises:
replacing the error words in the initial sentence to which the error words belong with the error correction words to obtain error correction sentences;
respectively acquiring a first vector A ═ A corresponding to the initial sentences through a Bert model1,A2,…,Am) And the second vector B corresponding to the error correction sentence is equal to (B)1,B2,…,Bn) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, AiA first score for characterizing the reasonableness of the ith character appearing in the initial sentence, i ═ 1,2, …, m, BjA second score for characterizing the reasonableness of the jth character of the error-corrected sentence appearing in the error-corrected sentence, j being 1,2, …, n;
and determining a first difference value between the average value of each second score in the second vector and the average value of each first score in the first vector as the semantic features of the error word and the error correction word.
5. The method according to any one of claims 1 to 3, wherein the obtaining semantic features of the error word and the error correction word comprises:
replacing the error words in the initial sentence to which the error words belong with the error correction words to obtain error correction sentences;
respectively acquiring a first vector A ═ A corresponding to the initial sentences through a Bert model1,A2,…,Am) And the second vector B corresponding to the error correction sentence is equal to (B)1,B2,…,Bn) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, AiA first score for characterizing the reasonableness of the ith character appearing in the initial sentence, i ═ 1,2, …, m, BjA second score for characterizing the reasonableness of the jth character of the error-corrected sentence appearing in the error-corrected sentence, j being 1,2, …, n;
sequentially converting the first vector and the second vector through each preset conversion function in a plurality of preset conversion functions to obtain a plurality of third vectors corresponding to the first vector and a plurality of fourth vectors corresponding to the second vector;
respectively calculating a second difference value of the average value of each third score in the third vector and the average value of each fourth score in the fourth vector aiming at the third vector and the fourth vector obtained by converting each preset conversion function;
determining a plurality of the second differences as semantic features of the error word and the error correction word.
6. The method of claim 5, further comprising:
marking a maximum value of the plurality of second difference values as 1, and marking second difference values except for the maximum value of the plurality of second difference values as 0;
the determining whether the error-correcting word is correct at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features includes:
and judging whether the error-correcting word is correct or not according to the mark of the second difference value, the first co-occurrence frequency, the second co-occurrence frequency and the semantic features.
7. A method according to any of claims 1-3, characterized in that, prior to the step of determining error correction information for each sentence in the text to be collated, the method further comprises:
preprocessing the text to be corrected to obtain a new text to be corrected;
the determining of the error correction information of each sentence in the text to be corrected includes:
and determining the error correction information of each sentence in the new text to be corrected.
8. A text proofing apparatus, characterized in that the apparatus comprises:
the device comprises a first determining module, a second determining module and a correcting module, wherein the first determining module is used for determining error correction information of each sentence in a text to be corrected, and the error correction information comprises error words and at least one error correction word corresponding to the error words;
a second determining module, configured to determine, for each error word determined by the first determining module, a first frequency of co-occurrence of the error word and a preceding word of the error word in a preset corpus and a second frequency of co-occurrence of the error word and a following word of the error word in the preset corpus, respectively;
an obtaining module, configured to obtain, for each error correction word corresponding to the error word determined by the first determining module, a semantic feature of the error correction word in a corresponding sentence;
and the judging module is used for judging whether the error-correcting word is correct or not at least according to the first co-occurrence frequency and the second co-occurrence frequency determined by the second determining module and the semantic features acquired by the acquiring module.
9. A computer-readable storage medium, on which a computer program is stored which, when being executed by a processor, carries out the steps of the method according to any one of claims 1 to 7.
10. An electronic device, comprising:
a memory having a computer program stored thereon;
a processor for executing the computer program in the memory to carry out the steps of the method of any one of claims 1 to 7.
CN201911144534.0A 2019-11-20 2019-11-20 Text collation method, text collation apparatus, computer-readable storage medium, and electronic device Active CN110929514B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911144534.0A CN110929514B (en) 2019-11-20 2019-11-20 Text collation method, text collation apparatus, computer-readable storage medium, and electronic device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911144534.0A CN110929514B (en) 2019-11-20 2019-11-20 Text collation method, text collation apparatus, computer-readable storage medium, and electronic device

Publications (2)

Publication Number Publication Date
CN110929514A true CN110929514A (en) 2020-03-27
CN110929514B CN110929514B (en) 2023-06-27

Family

ID=69851442

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911144534.0A Active CN110929514B (en) 2019-11-20 2019-11-20 Text collation method, text collation apparatus, computer-readable storage medium, and electronic device

Country Status (1)

Country Link
CN (1) CN110929514B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113241061A (en) * 2021-05-17 2021-08-10 北京字跳网络技术有限公司 Method and device for processing voice recognition result, electronic equipment and storage medium
CN114611524A (en) * 2022-02-08 2022-06-10 马上消费金融股份有限公司 Text error correction method and device, electronic equipment and storage medium

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6618697B1 (en) * 1999-05-14 2003-09-09 Justsystem Corporation Method for rule-based correction of spelling and grammar errors
JP2011227876A (en) * 2010-03-31 2011-11-10 Fujitsu Ltd Calibration support device and calibration support program
KR20150017507A (en) * 2013-08-07 2015-02-17 부산대학교 산학협력단 Method and system for Context-sensitive Spelling Correction Rules using Korean WordNet
US20150161096A1 (en) * 2012-08-23 2015-06-11 Sk Telecom Co., Ltd. Method for detecting grammatical errors, error detection device for same and computer-readable recording medium having method recorded thereon
CN105279149A (en) * 2015-10-21 2016-01-27 上海应用技术学院 Chinese text automatic correction method
CN105550173A (en) * 2016-02-06 2016-05-04 北京京东尚科信息技术有限公司 Text correction method and device
US20170293604A1 (en) * 2016-04-11 2017-10-12 Zhuhai Kingsoft Office Software Co., Ltd Methods and apparatus for spell checking
CN109948144A (en) * 2019-01-29 2019-06-28 汕头大学 A method of the Teachers ' Talk Intelligent treatment based on classroom instruction situation
CN110210029A (en) * 2019-05-30 2019-09-06 浙江远传信息技术股份有限公司 Speech text error correction method, system, equipment and medium based on vertical field
CN110348020A (en) * 2019-07-17 2019-10-18 杭州嘉云数据科技有限公司 A kind of English- word spelling error correction method, device, equipment and readable storage medium storing program for executing

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6618697B1 (en) * 1999-05-14 2003-09-09 Justsystem Corporation Method for rule-based correction of spelling and grammar errors
JP2011227876A (en) * 2010-03-31 2011-11-10 Fujitsu Ltd Calibration support device and calibration support program
US20150161096A1 (en) * 2012-08-23 2015-06-11 Sk Telecom Co., Ltd. Method for detecting grammatical errors, error detection device for same and computer-readable recording medium having method recorded thereon
KR20150017507A (en) * 2013-08-07 2015-02-17 부산대학교 산학협력단 Method and system for Context-sensitive Spelling Correction Rules using Korean WordNet
CN105279149A (en) * 2015-10-21 2016-01-27 上海应用技术学院 Chinese text automatic correction method
CN105550173A (en) * 2016-02-06 2016-05-04 北京京东尚科信息技术有限公司 Text correction method and device
US20170293604A1 (en) * 2016-04-11 2017-10-12 Zhuhai Kingsoft Office Software Co., Ltd Methods and apparatus for spell checking
CN109948144A (en) * 2019-01-29 2019-06-28 汕头大学 A method of the Teachers ' Talk Intelligent treatment based on classroom instruction situation
CN110210029A (en) * 2019-05-30 2019-09-06 浙江远传信息技术股份有限公司 Speech text error correction method, system, equipment and medium based on vertical field
CN110348020A (en) * 2019-07-17 2019-10-18 杭州嘉云数据科技有限公司 A kind of English- word spelling error correction method, device, equipment and readable storage medium storing program for executing

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
YE-CHAN AHN: "Natural Language Contents Evaluation System for Detecting Fake News using Deep Learning", 《2019 16TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER SCIENCE AND SOFTWARE ENGINEERING (JCSSE)》 *
李蓉: "一个用于OCR输出的中文文本的拼写校对系统", 《中文信息学报》 *
李蓉: "一个用于OCR输出的中文文本的拼写校对系统", 《中文信息学报》, no. 05, 15 September 2009 (2009-09-15) *
陶永才: "中文词语搭配特征提取及文本校对研究", 《小型微型计算机系统》 *
陶永才: "中文词语搭配特征提取及文本校对研究", 《小型微型计算机系统》, 30 November 2018 (2018-11-30) *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113241061A (en) * 2021-05-17 2021-08-10 北京字跳网络技术有限公司 Method and device for processing voice recognition result, electronic equipment and storage medium
CN113241061B (en) * 2021-05-17 2023-03-10 北京字跳网络技术有限公司 Method and device for processing voice recognition result, electronic equipment and storage medium
CN114611524A (en) * 2022-02-08 2022-06-10 马上消费金融股份有限公司 Text error correction method and device, electronic equipment and storage medium
CN114611524B (en) * 2022-02-08 2023-11-17 马上消费金融股份有限公司 Text error correction method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN110929514B (en) 2023-06-27

Similar Documents

Publication Publication Date Title
CN110489760B (en) Text automatic correction method and device based on deep neural network
US9069753B2 (en) Determining proximity measurements indicating respective intended inputs
CN102156551B (en) Method and system for correcting error of word input
CN112016310A (en) Text error correction method, system, device and readable storage medium
CN111523306A (en) Text error correction method, device and system
CN112149406A (en) Chinese text error correction method and system
CN111310440B (en) Text error correction method, device and system
CN106503231B (en) Search method and device based on artificial intelligence
CN111428474A (en) Language model-based error correction method, device, equipment and storage medium
KR20190133624A (en) A method and system for context sensitive spelling error correction using realtime candidate generation
TWI567569B (en) Natural language processing systems, natural language processing methods, and natural language processing programs
CN111767717B (en) Grammar error correction method, device and equipment for Indonesia and storage medium
CN111460793A (en) Error correction method, device, equipment and storage medium
CN115438650B (en) Contract text error correction method, system, equipment and medium fusing multi-source characteristics
CN110826301B (en) Punctuation mark adding method, punctuation mark adding system, mobile terminal and storage medium
CN110929514A (en) Text proofreading method and device, computer readable storage medium and electronic equipment
CN111401012A (en) Text error correction method, electronic device and computer readable storage medium
CN113225612B (en) Subtitle generating method, device, computer readable storage medium and electronic equipment
JP5097802B2 (en) Japanese automatic recommendation system and method using romaji conversion
CN113255329A (en) English text spelling error correction method and device, storage medium and electronic equipment
CN112949290A (en) Text error correction method and device and communication equipment
CN109670040B (en) Writing assistance method and device, storage medium and computer equipment
Hladek et al. Unsupervised spelling correction for Slovak
CN114492396A (en) Text error correction method for automobile proper nouns and readable storage medium
CN112530406A (en) Voice synthesis method, voice synthesis device and intelligent equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: No.27, 1st floor, building 14, Haidian Road, zhongjiancai District, Beijing 100096

Applicant after: Beijing PERCENT Technology Group Co.,Ltd.

Address before: No.27, 1st floor, building 14, Haidian Road, zhongjiancai District, Beijing 100096

Applicant before: BEIJING BAIFENDIAN INFORMATION SCIENCE & TECHNOLOGY Co.,Ltd.

GR01 Patent grant
GR01 Patent grant