CN110929514B - Text collation method, text collation apparatus, computer-readable storage medium, and electronic device - Google Patents

Text collation method, text collation apparatus, computer-readable storage medium, and electronic device Download PDF

Info

Publication number
CN110929514B
CN110929514B CN201911144534.0A CN201911144534A CN110929514B CN 110929514 B CN110929514 B CN 110929514B CN 201911144534 A CN201911144534 A CN 201911144534A CN 110929514 B CN110929514 B CN 110929514B
Authority
CN
China
Prior art keywords
word
error
error correction
vector
occurrence frequency
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911144534.0A
Other languages
Chinese (zh)
Other versions
CN110929514A (en
Inventor
苏海波
苏萌
刘译璟
姚震
檀玉飞
黄伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Percent Technology Group Co ltd
Original Assignee
Beijing Percent Technology Group Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Percent Technology Group Co ltd filed Critical Beijing Percent Technology Group Co ltd
Priority to CN201911144534.0A priority Critical patent/CN110929514B/en
Publication of CN110929514A publication Critical patent/CN110929514A/en
Application granted granted Critical
Publication of CN110929514B publication Critical patent/CN110929514B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Document Processing Apparatus (AREA)
  • Machine Translation (AREA)

Abstract

The present disclosure relates to a text collation method, apparatus, computer-readable storage medium, and electronic device. The method comprises the following steps: determining error correction information of each sentence in the text to be corrected, wherein the error correction information comprises error words and at least one error correction word corresponding to the error words; for each error word, determining a first co-occurrence frequency and a second co-occurrence frequency of the error word and the front word and the rear word thereof in a preset corpus respectively; aiming at each error correction word corresponding to the error word, acquiring semantic features; and judging whether the error correction word is correct or not at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features. And the correctness of the error correction word is judged, so that the accuracy of text correction can be improved. When the correctness of the error correction words is determined, the collocation of the front and rear words and the context semantic features are comprehensively considered, so that the accuracy of the correctness determination of the error correction words can be ensured, and the accuracy of text correction is further improved. In addition, the intelligent and automatic correction work is realized, the pressure of manual correction is reduced, the working efficiency is improved, and the labor cost is reduced.

Description

Text collation method, text collation apparatus, computer-readable storage medium, and electronic device
Technical Field
The present disclosure relates to the field of computer technology, and in particular, to a text collation method, apparatus, computer readable storage medium, and electronic device.
Background
In text processing, a quite mature computer application system exists in the process of inputting, editing and typesetting, but the intermediate link of text proofreading still mainly depends on a manual processing stage, and the intermediate link becomes a bottleneck for restricting the development of the whole industry and influencing the working efficiency in the fields of news, publishing, office marks and the like. The manual text correction is time-consuming and labor-consuming, and the correction accuracy is difficult to ensure.
Based on the problems, at the present stage, an N-gram model is mainly adopted to detect errors in a text and give error correction suggestions, but the method only considers the collocation problem of front and rear words, and the text correction accuracy is low.
Disclosure of Invention
In order to overcome the problems in the related art, the present disclosure provides a text collation method, apparatus, computer-readable storage medium and electronic device.
To achieve the above object, according to a first aspect of embodiments of the present disclosure, there is provided a text collation method, the method comprising:
determining error correction information of each sentence in a text to be corrected, wherein the error correction information comprises error words and at least one error correction word corresponding to the error words;
For each error word, respectively determining a first co-occurrence frequency of the error word and a preceding word of the error word in a preset corpus and a second co-occurrence frequency of the error word and a following word of the error word in the preset corpus;
for each error correction word corresponding to the error word, acquiring semantic features of the error word and the error correction word;
and judging whether the error correction word is correct or not at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature.
Optionally, the determining whether the error correction word is correct at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature includes:
at least inputting the first co-occurrence frequency, the second co-occurrence frequency and the semantic features into a preset xgboost model to judge whether the error correction words are correct or not.
Optionally, the method further comprises:
marking words belonging to single characters in the error words and the error correction words as 1, and marking words belonging to multiple characters in the error words and the error correction words as 0;
the determining whether the error correction word is correct at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature comprises:
And judging whether the error correction word is correct or not according to the mark of the error word, the mark of the error correction word, the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature.
Optionally, the acquiring the semantic features of the error word and the error correction word includes:
replacing the error word in the initial sentence to which the error word belongs with the error correction word to obtain an error correction sentence;
respectively acquiring a first vector A= (A) corresponding to the initial sentence through a Bert model 1 ,A 2 ,…,A m ) A second vector b= (B) corresponding to the error correction sentence 1 ,B 2 ,…,B n ) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, A i For a first score for characterizing the rationality of the occurrence of the ith character in the initial sentence, i=1, 2, …, m, B j J=1, 2, …, n for a second score for characterizing the rationality of the j-th character in the error correction sentence occurring in the error correction sentence;
and determining a first difference value between the average value of each second score in the second vector and the average value of each first score in the first vector as the semantic features of the error word and the error correction word.
Optionally, the acquiring the semantic features of the error word and the error correction word includes:
replacing the error word in the initial sentence to which the error word belongs with the error correction word to obtain an error correction sentence;
respectively acquiring a first vector A= (A) corresponding to the initial sentence through a Bert model 1 ,A 2 ,…,A m ) A second vector b= (B) corresponding to the error correction sentence 1 ,B 2 ,…,B n ) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, A i For a first score for characterizing the rationality of the occurrence of the ith character in the initial sentence, i=1, 2, …, m, B j J=1, 2, …, n for a second score for characterizing the rationality of the j-th character in the error correction sentence occurring in the error correction sentence;
transforming the first vector and the second vector through each preset transformation function in a plurality of preset transformation functions in sequence to obtain a plurality of third vectors corresponding to the first vector and a plurality of fourth vectors corresponding to the second vector;
respectively calculating a second difference value of the average value of each third score in the third vector and the average value of each fourth score in the fourth vector aiming at the third vector and the fourth vector obtained by the transformation of each preset transformation function;
And determining a plurality of second difference values as semantic features of the error word and the error correction word.
Optionally, the method further comprises:
marking a maximum value of the plurality of second difference values as 1, and marking a second difference value except for the maximum value of the plurality of second difference values as 0;
the determining whether the error correction word is correct at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature comprises:
and judging whether the error correction word is correct or not according to the mark of the second difference value, the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature.
Optionally, before the step of determining the error correction information of each sentence in the text to be collated, the method further includes:
preprocessing the text to be checked to obtain a new text to be checked;
the determining the error correction information of each sentence in the text to be checked comprises the following steps:
and determining error correction information of each sentence in the new text to be checked.
According to a second aspect of embodiments of the present disclosure, there is provided a text collation apparatus, the apparatus comprising:
the first determining module is used for determining error correction information of each sentence in the text to be corrected, wherein the error correction information comprises error words and at least one error correction word corresponding to the error words;
The second determining module is used for respectively determining a first co-occurrence frequency of the error word and a preceding word of the error word in a preset corpus and a second co-occurrence frequency of the error word and a following word of the error word in the preset corpus for each error word determined by the first determining module;
the acquisition module is used for acquiring semantic features of the error correction words in corresponding sentences according to each error correction word corresponding to the error word determined by the first determination module;
and the judging module is used for judging whether the error correction word is correct or not at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features acquired by the acquisition module, which are determined by the second determination module.
According to a third aspect of embodiments of the present disclosure, there is provided a computer readable storage medium having stored thereon a computer program which when executed by a processor implements the steps of the method provided by the first aspect of the present disclosure.
According to a fourth aspect of embodiments of the present disclosure, there is provided an electronic device, comprising:
a memory having a computer program stored thereon;
a processor for executing the computer program in the memory to implement the steps of the method provided by the first aspect of the present disclosure.
In the technical scheme, firstly, determining error words in each sentence in a text to be checked and at least one error correction word corresponding to each error word; then, for each error word, the first co-occurrence frequency and the second co-occurrence frequency of the error word and the preceding word and the following word of the error word can be respectively determined, and meanwhile, for each error word corresponding to the error word, corresponding semantic features are obtained; and finally, judging the correctness of the error correction word at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features. After the error word and the corresponding error correction word are obtained, the correctness of the error correction word is further judged, so that the accuracy of text correction can be improved. In addition, when the correctness of the error correction word is determined, the matching problem of the front word and the rear word is considered, and the context semantic features of the words are combined, so that the accuracy of the correctness determination of the error correction word can be ensured, and the accuracy of text correction is further improved. In addition, the text proofreading method enables the proofreading work to be intelligent and automatic, reduces the pressure of manual proofreading, improves the working efficiency and reduces the labor cost.
Additional features and advantages of the present disclosure will be set forth in the detailed description which follows.
Drawings
The accompanying drawings are included to provide a further understanding of the disclosure, and are incorporated in and constitute a part of this specification, illustrate the disclosure and together with the description serve to explain, but do not limit the disclosure. In the drawings:
FIG. 1 illustrates a flow chart of a text collation method according to an exemplary embodiment.
FIG. 2A is a flowchart illustrating a method of acquiring semantic features according to an example embodiment.
FIG. 2B is a flowchart illustrating a method of acquiring semantic features according to another exemplary embodiment.
FIG. 3 illustrates a flow chart of a text collation method according to another exemplary embodiment.
FIG. 4 illustrates a flow chart of a text collation method according to another exemplary embodiment.
FIG. 5 illustrates a flow chart of a text collation method according to another exemplary embodiment.
FIG. 6 is a block diagram illustrating a text collation apparatus according to an example embodiment.
Fig. 7 is a block diagram of an electronic device, according to an example embodiment.
Fig. 8 is a block diagram of an electronic device, according to an example embodiment.
Detailed Description
Specific embodiments of the present disclosure are described in detail below with reference to the accompanying drawings. It should be understood that the detailed description and specific examples, while indicating and illustrating the disclosure, are not intended to limit the disclosure.
FIG. 1 illustrates a flow chart of a text collation method according to an exemplary embodiment. As shown in fig. 1, the above method may include the following steps 101 to 104.
In step 101, error correction information of each sentence in the text to be checked is determined, wherein the error correction information includes an error word and at least one error correction word corresponding to the error word.
In the present disclosure, the above error correction information may include 0, 1, or a plurality of error words, and each error word may correspond to 1 or a plurality of error correction words (i.e., error correction suggestions). For example, the sentence "today's atmosphere is very good. "error correction information is: the error word "weather" and its corresponding error correction word "weather".
Also, in the present disclosure, the correction information of each sentence in the text to be collated may be determined in various ways. In one embodiment, the error correction information may be obtained through an N-gram model. Specifically, this can be achieved by: (1) After the text to be checked is obtained, word segmentation processing, part-of-speech tagging and other operations can be performed on the text; (2) An N-gram model is adopted to take a huge corpus as a basis, positioning operation of error words is carried out, and the position where errors are likely to occur is detected; (3) Further detecting the position which is possibly in error by a part-of-speech N-gram method, if the detection is unreasonable, determining that the position is in error, and defining the position as an error word; (4) And carrying out error correction processing on the error words and giving out corresponding error correction words. Therefore, the error correction information of each sentence in the text to be checked can be obtained.
In another embodiment, the error correction information may be obtained by reverse error correction (i.e., character matching). Specifically, after the text to be checked is obtained, word segmentation processing can be performed on the text; and then, for each word, respectively matching the word with each word in a preset word stock to obtain a plurality of matching degrees, if the maximum value of the plurality of matching degrees is smaller than a preset matching degree threshold value, determining the word as an error word, and determining the word corresponding to one or more matching degrees which are ranked at the front in the preset word stock and are ranked at the front in the plurality of matching degrees as an error correction word corresponding to the error word, wherein the matching degrees are ranked from large to small. Therefore, the error correction information of each sentence in the text to be checked can be obtained.
The predetermined word library may be a database of words, may be generated based on a huge expected library, or may be an existing word library, and is not particularly limited in this disclosure. The preset matching degree threshold may be a value set by a user, or may be a default empirical value, which is not particularly limited in the present disclosure.
In another embodiment, the error correction information can be obtained by an N-gram model and a reverse error correction mode, then the error correction information obtained by the N-gram model and the reverse error correction mode are de-overlapped, and the error correction information after de-overlapping and merging is used as the error correction information of each sentence in the text to be corrected. Therefore, the comprehensiveness of error word detection can be improved, and the accuracy of text proofreading is improved.
In step 102, for each wrong word, a first co-occurrence frequency of the wrong word and a preceding word of the wrong word in a preset corpus and a second co-occurrence frequency of the wrong word and a following word of the wrong word in the preset corpus are respectively determined.
In the present disclosure, the preset corpus may be a database composed of text sentences. Further, in order to make the first co-occurrence frequency and the second co-occurrence frequency determined according to the technical solution of the embodiment of the present application more fit the text to be checked, the embodiment of the present application preferably forms the preset corpus from text sentences in the field to which the text to be checked belongs. It should be noted that, in the embodiment of the present application, the generation manner of the preset corpus is not limited.
The preceding word of the incorrect word may be a word located before and immediately adjacent to the incorrect word in the initial sentence to which the incorrect word belongs, and the following word of the incorrect word may be a word located after and immediately adjacent to the incorrect word in the initial sentence to which the incorrect word belongs. For example, the sentence "today's atmosphere is very good. "error correction information is: the error word "qi tian" and the corresponding error correction word "weather", wherein "today" is the former word of the error word "qi tian" and "true" is the latter word of the error word "qi tian".
And, the first co-occurrence frequency and the second co-occurrence frequency may be determined in various ways. In one embodiment, this may be determined by means of mathematical statistics. Specifically, for each wrong word, the number of occurrences of the wrong word and its preceding word in the preset corpus (i.e., the first co-occurrence frequency) and the number of occurrences of the wrong word and its following word in the preset corpus (i.e., the second co-occurrence frequency) may be counted separately.
In another embodiment, the first co-occurrence frequency and the second co-occurrence frequency may be determined by an N-gram model. Since the specific manner in which the first co-occurrence and the second co-occurrence are determined by the N-gram model is well known to those skilled in the art, details will not be provided in this disclosure.
In step 103, for each error correction word corresponding to the error word, the semantic features of the error word and the error correction word are obtained.
In the present disclosure, after the error correction information is obtained through the above step 101, corresponding semantic features may be obtained for each error correction word corresponding to each error word in the error correction information. In particular, the above semantic features may be obtained in a variety of ways. In one embodiment, this may be achieved by steps 1031-1033 shown in FIG. 2A.
In step 1031, the error word in the initial sentence to which the error word belongs is replaced with the error correction word, so as to obtain an error correction sentence.
Illustratively, the sentence "today's atmosphere is very good". "error correction information is: the error word "qi tian" and the error correction word "weather" corresponding thereto, wherein the initial sentence to which the error word "qi tian" belongs is: "today's atmosphere is very good. ". Thus, the error word "weather" in the initial sentence is replaced by the error correction word "weather", so that an error correction sentence "weather today is good. ".
In step 1032, a first vector corresponding to the initial sentence and a second vector corresponding to the error correction sentence are obtained by the Bert model, respectively.
In the present disclosure, a first vector a= (a 1 ,A 2 ,…,A m ) Second vector b= (B 1 ,B 2 ,…,B n ) Wherein m and n are the number of characters (including punctuation marks) contained in the initial sentence and the number of characters (including punctuation marks) contained in the error correction sentence, respectively, A i For a first score for characterizing the rationality of the occurrence of the ith character in an initial sentence in that initial sentence, i=1, 2, …, m, B j J=1, 2, …, n, a second score for characterizing the rationality of the j-th character in the error correction sentence appearing in the error correction sentence. Note that the number of characters included in the error word may be the same as or different from the number of characters included in the error word corresponding to the error word, and similarly, m and n may be equal or unequal.
Bert (Bidirectional encoder representation from transformers, a bi-directional coded representation of a transducer) is a method of pre-training language representation, a model that can be downloaded and used for free. Wherein the model can be used to extract high quality language features from sentences in the text to be collated. In the present disclosure, the first vectors a= (a) corresponding to the above initial sentences can be obtained by the Bert model, respectively 1 ,A 2 ,…,A m ) Second vector b= (B) corresponding to error correction sentence 1 ,B 2 ,…,B n )。
In step 1033, a first difference between the average of the second scores in the second vector and the average of the first scores in the first vector is determined as the semantic feature of the error word and the error correction word.
In the present disclosure, the magnitude of the first difference may reflect the quality of the corresponding error correction word, where when the first difference is greater than 0, it indicates that the corresponding error correction word is relatively good, and if the first difference is less than or equal to 0, it indicates that the corresponding error correction word is relatively poor.
In the step 1032, the first vector a= (a) corresponding to the initial sentence to which the wrong word belongs is obtained 1 ,A 2 ,…,A m ) Second vector b= (B) corresponding to the corresponding error correction sentence 1 ,B 2 ,…,B n ) After that, the first vector a= (a) can be calculated separately 1 ,A 2 ,…,A m ) Each first score A of i The average value of (i.e.,
Figure BDA0002281811340000091
) Second vector b= (B) 1 ,B 2 ,…,B n ) Each second score B of j Average value (i.e.)>
Figure BDA0002281811340000092
) The method comprises the steps of carrying out a first treatment on the surface of the Thereafter, the second vector b= (B 1 ,B 2 ,…,B n ) Each second score B of j Average value and first vector a= (a 1 ,A 2 ,…,A m ) Each first score A of i Is a first difference in the average value of (i.e.,
Figure BDA0002281811340000093
) Semantic features that are error words and error correction words are determined.
In another embodiment, the steps 1031, 1032, 1034, 1035, 1036 shown in fig. 2B may be implemented.
In step 1031, the error word in the initial sentence to which the error word belongs is replaced with the error correction word, so as to obtain an error correction sentence.
In step 1032, a first vector corresponding to the initial sentence and a second vector corresponding to the error correction sentence are obtained by the Bert model, respectively.
In step 1034, the first vector and the second vector are transformed by each preset transformation function of the plurality of preset transformation functions in turn, so as to obtain a plurality of third vectors corresponding to the first vector and a plurality of fourth vectors corresponding to the second vector.
In the present disclosure, the above-mentioned preset transformation function may be used to transform the first vector and the second vector to obtain corresponding third vector and fourth vector. And, the plurality of preset transform functions may be preset by a user or default (for example, the i-th preset transform function of the plurality of preset transform functions is Y (X) =x+c) i X is a first vector or a second vector, Y (X) is a third vector corresponding to the first vector or a fourth vector corresponding to the second vector, C i Constant vector corresponding to the ith preset transformation function and constant vector C corresponding to each preset transformation function i Different), not specifically defined in the present disclosure.
After the first vector of the initial sentence to which each error word belongs and the second vector corresponding to the corresponding error correction sentence are obtained through the step 1032, each first vector may be transformed by using a plurality of preset transformation functions, so as to obtain a plurality of third vectors; meanwhile, for each second vector, the second vector may be transformed by using the plurality of preset transformation functions, so as to obtain a plurality of fourth vectors.
In step 1035, for each of the third vector and the fourth vector obtained by transforming the preset transformation function, a second difference value between the average value of the third scores in the third vector and the average value of the fourth scores in the fourth vector is calculated.
In the present disclosure, after obtaining a plurality of third vectors corresponding to the initial sentence to which the error word belongs and a plurality of fourth vectors corresponding to the corresponding error correction sentence through the step 1034, the second difference between the average value of the third scores in the third vectors and the average value of the fourth scores in the fourth vectors may be calculated for the third vectors and the fourth vectors obtained by transforming each preset transformation function. Thus, a plurality of second differences may be obtained.
Note that, the number of the preset transform functions may be set by the user or may be default (for example, 6), which is not specifically limited in the present disclosure.
In step 1036, a plurality of second differences are determined as semantic features of the error word and the error correction word.
After the plurality of second differences are obtained through step 1035, the plurality of second differences may be determined as semantic features of the error word and the error correction word.
Returning to fig. 1, in step 104, it is determined whether the error correction word is correct based at least on the first co-occurrence frequency, the second co-occurrence frequency, and the semantic feature.
In the present disclosure, for each error word, after the first co-occurrence frequency of the error word and its preceding word in the preset corpus and the second co-occurrence frequency of the error word and its following word in the preset corpus are obtained through the above step 102, and the corresponding semantic features are obtained through the above step 103, they may be input into a preset xgboost (extreme gradient boosting, extreme gradient lifting) model to determine whether the error word corresponding to the error word is correct. The xgboost model is a classification model, the output of the model can be 0 or 1,0 can be used for representing error correction word errors, and 1 can be used for representing error correction word correctness.
In addition, in the present disclosure, the construction of the xgboost model may be performed based on the manually collated text. Firstly, according to the same manner as in the step 102, for each sample error word obtained by manual correction, a third co-occurrence frequency of the sample error word and a preceding word thereof in a preset corpus and a fourth co-occurrence frequency of the sample error word and a following word thereof in the preset corpus are obtained; meanwhile, corresponding reference semantic features are acquired in the same way as in the step 103; and then, at least the third co-occurrence frequency, the fourth co-occurrence frequency and the reference semantic features are used as training samples and input into an initial xgboost model for training, so that the preset xgboost model is obtained. The specific construction mode of the xgboost model is well known to those skilled in the art, so that detailed description is not given in this disclosure.
In addition, the xgboost model can be optimized, for example, the model can be optimized by modifying model parameters according to training and testing effects.
In one embodiment, it may be determined whether the error correction word is correct based on the first co-occurrence frequency, the second co-occurrence frequency, and the semantic features. Specifically, the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature may be input into the preset xgboost model, so as to determine whether the error correction word corresponding to the error word is correct according to the output of the preset xgboost model. That is, when the output of the preset xgboost model is 0, indicating that the error correction word corresponding to the error word is wrong; and when the output of the preset xgboost model is 1, the error correction word corresponding to the error word is correct.
In another embodiment, in order to further improve the accuracy of text proofreading, when determining whether the error correction word is correct, the error word and the character number information contained in the error correction word may be considered in addition to the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature. Specifically, as shown in fig. 3, the above method may further include the following step 105.
In step 105, the word belonging to the single word of the error word and the error correction word is marked as 1, and the word belonging to the multiple word of the error word and the error correction word is marked as 0.
After the error correction information is obtained in the step 101, it may be determined whether each error word or error correction word is a single word, and if the error word is a single word, the error word is marked as 1, otherwise, the error word is marked as 0. In this way, whether the error correction word is correct or not can be determined together based on the flag of the error word, the flag of the error correction word, the first co-occurrence frequency, the second co-occurrence frequency, and the semantic feature. Specifically, the marking of the error word, the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature may be input into the preset xgboost model, so as to determine whether the error word corresponding to the error word is correct according to the output of the preset xgboost model.
Note that, the step 105 may be performed before the step 102 or the step 103, may be performed after the step 102 or the step 103, may be performed simultaneously with the step 102 or the step 103, and is not particularly limited in this disclosure.
In still another embodiment, in order to further improve the accuracy of text proofreading, the second difference information may be considered in addition to the first co-occurrence frequency, the second co-occurrence frequency, and the semantic feature when determining whether the error correction word is correct. Specifically, as shown in fig. 4, the above method may further include the following step 106.
In step 106, the maximum value of the plurality of second difference values is marked as 1, and the second difference values other than the maximum value of the plurality of second difference values are marked as 0.
After the plurality of second differences are obtained through the above step 103 (i.e., step 1035), a maximum value of the plurality of second differences may be marked as 1, and second differences other than the above maximum value of the plurality of second differences may be marked as 0. Thus, it is possible to collectively determine whether or not the error correction word is correct based on the flag of each second difference value, the first co-occurrence frequency, the second co-occurrence frequency, and the semantic feature. Specifically, the marks of the second differences, the first co-occurrence frequency, the second co-occurrence frequency and the semantic features may be input into the preset xgboost model, so as to determine whether the error correction word corresponding to the error word is correct according to the output of the preset xgboost model.
The step 103 may be performed before the step 102, may be performed after the step 102, may be performed simultaneously with the step 102, and is not particularly limited in this disclosure.
In the technical scheme, firstly, determining error words in each sentence in a text to be checked and at least one error correction word corresponding to each error word; then, for each error word, the first co-occurrence frequency and the second co-occurrence frequency of the error word and the preceding word and the following word of the error word can be respectively determined, and meanwhile, for each error word corresponding to the error word, corresponding semantic features are obtained; and finally, judging the correctness of the error correction word at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features. After the error word and the corresponding error correction word are obtained, the correctness of the error correction word is further judged, so that the accuracy of text correction can be improved. In addition, when the correctness of the error correction word is determined, the matching problem of the front word and the rear word is considered, and the context semantic features of the words are combined, so that the accuracy of the correctness determination of the error correction word can be ensured, and the accuracy of text correction is further improved. In addition, the text proofreading method enables the proofreading work to be intelligent and automatic, reduces the pressure of manual proofreading, improves the working efficiency and reduces the labor cost.
In addition, in order to further improve the accuracy of text proofreading, the text to be proofread may be preprocessed before error correction information is obtained according to the text. Specifically, as shown in fig. 5, the method may further include the following step 107 before the step 101.
In step 107, the text to be checked is preprocessed, and a new text to be checked is obtained.
In the present disclosure, the preprocessing may include filtering out illegal characters (e.g., spaces, blank lines, etc.). After preprocessing the text to be checked, a new text to be checked may be obtained, and then error correction information may be obtained based on the new text to be checked, that is, the above step 101 is performed.
FIG. 6 is a block diagram illustrating a text collation apparatus according to an example embodiment. Referring to fig. 6, the apparatus 600 may include: a first determining module 601, configured to determine error correction information of each sentence in a text to be checked, where the error correction information includes an error word and at least one error correction word corresponding to the error word; a second determining module 602, configured to determine, for each of the error words determined by the first determining module 601, a first co-occurrence frequency of the error word and a preceding word of the error word in a preset corpus, and a second co-occurrence frequency of the error word and a following word of the error word in the preset corpus; an obtaining module 603, configured to obtain, for each of the error correction words corresponding to the error word determined by the first determining module 601, a semantic feature of the error correction word in a corresponding sentence; the determining module is configured to determine whether the error correction word is correct at least according to the first co-occurrence frequency, the second co-occurrence frequency, and the semantic feature acquired by the acquiring module 603, which are determined by the second determining module 602.
Optionally, the determining module 604 is configured to input at least the first co-occurrence frequency, the second co-occurrence frequency, and the semantic feature into a preset xgboost model to determine whether the error correction word is correct.
Optionally, the apparatus 600 may further include: the first marking module is used for marking the word belonging to the single word in the error word and the error correction word as 1, and marking the word belonging to the multi-word in the error word and the error correction word as 0;
the determining module 604 is configured to: and judging whether the error correction word is correct or not according to the mark of the error word, the mark of the error correction word, the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature obtained by the first marking module.
Optionally, the obtaining module 603 includes: a replacing sub-module, configured to replace the error word in the initial sentence to which the error word belongs with the error correction word, so as to obtain an error correction sentence; a vector obtaining submodule, configured to obtain first vectors a= (a) corresponding to the initial sentences respectively through a Bert model 1 ,A 2 ,…,A m ) A second vector b= (B) corresponding to the error correction sentence 1 ,B 2 ,…,B n ) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, A i For a first score for characterizing the rationality of the occurrence of the ith character in the initial sentence, i=1, 2, …, m, B j J=1, 2, …, n for a second score for characterizing the rationality of the j-th character in the error correction sentence occurring in the error correction sentence; a first determination submodule for determining a first difference between an average value of each second score in the second vector and an average value of each first score in the first vectorAnd determining semantic features of the error word and the error correction word.
Optionally, the obtaining module 603 includes: a replacing sub-module, configured to replace the error word in the initial sentence to which the error word belongs with the error correction word, so as to obtain an error correction sentence; a vector obtaining submodule, configured to obtain first vectors a= (a) corresponding to the initial sentences respectively through a Bert model 1 ,A 2 ,…,A m ) A second vector b= (B) corresponding to the error correction sentence 1 ,B 2 ,…,B n ) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, A i For a first score for characterizing the rationality of the occurrence of the ith character in the initial sentence, i=1, 2, …, m, B j J=1, 2, …, n for a second score for characterizing the rationality of the j-th character in the error correction sentence occurring in the error correction sentence; the transformation submodule is used for transforming the first vector and the second vector through each preset transformation function in a plurality of preset transformation functions in sequence to obtain a plurality of third vectors corresponding to the first vector and a plurality of fourth vectors corresponding to the second vector; the calculation sub-module is used for respectively calculating a second difference value of the average value of each third score in the third vector and the average value of each fourth score in the fourth vector according to the third vector and the fourth vector obtained by transformation of each preset transformation function; and the second determining submodule is used for determining a plurality of second difference values as semantic features of the error word and the error correction word.
Optionally, the apparatus 600 may further include: the second marking module is used for marking the maximum value of the second differences obtained by the calculating submodule as 1, and marking the second differences except for the maximum value of the second differences as 0;
the determining module 604 is configured to: and judging whether the error correction word is correct or not according to the mark of the second difference value, the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature obtained by the second mark module.
Optionally, the apparatus 600 may further include: the preprocessing module is used for preprocessing the text to be checked before the first determining module determines the error correction information of each sentence in the text to be checked, so as to obtain a new text to be checked;
the first determining module is used for determining error correction information of each sentence in the new text to be checked.
The specific manner in which the various modules perform the operations in the apparatus of the above embodiments have been described in detail in connection with the embodiments of the method, and will not be described in detail herein.
The present disclosure also provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements the steps of the above-described text collation method provided by the present disclosure.
Fig. 7 is a block diagram of an electronic device 700, according to an example embodiment. As shown in fig. 7, the electronic device 700 may include: a processor 701, a memory 702. The electronic device 700 may also include one or more of a multimedia component 703, an input/output (I/O) interface 704, and a communication component 705.
Wherein the processor 701 is configured to control the overall operation of the electronic device 700 to perform all or part of the steps of the text collation method described above. The memory 702 is used to store various types of data to support operation on the electronic device 700, which may include, for example, instructions for any application or method operating on the electronic device 700, as well as application-related data, such as contact data, messages sent and received, pictures, audio, video, and so forth. The Memory 702 may be implemented by any type or combination of volatile or non-volatile Memory devices, such as static random access Memory (Static Random Access Memory, SRAM for short), electrically erasable programmable Read-Only Memory (Electrically Erasable Programmable Read-Only Memory, EEPROM for short), erasable programmable Read-Only Memory (Erasable Programmable Read-Only Memory, EPROM for short), programmable Read-Only Memory (Programmable Read-Only Memory, PROM for short), read-Only Memory (ROM for short), magnetic Memory, flash Memory, magnetic disk, or optical disk. The multimedia component 703 can include a screen and an audio component. Wherein the screen may be, for example, a touch screen, the audio component being for outputting and/or inputting audio signals. For example, the audio component may include a microphone for receiving external audio signals. The received audio signals may be further stored in the memory 702 or transmitted through the communication component 705. The audio assembly further comprises at least one speaker for outputting audio signals. The I/O interface 704 provides an interface between the processor 701 and other interface modules, which may be a keyboard, mouse, buttons, etc. These buttons may be virtual buttons or physical buttons. The communication component 705 is for wired or wireless communication between the electronic device 700 and other devices. Wireless communication, such as Wi-Fi, bluetooth, near field communication (Near Field Communication, NFC for short), 2G, 3G, 4G, NB-IOT, eMTC, or other 5G, etc., or one or a combination of more of them, is not limited herein. The corresponding communication component 705 may thus comprise: wi-Fi module, bluetooth module, NFC module, etc.
In an exemplary embodiment, the electronic device 700 may be implemented by one or more application specific integrated circuits (Application Specific Integrated Circuit, abbreviated ASIC), digital signal processor (Digital Signal Processor, abbreviated DSP), digital signal processing device (Digital Signal Processing Device, abbreviated DSPD), programmable logic device (Programmable Logic Device, abbreviated PLD), field programmable gate array (Field Programmable Gate Array, abbreviated FPGA), controller, microcontroller, microprocessor, or other electronic components for performing the text-checking method described above.
In another exemplary embodiment, a computer readable storage medium is also provided, comprising program instructions which, when executed by a processor, implement the steps of the text collation method described above. For example, the computer readable storage medium may be the memory 702 including program instructions described above that are executable by the processor 701 of the electronic device 700 to perform the text collation method described above.
Fig. 8 is a block diagram of an electronic device 800, according to an example embodiment. For example, the electronic device 800 may be provided as a server. Referring to fig. 8, the electronic device 800 includes a processor 822, which may be one or more in number, and a memory 832 for storing computer programs executable by the processor 822. The computer program stored in memory 832 may include one or more modules each corresponding to a set of instructions. Further, the processor 822 may be configured to execute the computer program to perform the text collation method described above.
In addition, the electronic device 800 may further include a power supply component 826 and a communication component 850, the power supply component 826 may be configured to perform power management of the electronic device 800, and the communication component 850 may be configured to enable communication of the electronic device 800, such as wired or wireless communication. In addition, the electronic device 800 may also include an input/output (I/O) interface 858. The electronic device 800 may operate based on an operating system stored in the memory 832, such as Windows Server, mac OS XTM, unixTM, linuxTM, etc.
In another exemplary embodiment, a computer readable storage medium is also provided, comprising program instructions which, when executed by a processor, implement the steps of the text collation method described above. For example, the computer readable storage medium may be the memory 832 including program instructions described above that are executable by the processor 822 of the electronic device 800 to perform the text collation method described above.
In another exemplary embodiment, a computer program product is also provided, comprising a computer program executable by a programmable apparatus, the computer program having code portions for performing the above-described text collation method when executed by the programmable apparatus.
The preferred embodiments of the present disclosure have been described in detail above with reference to the accompanying drawings, but the present disclosure is not limited to the specific details of the above embodiments, and various simple modifications may be made to the technical solutions of the present disclosure within the scope of the technical concept of the present disclosure, and all the simple modifications belong to the protection scope of the present disclosure.
In addition, the specific features described in the above embodiments may be combined in any suitable manner without contradiction. The various possible combinations are not described further in this disclosure in order to avoid unnecessary repetition.
Moreover, any combination between the various embodiments of the present disclosure is possible as long as it does not depart from the spirit of the present disclosure, which should also be construed as the disclosure of the present disclosure.

Claims (8)

1. A method of text collation, the method comprising:
determining error correction information of each sentence in a text to be corrected, wherein the error correction information comprises error words and at least one error correction word corresponding to the error words;
for each error word, respectively determining a first co-occurrence frequency of the error word and a preceding word of the error word in a preset corpus and a second co-occurrence frequency of the error word and a following word of the error word in the preset corpus;
For each error correction word corresponding to the error word, acquiring semantic features of the error word and the error correction word;
judging whether the error correction word is correct or not at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features;
the obtaining the semantic features of the error word and the error correction word includes:
replacing the error word in the initial sentence to which the error word belongs with the error correction word to obtain an error correction sentence;
respectively acquiring a first vector A= (A) corresponding to the initial sentence through a Bert model 1 ,A 2 ,…,A m ) A second vector b= (B) corresponding to the error correction sentence 1 ,B 2 ,…,B n ) Wherein m and n are the number of characters contained in the initial sentence respectivelyThe number of characters contained in the error correction sentence, A i For a first score for characterizing the rationality of the occurrence of the ith character in the initial sentence, i=1, 2, …, m, B j J=1, 2, …, n for a second score for characterizing the rationality of the j-th character in the error correction sentence occurring in the error correction sentence;
one of the following:
determining a first difference value between the average value of each second score in the second vector and the average value of each first score in the first vector as the semantic features of the error word and the error correction word;
Transforming the first vector and the second vector through each preset transformation function in a plurality of preset transformation functions in sequence to obtain a plurality of third vectors corresponding to the first vector and a plurality of fourth vectors corresponding to the second vector; respectively calculating a second difference value of the average value of each third score in the third vector and the average value of each fourth score in the fourth vector aiming at the third vector and the fourth vector obtained by the transformation of each preset transformation function; and determining a plurality of second difference values as semantic features of the error word and the error correction word.
2. The method of claim 1, wherein the determining whether the error correction word is correct based at least on the first co-occurrence frequency, the second co-occurrence frequency, and the semantic feature comprises:
at least inputting the first co-occurrence frequency, the second co-occurrence frequency and the semantic features into a preset xgboost model to judge whether the error correction words are correct or not.
3. The method according to claim 1, wherein the method further comprises:
marking words belonging to single characters in the error words and the error correction words as 1, and marking words belonging to multiple characters in the error words and the error correction words as 0;
The determining whether the error correction word is correct at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature comprises:
and judging whether the error correction word is correct or not according to the mark of the error word, the mark of the error correction word, the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature.
4. The method according to claim 1, wherein in case the step of obtaining semantic features of the error word and the error correction word comprises the step of determining a plurality of the second differences as the semantic features of the error word and the error correction word, the method further comprises:
marking a maximum value of the plurality of second difference values as 1, and marking a second difference value except for the maximum value of the plurality of second difference values as 0;
the determining whether the error correction word is correct at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature comprises:
and judging whether the error correction word is correct or not according to the mark of the second difference value, the first co-occurrence frequency, the second co-occurrence frequency and the semantic feature.
5. A method according to any one of claims 1-3, characterized in that before the step of determining error correction information for each sentence in the text to be collated, the method further comprises:
preprocessing the text to be checked to obtain a new text to be checked;
the determining the error correction information of each sentence in the text to be checked comprises the following steps:
and determining error correction information of each sentence in the new text to be checked.
6. A text collation apparatus, the apparatus comprising:
the first determining module is used for determining error correction information of each sentence in the text to be corrected, wherein the error correction information comprises error words and at least one error correction word corresponding to the error words;
the second determining module is used for respectively determining a first co-occurrence frequency of the error word and a preceding word of the error word in a preset corpus and a second co-occurrence frequency of the error word and a following word of the error word in the preset corpus for each error word determined by the first determining module;
the acquisition module is used for acquiring semantic features of the error correction words in corresponding sentences according to each error correction word corresponding to the error word determined by the first determination module;
The judging module is used for judging whether the error correction word is correct or not at least according to the first co-occurrence frequency, the second co-occurrence frequency and the semantic features acquired by the acquisition module, which are determined by the second determining module;
wherein, the acquisition module includes:
a replacing sub-module, configured to replace the error word in the initial sentence to which the error word belongs with the error correction word, so as to obtain an error correction sentence;
a vector obtaining submodule, configured to obtain first vectors a= (a) corresponding to the initial sentences respectively through a Bert model 1 ,A 2 ,…,A m ) A second vector b= (B) corresponding to the error correction sentence 1 ,B 2 ,…,B n ) Wherein m and n are the number of characters contained in the initial sentence and the number of characters contained in the error correction sentence, respectively, A i For a first score for characterizing the rationality of the occurrence of the ith character in the initial sentence, i=1, 2, …, m, B j J=1, 2, …, n for a second score for characterizing the rationality of the j-th character in the error correction sentence occurring in the error correction sentence;
one of the following:
a first determining submodule, configured to determine a first difference between an average value of each second score in the second vector and an average value of each first score in the first vector as a semantic feature of the error word and the error correction word;
The transformation submodule is used for transforming the first vector and the second vector through each preset transformation function in a plurality of preset transformation functions in sequence to obtain a plurality of third vectors corresponding to the first vector and a plurality of fourth vectors corresponding to the second vector; the calculation sub-module is used for respectively calculating a second difference value of the average value of each third score in the third vector and the average value of each fourth score in the fourth vector according to the third vector and the fourth vector obtained by transformation of each preset transformation function; and the second determining submodule is used for determining a plurality of second difference values as semantic features of the error word and the error correction word.
7. A computer readable storage medium, on which a computer program is stored, characterized in that the program, when being executed by a processor, implements the steps of the method according to any one of claims 1-5.
8. An electronic device, comprising:
a memory having a computer program stored thereon;
a processor for executing the computer program in the memory to implement the steps of the method of any one of claims 1-5.
CN201911144534.0A 2019-11-20 2019-11-20 Text collation method, text collation apparatus, computer-readable storage medium, and electronic device Active CN110929514B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911144534.0A CN110929514B (en) 2019-11-20 2019-11-20 Text collation method, text collation apparatus, computer-readable storage medium, and electronic device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911144534.0A CN110929514B (en) 2019-11-20 2019-11-20 Text collation method, text collation apparatus, computer-readable storage medium, and electronic device

Publications (2)

Publication Number Publication Date
CN110929514A CN110929514A (en) 2020-03-27
CN110929514B true CN110929514B (en) 2023-06-27

Family

ID=69851442

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911144534.0A Active CN110929514B (en) 2019-11-20 2019-11-20 Text collation method, text collation apparatus, computer-readable storage medium, and electronic device

Country Status (1)

Country Link
CN (1) CN110929514B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113241061B (en) * 2021-05-17 2023-03-10 北京字跳网络技术有限公司 Method and device for processing voice recognition result, electronic equipment and storage medium
CN114611524B (en) * 2022-02-08 2023-11-17 马上消费金融股份有限公司 Text error correction method and device, electronic equipment and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6618697B1 (en) * 1999-05-14 2003-09-09 Justsystem Corporation Method for rule-based correction of spelling and grammar errors
CN105279149A (en) * 2015-10-21 2016-01-27 上海应用技术学院 Chinese text automatic correction method

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5673265B2 (en) * 2010-03-31 2015-02-18 富士通株式会社 Calibration support apparatus and calibration support program
KR101636902B1 (en) * 2012-08-23 2016-07-06 에스케이텔레콤 주식회사 Method for detecting a grammatical error and apparatus thereof
KR101500617B1 (en) * 2013-08-07 2015-03-10 부산대학교 산학협력단 Method and system for Context-sensitive Spelling Correction Rules using Korean WordNet
CN105550173A (en) * 2016-02-06 2016-05-04 北京京东尚科信息技术有限公司 Text correction method and device
CN107291683A (en) * 2016-04-11 2017-10-24 珠海金山办公软件有限公司 A kind of spell checking methods and device
CN109948144B (en) * 2019-01-29 2022-12-06 汕头大学 Teacher utterance intelligent processing method based on classroom teaching situation
CN110210029B (en) * 2019-05-30 2020-06-19 浙江远传信息技术股份有限公司 Method, system, device and medium for correcting error of voice text based on vertical field
CN110348020A (en) * 2019-07-17 2019-10-18 杭州嘉云数据科技有限公司 A kind of English- word spelling error correction method, device, equipment and readable storage medium storing program for executing

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6618697B1 (en) * 1999-05-14 2003-09-09 Justsystem Corporation Method for rule-based correction of spelling and grammar errors
CN105279149A (en) * 2015-10-21 2016-01-27 上海应用技术学院 Chinese text automatic correction method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Natural Language Contents Evaluation System for Detecting Fake News using Deep Learning;Ye-Chan Ahn;《2019 16th International Joint Conference on Computer Science and Software Engineering (JCSSE)》;全文 *
一个用于OCR输出的中文文本的拼写校对系统;李蓉;《中文信息学报》;20090915(第05期);全文 *

Also Published As

Publication number Publication date
CN110929514A (en) 2020-03-27

Similar Documents

Publication Publication Date Title
CN106534548B (en) Voice error correction method and device
CN108052499B (en) Text error correction method and device based on artificial intelligence and computer readable medium
CN107341143B (en) Sentence continuity judgment method and device and electronic equipment
CN111753531A (en) Text error correction method and device based on artificial intelligence, computer equipment and storage medium
TWI567569B (en) Natural language processing systems, natural language processing methods, and natural language processing programs
CN113948066B (en) Error correction method, system, storage medium and device for real-time translation text
CN112613293B (en) Digest generation method, digest generation device, electronic equipment and storage medium
CN111274785A (en) Text error correction method, device, equipment and medium
CN110929514B (en) Text collation method, text collation apparatus, computer-readable storage medium, and electronic device
CN115438650B (en) Contract text error correction method, system, equipment and medium fusing multi-source characteristics
CN110826301B (en) Punctuation mark adding method, punctuation mark adding system, mobile terminal and storage medium
US20110229036A1 (en) Method and apparatus for text and error profiling of historical documents
CN112447172A (en) Method and device for improving quality of voice recognition text
CN110442843B (en) Character replacement method, system, computer device and computer readable storage medium
CN113255329A (en) English text spelling error correction method and device, storage medium and electronic equipment
CN112632956A (en) Text matching method, device, terminal and storage medium
CN112559725A (en) Text matching method, device, terminal and storage medium
CN115858776B (en) Variant text classification recognition method, system, storage medium and electronic equipment
CN111651961A (en) Voice-based input method and device
CN112733517B (en) Method for checking requirement template conformity, electronic equipment and storage medium
CN114510925A (en) Chinese text error correction method, system, terminal equipment and storage medium
CN114528824A (en) Text error correction method and device, electronic equipment and storage medium
CN114707489B (en) Method and device for acquiring annotation data set, electronic equipment and storage medium
CN111401011B (en) Information processing method and device and electronic equipment
CN113268977B (en) Text error correction method and device based on language model, terminal equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: No.27, 1st floor, building 14, Haidian Road, zhongjiancai District, Beijing 100096

Applicant after: Beijing PERCENT Technology Group Co.,Ltd.

Address before: No.27, 1st floor, building 14, Haidian Road, zhongjiancai District, Beijing 100096

Applicant before: BEIJING BAIFENDIAN INFORMATION SCIENCE & TECHNOLOGY Co.,Ltd.

GR01 Patent grant
GR01 Patent grant