CN116384388B - Method, device, equipment and medium for reverse identification AI intelligent writing - Google Patents

Method, device, equipment and medium for reverse identification AI intelligent writing Download PDF

Info

Publication number
CN116384388B
CN116384388B CN202310108714.3A CN202310108714A CN116384388B CN 116384388 B CN116384388 B CN 116384388B CN 202310108714 A CN202310108714 A CN 202310108714A CN 116384388 B CN116384388 B CN 116384388B
Authority
CN
China
Prior art keywords
text
clause
sentence
input
dimension
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310108714.3A
Other languages
Chinese (zh)
Other versions
CN116384388A (en
Inventor
陈卫海
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Xijin Information Technology Co ltd
Original Assignee
Shanghai Xijin Information Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Xijin Information Technology Co ltd filed Critical Shanghai Xijin Information Technology Co ltd
Priority to CN202310108714.3A priority Critical patent/CN116384388B/en
Publication of CN116384388A publication Critical patent/CN116384388A/en
Application granted granted Critical
Publication of CN116384388B publication Critical patent/CN116384388B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/216Parsing using statistical methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The application relates to a method, a device, equipment and a medium for reverse identification AI intelligent writing. The method comprises the following steps: segmenting the text, and dividing sentences of each segment of text according to punctuation marks; when the number of clauses of the target segment text is larger than or equal to a first threshold value, removing common high-frequency sentences to form an input clause data set formed by a plurality of clauses; extracting feature values from each dimension of sentence confusion, sentence presentation dispersion, knowledge span, topic cohesiveness, emotion richness and non-civilized expression of the input clause data set respectively; carrying out data normalization processing on the feature values extracted from each dimension, and optimizing a support vector machine classifier by utilizing a normalized feature group to obtain a classification model; and inputting the text to be identified into the classification model, and classifying each text segment of the text to be identified into human authoring or machine authoring, so as to assist human selectively putting time and energy into the artificial intelligent authored works.

Description

Method, device, equipment and medium for reverse identification AI intelligent writing
Technical Field
The application relates to the technical field of reverse identification AI intelligent writing, in particular to a method, a device, computer equipment and a storage medium for reverse identification AI intelligent writing.
Background
With the progress of technology, AI intelligent writing is more and more excellent, and meanwhile, a large amount of garbage data is generated. The garbage data not only has no substantial creative value, but also brings great interference and waste of time and energy to normal work of people.
The reverse identification AI intelligent writing algorithm can provide a means for intelligently identifying whether an article or a section of text is authored by human or machine, thereby assisting the human to selectively invest time and energy in the work authored by artificial intelligence.
Therefore, how to implement a method for effectively distinguishing whether an article or a text is created by human or machine is an urgent technical problem to be solved.
Disclosure of Invention
Based on the above, it is necessary to provide a method, apparatus, computer device and storage medium for reverse identification AI intelligent writing, which can solve the technical problem that it is not possible to effectively distinguish whether an article or a text is created by human or by machine at present.
In one aspect, a method for reverse identification AI smart writing is provided, the method comprising:
segmenting the text to form an input paragraph data set D= { D composed of multiple segments of characters 1 ,d 2 ,d 3 ,...,d i ,...,d m };
Initializing the input paragraph data set D, and dividing each paragraph of characters according to punctuation marks; word d when the target segment i When the number t of the clauses of the (2) is larger than or equal to a first threshold value, removing the common high-frequency sentence to form an input clause data set S formed by a plurality of clauses 1 ={s 1 ,s 2 ,s 3 ,...,s i ,...,s T -wherein T is less than or equal to T;
for the input clause dataset S 1 Extracting characteristic values from each dimension of sentence confusion, sentence presentation dispersion, knowledge span, theme cohesiveness, emotion richness and non-civilized expression respectively;
carrying out data normalization processing on the feature values extracted by each dimension, mapping the feature values extracted by each dimension into the same interval (0, 1) to form a normalized feature group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 );
Optimizing a support vector machine classifier by utilizing the normalized feature set X, and finding out the most suitable kernel function through multiple rounds of training to obtain a classification model;
and inputting the text to be identified into the classification model, and classifying each text segment of the text to be identified as human authoring or machine authoring.
In one embodiment, after the initializing the input paragraph data set, the sentence dividing step is performed on each paragraph of text according to punctuation marks, the method further includes: and when the number of the clauses of the target segment text is smaller than a first threshold value, whether the target segment text is machine creation is not judged, and otherwise, the next step is executed.
In one embodiment, the step of extracting feature values from sentence confusion dimensions for the input clause dataset comprises:
using a crust word segmentation device to segment each text D in the input paragraph data set D i Chinese word segmentation is carried out on each sentence of the number (1) to generate a word sequence, the length of the word sequence is recorded as N, the word element is recorded as WD, and the current sentence is recorded as S= { WD 1 ,WD 2 ,...,WD N };
Calculating the confusion of each sentence by a statistical-based language model confusion algorithm PP (S), wherein the calculation formula of the PP (S) is as followsWherein S represents the current sentence; n word sequence length; p (w) i ) A probability representing an i-th word; p (w) i |w 1 w 2 w 3 …w i-1 ) Representing the probability of calculating the ith word based on the previous i-1 words;
calculation of input clause dataset S Using PP (S) 1 Each sentence s i Is the confusion p of (2) i Generating a sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T };
Obtaining the sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T Maximum degree of confusion p in } max If p max If the text is smaller than or equal to the second threshold value, judging that the text is created by the machine; otherwise, entering the next step;
arranging the elements of the sentence confusion degree set P from small to large according to the confusion degree value, and taking the elements from 3T/4 to the tail to form a set Pa= { P 3T/4 ,...,p T -a }; calculating an arithmetic mean of the set PaConfusion of the text >
In one embodiment, the step of extracting feature values from a statement presentation dispersion dimension for the input clause dataset comprises:
for the input clause dataset S 1 Each clause s of (a) i Calculating character length l i Generating a clause length set L s ={l 1 ,l 2 ,...,l T };
Calculating the set of clause lengths L s Is of the dispersion of (2)Wherein->For the set L of clause lengths s An average of all clause character lengths;
marking the dispersion of text paragraphs as Q d ,Q d =q d
In one embodiment, the step of extracting feature values from the knowledge span dimension for the input clause dataset comprises:
using a pre-trained knowledge classification model to classify the input clause data set S 1 Each clause s of (a) i Classifying, and the corresponding classification set is C s ={c 1 ,.c 2 ,...c r -wherein r.ltoreq.T;
for class set C s The element in the mixture is subjected to de-duplication treatment to obtain C d ={c 1 ,.c 2 ,...c R R.ltoreq.r;
calculating knowledge span K of text passage d =card(C d ) Wherein card (C d ) Representing set C d The number of elements in the list.
In one embodiment, the step of extracting feature values from the topic cohesiveness dimension for the input clause dataset comprises:
using word2vec model to input clause data set S 1 Each clause s of (a) i Calculating the semantic similarity of the sentence with other sentences;
Judging sentences with semantic similarity larger than a third threshold as a theme;
calculating the input clause data set S 1 The number k of the topics in the list;
calculating topic cohesiveness T of text paragraphs d =e 1-k Wherein k is>0。
In one embodiment, the step of extracting feature values from the emotion richness dimension for the input clause dataset comprises:
calculating character richness, representing all characters except Chinese characters and English letters as a set Then calculate the target segment word d i Contains a set F 1 The number of characters in (a) is recorded as N f1
Calculating the richness of the language-gas assisted words and setting a language-gas assisted word set F 2 = { A, o, la, w, o, bar, have, ou, if, mock, ou, ha, have, no }, calculate the target segment word d i Contains a set F 2 The number of the auxiliary words is recorded as N f2
Calculating emotion richness E of text paragraph d =αN f1 +βN f2 Wherein alpha and beta are weight factors.
In one embodiment, the step of extracting feature values from the non-civilized term dimension for the input clause dataset comprises:
based on rule engine and knowledge base for said input clause data set S 1 Each clause s of (a) i Performing non-civilization calculation, and judging the target segment text d when any clause accords with the non-civilization rule i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Based on rule engine and knowledge base for said input clause data set S 1 Each clause s of (a) i Performing reverse blacking or quoting calculation on positive characters or events, and judging the target segment text d when any clause accords with the definition of a rule engine i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Non-civilized expression of a text paragraph is denoted as D d ,D d Is f 1 And f 2 Of (f) 1 And f 2 One of them is 1, D d =1, otherwise D d =0。
In one embodiment, the data normalization processing is performed on the feature values extracted from each dimension, and the feature values extracted from each dimension are mapped into the same interval (0, 1) to form a normalized feature set x= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 ) The method comprises the following steps:
obtaining the input clause data set S 1 Feature values respectively extracted from each dimension of sentence confusion, sentence presentation dispersion, knowledge span, topic cohesiveness, emotion richness and non-civilized expression to form feature dimension X' = (P) d ,Q d ,K d ,T d ,E d ,D d );
Using the formulaNormalizing the feature values extracted from each dimension, wherein x is min For the minimum value in the corresponding feature set, x max For the maximum value in the corresponding characteristic value set, x is the corresponding current characteristic value, x norm Normalizing the processed value for the corresponding current characteristic value;
after processing, mapping each dimension characteristic value in a (0, 1) interval to form a normalized characteristic group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 )。
On the other hand, provided is a reverse identification AI smart writing device, the device includes:
the segmentation processing module is used for carrying out segmentation processing on the text to form an input paragraph data set D= { D formed by a plurality of sections of characters 1 ,d 2 ,d 3 ,...,d i ,...,d m };
The clause processing module is used for initializing the input paragraph data set D and carrying out clause on each paragraph of characters according to punctuation marks; word d when the target segment i When the number t of the clauses of the (2) is larger than or equal to a first threshold value, removing the common high-frequency sentence to form an input clause data set S formed by a plurality of clauses 1 ={s 1 ,s 2 ,s 3 ,...,s i ,...,s T -wherein T is less than or equal to T;
a module for extracting characteristic values for the input clause data set S 1 Extracting characteristic values from each dimension of sentence confusion, sentence presentation dispersion, knowledge span, theme cohesiveness, emotion richness and non-civilized expression respectively;
the data normalization processing module is used for performing data normalization processing on the feature values extracted by each dimension, mapping the feature values extracted by each dimension into the same interval (0, 1) to form a normalized feature group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 );
The classification model acquisition module is used for optimizing the support vector machine classifier by utilizing the normalized feature set X, and finding out the most suitable kernel function through multiple rounds of training to obtain a classification model;
and the text recognition module is used for inputting the text to be recognized into the classification model, and classifying each text segment of the text to be recognized into human authoring or machine authoring.
In yet another aspect, a computer device is provided comprising a memory, a processor, and a computer program stored on the memory and executable on the processor, the processor implementing the steps of:
segmenting the text to form an input paragraph data set D= { D composed of multiple segments of characters 1 ,d 2 ,d 3 ,...,d i ,...,d m };
Initializing the input paragraph data set D, and dividing each paragraph of characters according to punctuation marks; word d when the target segment i When the number t of the clauses of the (2) is larger than or equal to a first threshold value, removing the common high-frequency sentence to form an input clause data set S formed by a plurality of clauses 1 ={s 1 ,s 2 ,s 3 ,...,s i ,...,s T -wherein T is less than or equal to T;
for the input clause dataset S 1 Extracting characteristic values from each dimension of sentence confusion, sentence presentation dispersion, knowledge span, theme cohesiveness, emotion richness and non-civilized expression respectively;
Carrying out data normalization processing on the feature values extracted by each dimension, mapping the feature values extracted by each dimension into the same interval (0, 1) to form a normalized feature group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 );
Optimizing a support vector machine classifier by utilizing the normalized feature set X, and finding out the most suitable kernel function through multiple rounds of training to obtain a classification model;
and inputting the text to be identified into the classification model, and classifying each text segment of the text to be identified as human authoring or machine authoring.
In yet another aspect, a computer readable storage medium is provided, having stored thereon a computer program which when executed by a processor performs the steps of:
segmenting the text to form an input paragraph data set D= { D composed of multiple segments of characters 1 ,d 2 ,d 3 ,...,d i ,...,d m };
Initializing the input paragraph data set D, and dividing each paragraph of characters according to punctuation marks; word d when the target segment i The number t of clauses of (2) is greater than or equal toRemoving common high-frequency sentences to form an input clause data set S formed by a plurality of clauses when the first threshold value is reached 1 ={s 1 ,s 2 ,s 3 ,...,s i ,...,s T -wherein T is less than or equal to T;
for the input clause dataset S 1 Extracting characteristic values from each dimension of sentence confusion, sentence presentation dispersion, knowledge span, theme cohesiveness, emotion richness and non-civilized expression respectively;
Carrying out data normalization processing on the feature values extracted by each dimension, mapping the feature values extracted by each dimension into the same interval (0, 1) to form a normalized feature group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 );
Optimizing a support vector machine classifier by utilizing the normalized feature set X, and finding out the most suitable kernel function through multiple rounds of training to obtain a classification model;
and inputting the text to be identified into the classification model, and classifying each text segment of the text to be identified as human authoring or machine authoring.
According to the method, the device, the computer equipment and the storage medium for reversely identifying the AI intelligent writing, based on the characteristics of the AI intelligent writing, the characteristic values of the data extracted by each dimension are reversely judged, so that whether an article or a section of text is created by a human or by a machine can be intelligently identified, and the human can be assisted to selectively input time and energy to the work created by the artificial intelligence.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings required for the description of the embodiments will be briefly described below, and it is apparent that the drawings in the following description are only some embodiments of the present invention, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is an algorithm flow diagram of a method of reverse identification AI smart writing in one embodiment;
FIG. 2 is a flow chart of a method for reverse identification AI smart writing in another embodiment;
FIG. 3 is a flow diagram of the step of extracting feature values from sentence confusion dimensions for the input clause dataset, in one embodiment;
FIG. 4 is a flow diagram of the step of extracting feature values from a statement presentation dispersion dimension for the input clause dataset in one embodiment;
FIG. 5 is a flow diagram of the step of extracting feature values from knowledge span dimensions for the input clause dataset in one embodiment;
FIG. 6 is a flow diagram of the step of extracting feature values from a topic cohesiveness dimension for the input clause dataset, in one embodiment;
FIG. 7 is a flow diagram of a step of extracting feature values from emotion richness dimensions for the input clause dataset, in one embodiment;
FIG. 8 is a flow diagram of a step of extracting feature values from non-civilized phrase dimensions for the input phrase dataset in one embodiment;
fig. 9 illustrates a data normalization process for feature values extracted from each dimension, and maps the feature values extracted from each dimension into the same interval (0, 1) to form a normalized feature set x= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 ) A flow diagram of the steps;
FIG. 10 is a block diagram of a reverse identification AI smart writing apparatus in one embodiment;
FIG. 11 is an internal block diagram of a computer device in one embodiment.
Detailed Description
In order to make the objects, technical solutions and advantages of the present application more apparent, the present application will be further described in detail with reference to the accompanying drawings and examples. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the present application.
Example 1
In order to solve the problems pointed out by the background technology, the embodiment 1 of the invention creatively provides a method for reversely identifying the intelligent writing of the AI, wherein the intelligent writing algorithm for reversely identifying the AI is a means for intelligently identifying whether an article or a section of text is created by human or by a machine, thereby assisting the human to selectively input time and energy to the work created by the artificial intelligence.
Reverse recognition AI intelligent writing algorithm PQKTED (Ppl-QTd-knowledgespan-Theme-motion-Deviat ion from social core values) judges whether text is written by a machine based on artificial intelligent writing principle and feature reverse calculation. The PQKTE algorithm adopts six dimensions of sentence confusion (Ppl), sentence presentation dispersion (QTd), knowledge span (knowledgespan), theme cohesiveness (Theme), emotion richness (Emotion) and non-civilized term (Deviation from social core values) as characteristics, and adopts a supervised classifier to carry out model training and classification tasks, namely:
y=f(x 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 )
Wherein:
y is a classification result;
f(x 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 ) For the classifier (SVM, LSTM and other traditional machine learning or deep learning methods), different final accuracy and recall rate of the classifier are selected, and in the experiment, the SVM is adopted as the classifier;
x 1 to the degree of confusion (Ppl), x 2 As a dispersion (QTd), x 3 For Knowledge span (knowledgespan), x 4 Is the topic cohesiveness (Theme), x 5 Is emotion richness (Emotion), x 6 Is a non-civilized term (Deviat ion from social core values).
Specifically, as shown in fig. 1, fig. 1 is an algorithm flow chart of the method for reverse identification AI smart writing.
The algorithm has seven links, namely, segmentation processing, data preprocessing, data feature extraction, feature normalization processing, support vector machine classifier optimization and model training are respectively carried out on the text.
1. Text segmentation processing. First through warpSorting out a segmentation rule base by different presentation modes of test data or specific data, and then segmenting the text according to paragraphs, wherein the paragraphs are marked as D= { D 1 ,d 2 ,d 3 ,...,d m }。
2. And (5) preprocessing data.
a) Each d in pair 1 i Clauses are carried out according to punctuation marks and are marked as S= { S 1 ,s 2 ,s 3 ,...,s T }. Setting a threshold value a, when n<a is not d i A calculation is made of whether to author for the machine. Otherwise, the next step is carried out.
b) For s using the common high-frequency sentence database i Filtering to remove sentences familiar to both human and machine, and generating new S 1 ={s 1 ,s 2 ,s 3 ,...,s T And T.ltoreq.t.
3. And (5) extracting data characteristics. Data features are extracted from the following six dimensions, with no precedence order.
a) Confusion (Ppl). The AI intelligent writing judges whether the index of whether the sentence is smooth or not is the confusion degree of the sentence. Lower sentence confusion indicates more smooth sentences. In this model, the higher the confusion is considered to be the more likely to be human authoring, the lower the confusion is or the confusion always approaches a threshold b, but the threshold b is not always exceeded, and the AI intelligent authoring is considered to be possible. The confusion degree calculating process comprises the following steps: firstly, chinese word segmentation is carried out on sentences by using a barker word segmentation device to generate word sequences, the length of the word sequences is marked as N, and the word elements are marked as WD, namely
S={WD 1 ,WD 2 ,...,WD N }。
Secondly, calculating sentence confusion through a statistical-based language model confusion algorithm PP (S), wherein the calculation formula is as follows:
wherein S represents the current sentence; n word sequence length; p (w) i ) A probability representing an i-th word;
p(w i |w 1 w 2 w 3 …w i-1 ) The probability of the i-th word is calculated based on the first i-1 words.
And a third step of: calculating set S using PP (S) 1 Each s of (3) i Is the confusion p of (2) i Generating a sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T }. Then calculate the confusion P of a section of characters d
A) If p is max <=b, then d i The method is intelligent creation for AI. Otherwise, the next step is carried out.
B) The set P is processed as follows:
(1) arranging elements in the P set from small to large according to values;
(2) taking 3T/4 to last element to generate new set P a ={p 3T/4 ,...,p T };
(3) Calculation of P a Is the arithmetic mean of (a):
(4) meter with a meter body
b) Dispersion (QTd). The dispersion of the article paragraphs is noted as Q d . Calculate Q d The algorithm process is as follows:
(1) for set S 1 Each s of (3) i Calculating character length l i Generating a new set L s ={l 1 ,l 2 ,...,l T }。
(2) Calculate L s Is of the dispersion Q of (2) d
Wherein the method comprises the steps ofIs the average.
(3) Meter Q d =q d
c) Knowledge span (knowledgespan). Record paragraph knowledge span as K d
(1) The set S is obtained by using a Knowledge classification model (M-knowledges, including 22 classifications of sociology, mankind, economics, politics, histories, law and the like) pre-trained by oneself 1 Classifying each sentence in the list, and classifying the corresponding classification set as follows:
C s ={c 1 ,.c 2 ,...c r -where r<=n。
(2) For C s The elements in the mixture are subjected to de-duplication treatment to obtain
C d ={c 1 ,.c 2 ,...c R (wherein R is)<=r。
(3) Knowledge span K d =card(C d ) Wherein card (C d ) Representing set C d The number of elements in the list. The number of classifications involved in the text is used as the knowledge span.
d) Theme cohesiveness (Theme). The topic cohesiveness was noted as T d . Calculate T d The process of (2) is as follows:
(1) pair set S using word2vec model 1 Each s of (3) i And calculating the semantic similarity of the sentence with other sentences.
(2) Sentences with similarity greater than threshold B are treated as a topic
(3) S is calculated by (1) and (2) 1 Number of topics k
(4) Subject cohesiveness is noted as T d
T d =e 1-k Wherein k is>0;
e) Emotion richness (Emotion). Emotion richness is marked as E d . Text condition richness is measured here in terms of character richness and word-of-speech richness in both dimensions.
(1) And (5) calculating character richness. All characters except Chinese and English letters, expressed as a set
Then calculate paragraph d i Contains F 1 The number of characters in the book is recorded as N f1
(2) And (5) calculating the richness of the language and the auxiliary words. Word-aid set F 2 = { a, o, la, w, bar, have, wa, if, do, ha, no. Then calculate paragraph d i Contains F 2 The number of the auxiliary words in the book is recorded as N f2
(3) Emotion richness is recorded as
E d =αN f1 +βN f2 Wherein alpha, beta is a weight factor
f) Non-civilized term (Deviation from social core values), denoted as D d . The selection of non-civilizing and positive characters or events in the present model is either reverse blackened or calculated by referencing both dimensions.
(1) Non-civilized calculation. Based on the rule engine and knowledge base, when S 1 Any one s i The text d of the segment is processed as long as the non-civilized rule is met i And (2) is marked as 1, otherwise as 0.
(2) Positive characters or events are reverse blacked out or quoted. Based on the rule engine and knowledge base, when S 1 Any one s i The non-civilized value of the segment of text di is marked as 1 as long as the rule engine definition is met, otherwise marked as 0.
③D d Is f 1 And f 2 Of (1), i.e. as long as f 1 And f 2 One is 1, then D d =1, otherwise D d =0。
4. Data normalization
Feature dimension X' extracted in pair 3 = (P) d ,Q d ,K d ,T d ,E d ,D d ) Using the formula:
the feature of each dimension is normalized, after the normalization, the 6-dimension feature values are mapped in the (0, 1) interval, so that normalized feature group X= (X) is recorded 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 ). Wherein x is min For the minimum value in the corresponding feature set, x max Is the maximum value in the corresponding eigenvalue set. X is the corresponding current characteristic value.
5. Model training and classifier tuning
Manual annotation data is obtained according to the following steps: 1:1 into a training set, a validation set and a test set. Wherein the training set is used for training the model and determining parameters; the verification set is used for adjusting parameters, monitoring whether the model is fitted or not, and being used for multiple times in the training process so as to continuously adjust the parameters; the test set is used to evaluate the generalization ability of the final model.
The support vector machine is a binary model algorithm whose basic model is the most widely spaced linear classifier defined in feature space.
The data of this task is linearly inseparable in two dimensions, but the problem of linear inseparability may be non-linearly inseparable, i.e. there is a hyperplane in the feature space separating the positive and negative classes. The use of nonlinear functions can map nonlinear separable problems from the original feature space to higher dimensional space to translate into linear separable problems, where the hyperplane as a decision boundary is represented as follows:
ω·φ(x)+b=0
y(x)=sign(w·φ(x)+b)
since the mapping function has a complex form, it is difficult to calculate its inner product, a kernel method can be used. Phi (x) in the formula is a kernel function.
The kernel function adopted by the task is a polynomial kernel, and the analytic formula is as follows:
when the order of the polynomial kernel is 1, it is called a linear kernel, and the corresponding nonlinear classifier is degenerated to a linear classifier.
6. Model evaluation
Models are typically evaluated from several dimensions, such as accuracy, precision, recall, F1 values, etc. The accuracy is that the number of correctly classified samples accounts for the number of total samples; the accuracy rate (P) is the proportion of the correct predicted positive example data to the predicted positive example data; the recall ratio (R) is the ratio of the predicted correct positive example data to the actual positive example data. F1 values are harmonic mean, f1=2/(1/p+1/R).
The dimensions to be concerned in the task are the precision, recall rate and F1 value, and particularly different dimensions can be concerned according to different usage scenes.
Model generalization evaluation is carried out by using the test set, and the following results are obtained:
accuracy rate of Recall rate of recall F1-score
PQKTED 0.76 0.93 0.803
From the statistical result, the accuracy is 76%, which represents that 76% of the results of all models are judged to be AI writing, and the actual results are AI writing; the recall rate was 93%, representing samples of AI writing in the test set, with 93% recalled. Therefore, the model has good generalization, and the PQKTED algorithm can effectively identify whether the text is authored by the AI.
Example 2
All the technical features of example 1 are included in example 2. Specifically, as shown in fig. 2, in embodiment 2, a method for reverse identification AI smart writing is provided, which includes the following steps:
step S1, segmenting the text to form an input paragraph data set D= { D composed of multiple segments of characters 1 ,d 2 ,d 3 ,...,d i ,...,d m };
Step S2, initializing the input paragraph data set D, and carrying out clause on each paragraph of characters according to punctuation marks; word d when the target segment i When the number t of the clauses of the (2) is larger than or equal to a first threshold value, removing the common high-frequency sentence to form an input clause data set S formed by a plurality of clauses 1 ={s 1 ,s 2 ,s 3 ,...,s i ,...,s T -wherein T is less than or equal to T;
step S3, for the input clause dataset S 1 Extracting feature values from each dimension of sentence confusion (Ppl), sentence presentation dispersion (QTd), knowledge span (knowledgespan), topic cohesiveness (Theme), emotion richness (motion) and non-civilization term (Deviation from social core values);
step S4, carrying out data normalization processing on the feature values extracted from each dimension, mapping the feature values extracted from each dimension into the same interval (0, 1) to form a normalizationNormalized feature set x= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 );
Step S5, optimizing a support vector machine classifier by utilizing the normalized feature set X, and finding out the most suitable kernel function through multiple rounds of training to obtain a classification model;
and S6, inputting the text to be recognized into the classification model, and classifying each text segment of the text to be recognized as human authoring or machine authoring.
In this embodiment, after the initializing the input paragraph data set, the sentence dividing step is performed on each paragraph of text according to punctuation marks, the method further includes: and when the number of the clauses of the target segment text is smaller than a first threshold value, whether the target segment text is machine creation is not judged, and otherwise, the next step is executed.
As shown in fig. 3, in this embodiment, the step of extracting feature values from sentence confusion dimensions for the input clause data set includes:
step S11, using a crust word segmentation device to segment each text D in the input paragraph data set D i Chinese word segmentation is carried out on each sentence of the number (1) to generate a word sequence, the length of the word sequence is recorded as N, the word element is recorded as WD, and the current sentence is recorded as S= { WD 1 ,WD 2 ,...,WD N };
Step S12, calculating the confusion of each sentence by a statistical-based language model confusion algorithm PP (S), wherein the calculation formula of the PP (S) is as followsWherein S represents the current sentence; n word sequence length; p (w) i ) A probability representing an i-th word; p (w) i |w 1 w 2 w 3 …w i-1 ) Representing the probability of calculating the ith word based on the previous i-1 words;
step S13, calculating the input clause data set S by using the PP (S) 1 Each sentence s i Is the confusion p of (2) i Generating a sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T };
Step S14, obtaining the sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T Maximum degree of confusion p in } max If p max If the text is smaller than or equal to the second threshold value, judging that the text is created by the machine; otherwise, entering the next step;
step S15, arranging the elements of the sentence confusion degree set P from small to large according to the confusion degree value, and taking the elements from 3T/4 to the tail to form a set Pa= { P 3T/4 ,...,p T -a }; calculating an arithmetic mean of the set PaAs confusion of the text, i.e. +.>
As shown in fig. 4, in the present embodiment, the step of extracting feature values from a statement presentation dispersion dimension for the input phrase data set includes:
step S21, for the input clause dataset S 1 Each clause s of (a) i Calculating character length l i Generating a clause length set L s ={l 1 ,l 2 ,...,l T };
Step S22, calculating the sentence length set L s Is of the dispersion of (2)Wherein->For the set L of clause lengths s An average of all clause character lengths;
step S23, marking the dispersion of the text passage as Q d ,Q d =q d
As shown in fig. 5, in the present embodiment, the step of extracting the feature value from the knowledge span dimension for the input clause dataset includes:
step S31, utilizing a pre-trained knowledge classification modelFor the input clause dataset S 1 Each clause s of (a) i Classifying, and the corresponding classification set is C s ={c 1 ,.c 2 ,...c r -wherein r.ltoreq.T;
step S32, classifying the set C s The element in the mixture is subjected to de-duplication treatment to obtain C d ={c 1 ,.c 2 ,...c R R.ltoreq.r;
step S33, calculating knowledge span K of text paragraph d =card(C d ) Wherein card (C d ) Representing set C d The number of elements in the list.
As shown in fig. 6, in the present embodiment, the step of extracting a feature value from the topic cohesiveness dimension for the input clause dataset includes:
Step S41, utilizing word2vec model to input clause data set S 1 Each clause s of (a) i Calculating the semantic similarity of the sentence with other sentences;
step S42, judging sentences with semantic similarity larger than a third threshold as a theme;
step S43, calculating the input clause dataset S 1 The number k of the topics in the list;
step S44, calculating the topic cohesiveness T of the text paragraph d =e 1-k Wherein k is>0。
As shown in fig. 7, in this embodiment, the step of extracting a feature value from the emotion richness dimension for the input clause dataset includes:
step S51, calculating character richness, representing all characters except Chinese characters and English letters as a set Then calculate the target segment word d i Contains a set F 1 The number of characters in (a) is recorded as N f1
Step S52, calculating the richness of the language and the auxiliary words, and settingWord-aid set F 2 = { A, o, la, w, o, bar, have, ou, if, mock, ou, ha, have, no }, calculate the target segment word d i Contains a set F 2 The number of the auxiliary words is recorded as N f2
Step S53, calculating emotion richness E of text paragraph d =αN f1 +βN f2 Wherein alpha and beta are weight factors.
As shown in fig. 8, in the present embodiment, the step of extracting the feature value from the non-civilized term dimension for the input clause dataset includes:
Step S61, inputting a clause data set S based on a rule engine and a knowledge base 1 Each clause s of (a) i Performing non-civilization calculation, and judging the target segment text d when any clause accords with the non-civilization rule i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Step S62, inputting the clause dataset S based on a rule engine and a knowledge base 1 Each clause s of (a) i Performing reverse blacking or quoting calculation on positive characters or events, and judging the target segment text d when any clause accords with the definition of a rule engine i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Step S63, the non-civilized expression of the text paragraph is denoted as D d ,D d Is f 1 And f 2 Of (f) 1 And f 2 One of them is 1, D d =1, otherwise D d =0。
As shown in fig. 9, in this embodiment, the data normalization processing is performed on the feature values extracted in each dimension, and the feature values extracted in each dimension are mapped into the same interval (0, 1) to form a normalized feature group x= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 ) The method comprises the following steps:
step S71, obtaining the input clause data set S 1 Feature values respectively extracted from each dimension of sentence confusion (Ppl), sentence presentation dispersion (QTd), knowledge span (knowledgespan), topic cohesiveness (Theme), emotion richness (motion) and non-civilization term (Deviation from social core values) to form feature dimension X' = (P) d ,Q d ,K d ,T d ,E d ,D d );
Step S72, using the formulaNormalizing the feature values extracted from each dimension, wherein x is min For the minimum value in the corresponding feature set, x max For the maximum value in the corresponding characteristic value set, x is the corresponding current characteristic value, x norm Normalizing the processed value for the corresponding current characteristic value;
step S73, mapping each dimension characteristic value in the (0, 1) interval after processing to form a normalized characteristic group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 )。
In this embodiment, the step of optimizing the support vector machine classifier by using the normalized feature set X, finding out the most suitable kernel function through multiple rounds of training, and obtaining the classification model includes:
manual annotation data is obtained according to the following steps: 1:1 into a training set, a validation set and a test set. Wherein the training set is used for training the model and determining parameters; the verification set is used for adjusting parameters, monitoring whether the model is fitted or not, and being used for multiple times in the training process so as to continuously adjust the parameters; the test set is used to evaluate the generalization ability of the final model.
The support vector machine is a binary model algorithm whose basic model is the most widely spaced linear classifier defined in feature space.
The data of this task is linearly inseparable in two dimensions, but the problem of linear inseparability may be non-linearly inseparable, i.e. there is a hyperplane in the feature space separating the positive and negative classes. The use of nonlinear functions can map nonlinear separable problems from the original feature space to higher dimensional space to translate into linear separable problems, where the hyperplane as a decision boundary is represented as follows:
ω·φ(x)+b=0
y(x)=sign(w·φ(x)+b)
Since the mapping function has a complex form, it is difficult to calculate its inner product, a kernel method can be used. Phi (x) in the formula is a kernel function.
The kernel function adopted by the task is a polynomial kernel, and the analytic formula is as follows:
when the order of the polynomial kernel is 1, it is called a linear kernel, and the corresponding nonlinear classifier is degenerated to a linear classifier.
In this embodiment, the text to be identified is input into the classification model, and the classification model is further evaluated after each text segment of the text to be identified is classified as human authoring or machine authoring.
Models are typically evaluated from several dimensions, such as accuracy, precision, recall, F1 values, etc. The accuracy is that the number of correctly classified samples accounts for the number of total samples; the accuracy rate (P) is the proportion of the correct predicted positive example data to the predicted positive example data; the recall ratio (R) is the ratio of the predicted correct positive example data to the actual positive example data. F1 values are harmonic mean, f1=2/(1/p+1/R).
The dimensions to be concerned in the task are the precision, recall rate and F1 value, and particularly different dimensions can be concerned according to different usage scenes.
Model generalization evaluation is carried out by using the test set, and the following results are obtained:
Accuracy rate of Recall rate of recall F1-score
PQKTED 0.76 0.93 0.803
From the statistical result, the accuracy is 76%, which represents that 76% of the results of all models are judged to be AI writing, and the actual results are AI writing; the recall rate was 93%, representing samples of AI writing in the test set, with 93% recalled. Therefore, the model has good generalization, and the PQKTED algorithm can effectively identify whether the text is authored by the AI.
According to the method for reversely identifying the AI intelligent writing, based on the characteristics of the AI intelligent writing, the characteristic values of the data extracted by each dimension are reversely judged, so that whether an article or a section of text is created by a human or by a machine can be intelligently identified, and the human is assisted to selectively input time and energy to the work created by the artificial intelligence.
It should be understood that, although the steps in the flowcharts of fig. 2-9 are shown in order as indicated by the arrows, these steps are not necessarily performed in order as indicated by the arrows. The steps are not strictly limited to the order of execution unless explicitly recited herein, and the steps may be executed in other orders. Moreover, at least some of the steps of fig. 2-9 may include multiple sub-steps or stages that are not necessarily performed at the same time, but may be performed at different times, nor does the order in which the sub-steps or stages are performed necessarily occur in sequence, but may be performed alternately or alternately with at least a portion of other steps or sub-steps or stages of other steps.
In one embodiment, as shown in fig. 10, there is provided a reverse identification AI-smart writing apparatus 10, including: the system comprises a segmentation processing module 1, a clause processing module 2, an extracting characteristic value module 3, a data normalization processing module 4, a classification model obtaining module 5 and a text recognition module 6.
The segmentation processing module 1 is configured to perform segmentation processing on a text to form an input paragraph data set d= { D composed of multiple segments of text 1 ,d 2 ,d 3 ,...,d i ,...,d m }。
The clause processing module 2 is used for initializing the input paragraph data set D and carrying out clause on each paragraph of characters according to punctuation marks; word d when the target segment i When the number t of the clauses of the (2) is larger than or equal to a first threshold value, removing the common high-frequency sentence to form an input clause data set S formed by a plurality of clauses 1 ={s 1 ,s 2 ,s 3 ,...,s i ,...,s T And T.ltoreq.t.
The extraction feature value module 3 is used for inputting the clause data set S 1 Characteristic values are extracted from each dimension of sentence confusion (Ppl), sentence presentation dispersion (QTd), knowledge span (knowledgespan), topic cohesiveness (Theme), emotion richness (Emotion) and non-civilization expression (Deviation from social core values) respectively.
The data normalization processing module 4 is configured to perform data normalization processing on feature values extracted in each dimension, map the feature values extracted in each dimension into the same interval (0, 1), and form a normalized feature set x= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 )。
The module 5 for obtaining the classification model is used for optimizing the support vector machine classifier by utilizing the normalized feature set X, and finding out the most suitable kernel function through multiple rounds of training to obtain the classification model.
The text recognition module 6 is configured to input the text to be recognized into the classification model, where each text segment of the text to be recognized is classified as human authored or machine authored.
In this embodiment, after the initializing the input paragraph data set, the sentence dividing step is performed on each paragraph of text according to punctuation marks, the method further includes: and when the number of the clauses of the target segment text is smaller than a first threshold value, whether the target segment text is machine creation is not judged, and otherwise, the next step is executed.
In this embodiment, the step of extracting feature values from sentence confusion dimensions for the input clause dataset includes:
using a crust word segmentation device to segment each text D in the input paragraph data set D i Chinese word segmentation is carried out on each sentence of the number (1) to generate a word sequence, the length of the word sequence is recorded as N, the word element is recorded as WD, and the current sentence is recorded as S= { WD 1 ,WD 2 ,...,WD N };
Calculating the confusion of each sentence by a statistical-based language model confusion algorithm PP (S), wherein the calculation formula of the PP (S) is as followsWherein S represents the current sentence; n word sequence length; p (w) i ) A probability representing an i-th word; p (w) i |w 1 w 2 w 3 …w i-1 ) Representing the probability of calculating the ith word based on the previous i-1 words;
calculation of input clause dataset S Using PP (S) 1 Each sentence s i Is the confusion p of (2) i Generating a sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T };
Obtaining the sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T Maximum degree of confusion p in } max If p max If the text is smaller than or equal to the second threshold value, judging that the text is created by the machine; otherwise, entering the next step;
the elements of the sentence confusion degree set P are confusedThe degree values are arranged from small to large, and elements from 3T/4 to the tail are taken to form a set Pa= { p 3T/4 ,...,p T -a }; calculating an arithmetic mean of the set PaConfusion of the text>
In this embodiment, the step of extracting feature values from the statement presentation dispersion dimension for the input phrase dataset includes:
for the input clause dataset S 1 Each clause s of (a) i Calculating character length l i Generating a clause length set L s ={l 1 ,l 2 ,...,l T };
Calculating the set of clause lengths L s Is of the dispersion of (2)Wherein->For the set L of clause lengths s An average of all clause character lengths;
marking the dispersion of text paragraphs as Q d ,Q d =q d
In this embodiment, the step of extracting feature values from the knowledge span dimension for the input clause dataset includes:
Using a pre-trained knowledge classification model to classify the input clause data set S 1 Each clause s of (a) i Classifying, and the corresponding classification set is C s ={c 1 ,.c 2 ,...c r -wherein r.ltoreq.T;
for class set C s The element in the mixture is subjected to de-duplication treatment to obtain C d ={c 1 ,.c 2 ,...c R R.ltoreq.r;
calculating knowledge span K of text passage d =card(C d ) Wherein card (C d ) Representing set C d The number of elements in the list.
In this embodiment, the step of extracting feature values from the topic cohesiveness dimension for the input clause dataset includes:
using word2vec model to input clause data set S 1 Each clause s of (a) i Calculating the semantic similarity of the sentence with other sentences;
judging sentences with semantic similarity larger than a third threshold as a theme;
calculating the input clause data set S 1 The number k of the topics in the list;
calculating topic cohesiveness T of text paragraphs d =e 1-k Wherein k is>0。
In this embodiment, the step of extracting a feature value from the emotion richness dimension for the input clause dataset includes:
calculating character richness, representing all characters except Chinese characters and English letters as a set
Then calculate the target segment word d i Contains a set F 1 The number of characters in (a) is recorded as N f1
Calculating the richness of the language-gas assisted words and setting a language-gas assisted word set F 2 = { A, o, la, w, o, bar, have, ou, if, mock, ou, ha, have, no }, calculate the target segment word d i Contains a set F 2 The number of the auxiliary words is recorded as N f2
Calculating emotion richness E of text paragraph d =αN f1 +βN f2 Wherein alpha and beta are weight factors.
In this embodiment, the step of extracting the feature value from the non-civilized term dimension for the input clause dataset includes:
based on rule engine and knowledge base for said input clause data set S 1 Each clause s of (a) i Performing non-civilization calculation, and judging the target segment text d when any clause accords with the non-civilization rule i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Based on rule engine and knowledge base for said input clause data set S 1 Each clause s of (a) i Performing reverse blacking or quoting calculation on positive characters or events, and judging the target segment text d when any clause accords with the definition of a rule engine i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Non-civilized expression of a text paragraph is denoted as D d ,D d Is f 1 And f 2 Of (f) 1 And f 2 One of them is 1, D d =1, otherwise D d =0。
In this embodiment, the data normalization processing is performed on the feature values extracted in each dimension, and the feature values extracted in each dimension are mapped into the same interval (0, 1) to form a normalized feature set x= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 ) The method comprises the following steps:
obtaining the input clause data set S 1 Feature values respectively extracted from each dimension of sentence confusion (Ppl), sentence presentation dispersion (QTd), knowledge span (knowledgespan), topic cohesiveness (Theme), emotion richness (motion) and non-civilization term (Deviation from social core values) to form feature dimension X' = (P) d ,Q d ,K d ,T d ,E d ,D d );
Using the formulaNormalizing the feature values extracted from each dimension, wherein x is min For the minimum value in the corresponding feature set, x max For the maximum value in the corresponding characteristic value set, x is the corresponding current characteristic value, x norm Normalizing the processed value for the corresponding current characteristic value; />
After processing, mapping each dimension characteristic value in a (0, 1) interval to form a normalized characteristic group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 )。
According to the reverse identification AI intelligent writing device, based on the characteristics of AI intelligent writing, through carrying out reverse judgment on the data characteristic values extracted by each dimension, whether an article or a section of text is created by human or machine can be intelligently identified, so that the human can be assisted in selectively inputting time and energy to the work created by the artificial intelligence.
For specific limitations on the reverse-identification AI smart writing device, reference may be made to the above limitations on the method of reverse-identification AI smart writing, and no further description is given here. The above-mentioned each module in the reverse identification AI intelligent writing apparatus may be implemented in whole or in part by software, hardware, and combinations thereof. The above modules may be embedded in hardware or may be independent of a processor in the computer device, or may be stored in software in a memory in the computer device, so that the processor may call and execute operations corresponding to the above modules.
In one embodiment, a computer device is provided, which may be a server, and the internal structure of which may be as shown in fig. 11. The computer device includes a processor, a memory, a network interface, and a database connected by a system bus. Wherein the processor of the computer device is configured to provide computing and control capabilities. The memory of the computer device includes a non-volatile storage medium and an internal memory. The non-volatile storage medium stores an operating system, computer programs, and a database. The internal memory provides an environment for the operation of the operating system and computer programs in the non-volatile storage media. The database of the computer device is used for storing reverse identification AI intelligent writing data. The network interface of the computer device is used for communicating with an external terminal through a network connection. The computer program, when executed by the processor, implements a method of reverse identification AI smart writing.
It will be appreciated by those skilled in the art that the structure shown in fig. 11 is merely a block diagram of a portion of the structure associated with the present application and is not limiting of the computer device to which the present application applies, and that a particular computer device may include more or fewer components than shown, or may combine some of the components, or have a different arrangement of components.
In one embodiment, a computer device is provided comprising a memory, a processor, and a computer program stored on the memory and executable on the processor, the processor implementing the steps of when executing the computer program:
segmenting the text to form an input paragraph data set D= { D composed of multiple segments of characters 1 ,d 2 ,d 3 ,...,d i ,...,d m };
Initializing the input paragraph data set D, and dividing each paragraph of characters according to punctuation marks; word d when the target segment i When the number t of the clauses of the (2) is larger than or equal to a first threshold value, removing the common high-frequency sentence to form an input clause data set S formed by a plurality of clauses 1 ={s 1 ,s 2 ,s 3 ,...,s i ,...,s T -wherein T is less than or equal to T;
for the input clause dataset S 1 Extracting feature values from each dimension of sentence confusion (Ppl), sentence presentation dispersion (QTd), knowledge span (knowledgespan), topic cohesiveness (Theme), emotion richness (motion) and non-civilization term (Deviation from social core values);
carrying out data normalization processing on the feature values extracted from each dimension, and mapping the feature values extracted from each dimension into the same interval (0, 1)Form normalized feature set x= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 );
Optimizing a support vector machine classifier by utilizing the normalized feature set X, and finding out the most suitable kernel function through multiple rounds of training to obtain a classification model;
And inputting the text to be identified into the classification model, and classifying each text segment of the text to be identified as human authoring or machine authoring.
In this embodiment, after the initializing the input paragraph data set, the sentence dividing step is performed on each paragraph of text according to punctuation marks, the method further includes: and when the number of the clauses of the target segment text is smaller than a first threshold value, whether the target segment text is machine creation is not judged, and otherwise, the next step is executed.
In this embodiment, the step of extracting feature values from sentence confusion dimensions for the input clause dataset includes:
using a crust word segmentation device to segment each text D in the input paragraph data set D i Chinese word segmentation is carried out on each sentence of the number (1) to generate a word sequence, the length of the word sequence is recorded as N, the word element is recorded as WD, and the current sentence is recorded as S= { WD 1 ,WD 2 ,...,WD N };
Calculating the confusion of each sentence by a statistical-based language model confusion algorithm PP (S), wherein the calculation formula of the PP (S) is as followsWherein S represents the current sentence; n word sequence length; p (w) i ) A probability representing an i-th word; p (w) i |w 1 w 2 w 3 …w i-1 ) Representing the probability of calculating the ith word based on the previous i-1 words;
calculation of input clause dataset S Using PP (S) 1 Each sentence s i Is the confusion p of (2) i Generating a sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T };
Obtaining the sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T Maximum degree of confusion p in } max If p max If the text is smaller than or equal to the second threshold value, judging that the text is created by the machine; otherwise, entering the next step;
arranging the elements of the sentence confusion degree set P from small to large according to the confusion degree value, and taking elements from 3n/4 to the tail to form a set Pa= { P 3T/4 ,...,p T -a }; calculating an arithmetic mean of the set PaConfusion of the text>
In this embodiment, the step of extracting feature values from the statement presentation dispersion dimension for the input phrase dataset includes:
for the input clause dataset S 1 Each clause s of (a) i Calculating character length l i Generating a clause length set L s ={l 1 ,l 2 ,...,l T };
Calculating the set of clause lengths L s Is of the dispersion of (2)Wherein->For the set L of clause lengths s An average of all clause character lengths;
marking the dispersion of text paragraphs as Q d ,Q d =q d
In this embodiment, the step of extracting feature values from the knowledge span dimension for the input clause dataset includes:
using a pre-trained knowledge classification model to classify the input clause data set S 1 Each clause s of (a) i Classifying, and the corresponding classification set is C s ={c 1 ,.c 2 ,...c r },Wherein r is less than or equal to T;
For class set C s The element in the mixture is subjected to de-duplication treatment to obtain C d ={c 1 ,.c 2 ,...c R R.ltoreq.r;
calculating knowledge span K of text passage d =card(C d ) Wherein card (C d ) Representing set C d The number of elements in the list.
In this embodiment, the step of extracting feature values from the topic cohesiveness dimension for the input clause dataset includes:
using word2vec model to input clause data set S 1 Each clause s of (a) i Calculating the semantic similarity of the sentence with other sentences;
judging sentences with semantic similarity larger than a third threshold as a theme;
calculating the input clause data set S 1 The number k of the topics in the list;
calculating topic cohesiveness T of text paragraphs d =e 1-k Wherein k is>0。
In this embodiment, the step of extracting a feature value from the emotion richness dimension for the input clause dataset includes:
calculating character richness, representing all characters except Chinese characters and English letters as a set Then calculate the target segment word d i Contains a set F 1 The number of characters in (a) is recorded as N f1
Calculating the richness of the language-gas assisted words and setting a language-gas assisted word set F 2 = { A, o, la, w, o, bar, have, ou, if, mock, ou, ha, have, no }, calculate the target segment word d i Contains a set F 2 The number of the auxiliary words is recorded as N f2
Calculating emotion richness E of text paragraph d =αN f1 +βN f2 Wherein alpha and beta are weight factors.
In this embodiment, the step of extracting the feature value from the non-civilized term dimension for the input clause dataset includes:
based on rule engine and knowledge base for said input clause data set S 1 Each clause s of (a) i Performing non-civilization calculation, and judging the target segment text d when any clause accords with the non-civilization rule i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Based on rule engine and knowledge base for said input clause data set S 1 Each clause s of (a) i Performing reverse blacking or quoting calculation on positive characters or events, and judging the target segment text d when any clause accords with the definition of a rule engine i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Non-civilized expression of a text paragraph is denoted as D d ,D d Is f 1 And f 2 Of (f) 1 And f 2 One of them is 1, D d =1, otherwise D d =0。
In this embodiment, the data normalization processing is performed on the feature values extracted in each dimension, and the feature values extracted in each dimension are mapped into the same interval (0, 1) to form a normalized feature set x= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 ) The method comprises the following steps:
Obtaining the input clause data set S 1 From sentence confusion (Ppl), sentence presentation dispersion (QTd), knowledge span (knowledgespan), topic cohesiveness (Theme), emotion richness (Emotio)n) and the feature values respectively extracted from each dimension of the non-civilized term (Deviation from social core values) to form a feature dimension X' = (P) d ,Q d ,K d ,T d ,E d ,D d );
Using the formulaNormalizing the feature values extracted from each dimension, wherein x is min For the minimum value in the corresponding feature set, x max For the maximum value in the corresponding characteristic value set, x is the corresponding current characteristic value, x norm Normalizing the processed value for the corresponding current characteristic value;
after processing, mapping each dimension characteristic value in a (0, 1) interval to form a normalized characteristic group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 )。
For specific limitations regarding implementation steps when the processor executes the computer program, reference may be made to the above limitation on the method for intelligently writing the reverse identification AI, which is not described herein.
In one embodiment, a computer readable storage medium is provided having a computer program stored thereon, which when executed by a processor, performs the steps of:
segmenting the text to form an input paragraph data set D= { D composed of multiple segments of characters 1 ,d 2 ,d 3 ,...,d i ,...,d m };
Initializing the input paragraph data set D, and dividing each paragraph of characters according to punctuation marks; word d when the target segment i When the number t of the clauses of the (2) is larger than or equal to a first threshold value, removing the common high-frequency sentence to form an input clause data set S formed by a plurality of clauses 1 ={s 1 ,s 2 ,s 3 ,...,s i ,...,s T -wherein T is less than or equal to T;
for the input clause dataset S 1 From sentence confusion (Ppl), sentence presentation dispersion (QTd), knowledge span (knowledgespan), topicsExtracting feature values from each dimension of the aggregation (Theme), emotion richness (Emotion) and non-civilization term (Deviation from social core values);
carrying out data normalization processing on the feature values extracted by each dimension, mapping the feature values extracted by each dimension into the same interval (0, 1) to form a normalized feature group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 );
Optimizing a support vector machine classifier by utilizing the normalized feature set X, and finding out the most suitable kernel function through multiple rounds of training to obtain a classification model;
and inputting the text to be identified into the classification model, and classifying each text segment of the text to be identified as human authoring or machine authoring.
In this embodiment, after the initializing the input paragraph data set, the sentence dividing step is performed on each paragraph of text according to punctuation marks, the method further includes: and when the number of the clauses of the target segment text is smaller than a first threshold value, whether the target segment text is machine creation is not judged, and otherwise, the next step is executed.
In this embodiment, the step of extracting feature values from sentence confusion dimensions for the input clause dataset includes:
using a crust word segmentation device to segment each text D in the input paragraph data set D i Chinese word segmentation is carried out on each sentence of the number (1) to generate a word sequence, the length of the word sequence is recorded as N, the word element is recorded as WD, and the current sentence is recorded as S= { WD 1 ,WD 2 ,...,WD N };
Calculating the confusion of each sentence by a statistical-based language model confusion algorithm PP (S), wherein the calculation formula of the PP (S) is as followsWherein S represents the current sentence; n word sequence length; p (w) i ) A probability representing an i-th word; p (w) i |w 1 w 2 w 3 …w i-1 ) Representing the probability of calculating the ith word based on the previous i-1 words;
by PP (S) calculating the input clause dataset S 1 Each sentence s i Is the confusion p of (2) i Generating a sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T };
Obtaining the sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T Maximum degree of confusion p in } max If p max If the text is smaller than or equal to the second threshold value, judging that the text is created by the machine; otherwise, entering the next step;
arranging the elements of the sentence confusion degree set P from small to large according to the confusion degree value, and taking the elements from 3T/4 to the tail to form a set Pa= { P 3T/4 ,...,p T -a }; calculating an arithmetic mean of the set PaConfusion of the text >/>
In this embodiment, the step of extracting feature values from the statement presentation dispersion dimension for the input phrase dataset includes:
for the input clause dataset S 1 Each clause s of (a) i Calculating character length l i Generating a clause length set L s ={l 1 ,l 2 ,...,l T };
Calculating the set of clause lengths L s Is of the dispersion of (2)Wherein->For the set L of clause lengths s An average of all clause character lengths;
marking the dispersion of text paragraphs as Q d ,Q d =q d
In this embodiment, the step of extracting feature values from the knowledge span dimension for the input clause dataset includes:
using a pre-trained knowledge classification model to classify the input clause data set S 1 Each clause s of (a) i Classifying, and the corresponding classification set is C s ={c 1 ,.c 2 ,...c r -wherein r.ltoreq.T;
for class set C s The element in the mixture is subjected to de-duplication treatment to obtain C d ={c 1 ,.c 2 ,...c R R.ltoreq.r;
calculating knowledge span K of text passage d =card(C d ) Wherein card (C d ) Representing set C d The number of elements in the list.
In this embodiment, the step of extracting feature values from the topic cohesiveness dimension for the input clause dataset includes:
using word2vec model to input clause data set S 1 Each clause s of (a) i Calculating the semantic similarity of the sentence with other sentences;
Judging sentences with semantic similarity larger than a third threshold as a theme;
calculating the input clause data set S 1 The number k of the topics in the list;
calculating topic cohesiveness T of text paragraphs d =e 1-k Wherein k is>0。
In this embodiment, the step of extracting a feature value from the emotion richness dimension for the input clause dataset includes:
calculating character richness, representing all characters except Chinese characters and English letters as a set Then calculate the target segment word d i Contains a set F 1 The number of characters in (a) is recorded as N f1
Calculating the richness of the language-gas assisted words and setting a language-gas assisted word set F 2 = { A, o, la, w, o, bar, have, ou, if, mock, ou, ha, have, no }, calculate the target segment word d i Contains a set F 2 The number of the auxiliary words is recorded as N f2
Calculating emotion richness E of text paragraph d =αN f1 +βN f2 Wherein alpha and beta are weight factors.
Extracting feature values from the non-civilized term dimension for the input clause dataset, comprising:
based on rule engine and knowledge base for said input clause data set S 1 Each clause s of (a) i Performing non-civilization calculation, and judging the target segment text d when any clause accords with the non-civilization rule i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Based on rule engine and knowledge base for said input clause data set S 1 Each clause s of (a) i Performing reverse blacking or quoting calculation on positive characters or events, and judging the target segment text d when any clause accords with the definition of a rule engine i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Non-civilized expression of a text paragraph is denoted as D d ,D d Is f 1 And f 2 Of (f) 1 And f 2 One of them is 1, D d =1, otherwise D d =0。
The method for reverse identification AI intelligent authoring according to claim 1, wherein the feature values extracted from each dimension are subjected to data normalization processing, and the feature values extracted from each dimension are mapped into the same interval (0, 1) to formNormalized feature set x= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 ) The method comprises the following steps:
obtaining the input clause data set S 1 Feature values respectively extracted from each dimension of sentence confusion (Ppl), sentence presentation dispersion (QTd), knowledge span (knowledgespan), topic cohesiveness (Theme), emotion richness (motion) and non-civilization term (Deviation from social core values) to form feature dimension X' = (P) d ,Q d ,K d ,T d ,E d ,D d );
Using the formulaNormalizing the feature values extracted from each dimension, wherein x is min For the minimum value in the corresponding feature set, x max For the maximum value in the corresponding characteristic value set, x is the corresponding current characteristic value, x norm Normalizing the processed value for the corresponding current characteristic value;
after processing, mapping each dimension characteristic value in a (0, 1) interval to form a normalized characteristic group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 )。
For specific limitations regarding implementation steps of the computer program when executed by the processor, reference may be made to the above limitation of the method for intelligent writing of reverse identification AI, which is not described in detail herein.
Those skilled in the art will appreciate that implementing all or part of the above described methods may be accomplished by way of a computer program stored on a non-transitory computer readable storage medium, which when executed, may comprise the steps of the embodiments of the methods described above. Any reference to memory, storage, database, or other medium used in the various embodiments provided herein may include non-volatile and/or volatile memory. The nonvolatile memory can include Read Only Memory (ROM), programmable ROM (PROM), electrically Programmable ROM (EPROM), electrically Erasable Programmable ROM (EEPROM), or flash memory. Volatile memory can include Random Access Memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in a variety of forms such as Static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double Data Rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous Link DRAM (SLDRAM), memory bus direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), among others.
The technical features of the above embodiments may be arbitrarily combined, and all possible combinations of the technical features in the above embodiments are not described for brevity of description, however, as long as there is no contradiction between the combinations of the technical features, they should be considered as the scope of the description.
The above examples merely represent a few embodiments of the present application, which are described in more detail and are not to be construed as limiting the scope of the invention. It should be noted that it would be apparent to those skilled in the art that various modifications and improvements could be made without departing from the spirit of the present application, which would be within the scope of the present application. Accordingly, the scope of protection of the present application is to be determined by the claims appended hereto.

Claims (10)

1. A method for reverse identification AI intelligent authoring, comprising:
segmenting the text to form an input paragraph data set D= { D composed of multiple segments of characters 1 ,d 2 ,d 3 ,...,d i ,...,d m };
Initializing the input paragraph data set D, and dividing each paragraph of characters according to punctuation marks; word d when the target segment i When the number t of the clauses of the (2) is larger than or equal to a first threshold value, removing the common high-frequency sentence to form an input clause data set S formed by a plurality of clauses 1 ={s 1 ,s 2 ,s 3 ,...,s i ,...,s T -wherein T is less than or equal to T;
for the input clause dataset S 1 Extracting characteristic values from each dimension of sentence confusion, sentence presentation dispersion, knowledge span, theme cohesiveness, emotion richness and non-civilized expression respectively;
carrying out data normalization processing on the feature values extracted by each dimension, mapping the feature values extracted by each dimension into the same interval (0, 1) to form a normalized feature group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 );
Optimizing a support vector machine classifier by utilizing the normalized feature set X, and finding out the most suitable kernel function through multiple rounds of training to obtain a classification model;
inputting the text to be identified into the classification model, and classifying each text segment of the text to be identified into human authoring or machine authoring;
the step of extracting the feature value from the sentence confusion degree dimension of the input clause data set comprises the following steps:
using a crust word segmentation device to segment each text D in the input paragraph data set D i Chinese word segmentation is carried out on each sentence of the number (1) to generate a word sequence, the length of the word sequence is recorded as N, the word element is recorded as WD, and the current sentence is recorded as S= { WD 1 ,WD 2 ,...,WD N };
Calculating the confusion of each sentence by a statistical-based language model confusion algorithm PP (S), wherein the calculation formula of the PP (S) is as followsWherein S represents the current sentence; n word sequence length; p (w) i ) A probability representing an i-th word; p (w) i |w 1 w 2 w 3 …w i-1 ) Representing the probability of calculating the ith word based on the previous i-1 words;
calculation of input clause dataset S Using PP (S) 1 Each sentence s i Is the confusion p of (2) i Generating a sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T };
Obtaining the sentence confusion degree setP={p 1 ,p 2 ,p 3 ,...,p T Maximum degree of confusion p in } max If p max If the text is smaller than or equal to the second threshold value, judging that the text is created by the machine; otherwise, entering the next step;
arranging the elements of the sentence confusion degree set P from small to large according to the confusion degree value, and taking the elements from 3T/4 to the tail to form the set P a ={p 3T/4 ,...,p T -a }; calculating an arithmetic mean of the set PaConfusion of the text
Wherein the step of extracting feature values from a statement presentation dispersion dimension for the input clause dataset comprises: for the input clause dataset S 1 Each clause s of (a) i Calculating character length l i Generating a clause length set L s ={l 1 ,l 2 ,...,l T };
Calculating the set of clause lengths L s Is of the dispersion of (2)Wherein->For the set L of clause lengths s An average of all clause character lengths;
marking the dispersion of text paragraphs as Q d ,Q d =q d
2. The method for reverse-identification AI-intelligent authoring according to claim 1, further comprising, after said initializing said input paragraph dataset, a sentence-dividing step for each paragraph of text according to punctuation marks: and when the number of the clauses of the target segment text is smaller than a first threshold value, whether the target segment text is machine creation is not judged, and otherwise, the next step is executed.
3. The method of reverse recognition AI-smart authoring of claim 1, wherein extracting feature values from a knowledge span dimension for said input clause dataset comprises:
using a pre-trained knowledge classification model to classify the input clause data set S 1 Each clause s of (a) i Classifying, and the corresponding classification set is C s ={c 1 ,.c 2 ,...c r -wherein r.ltoreq.T;
for class set C s The element in the mixture is subjected to de-duplication treatment to obtain C d ={c 1 ,.c 2 ,...c R R.ltoreq.r;
calculating knowledge span K of text passage d =card(C d ) Wherein card (C d ) Representing set C d The number of elements in the list.
4. The method of reverse recognition AI-smart authoring of claim 1, wherein extracting feature values from a topic-cohesiveness dimension for said input clause dataset comprises:
using word2vec model to input clause data set S 1 Each clause s of (a) i Calculating the semantic similarity of the sentence with other sentences;
judging sentences with semantic similarity larger than a third threshold as a theme;
calculating the input clause data set S 1 The number k of the topics in the list;
calculating topic cohesiveness T of text paragraphs d =e 1-k Wherein k is>0。
5. The method of reverse recognition AI-smart authoring of claim 1, wherein extracting feature values from emotion-rich dimensions for said input clause dataset comprises:
Calculating character richness, except for Chinese characters and EnglishAll characters other than letters, denoted as set F 1 The method comprises the steps of carrying out a first treatment on the surface of the Then calculate the target segment word d i Contains a set F 1 The number of characters in (a) is recorded as N f1
Calculating the richness of the language-gas assisted words and setting a language-gas assisted word set F 2 = { A, o, la, w, o, bar, have, ou, if, mock, ou, ha, have, no }, calculate the target segment word d i Contains a set F 2 The number of the auxiliary words is recorded as N f2
Calculating emotion richness E of text paragraph d =αN f1 +βN f2 Wherein alpha and beta are weight factors.
6. The method of reverse recognition AI-smart authoring of claim 1, wherein extracting feature values from a non-textual term dimension for said input clause dataset comprises:
based on rule engine and knowledge base for said input clause data set S 1 Each clause s of (a) i Performing non-civilization calculation, and judging the target segment text d when any clause accords with the non-civilization rule i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Based on rule engine and knowledge base for said input clause data set S 1 Each clause s of (a) i Performing reverse blacking or quoting calculation on positive characters or events, and judging the target segment text d when any clause accords with the definition of a rule engine i The non-civilized value of (2) is marked as 1, otherwise marked as 0; recording device
Non-civilized expression of a text paragraph is denoted as D d ,D d Is f 1 And f 2 Of (f) 1 And f 2 One of them is 1, D d =1, otherwise D d =0。
7. The method for reverse identification AI intelligent authoring according to claim 1, wherein the data normalization processing is performed on the feature values extracted from each dimension, and the feature values extracted from each dimension are mapped into the same interval (0, 1) to form a normalized feature group x= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 ) The method comprises the following steps:
obtaining the input clause data set S 1 Feature values respectively extracted from each dimension of sentence confusion, sentence presentation dispersion, knowledge span, topic cohesiveness, emotion richness and non-civilized expression to form feature dimension X' = (P) d ,Q d ,K d ,T d ,E d ,D d );
Using the formulaNormalizing the feature values extracted from each dimension, wherein x is min For the minimum value in the corresponding feature set, x max For the maximum value in the corresponding characteristic value set, x is the corresponding current characteristic value, x norm Normalizing the processed value for the corresponding current characteristic value;
after processing, mapping each dimension characteristic value in a (0, 1) interval to form a normalized characteristic group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 )。
8. An intelligent reverse identification AI writing device, comprising:
The segmentation processing module is used for carrying out segmentation processing on the text to form an input paragraph data set D= { D formed by a plurality of sections of characters 1 ,d 2 ,d 3 ,...,d i ,...,d m };
The clause processing module is used for initializing the input paragraph data set D and carrying out clause on each paragraph of characters according to punctuation marks; word d when the target segment i When the number t of the clauses of the (2) is larger than or equal to a first threshold value, removing the common high-frequency sentence to form an input clause data set S formed by a plurality of clauses 1 ={s 1 ,s 2 ,s 3 ,...,s i ,...,s T -wherein T is less than or equal to T;
a module for extracting characteristic values for the input clause data set S 1 Extracting characteristic values from each dimension of sentence confusion, sentence presentation dispersion, knowledge span, theme cohesiveness, emotion richness and non-civilized expression respectively; the step of extracting the feature value from the sentence confusion degree dimension of the input clause data set comprises the following steps: using a crust word segmentation device to segment each text D in the input paragraph data set D i Chinese word segmentation is carried out on each sentence of the number (1) to generate a word sequence, the length of the word sequence is recorded as N, the word element is recorded as WD, and the current sentence is recorded as S= { WD 1 ,WD 2 ,...,WD N -a }; calculating the confusion of each sentence by a statistical-based language model confusion algorithm PP (S), wherein the calculation formula of the PP (S) is as follows Wherein S represents the current sentence; n word sequence length; p (w) i ) A probability representing an i-th word; p (w) i |w 1 w 2 w 3 …w i-1 ) Representing the probability of calculating the ith word based on the previous i-1 words; calculation of input clause dataset S Using PP (S) 1 Each sentence s i Is the confusion p of (2) i Generating a sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T -a }; obtaining the sentence confusion degree set P= { P 1 ,p 2 ,p 3 ,...,p T Maximum degree of confusion p in } max If p max If the text is smaller than or equal to the second threshold value, judging that the text is created by the machine; otherwise, entering the next step; arranging the elements of the sentence confusion degree set P from small to large according to the confusion degree value, and taking the elements from 3T/4 to the tail to form the set P a ={p 3T/4 ,...,p T -a }; calculating the arithmetic mean of set Pa>Confusion of the text>Wherein the step of extracting feature values from a statement presentation dispersion dimension for the input clause dataset comprises: for the input clause dataset S 1 Each clause s of (a) i Calculating character length l i Generating a clause length set L s ={l 1 ,l 2 ,...,l T -a }; calculating the set of clause lengths L s Is>Wherein->For the set L of clause lengths s An average of all clause character lengths; marking the dispersion of text paragraphs as Q d ,Q d =q d
The data normalization processing module is used for performing data normalization processing on the feature values extracted by each dimension, mapping the feature values extracted by each dimension into the same interval (0, 1) to form a normalized feature group X= (X) 1 ,x 2 ,x 3 ,x 4 ,x 5 ,x 6 );
The classification model acquisition module is used for optimizing the support vector machine classifier by utilizing the normalized feature set X, and finding out the most suitable kernel function through multiple rounds of training to obtain a classification model;
and the text recognition module is used for inputting the text to be recognized into the classification model, and classifying each text segment of the text to be recognized into human authoring or machine authoring.
9. A computer device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, characterized in that the processor implements the steps of the method according to any one of claims 1 to 7 when the computer program is executed by the processor.
10. A computer readable storage medium, on which a computer program is stored, characterized in that the computer program, when being executed by a processor, implements the steps of the method of any of claims 1 to 7.
CN202310108714.3A 2023-02-14 2023-02-14 Method, device, equipment and medium for reverse identification AI intelligent writing Active CN116384388B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310108714.3A CN116384388B (en) 2023-02-14 2023-02-14 Method, device, equipment and medium for reverse identification AI intelligent writing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310108714.3A CN116384388B (en) 2023-02-14 2023-02-14 Method, device, equipment and medium for reverse identification AI intelligent writing

Publications (2)

Publication Number Publication Date
CN116384388A CN116384388A (en) 2023-07-04
CN116384388B true CN116384388B (en) 2024-02-02

Family

ID=86973883

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310108714.3A Active CN116384388B (en) 2023-02-14 2023-02-14 Method, device, equipment and medium for reverse identification AI intelligent writing

Country Status (1)

Country Link
CN (1) CN116384388B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117113977B (en) * 2023-10-09 2024-04-16 北京信诺软通信息技术有限公司 Method, medium and system for identifying text generated by AI contained in test paper

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107153664A (en) * 2016-03-04 2017-09-12 同方知网(北京)技术有限公司 A kind of method flow that research conclusion is simplified based on the scientific and technical literature mark that assemblage characteristic is weighted
CN109992769A (en) * 2018-12-06 2019-07-09 平安科技(深圳)有限公司 Sentence reasonability judgment method, device, computer equipment based on semanteme parsing
CN111737475A (en) * 2020-07-21 2020-10-02 南京擎盾信息科技有限公司 Unsupervised network public opinion spam long text recognition method
CN112580351A (en) * 2020-12-31 2021-03-30 成都信息工程大学 Machine-generated text detection method based on self-information loss compensation
CN115081437A (en) * 2022-07-20 2022-09-20 中国电子科技集团公司第三十研究所 Machine-generated text detection method and system based on linguistic feature contrast learning
CN115630640A (en) * 2022-12-23 2023-01-20 苏州浪潮智能科技有限公司 Intelligent writing method, device, equipment and medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2009049262A1 (en) * 2007-10-11 2009-04-16 Honda Motor Co., Ltd. Text categorization with knowledge transfer from heterogeneous datasets

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107153664A (en) * 2016-03-04 2017-09-12 同方知网(北京)技术有限公司 A kind of method flow that research conclusion is simplified based on the scientific and technical literature mark that assemblage characteristic is weighted
CN109992769A (en) * 2018-12-06 2019-07-09 平安科技(深圳)有限公司 Sentence reasonability judgment method, device, computer equipment based on semanteme parsing
CN111737475A (en) * 2020-07-21 2020-10-02 南京擎盾信息科技有限公司 Unsupervised network public opinion spam long text recognition method
CN112580351A (en) * 2020-12-31 2021-03-30 成都信息工程大学 Machine-generated text detection method based on self-information loss compensation
CN115081437A (en) * 2022-07-20 2022-09-20 中国电子科技集团公司第三十研究所 Machine-generated text detection method and system based on linguistic feature contrast learning
CN115630640A (en) * 2022-12-23 2023-01-20 苏州浪潮智能科技有限公司 Intelligent writing method, device, equipment and medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
面向词权重的主题识别应用研究;阮光册等;《情报理论与实践》;第42卷(第12期);第144-149页 *

Also Published As

Publication number Publication date
CN116384388A (en) 2023-07-04

Similar Documents

Publication Publication Date Title
CN112163426B (en) Relationship extraction method based on combination of attention mechanism and graph long-time memory neural network
CN108399158B (en) Attribute emotion classification method based on dependency tree and attention mechanism
Boenninghoff et al. Similarity learning for authorship verification in social media
CN111914558A (en) Course knowledge relation extraction method and system based on sentence bag attention remote supervision
CN110263325B (en) Chinese word segmentation system
CN106991085B (en) Entity abbreviation generation method and device
CN110502742B (en) Complex entity extraction method, device, medium and system
CN112711953A (en) Text multi-label classification method and system based on attention mechanism and GCN
CN110097096B (en) Text classification method based on TF-IDF matrix and capsule network
CN112052684A (en) Named entity identification method, device, equipment and storage medium for power metering
CN109446423B (en) System and method for judging sentiment of news and texts
CN116384388B (en) Method, device, equipment and medium for reverse identification AI intelligent writing
CN112417153A (en) Text classification method and device, terminal equipment and readable storage medium
CN112100377A (en) Text classification method and device, computer equipment and storage medium
CN111507093A (en) Text attack method and device based on similar dictionary and storage medium
CN112380346B (en) Financial news emotion analysis method and device, computer equipment and storage medium
CN111680132B (en) Noise filtering and automatic classifying method for Internet text information
CN111611796A (en) Hypernym determination method and device for hyponym, electronic device and storage medium
CN116257601A (en) Illegal word stock construction method and system based on deep learning
CN116245107A (en) Electric power audit text entity identification method, device, equipment and storage medium
CN111191033A (en) Open set classification method based on classification utility
CN116089605A (en) Text emotion analysis method based on transfer learning and improved word bag model
CN115599915A (en) Long text classification method based on TextRank and attention mechanism
CN114648029A (en) Electric power field named entity identification method based on BiLSTM-CRF model
CN111199154B (en) Fault-tolerant rough set-based polysemous word expression method, system and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20240111

Address after: Room W-3984, No. 559 Yueluo Road, Baoshan District, Shanghai, 2019

Applicant after: Shanghai Xijin Information Technology Co.,Ltd.

Address before: E1-031, artificial intelligence Industrial Park, No. 88, Jinjihu Avenue, Suzhou Industrial Park, Jiangsu 215000

Applicant before: Suzhou Xinsi Technology Co.,Ltd.

TA01 Transfer of patent application right
GR01 Patent grant
GR01 Patent grant