CN110110318A - Text Stego-detection method and system based on Recognition with Recurrent Neural Network - Google Patents

Text Stego-detection method and system based on Recognition with Recurrent Neural Network Download PDF

Info

Publication number
CN110110318A
CN110110318A CN201910058680.5A CN201910058680A CN110110318A CN 110110318 A CN110110318 A CN 110110318A CN 201910058680 A CN201910058680 A CN 201910058680A CN 110110318 A CN110110318 A CN 110110318A
Authority
CN
China
Prior art keywords
text
neural network
recognition
recurrent neural
detected
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910058680.5A
Other languages
Chinese (zh)
Other versions
CN110110318B (en
Inventor
黄永峰
杨忠良
王颗
杨震
胡雨婷
武楚涵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN201910058680.5A priority Critical patent/CN110110318B/en
Publication of CN110110318A publication Critical patent/CN110110318A/en
Application granted granted Critical
Publication of CN110110318B publication Critical patent/CN110110318B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24133Distances to prototypes
    • G06F18/24137Distances to cluster centroïds
    • G06F18/2414Smoothing the distance, e.g. radial basis function networks [RBFN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/216Parsing using statistical methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The text Stego-detection method and system based on Recognition with Recurrent Neural Network that the invention discloses a kind of, wherein this method comprises: obtaining term vector matrix, input word sequence vector is converted for text to be detected according to term vector matrix;Input word sequence vector is input in the Recognition with Recurrent Neural Network model constructed in advance, the feature vector of correlativity between indicating text word to be detected is generated;Classified by classifier to feature vector, judges whether text to be detected includes hiding information;If text to be detected includes hiding information, the information insertion rate of text to be detected is estimated according to the difference of steganography Text eigenvector under different insertion rates.This method applies to Recognition with Recurrent Neural Network in text Stego-detection, can effectively identify whether text carrier contains hiding information, and the capacity of hiding information is accurately estimated according to the statistical distribution of extraction feature.

Description

Text Stego-detection method and system based on Recognition with Recurrent Neural Network
Technical field
The present invention relates to text information field of communication technology, in particular to a kind of text steganography based on Recognition with Recurrent Neural Network Detection method and system.
Background technique
In the monograph about information security, Shannon summarizes three basic information safety systems: encryption system, Privacy system and hiding system.
1, encryption system in a special manner encodes information, to only have authorized party that could be decoded to it, and Unwarranted side can be decoded it.It is by making message be difficult to read safety to ensure information.
2, privacy system is mainly the access limited to information, to only have authorized user that could access important information, not Authorized user can not access it in any way under any circumstance.However, although the two systems ensure information Safety, but they also expose the presence and importance of information, make it easier under attack, such as intercept and crack.
3, it hides system and the two secrecy systems is very different.It by being embedded in confidential information in common vector, Then it is transmitted using common signal channel, the presence of Hiding Secret information, to achieve the purpose that be not easy under a cloud and attack.
Wherein, Steganography is the key technology in system of hiding, and has the carrier of various media formats to can be used for Information hiding, Including image, audio, text etc..Text attracts as information carrier most popular in daily life, text Steganography The interest of numerous studies personnel.In recent years, more and more text based information concealing methods have already appeared.
J.Fridrich concludes that in general, it is embedding to determine that steganographic algorithm can use three kinds of different basic frameworks Enter the internal mechanism with extraction algorithm: being retrieved by carrier, the Steganography of carrier modification and carrier generation.
1, in the Steganography retrieved by carrier, all carriers concentrated first to carrier are encoded, and are then selected Different carriers is transmitted to realize hidden message transmission.The advantages of this method is carrier always " 100% is natural ", but bright Aobvious disadvantage is that transmitting information content is considerably less.
2, the steganography method of current most study is the Steganography based on carrier modification.By modifying given carrier come real The insertion of existing confidential information.This method has a wide range of applications on multiple carriers, such as image, voice and text.But it is total For, the redundancy space of image and voice is relatively large, therefore suitable modification not will cause very big visual effect Or sense of hearing variation.For text, it has higher information coding degree, and information redundancy amount is less, and space can be modified by limiting Size.Therefore, the method based on modification is difficult to realize sufficiently high hidden capacity on text carrier.
3, the third method is to generate to carry out steganography by carrier.The confidential information transmitted as needed automatically generates load Body, and hidden information is embedded in generating process.This method hidden capacity usually with higher, however, pervious model It is difficult to generate the readable text of high quality, therefore the concealment of such methods is limited.With deep learning (deep Learning) in the extensive use of natural language processing field (natural language processing, abbreviation NLP), out Many text generation formula steganography methods based on deep learning are showed.By the powerful feature learning of neural network and characterization energy The readable text of high quality can be generated in power, and by coding insertion hiding information in generating process, so that generating The text with hiding information be provided simultaneously with highly concealed type and high insertion rate.
However, above-mentioned various text steganography methods, such as utilized by terrorist, hacker criminal, for transmitting The hidden information of the public may be endangered, potential threat will be caused to public safety.This potential threat is fought, needs to identify text In whether include hiding information, key technology is text Stego-detection.
Traditional text Stego-detection technology, is typically based on such as lower frame: firstly, extracting training set by specific method Some statistical natures of middle steganography text and plain text;The above-mentioned statistical nature for analyzing steganography text and plain text again exists Difference;Discriminator is finally designed based on such difference, for determining whether given text includes hiding information.However it passes The text Stego-detection technology of system, some artificially specified plain text features of analysis and utilization, these features be easy to by Deep neural network with powerful feature learning and ability to express is imitated, to generate the steganography for meeting these feature distributions Text.Therefore, for these conventional methods when facing the newest production text steganographic algorithm based on deep learning, detection is accurate Rate is greatly reduced.This kind of powerful production text steganographic algorithm, is such as utilized by criminal, will pacify for cyberspace and the public Potential threat is brought entirely.Therefore, identify production steganography text and have become the urgent problem to be solved for being related to public safety.
Summary of the invention
The present invention is directed to solve at least some of the technical problems in related technologies.
For this purpose, an object of the present invention is to provide a kind of text Stego-detection method based on Recognition with Recurrent Neural Network, This method applies to Recognition with Recurrent Neural Network in text Stego-detection, can effectively identify whether text carrier contains hiding information, And the capacity of hiding information is accurately estimated according to the statistical distribution of extraction feature.
It is another object of the present invention to propose a kind of text Stego-detection system based on Recognition with Recurrent Neural Network.
In order to achieve the above objectives, one aspect of the present invention embodiment proposes a kind of text steganography based on Recognition with Recurrent Neural Network Detection method, comprising: obtain term vector matrix, input term vector sequence is converted for text to be detected according to the term vector matrix Column;The input word sequence vector is input in the Recognition with Recurrent Neural Network model constructed in advance, generating indicates described to be detected The feature vector of correlativity between text word;Classified by classifier to described eigenvector, is judged described to be detected Whether text includes hiding information;If the text to be detected includes hiding information, according to steganography text under different insertion rates The information insertion rate of the text to be detected is estimated in the difference of feature vector.
The text Stego-detection method based on Recognition with Recurrent Neural Network of the embodiment of the present invention, by transporting Recognition with Recurrent Neural Network It uses in text Stego-detection, can effectively identify whether text carrier contains hiding information, and according to the statistical of extraction feature Cloth accurately estimates the capacity of hiding information.It can effectively overcome and be lacked possessed by traditional Stego-detection method based on simple feature It falls into, can also obtain very high detection accuracy in face of production text steganography method neural network based;Based on LSTM (Long Short-Term Memory, shot and long term memory network) statistical nature of correlativity between model extraction word, it can effectively model Long distance dependent relation in text;Based on the feature distribution difference of text under different insertion rates, text to be detected is accurately estimated Implicit information capacity.
In addition, the text Stego-detection method according to the above embodiment of the present invention based on Recognition with Recurrent Neural Network can also have There is following additional technical characteristic:
Further, in one embodiment of the invention, the acquisition term vector matrix, comprising: obtain a large amount of generate Text is segmented and is pre-processed to the generation text, and is filtered based on word frequency, and the common word finder for generating text is obtained It closes;Common vocabulary in the common lexical set is sequentially inputted in the module pre-established, each common word is obtained Converge corresponding term vector;Form term vector sample according to the corresponding term vector of each described common vocabulary, according to institute's predicate to It measures sample and generates the term vector matrix.
Further, in one embodiment of the invention, described to be turned text to be detected according to the term vector matrix Turn to input word sequence vector, comprising: the text to be detected is subjected to low-frequency word filtering, screening and participle, generate it is described to Detect the sequence of words of text;The term vector of all vocabulary in the sequence of words is obtained according to the term vector matrix, wherein The term vector matrix includes the corresponding relationship between common vocabulary and term vector;According to all vocabulary in the text to be detected Term vector, generate the input word sequence vector of the text to be detected.
Further, in one embodiment of the invention, the input word sequence vector is input in advance described Before in the Recognition with Recurrent Neural Network model of building, further includes: obtain the corresponding institute of generation text of multiple Given information insertion rates Input word sequence vector is stated, according to the corresponding input word sequence vector of the generation text of the multiple Given information insertion rate Construct the Recognition with Recurrent Neural Network model.
Further, in one embodiment of the invention, after the building Recognition with Recurrent Neural Network model, also It include: the character pair vector that the generation text of the Given information insertion rate is generated according to the Recognition with Recurrent Neural Network model; Classified using feature vector of the classifier to the generation text of the known insertion rate, generates the known insertion rate Generation text insertion rate estimated probability;Using preset algorithm to the generation text of the estimated probability and the known insertion rate This rate that is actually embedded in carries out analysis and generates analysis as a result, the Recognition with Recurrent Neural Network model according to the analysis modified result The weight parameter of parameter and the classifier.
In order to achieve the above objectives, it is hidden to propose a kind of text based on Recognition with Recurrent Neural Network for another aspect of the present invention embodiment Write detection system, comprising: module is obtained, for obtaining term vector matrix;First generation module, for according to the term vector square Text to be detected is converted input word sequence vector by battle array;Second generation module, for inputting the input word sequence vector Into the Recognition with Recurrent Neural Network model constructed in advance, generate the feature of correlativity between indicating the text word to be detected to Amount;Judgment module judges whether the text to be detected includes hidden for classifying by classifier to described eigenvector Hide information;Estimation module, for the text to be detected include hiding information, according to steganography text feature under different insertion rates to The information insertion rate of the text to be detected is estimated in the difference of amount.
The decentralization Internet of Things safety detecting system based on block chain of the embodiment of the present invention, by the way that nerve net will be recycled Network applies in text Stego-detection, can effectively identify whether text carrier contains hiding information, and according to the system of extraction feature Score cloth accurately estimates the capacity of hiding information.Can effectively it overcome possessed by traditional Stego-detection method based on simple feature Defect can also obtain very high detection accuracy in face of production text steganography method neural network based;Based on LSTM model The statistical nature of correlativity, can effectively model the long distance dependent relation in text between extraction word.
In addition, the decentralization Internet of Things safety detecting system according to the above embodiment of the present invention based on block chain may be used also With following additional technical characteristic:
Further, in one embodiment of the invention, the acquisition module, is specifically used for: obtaining a large amount of generation texts This, is segmented and is pre-processed to the generation text, and is filtered based on word frequency, and the common lexical set for generating text is obtained; Common vocabulary in the common lexical set is sequentially inputted in the module pre-established, each common vocabulary pair is obtained The term vector answered;Term vector sample is formed according to the corresponding term vector of each described common vocabulary, according to the term vector sample This generation term vector matrix.
Further, in one embodiment of the invention, first generation module, is specifically used for: will be described to be checked It surveys text and carries out low-frequency word filtering, screening and participle, generate the sequence of words of the text to be detected;According to the term vector square Battle array obtains the term vector of all vocabulary in the sequence of words, wherein the term vector matrix includes common vocabulary and term vector Between corresponding relationship;According to the term vector of all vocabulary in the text to be detected, the described of the text to be detected is generated Input word sequence vector.
Further, in one embodiment of the invention, further includes: building module;The building module, for obtaining The corresponding input word sequence vector of the generation text of multiple Given information insertion rates is embedded according to the multiple Given information The corresponding input word sequence vector of the generation text of rate constructs the Recognition with Recurrent Neural Network model.
Further, in one embodiment of the invention, further includes: correction module;
The correction module, is specifically used for: generating the Given information insertion rate according to the Recognition with Recurrent Neural Network model Generation text character pair vector;Using the classifier to the known insertion rate generation text feature vector into Row classification generates the insertion rate estimated probability of the generation text of the known insertion rate;It is general to the estimation using preset algorithm The rate that is actually embedded in of the generation text of rate and the known insertion rate carries out analysis generation analysis as a result, according to the analysis result Correct the parameter of the Recognition with Recurrent Neural Network model and the weight parameter of the classifier.
The additional aspect of the present invention and advantage will be set forth in part in the description, and will partially become from the following description Obviously, or practice through the invention is recognized.
Detailed description of the invention
Above-mentioned and/or additional aspect and advantage of the invention will become from the following description of the accompanying drawings of embodiments Obviously and it is readily appreciated that, in which:
Fig. 1 is the text Stego-detection method flow diagram based on Recognition with Recurrent Neural Network according to one embodiment of the invention;
Fig. 2 is to extract word association characteristic pattern using LSTM according to one embodiment of the invention;
Fig. 3 is special according to the potential association of one embodiment of the invention extracted between word using bidirectional circulating neural network Sign figure;
Fig. 4 is to be illustrated according to the text Stego-detection system structure based on Recognition with Recurrent Neural Network of one embodiment of the invention Figure.
Specific embodiment
The embodiment of the present invention is described below in detail, examples of the embodiments are shown in the accompanying drawings, wherein from beginning to end Same or similar label indicates same or similar element or element with the same or similar functions.Below with reference to attached The embodiment of figure description is exemplary, it is intended to is used to explain the present invention, and is not considered as limiting the invention.
The text Stego-detection based on Recognition with Recurrent Neural Network proposed according to embodiments of the present invention is described with reference to the accompanying drawings Method and system describe the text steganography based on Recognition with Recurrent Neural Network proposed according to embodiments of the present invention with reference to the accompanying drawings first Detection method.
Fig. 1 is the text Stego-detection method flow diagram based on Recognition with Recurrent Neural Network according to one embodiment of the invention.
As shown in Figure 1, should text Stego-detection method based on Recognition with Recurrent Neural Network the following steps are included:
In step s101, obtain term vector matrix, according to term vector matrix by text to be detected be converted into input word to Measure sequence.
Further, in one embodiment of the invention, term vector matrix is obtained, comprising: a large amount of generation texts are obtained, It is segmented and is pre-processed to text is generated, and filtered based on word frequency, obtain the common lexical set for generating text;By common word The common vocabulary collected in closing is sequentially inputted in the module pre-established, obtains the corresponding term vector of each common vocabulary; Term vector sample is formed according to the corresponding term vector of each common vocabulary, term vector matrix is generated according to term vector sample.
Further, in one embodiment of the invention, input is converted for text to be detected according to term vector matrix Term vector sequence, comprising: text to be detected is subjected to low-frequency word filtering, screening and participle, generates the vocabulary sequence of text to be detected Column;According to term vector matrix obtain sequence of words in all vocabulary term vector, wherein term vector matrix include common vocabulary and Corresponding relationship between term vector;According to the term vector of vocabulary all in text to be detected, the input word of text to be detected is generated Sequence vector.
Specifically, the text that a large amount of common types are collected on the net, natural text and special text including Chinese and English, with Based on existing text generation formula steganography method, the text of a large amount of different insertion rates, the training number as Stego-detection are generated According to collection.
In step s 102, input word sequence vector is input in the Recognition with Recurrent Neural Network model constructed in advance, is generated Indicate the feature vector of correlativity between text word to be detected.
Further, in one embodiment of the invention, input word sequence vector is being input to following of constructing in advance Before in ring neural network model, further includes: obtain the corresponding input term vector of generation text of multiple Given information insertion rates Sequence constructs Recognition with Recurrent Neural Network mould according to the corresponding input word sequence vector of the generation text of multiple Given information insertion rates Type.
Specifically, Recognition with Recurrent Neural Network model is formed by 1 to multiple hidden layers, and each hidden layer includes 1 to multiple LSTM (Long Short-Term Memory) unit, wherein LSTM unit can be it is unidirectional be also possible to it is two-way.
Further, in one embodiment of the invention, after constructing Recognition with Recurrent Neural Network model, further includes: root The character pair vector of the generation text of Given information insertion rate is generated according to Recognition with Recurrent Neural Network model;Using classifier to known The feature vector of the generation text of insertion rate is classified, and the insertion rate estimated probability of the generation text of known insertion rate is generated; Using preset algorithm to the generation text of estimated probability and known insertion rate be actually embedded in rate carry out analysis generate analysis as a result, The parameter of Recognition with Recurrent Neural Network model and the weight parameter of classifier are corrected based on the analysis results.
Optionally, preset algorithm can be back-propagation algorithm, and classifier can be softmax classifier.
Specifically, the word correlated characteristic of the above-mentioned natural text collected is modeled, and constructs corresponding circulation The classifier of neural network model and diagnostic characteristics uses the steganography text of a large amount of plain text and different insertion rates as instruction Practice sample, training neural network model and classifier, continue to optimize model parameter and performance using back-propagation algorithm, to improve The feature extraction ability and characteristic differentiation ability of model, at interval of the loss value of a period of time test model, according to loss value tune Integral mould Training strategy, such as learning rate, until neural network model parameter and performance reach stable.
In step s 103, classified by classifier to feature vector, judge whether text to be detected includes hiding Information.
In step S104, if text to be detected includes hiding information, according to steganography text feature under different insertion rates The information insertion rate of text to be detected is estimated in the difference of vector.
Feature distribution difference of the embodiment of the present invention based on text under different insertion rates, accurately estimates that text to be detected is hidden The information capacity contained;It can be for different types of text (Chinese and English text, natural text and special format text) unified structure Then established model frame uses the text training pattern of corresponding types, can be to different types of hidden as long as reaching a model The effect that text is detected is write, there is good scalability;It does not need to think design rule and feature, what model learnt All features can more accurately extract the word in text compared to conventional method from a large amount of existing production steganography texts Related linked character between language, to obtain accurate identification result.
Further, the method for the embodiment of the present invention is different from previous steganalysis method, utilizes Recognition with Recurrent Neural Network Modeling and feature extraction are carried out to text to be detected, identify text to be detected and whether contain hiding information, and can be based on taking out The capacity of the feature assessment hiding information taken.It is superior in the past from the precision of steganographic detection, recall rate, accuracy rate various aspects Correlation technique, reached current optimal detection performance, be of great significance in terms of the Stego-detection of production text.
In an embodiment of the present invention, identify production steganography by designing better model to extract reasonable feature Text, is described in detail below how the model constructed in the present embodiment identifies steganography text.
The Recognition with Recurrent Neural Network model of the present embodiment building, including two main modulars: textual association analysis module and spy Levy identification module.Textual association analysis module is characteristic extracting module, using Recognition with Recurrent Neural Network to the modeling energy of sequence signal Power and feature extraction ability extract the statistical distribution of the Relating Characteristic between the word in text;Characteristic differentiation module by pair The analysis of the above-mentioned feature extracted identifies whether text to be detected includes hiding information, and according to text under different insertion rates The nuance of feature estimates the hiding information capacity of text insertion to be detected.
One, textual association analysis module
For plain text, there are stronger correlativities between word.If being embedded in hiding information in word, phase Pass relationship may be weakened.Therefore, correlativity can be considered as the index of steganography, extract the linked character between word, can be with For Stego-detection.The current output of Recognition with Recurrent Neural Network is always related with previous input data, therefore can be used to extract The linked character of word.
Linked character between word can be defined as follows: T={ X1,X2,…,XmIndicate one section by the molecular text of m sentence This, wherein XiIndicate i-th of sentence.Each sentence is made of several words:Wherein niIt is The length of a sentence, xi,jIt is j-th of word of i-th of sentence.Assuming that being statistical iteration between all words, i.e., without appointing What correlativity, then following formula is set up:
P(xi,j,xk,l)=P (xi,j)·P(xk,l)
So, the case where violating above-mentioned formula is likely to imply there is association between word.Further, this method is by text Middle word association is substantially divided into three classes:
(1) adjacent word is associated with
One sentence Xi, a sequence signal can be modeled asFor such a by word The sequence signal of composition, adjacent word may include stronger semantic coherence and correlation as the signal continuously occurred. This kind of word association is defined as adjacent word association, is represented by P (xi,j,xk,l| i=k, | l-j |=1)
(2) across word association
Include the sentence of multiple words for one in view of syntactic rule, word may not only with adjacent word There is semantic association, also there are semantic associations with other farther away words of distance in same sentence.Long distance in this kind of same sentence It is defined as being represented by P (x across word association from word associationi,j,xk,l| i=k, 1 < | l-j | < ni)。
(3) across sentence association
More generally, for the text comprising multiple sentences, may also there be certain semantic association between different sentences.Cause This, even if two words, not in same sentence, still there may be potential incidence relations for they.In this kind of difference sentence Remote word association is defined as across sentence association, is represented by P (xi,j,xk,l|i≠ k,j∈[1,ni],l∈[1,nk])。
In textual association characteristic extraction procedure, the embodiment of the present invention mainly utilizes LSTM (Long Short-Term Memory, shot and long term memory network) great ability in terms of sequence signal feature extraction and expression and LSTM be to long range The modeling ability of dependence carries out the analysis and extraction of textual association feature.One LSTM unit is combined in the output of t moment Current input and current hidden layer state, and the input information at preceding t-1 moment is contained in hidden state, therefore the output of t moment It may be expressed as:
yt=fLSTM(xt|x1,x2,…,xt-1)
One sentence X comprising L word can be considered as the sequence signal that a length is L, wherein i-th of word xiThe input signal of moment i can be regarded as.The first part of the neural network model of this method is word coding module, by a word The dense vector that language is mapped to a d dimension semantic space indicates.Therefore for the sentence X of input, matrix can be expressed asWherein the i-th row indicates i-th of word in sentence, it may be assumed that
In general, Recognition with Recurrent Neural Network is made of multilayer, every layer has multiple LSTM units.N is used in the embodiment of the present inventionj To indicate j-th of hidden layer UjLSTM unit number, the unit of jth layer can indicate are as follows:
For first hidden layer, unitIt is x in the input at t-th of momenttThe weighted sum of middle element, it may be assumed that
WhereinWithThe weight and deviation for being.In t moment,Output valve be:
Vector can be usedIndicate the output in j-th of hidden layer at t-th of moment,In each list of elements Show the output valve of the unit in j-th of hidden layer at t-th of moment, it may be assumed that
Previous work shows that in a certain range the neural network number of plies is more in space, model extraction and expression characteristic Ability is stronger.Therefore, the embodiment of the present invention builds network model using multilayer LSTM unit.Adjacent hidden layer can pass through Transfer matrix connection.For example, the l layers of transfer matrix between (l+1) layer can be expressed as matrix
Each unit in first of hidden layer at t-th of momentInput be in preceding layer the output valve of unit plus Quan He, it may be assumed that
The output that l layers of t moment is as follows:
To sum up, the output at t-th of moment depends not only on current time xtInput vector, and depend on previous t-1 Vector in the unit at a moment.Therefore, the output of first of hidden layer at t-th of moment can be considered as the total of preceding t moment With, that is, t previous word { x1,x2,…,xtInformation fusion.
Textual association analysis module has the potentiality for modeling aforementioned three kinds of word associations.LSTM has the energy of memory history input The LSTM unit of power, first layer can directly remember adjacent previous word feature, and the LSTM unit of later layer is with previous The LSTM unit output of layer is input, can remember increasingly complex history feature.Therefore, LSTM can easily be modeled across the time Correlated characteristic.And adjacent word is associated with, is only across the time pass on different time scales across word association, across sentence association Connection, therefore these linked characters can be extracted by textual association analysis module.Above-mentioned textual association analysis module such as Fig. 2 institute Show.
Two, characteristic differentiation module
Model shown in Fig. 2 can only extract each word and the word correlation before it.It is every in order to further extract The potentially relevant property of a word and all words (word and subsequent word including front) nearby, the embodiment of the present invention is into one Step is added to a reversed RNN (Recurrent Neural Network, Recognition with Recurrent Neural Network), therefore, entire textual association Analysis module is made of a two-way RNN, as shown in Figure 3.?
In Fig. 3, using bidirectional circulating neural network (BiRNN) extract word between potential linked character (including with forward direction The relationship of word and relationship with backward word), also identify whether text to be detected includes hiding information using linked character. Difference between reversed RNN and forward direction RNN be the input of reversed RNN be from the last one word of text to first word, because This, forward direction RNN is more likely to extract the correlation between each word and word before, and reversed RNN is more focused on extraction each Correlation between word and subsequent word.
WithWithRespectively indicate the word association feature extracted by forward direction RNN and the word association extracted by reversed RNN Feature.In order to merge the two features, first the feature vector of their last moments is spliced, obtains vector Z:
Then a Fusion Features matrix is definedFeatures described above Z is merged in the following way:
Wherein h is characteristic vector F=[f1,f2,…,fh] dimension.In order to identify text to be detected with obtained feature F Whether this includes hiding information, defines the identification weight vectors C that length is h, the inner product of C and F is calculated, with amount of biasIt is added Afterwards, scalar output is obtainedIt is as follows:
And output is normalized into section [0,1] using sigmoid activation primitive S:
Final output are as follows:
Due to the memory function of LSTM, final output has contained the output information at first layer all moment.Therefore, most The value exported eventually reflects model and thinks the probability that text to be detected contains hiding information.By setting a detection threshold value, most Whole testing result is as follows:
That is, model attempts to provide a prediction label to input text (0 represents plain text, and 1 represents steganography text).
In addition, capacity of the also adjustable integral mould to estimate hiding information.After obtaining characteristic vector F, addition Softmax layers calculate the probability distribution of text hidden capacity.Prediction weight (PW) is defined as matrix That is:
Wherein n indicates the possible value type of insertion rate, and h is characterized the dimension of vector F.Use known matrix WPTo count Calculate score a possibility that text to be detected belongs to the steganography text of n class difference insertion rate, it may be assumed that
Wherein WPAnd bpIt is known weight matrix and deviation, weight matrix WPIn value reflect the weight of each feature in F The property wanted, the dimension of output vector Y are n.In order to estimate the probability distribution of text insertion rate to be detected, the method for the embodiment of the present invention With reference to the work of forefathers, softmax classifier is added to output layer to calculate the possibility probability of every kind of insertion rate:
piText as to be detected belongs to the probability of the i-th class insertion rate text.The insertion rate for remembering the i-th class insertion rate text is ri, it thus can estimate the insertion rate of text to be detected:
R=rk, wherein
All parameters of neural network, including each term vector require to obtain by training.The embodiment of the present invention follows Supervised learning frame updates network parameter using back-propagation algorithm, minimizes loss by the iteration optimization of network Function, to obtain optimal models.Wherein, loss function consists of two parts, and a part is error term, and another part is canonical Constraint.
For whether include hiding information two taxonomic histories, LOSS function is defined as follows:
Wherein N is the quantity of single batch of training sample, yiIt includes hidden for representing i-th of the sample obtained through neural computing Hide the probability of information, tiThe physical tags of i-th of sample are represented, C is aforementioned identification weight vectors.
For estimating the identification of hiding information capacity, LOSS function is defined as follows:
Wherein, YiIndicate the insertion rate probability distribution for the output of i-th of sample classification device, Yi,jRepresent i-th of sample category In the probability of jth class insertion rate steganography text.TiThe rate that is actually embedded in for i-th of sample indicates, if the sample belongs to kth class The steganography text of insertion rate, then TiKth dimension is 1, residual value 0.WPFor aforementioned Weight prediction matrix.
Error term in loss function calculates the average cross entropy between prediction probability value and true tag.Pass through model Self study, prediction error can become smaller and smaller, that is to say, that prediction result becomes closer to true tag.In order to add Strong regularization and prevent overfitting, the method for the embodiment of the present invention use in the training process loss mechanisms and to weight to The constraint of two norms of amount.
Deep learning (deep learning) is a branch of machine learning, is a kind of based on logarithm in machine learning According to the method for carrying out representative learning, it attempts using comprising labyrinth or the multiple process layers being made of multiple nonlinear transformation Higher level of abstraction is carried out to data.The benefit of deep learning is that the feature learning and layered characteristic with non-supervisory formula or Semi-supervised mention Highly effective algorithm is taken to substitute and obtain feature by hand.The target of representative learning is to seek better representation method and create better mould Type learns these representation methods from extensive Unlabeled data.Expression way similar to Neuscience progress, and loosely It creates in the understanding of information processing and communication pattern in similar nervous system, such as neural coding, it is intended to which definition pulls nerve Relationship between the reaction of member and the relationship between the electrical activity of the neuron in brain.So far have several deep learning frames Frame, as deep neural network, convolutional neural networks and depth confidence network and Recognition with Recurrent Neural Network by Appliance computer vision, The domain variabilities such as speech recognition, natural language processing, audio identification and bioinformatics obtain fabulous effect.
Recognition with Recurrent Neural Network (Recurrent Neural Network, RNN) is a kind of deep learning frame, it is a kind of For the neural network of processing sequence data, it is made of input layer, hidden layer and output layer.Essential characteristic is network in each step It all include feedback link, therefore it can extend on time dimension and form " depth " neural network on time dimension.This One structure makes Recognition with Recurrent Neural Network be capable of handling sequence data.Compare other depth, feedforward neural network, recycles nerve net Network is capable of handling the sequence data of random length by using the neural unit with self feed back, is a kind of depth for having much attraction Learning structure is spent, and there are many improved variants.Wherein, shot and long term memory models (Long Short-Term Memory, LSTM), using the Hidden unit structure of smart design, long distance dependent relation has more effectively been handled than common RNN, so as to Preferably to model nature text.
Recognition with Recurrent Neural Network is widely used in the tasks such as speech recognition, language model and language generation, is had very Strong feature extraction, expression, semantic understanding ability, it does not need artificial design feature, but the self-teaching from mass data To various features.It is core of the invention by the text vocabulary association feature extraction that Recognition with Recurrent Neural Network applies in Stego-detection The heart.RNN preferably can model text than Markov model, be drawn into more reasonable vocabulary association feature, from And more effectively identify potential steganography text.
To sum up, carrying out text generation formula Stego-detection using Recognition with Recurrent Neural Network has incomparable excellent of existing method Gesture can effectively solve the problems, such as that existing method, various aspects of performance are better than existing method.
The text Stego-detection method based on Recognition with Recurrent Neural Network proposed according to embodiments of the present invention, by the way that mind will be recycled Through network application into text Stego-detection, it can effectively identify whether text carrier contains hiding information, and according to extraction feature Statistical distribution accurately estimate the capacity of hiding information.Traditional Stego-detection method based on simple feature can be effectively overcome to be had Some defects can also obtain very high detection accuracy in face of production text steganography method neural network based;Based on LSTM The statistical nature of correlativity between model extraction word can effectively model the long distance dependent relation in text.
The text Stego-detection based on Recognition with Recurrent Neural Network proposed according to embodiments of the present invention is described referring next to attached drawing System.
Fig. 4 is to be illustrated according to the text Stego-detection system structure based on Recognition with Recurrent Neural Network of one embodiment of the invention Figure.
As shown in figure 4, should text Stego-detection system 10 based on Recognition with Recurrent Neural Network include: to obtain module 100, the One generation module 200, the second generation module 300, judgment module 400 and estimation module 500.
Wherein, module 100 is obtained for obtaining term vector matrix.First generation module 200 is used for according to term vector matrix Input word sequence vector is converted by text to be detected.Second generation module 300 is pre- for input word sequence vector to be input to In the Recognition with Recurrent Neural Network model first constructed, the feature vector of correlativity between indicating text word to be detected is generated.Judge mould Block 400 judges whether text to be detected includes hiding information for classifying by classifier to feature vector.Estimation module 500 include hiding information for text to be detected, and the difference estimation according to steganography Text eigenvector under different insertion rates is to be checked Survey the information insertion rate of text.The detection system 10 applies to Recognition with Recurrent Neural Network in text Stego-detection, can effectively identify Whether text carrier contains hiding information, and the capacity of hiding information is accurately estimated according to the statistical distribution of extraction feature.
Further, in one embodiment of the invention, module 100 is obtained, is specifically used for: obtaining a large amount of generation texts This, is segmented and is pre-processed to text is generated, and filtered based on word frequency, and the common lexical set for generating text is obtained;It will be normal The common vocabulary seen in lexical set is sequentially inputted in the module pre-established, obtain the corresponding word of each common vocabulary to Amount;Term vector sample is formed according to the corresponding term vector of each common vocabulary, term vector matrix is generated according to term vector sample.
Further, in one embodiment of the invention, the first generation module 200, is specifically used for: by text to be detected Low-frequency word filtering, screening and participle are carried out, the sequence of words of text to be detected is generated;Sequence of words is obtained according to term vector matrix In all vocabulary term vector, wherein term vector matrix includes corresponding relationship between common vocabulary and term vector;According to be checked The term vector for surveying all vocabulary in text, generates the input word sequence vector of text to be detected.
Further, in one embodiment of the invention, further includes: building module;
Construct module, the corresponding input word sequence vector of generation text for obtaining multiple Given information insertion rates, root Recognition with Recurrent Neural Network model is constructed according to the corresponding input word sequence vector of the generation text of multiple Given information insertion rates.
Further, in one embodiment of the invention, further includes: correction module;
Correction module is specifically used for: generating the generation text of Given information insertion rate according to Recognition with Recurrent Neural Network model Character pair vector;Classified using feature vector of the classifier to the generation text of known insertion rate, generates known insertion The insertion rate estimated probability of the generation text of rate;Using preset algorithm to the reality of the generation text of estimated probability and known insertion rate Border insertion rate carries out analysis and generates analysis as a result, correcting the parameter and classifier of Recognition with Recurrent Neural Network model based on the analysis results Weight parameter.
It should be noted that the aforementioned explanation to the text Stego-detection embodiment of the method based on Recognition with Recurrent Neural Network The system for being also applied for the embodiment, details are not described herein again.
The text Stego-detection system based on Recognition with Recurrent Neural Network proposed according to embodiments of the present invention, passes through
In addition, term " first ", " second " are used for descriptive purposes only and cannot be understood as indicating or suggesting relative importance Or implicitly indicate the quantity of indicated technical characteristic.Define " first " as a result, the feature of " second " can be expressed or Implicitly include at least one this feature.In the description of the present invention, the meaning of " plurality " is at least two, such as two, three It is a etc., unless otherwise specifically defined.
In the description of this specification, reference term " one embodiment ", " some embodiments ", " example ", " specifically show The description of example " or " some examples " etc. means specific features, structure, material or spy described in conjunction with this embodiment or example Point is included at least one embodiment or example of the invention.In the present specification, schematic expression of the above terms are not It must be directed to identical embodiment or example.Moreover, particular features, structures, materials, or characteristics described can be in office It can be combined in any suitable manner in one or more embodiment or examples.In addition, without conflicting with each other, the skill of this field Art personnel can tie the feature of different embodiments or examples described in this specification and different embodiments or examples It closes and combines.
Although the embodiments of the present invention has been shown and described above, it is to be understood that above-described embodiment is example Property, it is not considered as limiting the invention, those skilled in the art within the scope of the invention can be to above-mentioned Embodiment is changed, modifies, replacement and variant.

Claims (10)

1. a kind of text Stego-detection method based on Recognition with Recurrent Neural Network, which comprises the following steps:
Term vector matrix is obtained, input word sequence vector is converted for text to be detected according to the term vector matrix;
The input word sequence vector is input in the Recognition with Recurrent Neural Network model constructed in advance, generating indicates described to be detected The feature vector of correlativity between text word;
Classified by classifier to described eigenvector, judges whether the text to be detected includes hiding information;
If the text to be detected includes hiding information, estimated according to the difference of steganography Text eigenvector under different insertion rates The information insertion rate of the text to be detected.
2. the text Stego-detection method according to claim 1 based on Recognition with Recurrent Neural Network, which is characterized in that described to obtain Take term vector matrix, comprising:
A large amount of generation texts are obtained, the generation text is segmented and pre-processed, and is filtered based on word frequency, obtain generating text This common lexical set;
Common vocabulary in the common lexical set is sequentially inputted in the module pre-established, each common word is obtained Converge corresponding term vector;
Term vector sample is formed according to the corresponding term vector of each described common vocabulary, institute is generated according to the term vector sample Predicate vector matrix.
3. the text Stego-detection method according to claim 1 based on Recognition with Recurrent Neural Network, which is characterized in that described Input word sequence vector is converted by text to be detected according to the term vector matrix, comprising:
The text to be detected is subjected to low-frequency word filtering, screening and participle, generates the sequence of words of the text to be detected;
The term vector of all vocabulary in the sequence of words is obtained according to the term vector matrix, wherein the term vector matrix Include the corresponding relationship between common vocabulary and term vector;
According to the term vector of all vocabulary in the text to be detected, the input term vector sequence of the text to be detected is generated Column.
4. the text Stego-detection method according to claim 1 based on Recognition with Recurrent Neural Network, which is characterized in that described Before the input word sequence vector is input in the Recognition with Recurrent Neural Network model constructed in advance, further includes:
The corresponding input word sequence vector of generation text for obtaining multiple Given information insertion rates, according to the multiple known The corresponding input word sequence vector of the generation text of information insertion rate constructs the Recognition with Recurrent Neural Network model.
5. the text Stego-detection method according to claim 4 based on Recognition with Recurrent Neural Network, which is characterized in that described After constructing the Recognition with Recurrent Neural Network model, further includes:
The character pair vector of the generation text of the Given information insertion rate is generated according to the Recognition with Recurrent Neural Network model;
Classified using feature vector of the classifier to the generation text of the known insertion rate, is generated described known embedding Enter the insertion rate estimated probability of the generation text of rate;
It is analyzed using be actually embedded in rate of the preset algorithm to the generation text of the estimated probability and the known insertion rate Analysis is generated as a result, the parameter of Recognition with Recurrent Neural Network model and the weight of the classifier according to the analysis modified result Parameter.
6. a kind of text Stego-detection system based on Recognition with Recurrent Neural Network characterized by comprising
Module is obtained, for obtaining term vector matrix;
First generation module, for converting input word sequence vector for text to be detected according to the term vector matrix;
Second generation module, for the input word sequence vector to be input in the Recognition with Recurrent Neural Network model constructed in advance, Generate the feature vector of correlativity between indicating the text word to be detected;
Judgment module, for being classified by classifier to described eigenvector, judge the text to be detected whether include Hiding information;
Estimation module includes hiding information for the text to be detected, according to steganography Text eigenvector under different insertion rates Difference estimate the information insertion rate of the text to be detected.
7. the text Stego-detection system according to claim 6 based on Recognition with Recurrent Neural Network, which is characterized in that described to obtain Modulus block, is specifically used for:
A large amount of generation texts are obtained, the generation text is segmented and pre-processed, and is filtered based on word frequency, obtain generating text This common lexical set;
Common vocabulary in the common lexical set is sequentially inputted in the module pre-established, each common word is obtained Converge corresponding term vector;
Term vector sample is formed according to the corresponding term vector of each described common vocabulary, institute is generated according to the term vector sample Predicate vector matrix.
8. the text Stego-detection system according to claim 6 based on Recognition with Recurrent Neural Network, which is characterized in that described One generation module, is specifically used for:
The text to be detected is subjected to low-frequency word filtering, screening and participle, generates the sequence of words of the text to be detected;
The term vector of all vocabulary in the sequence of words is obtained according to the term vector matrix, wherein the term vector matrix Include the corresponding relationship between common vocabulary and term vector;
According to the term vector of all vocabulary in the text to be detected, the input term vector sequence of the text to be detected is generated Column.
9. the text Stego-detection system according to claim 6 based on Recognition with Recurrent Neural Network, which is characterized in that also wrap It includes: building module;
The building module, the corresponding input term vector sequence of generation text for obtaining multiple Given information insertion rates Column construct the circulation mind according to the corresponding input word sequence vector of the generation text of the multiple Given information insertion rate Through network model.
10. the text Stego-detection system according to claim 6 based on Recognition with Recurrent Neural Network, which is characterized in that also wrap It includes: correction module;
The correction module, is specifically used for:
The character pair vector of the generation text of the Given information insertion rate is generated according to the Recognition with Recurrent Neural Network model;
Classified using feature vector of the classifier to the generation text of the known insertion rate, is generated described known embedding Enter the insertion rate estimated probability of the generation text of rate;
It is analyzed using be actually embedded in rate of the preset algorithm to the generation text of the estimated probability and the known insertion rate Analysis is generated as a result, the parameter of Recognition with Recurrent Neural Network model and the weight of the classifier according to the analysis modified result Parameter.
CN201910058680.5A 2019-01-22 2019-01-22 Text steganography detection method and system based on cyclic neural network Active CN110110318B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910058680.5A CN110110318B (en) 2019-01-22 2019-01-22 Text steganography detection method and system based on cyclic neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910058680.5A CN110110318B (en) 2019-01-22 2019-01-22 Text steganography detection method and system based on cyclic neural network

Publications (2)

Publication Number Publication Date
CN110110318A true CN110110318A (en) 2019-08-09
CN110110318B CN110110318B (en) 2021-02-05

Family

ID=67483357

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910058680.5A Active CN110110318B (en) 2019-01-22 2019-01-22 Text steganography detection method and system based on cyclic neural network

Country Status (1)

Country Link
CN (1) CN110110318B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110830489A (en) * 2019-11-14 2020-02-21 国网江苏省电力有限公司苏州供电分公司 Method and system for detecting counterattack type fraud website based on content abstract representation
CN111477320A (en) * 2020-03-11 2020-07-31 北京大学第三医院(北京大学第三临床医学院) Construction system of treatment effect prediction model, treatment effect prediction system and terminal
CN111859897A (en) * 2019-10-16 2020-10-30 沈阳工业大学 Text steganalysis method based on dynamic routing capsule network
CN111931514A (en) * 2020-07-28 2020-11-13 薛杨杨 Information processing method based on deep learning and big data and block chain service platform
CN112132262A (en) * 2020-09-08 2020-12-25 西安交通大学 Recurrent neural network backdoor attack detection method based on interpretable model
CN112749274A (en) * 2019-10-30 2021-05-04 中南大学 Chinese text classification method based on attention mechanism and interference word deletion
CN113096671A (en) * 2020-01-09 2021-07-09 齐鲁工业大学 Reversible information hiding method and system for high-capacity audio file
CN113743110A (en) * 2021-11-08 2021-12-03 京华信息科技股份有限公司 Word missing detection method and system based on fine-tuning generation type confrontation network model
CN115169293A (en) * 2022-09-02 2022-10-11 南京信息工程大学 Text steganalysis method, system, device and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1668995A (en) * 2002-06-06 2005-09-14 克瑞迪科公司 Method for improving unpredictability of output of pseudo-random number generators
CN102292700A (en) * 2009-01-24 2011-12-21 惠普开发有限公司 System and method for enhancing security printing
CN106055673A (en) * 2016-06-06 2016-10-26 中国人民解放军国防科学技术大学 Chinese short-text sentiment classification method based on text characteristic insertion
WO2018212811A1 (en) * 2017-05-19 2018-11-22 Google Llc Hiding information and images via deep learning
CN108923922A (en) * 2018-07-26 2018-11-30 北京工商大学 A kind of text steganography method based on generation confrontation network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1668995A (en) * 2002-06-06 2005-09-14 克瑞迪科公司 Method for improving unpredictability of output of pseudo-random number generators
CN102292700A (en) * 2009-01-24 2011-12-21 惠普开发有限公司 System and method for enhancing security printing
CN106055673A (en) * 2016-06-06 2016-10-26 中国人民解放军国防科学技术大学 Chinese short-text sentiment classification method based on text characteristic insertion
WO2018212811A1 (en) * 2017-05-19 2018-11-22 Google Llc Hiding information and images via deep learning
CN108923922A (en) * 2018-07-26 2018-11-30 北京工商大学 A kind of text steganography method based on generation confrontation network

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
GUOMING CHEN ET AL: "Discriminative multimodal for steganalysis", 《2015 8TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING》 *
LINGYUN XIANG ET AL: "A Word-Embedding-Based Steganalysis Method for Linguistic Steganography via Synonym Substitution", 《IEEE ACCESS》 *
YUBO LUO ET AL: "Text Steganography with High Embedding Rate: Using Recurrent Neural Networks to Generate Chinese Classic Poetry", 《PROCEEDINGS OF THE 5TH ACM WORKSHOP ON INFORMATION HIDING AND MULTIMEDIA SECURITY》 *
ZHONG-LIANG YANG ET AL: "RNN-Stega: Linguistic Steganography Based on Recurrent Neural Networks", 《IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY》 *
ZHONGLIANG YANG ET AL: "TS-CNN-Text Steganalysis from Semantic Space Based on Convolutional Neural Network", 《ARXIV.ORG > CS > ARXIV:1810.08136》 *
刘明明 等: "一种基于浅层卷积神经网络的隐写分析方法", 《山东大学学报(理学版)》 *
金鹏 等: "基于卷积神经网络的诗词隐写检测方法", 《电子技术应用》 *
高培贤 等: "针对图像隐写分析的卷积神经网络结构改进", 《计算机工程》 *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111859897A (en) * 2019-10-16 2020-10-30 沈阳工业大学 Text steganalysis method based on dynamic routing capsule network
CN112749274A (en) * 2019-10-30 2021-05-04 中南大学 Chinese text classification method based on attention mechanism and interference word deletion
CN110830489A (en) * 2019-11-14 2020-02-21 国网江苏省电力有限公司苏州供电分公司 Method and system for detecting counterattack type fraud website based on content abstract representation
CN110830489B (en) * 2019-11-14 2022-09-13 国网江苏省电力有限公司苏州供电分公司 Method and system for detecting counterattack type fraud website based on content abstract representation
CN113096671A (en) * 2020-01-09 2021-07-09 齐鲁工业大学 Reversible information hiding method and system for high-capacity audio file
CN113096671B (en) * 2020-01-09 2022-05-13 齐鲁工业大学 Reversible information hiding method and system for high-capacity audio file
CN111477320A (en) * 2020-03-11 2020-07-31 北京大学第三医院(北京大学第三临床医学院) Construction system of treatment effect prediction model, treatment effect prediction system and terminal
CN111477320B (en) * 2020-03-11 2023-05-30 北京大学第三医院(北京大学第三临床医学院) Treatment effect prediction model construction system, treatment effect prediction system and terminal
CN111931514A (en) * 2020-07-28 2020-11-13 薛杨杨 Information processing method based on deep learning and big data and block chain service platform
CN112132262A (en) * 2020-09-08 2020-12-25 西安交通大学 Recurrent neural network backdoor attack detection method based on interpretable model
CN113743110A (en) * 2021-11-08 2021-12-03 京华信息科技股份有限公司 Word missing detection method and system based on fine-tuning generation type confrontation network model
CN115169293A (en) * 2022-09-02 2022-10-11 南京信息工程大学 Text steganalysis method, system, device and storage medium

Also Published As

Publication number Publication date
CN110110318B (en) 2021-02-05

Similar Documents

Publication Publication Date Title
CN110110318A (en) Text Stego-detection method and system based on Recognition with Recurrent Neural Network
Yang et al. RNN-stega: Linguistic steganography based on recurrent neural networks
US11443178B2 (en) Deep neural network hardening framework
Cisse et al. Houdini: Fooling deep structured prediction models
CN111061843B (en) Knowledge-graph-guided false news detection method
CN110213244A (en) A kind of network inbreak detection method based on space-time characteristic fusion
CN112084331A (en) Text processing method, text processing device, model training method, model training device, computer equipment and storage medium
CN109299657B (en) Group behavior identification method and device based on semantic attention retention mechanism
CN112861945B (en) Multi-mode fusion lie detection method
Yang et al. Rits: Real-time interactive text steganography based on automatic dialogue model
CN111325319A (en) Method, device, equipment and storage medium for detecting neural network model
CN112527966B (en) Network text emotion analysis method based on Bi-GRU neural network and self-attention mechanism
CN112015901A (en) Text classification method and device and warning situation analysis system
CN111985207B (en) Access control policy acquisition method and device and electronic equipment
CN113505307B (en) Social network user region identification method based on weak supervision enhancement
CN108985382A (en) The confrontation sample testing method indicated based on critical data path
CN113888368A (en) Feature selection method for criminal case detention risk assessment based on image coding
CN113627233A (en) Visual semantic information-based face counterfeiting detection method and device
CN112712099A (en) Double-layer knowledge-based speaker model compression system and method
Zhang et al. Self-supervised adversarial example detection by disentangled representation
CN111737688A (en) Attack defense system based on user portrait
CN116561272A (en) Open domain visual language question-answering method and device, electronic equipment and storage medium
CN113205044B (en) Deep fake video detection method based on characterization contrast prediction learning
CN110049034A (en) A kind of real-time Sybil attack detection method of complex network based on deep learning
CN114818719A (en) Community topic classification method based on composite network and graph attention machine mechanism

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant