CN108566627A - A kind of method and system identifying fraud text message using deep learning - Google Patents

A kind of method and system identifying fraud text message using deep learning Download PDF

Info

Publication number
CN108566627A
CN108566627A CN201711205293.7A CN201711205293A CN108566627A CN 108566627 A CN108566627 A CN 108566627A CN 201711205293 A CN201711205293 A CN 201711205293A CN 108566627 A CN108566627 A CN 108566627A
Authority
CN
China
Prior art keywords
deep learning
short message
vector
word
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201711205293.7A
Other languages
Chinese (zh)
Inventor
陈晓莉
刘亭
丁帆
丁一帆
徐菁
林建洪
徐佳丽
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Ponshine Information Technology Co Ltd
Original Assignee
Zhejiang Ponshine Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Ponshine Information Technology Co Ltd filed Critical Zhejiang Ponshine Information Technology Co Ltd
Priority to CN201711205293.7A priority Critical patent/CN108566627A/en
Publication of CN108566627A publication Critical patent/CN108566627A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W12/00Security arrangements; Authentication; Protecting privacy or anonymity
    • H04W12/12Detection or prevention of fraud
    • H04W12/128Anti-malware arrangements, e.g. protection against SMS fraud or mobile malware
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/12Messaging; Mailboxes; Announcements
    • H04W4/14Short messaging services, e.g. short message services [SMS] or unstructured supplementary service data [USSD]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W12/00Security arrangements; Authentication; Protecting privacy or anonymity
    • H04W12/12Detection or prevention of fraud

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computer Security & Cryptography (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • General Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses a kind of method and system identifying fraud text message using deep learning, cause algorithm recognition accuracy not high to solve the problems, such as that the feature of fraud text message is ever-changing.This method includes:It obtains the text data of short message sample and carries out word segmentation processing;Term vector is converted into the text data after participle using Word2Vec;The term vector is converted by sentence vector using LSTM algorithms;Using the sentence vector as the input vector of softmax graders to train deep learning model;Fraud text message is identified according to the output result of the softmax graders of the deep learning after training.The present invention provides a kind of method and system that fraud text message is identified using deep learning, improves the ability accurately identified to fraud text message.

Description

Method and system for recognizing fraud short messages by utilizing deep learning
Technical Field
The invention relates to the technical field of communication security, in particular to a method and a system for recognizing fraud short messages by utilizing deep learning.
Background
The short message is used as a carrier for transmitting information among massive clients, and an effective channel is established for communication among the clients. Along with the wide use of short messages, the phenomenon of fraud information content transmitted through the short messages is more and more serious, great inconvenience is brought to normal information communication of mobile phone users, and more users are consequently cheated and cannot be read.
The short message brings serious social cost and cultural loss to operators while creating economic benefit.
At present, the research on the identification of the fraud short messages mainly adopts a mode of combining feature extraction and a traditional machine learning algorithm, but a large amount of manpower and material resources are needed in the aspect of feature extraction, and along with the rapid updating and iteration of information, the features of the fraud short messages are also changed greatly, so that the accuracy of algorithm identification is seriously reduced. In addition, short messages belong to short texts, are short in length and sparse in characteristics, and especially, the structure and content of fraud short messages serving as an escape filtering mechanism are often not standardized, so that the traditional text characteristic extraction method cannot be completely suitable for short message classification.
Patent publication No. CN102547709A provides a method and a mobile phone for confirming fraud short messages, the method includes: comparing the received short message content with a preset keyword database, if the keywords are matched or partially matched, comparing the short message content with a preset fraud short message database, if the keywords are matched, determining that the short message is a fraud short message, filtering the short message, and if the keywords are partially matched, determining that the short message is a suspected fraud short message, and prompting a user to prevent the short message. The mobile phone comprises a first comparison unit, a second comparison unit, a filtering unit and a display unit. The method for confirming the fraud short messages and the mobile phone can ensure that the mobile phone can confirm whether the short messages are fraud short messages and can filter and prevent the fraud short messages. However, the recognition rate of the invention is very low when the invention faces the varied irregular fraud short messages.
Disclosure of Invention
The invention aims to provide a method and a system for recognizing fraud short messages by utilizing deep learning, which are used for solving the problem of low algorithm recognition accuracy caused by the fact that the characteristics of the fraud short messages are changeable.
In order to achieve the purpose, the invention adopts the technical scheme that:
a method for recognizing fraud short messages by deep learning comprises the following steps:
acquiring text data of a short message sample and performing word segmentation processing;
converting the text data after Word segmentation into Word vectors by adopting Word2 Vec;
converting the word vector into a sentence vector by adopting an LSTM algorithm;
taking the sentence vector as an input vector of a softmax classifier to train a deep learning model;
and recognizing fraud short messages according to the output result of the trained softmax classifier.
Further, the step of performing word segmentation processing by using the acquired text data of the short message specifically includes:
collecting text data of all short message samples;
removing a non-text part in the text data by adopting a regularization method;
dividing the short message sample into a negative sample and a positive sample, and dividing the short message sample into a training sample and a testing sample according to a preset proportion;
performing word segmentation processing on the short message sample by adopting a crust word segmentation tool;
stop words are introduced to remove invalid words from the text.
Further, the step of training the deep learning model specifically includes:
selecting the result with the maximum probability value as the output short message category;
calculating a loss function and performing back propagation;
adjusting the weight of the deep learning model to a preset threshold value;
inputting the test sample into the deep learning model, and calculating accuracy, recall rate and F value;
and optimizing the deep learning model by utilizing the self-learning capability of the neural network.
Further, the step of converting the text data after Word segmentation into Word vectors by using Word2Vec specifically includes:
counting the number m of the keywords in the fraud short message special library;
converting a word into an n-dimensional word vector x by using a one-hot-vector;
establishing a weight matrix w of n m to map the vector of n dimensions to the hidden neuron with dimension 1 m;
obtaining a vector W by back-propagation and a word vector W (i) of 1 m by multiplication with the word vector x;
and adding the word vectors corresponding to the fraud keywords appearing in each short message to obtain a text vector d of the short message.
Further, the step of converting the word vector into the sentence vector by using the LSTM algorithm specifically includes:
arranging the word vectors in a preset sequence, assuming that there are m word vectors xtForming;
initializing model parameters Wf,Uf,bf,Wa,Ua,ba,Wi,Ui,bi,Wo,Uo,bo
X is to betPass into forgetting door ft+1And updating the weight W of the forgetting gatet+1,Ut+1,bt+1(ii) a Wherein,
ft+1=σ(Wt+1ht+Ut+1xt+1+bt+1);
wherein Wf,Uf,bfCoefficients and biases that are linear relationships; sigma is a sigmoid activation function;
updating input Gate parameter itAnd at(ii) a Wherein,
it=σ(Wiht-1+Uixt+bi);
at=tanh(Waht-1+Uaxt+ba);
Wa,Ua,ba,Wi,Ui,bicoefficients and biases that are linear relationships; sigma is a sigmoid activation function;
updating model output State CtWherein
Ct=Ct-1⊙ft+it⊙at
as Hadamard product;
updating output gate parameter σtAnd htAnd outputs the predicted value of the current sequence indexWherein,
σt=σ(Woht-1+Uoxt+bo);
ht=ot⊙tanh(Ct);
further, the softmax function is formulated as:
wherein,is the output of the jth neuron in the lth layer,an output representing a jth neuron of a current layer; wherein,
further, the method also comprises the following steps:
and if the fraud short message is determined, intercepting the short message.
A system for recognizing fraud messages using deep learning, comprising:
the processing module is used for acquiring text data of the short message sample and performing word segmentation processing;
the Word vector module is used for converting the text data after Word segmentation into Word vectors by adopting Word2 Vec;
the sentence vector module is used for converting the word vector into a sentence vector by adopting an LSTM algorithm;
the training module is used for taking the sentence vector as an input vector of a softmax classifier so as to train a deep learning model;
and the recognition module is used for recognizing the fraud short messages according to the output result of the deep-learning softmax classifier after training.
Further, the processing module specifically includes:
the acquisition unit is used for acquiring text data of all short message samples;
the removing unit is used for removing the non-text part in the text data by adopting a regularization method;
the dividing unit is used for dividing the short message sample into a negative sample and a positive sample and dividing the short message sample into a training sample and a test sample according to a preset proportion;
the word segmentation unit is used for performing word segmentation processing on the short message sample by adopting a crust word segmentation tool;
and the introducing unit is used for introducing stop words to remove invalid words in the text.
Further, the training module specifically includes:
the selection unit is used for selecting the result with the maximum probability value as the output short message category;
the computing unit is used for computing a loss function and performing back propagation;
the adjusting unit is used for adjusting the weight of the deep learning model to a preset threshold value;
the input unit is used for inputting the test sample into the deep learning model and calculating the accuracy, the recall rate and the F value;
and the tuning unit is used for tuning the deep learning model by utilizing the self-learning capability of the neural network.
Compared with the traditional technology, the invention has the following advantages:
the invention provides a method and a system for identifying fraud short messages by utilizing deep learning, which improve the capability of accurately identifying the fraud short messages.
Drawings
FIG. 1 is a flowchart of a method for identifying fraud short messages by deep learning according to an embodiment;
FIG. 2 is a diagram of the LSTM algorithm provided by the first embodiment for fraud message identification;
FIG. 3 is a block diagram of a system for recognizing fraud messages by deep learning according to the second embodiment.
Detailed Description
The following are specific embodiments of the present invention and are further described with reference to the drawings, but the present invention is not limited to these embodiments.
Example one
The embodiment provides a method for identifying fraud short messages by deep learning, as shown in fig. 1, comprising the steps of:
s11: acquiring text data of a short message sample and performing word segmentation processing;
s12: converting the text data after halving by adopting Word2Vec into Word vectors;
s13: converting the word vector into a sentence vector by adopting an LSTM algorithm;
s14: taking the sentence vector as an input vector of a softmax classifier to train a deep learning model;
s15: recognizing fraud short messages according to the output result of the trained deep learning softmax classifier;
s16: and if the fraud short message is determined, intercepting the short message.
In recent years, deep learning algorithms have been applied to the field of natural language processing, with superior results compared to traditional models. In natural language processing, a Recurrent Neural Network (RNNs) or a Recurrent Neural Network (RNNs) is a commonly used method. Their role is to encode the input in matrix form as a one-dimensional vector at a lower latitude, while retaining most of the useful information. RNNs are in many variations, such as common RNNs, as well as GRU, LSTM, and the like.
The embodiment provides a method for effectively identifying fraud messages by utilizing a deep learning algorithm under mass data. When the short message sender sends the short message to the short message network gateway, the deep learning analysis system is utilized to identify the content of the short message, if the deep learning model judges that the short message is a fraud short message, the fraud short message is intercepted, and the short message cannot reach the receiver. Through the deep learning algorithm, the accuracy rate of recognizing the fraud short messages is greatly improved. The problem of low recognition rate caused by the diversification of the fraud short messages is avoided.
In this embodiment, step S11 is to obtain text data of the short message sample and perform word segmentation processing.
Wherein, step S11 specifically includes:
collecting text data of all short message samples;
removing a non-text part in the text data by adopting a regularization method;
dividing the short message sample into a negative sample and a positive sample, and dividing the short message sample into a training sample and a test sample according to a preset proportion;
performing word segmentation processing on the short message sample by adopting a crust word segmentation tool;
stop words are introduced to take out invalid words in the text.
Specifically, regularization means that in linear algebra theory, an ill-defined problem is usually defined by a set of linear algebra equations, and the set of equations usually results from an ill-defined inverse problem with a large condition number. The large condition number means that rounding errors or other errors can severely impact the outcome of the problem.
The ending word segmentation tool is used for realizing efficient word graph scanning based on a prefix dictionary and generating a directed acyclic graph formed by all possible word forming conditions of Chinese characters in a sentence; a maximum probability path is searched by adopting dynamic programming, and a maximum segmentation combination based on word frequency is found out; for unknown words, an HMM model based on Chinese character word forming capability is adopted, and a Viterbi algorithm is used.
Firstly, collecting sample text data containing fraud short messages and normal short messages; and removing the non-text part of the text data by adopting a regularization method, and only keeping the text part. And dividing the short message sample into a positive sample and a negative sample and dividing the samples into a training sample and a test sample according to a preset proportion. For example, the sample is divided into a training sample and a test sample in a ratio of 3: 1. And then, performing word segmentation processing on the short message sample by adopting a word segmentation tool, and finally introducing stop words to take out invalid words in the text.
In this embodiment, step S12 is to convert Word2Vec into Word vectors.
Wherein, step S12 specifically includes:
counting the number m of the keywords in the fraud short message special library;
converting a word into an n-dimensional word vector x by using a one-hot-vector;
establishing a weight matrix w of n m to map the vector of n dimensions to the hidden neuron with dimension 1 m;
obtaining a vector W by inverse transfer and a word vector W (i) of 1 m by multiplication with the word vector x;
and adding the word vectors corresponding to the fraud keywords appearing in each short message to obtain a text vector d of the short message.
Specifically, Word2Vec is a Word vector generation tool for deep learning, essentially utilizes a neural network language model and simplifies the neural network language model, thereby ensuring the effect and improving the computational complexity. Word2Vec represents words with high dimensional vectors and places words of similar meaning in similar positions, and uses real vectors (not limited to integers). Only a large amount of linguistic data of a certain language is needed to train the model and obtain the word vector. There are two algorithms commonly used for this model: CBOW and Skip-gram. The CBOW model predicts a current word by using k words before and after the word W (t); the Skip-gram model is just the opposite, and uses word W (t) to predict k words before and after the word W (t).
Before training the model, words in the fraud message feature library need to be quantized and converted into word vectors. The number of the fraud short message words in the feature word library is the dimension of the vector, each word is endowed with a code by using a one-hot-vector, the position of the word is marked as '1', and the other positions are marked as '0'. For example, the word vector [0,0,0,0,1, …,0,0] for "congratulation", and the word vector for "winning" is [0,1,0,0,0, …,0,0 ].
Word2Vec is a neural network with one hidden layer. The input and output of the method are word vectors, and after the trained neural network converges, the weight from an input layer to a hidden layer is assigned to each word vector, so that each word can obtain a new vector with semantic meaning.
The specific implementation process of the Word2Vec algorithm can be further subdivided into:
counting the keywords in the fraud short message feature library, and assuming that m keywords exist;
firstly, converting a word into an n-dimensional word vector x by utilizing a one-hot-vector; take "winning" as an example:
"winning" → [0,0,0,0,1, …,0,0]
The hidden layer is provided with m neurons, and the known input layer is an n-dimensional vector and is fully connected with the hidden layer;
the hidden layer is connected with the output layer in a full mode, a softmax classifier is added in the computer output unit, a final vector W can be obtained through reverse transfer, and the final word vector, namely the vector W (i) of 1 m, can be obtained through multiplication of the final word vector and the initial word vector, namely x W;
x*w=W(i)=[Wi1Wi2… Wim]
and adding the word vectors corresponding to the fraud keywords appearing in each short message to obtain a text vector d belonging to the short message.
In this embodiment, step S13 is to convert the word vector into the sentence vector by using the LSTM algorithm.
Wherein, step S13 specifically includes:
arranging the word vectors in a preset sequence, assuming that there are m word vectors xtForming;
initializing model parameters Wf,Uf,bf,Wa,Ua,ba,Wi,Ui,bi,Wo,Uo,bo
X is to betPass into forgetting door ft+1And updating the weight W of the forgetting gatet+1,Ut+1,bt+1(ii) a Wherein,
ft+1=σ(Wt+1ht+Ut+1xt+1+bt+1);
wherein Wf,Uf,bfCoefficients and biases that are linear relationships; sigma is a sigmoid activation function;
updating input Gate parameter itAnd at(ii) a Wherein,
it=σ(Wiht-1+Uixt+bi);at=tanh(Waht-1+Uaxt+ba);
Wa,Ua,ba,Wi,Ui,bicoefficients and biases that are linear relationships; sigma is a sigmoid activation function;
updating model output State CtWherein
Ct=Ct-1⊙ft+it⊙at
as Hadamard product;
updating output gate parameter σtAnd htAnd outputs the predicted value of the current sequence indexWherein,
σt=σ(Woht-1+Uoxt+bo);
ht=ot⊙tanh(Ct);
specifically, a deep learning model for fraud short message recognition is built based on LSTM (long-short term memory artificial neural network), as shown in fig. 2.
LSTM, as a special case of RNN, can avoid the problem of gradient disappearance of conventional RNN. RNN assumes that the samples are sequence based. Such as from sequence index 1 to sequence index t. For any sequence index number t, its corresponding input is x in the corresponding sample sequence numberiAnd hidden state h at t-1 positiontThe determination is made in common. At any sequence index t, there is a corresponding model prediction output ot. By predicting the output otAnd training sequence true output ytAnd a loss function LtThe model can be trained in a similar way to softmax and then used to predict the output for some positions in the test sequence. Due to the problem of RNN gradient disappearance, a hidden structure of a sequence index position t is improved to avoid the problem of gradient disappearance, and the special RNN is LSTM.
Hidden states h propagated forward at each sequence index position t instant, except as RNNtThere is another hidden state, generally called the cellular state, denoted Ct
In addition to cellular states, there are other structures in LSTM, which are generally referred to as gated structures. The gates of the LSTM at each sequence index position t typically include three types, a forgetting gate, an input gate, and an output gate.
The forgetting gate controls whether to forget, and in the LSTM, whether to forget the state of the hidden cell in the previous layer is controlled with a certain probability. Hidden state h of last sequence of inputt-1And the present sequence data xiThe output f of the forgetting gate is obtained by an activation function, generally sigmoidt. Output f due to sigmoidtIn [0,1 ]]And hence the output f heretRepresenting the probability of forgetting the state of a layer of hidden cells. Using the mathematical expression as:
ft+1=σ(Wt+1ht+Ut+1xt+1+bt+1);
wherein Wf,Uf,bfCoefficients and biases that are linear relationships; sigma is sigmoid activation function.
The input gate is responsible for processing the input of the current sequence position and consists of two parts, wherein the first part uses a sigmoid activation function, and the output is itThe second part uses the tanh activation function and outputs a bit atAnd multiplying the results of the two methods to update the cell state, wherein the mathematical expression is as follows:
it=σ(Wiht-1+Uixt+bi);
at=tanh(Waht-1+Uaxt+ba);
wherein, Wa,Ua,ba,Wi,Ui,biCoefficients and biases that are linear relationships; sigma is sigmoid activation function.
The results of both the forgetting gate and the entry gate contribute to the cell state Ct. Cell State CtIs composed of two parts, the first part is Ci-1And forget gate output ftThe second part is the input gate itAnd atThe product of (a) and (b), namely:
Ct=Ct-1⊙ft+it⊙at
wherein, it is Hadamard product.
Hidden state h of output gatetThe update of (2) is composed of two parts, the first part being otFrom the previous sequence of hidden states ht-1And the present sequence data xtAnd the activation function sigmoid, the second part being derived from the hidden state CtAnd tanh activation function, i.e.:
σt=σ(Woht-1+Uoxt+bo);
ht=ot⊙tanh(Ct)。
LSTM model with hidden state ht,CtParameter has Wf,Uf,bfWa,Ua,ba,Wi,Ui, bi,Wo,Uo,boAnd sequentially updating parameters according to the sequence of the forgetting gate, the input gate and the output gate, and outputting a current sequence index predicted value:
in this embodiment, step S14 is to use the vector as an input vector of the softmax classifier to train the deep learning model.
Wherein the softmax function formula is as follows:
wherein,is the output of the jth neuron in the lth layer,an output representing a jth neuron of a current layer; wherein,
in particular, one feature of the softmax function is that it takes as the output of each neuron the ratio of the input of that neuron to the sum of the inputs of all neurons in the current layer. This makes the output easier to interpret. The larger the output value of a neuron is, the higher the probability that the class corresponding to the neuron is a true class is.
Step S14 specifically includes:
selecting the result with the maximum probability value as the output short message category;
calculating a loss function and performing back propagation;
adjusting the weight of the deep learning model to a preset threshold;
inputting the test sample into a deep learning model, and calculating accuracy, recall rate and F value;
and (4) optimizing the deep learning model by utilizing the self-learning capability of the neural network.
Specifically, the idea of the back propagation algorithm is to iteratively update all parameters by a gradient descent method, and the key point is to calculate partial derivatives of all parameters based on a loss function. There are two hidden states h in the LSTMtAnd CtTwo deltas are defined to propagate in opposite directions, i.e.
In this embodiment, the result with the maximum probability value of the hidden cell state is selected as the output of the short message, and then the loss function is calculated and the reverse propagation is performed. And then, adjusting the weight of the deep learning model to a preset threshold value, inputting the test sample into the deep learning model, calculating the accuracy, the recall rate and the F value, and then optimizing the deep learning model according to the calculated record. And training by the method to obtain the trained deep learning model.
In this embodiment, step S15 is to identify fraud short messages according to the output result of the trained deep learning softmax classifier.
Specifically, after the deep learning model is obtained through training, data is input into the deep learning model to identify the fraud short messages.
In this embodiment, step S16 is to intercept the short message if the fraud short message is determined.
Specifically, if the fraud short message is judged through the deep learning model, the short message gateway intercepts the short message, and the short message receiver does not receive the short message.
Example two
The embodiment provides a system for classifying fraud short messages by deep learning, as shown in fig. 3, including:
the processing module 21 is configured to obtain text data of a short message sample and perform word segmentation processing;
the Word vector module 22 is used for converting the text data after halving by Word2Vec into Word vectors;
a sentence vector module 23, configured to convert the word vector into a sentence vector by using an LSTM algorithm;
a training module 24, configured to use the sentence vector as an input vector of the softmax classifier to train the deep learning model;
and the identification module 25 is used for identifying the fraud short messages according to the output result of the trained deep learning softmax classifier.
The embodiment provides a system for effectively identifying fraud messages by utilizing a deep learning algorithm under mass data. When the short message sender sends the short message to the short message network gateway, the deep learning analysis system is utilized to identify the content of the short message, if the deep learning model judges that the short message is a fraud short message, the fraud short message is intercepted, and the short message cannot reach the receiver. Through the deep learning algorithm, the accuracy rate of recognizing the fraud short messages is greatly improved. The problem of low recognition rate caused by the diversification of the fraud short messages is avoided.
In this embodiment, the processing module 21 is configured to obtain text data of a short message sample and perform word segmentation processing.
Wherein, the processing module 21 specifically includes:
the acquisition unit is used for acquiring text data of all short message samples;
the removing unit is used for removing the non-text part in the text data by adopting a regularization method;
the dividing unit is used for dividing the short message sample into a negative sample and a positive sample and dividing the short message sample into a training sample and a test sample according to a preset proportion;
the word segmentation unit is used for performing word segmentation processing on the short message sample by adopting a crust word segmentation tool;
and the introducing unit is used for introducing stop words to extract invalid words in the text.
Specifically, the collecting unit collects sample text data containing the fraud short messages and the normal short messages; and the removing unit removes the non-text part of the text data by adopting a regularization method and only keeps the text part. The dividing unit divides the short message sample into a positive sample and a negative sample and divides the samples into a training sample and a testing sample according to a preset proportion. For example, the sample is divided into a training sample and a test sample in a ratio of 3: 1. And finally, the word segmentation unit introduces stop words to take out invalid words in the text.
In this embodiment, the Word vector module 22 converts the text data after Word segmentation into Word vectors by using Word2 Vec.
The word vector module 22 specifically includes:
the statistical unit is used for counting the number m of the key words in the fraud short message special library;
the conversion unit is used for converting a word into an n-dimensional word vector x by utilizing a one-hot-vector;
the establishing unit is used for establishing a weight matrix w of n m so as to map the vector of n dimensions to the hidden neuron with the dimension of 1 m;
a backward transfer unit for finding a vector W by backward transfer and obtaining a word vector W (i) of 1 × m by multiplying with the word vector x;
and the adding unit is used for adding the word vectors corresponding to the fraud keywords appearing in each short message to obtain a text vector d of the short message.
Specifically, Word2Vec is a Word vector generation tool for deep learning, essentially utilizes a neural network language model and simplifies the neural network language model, thereby ensuring the effect and improving the computational complexity. Word2Vec represents words with high dimensional vectors and places words of similar meaning in similar positions, and uses real vectors (not limited to integers). Only a large amount of linguistic data of a certain language is needed to train the model and obtain the word vector. There are two algorithms commonly used for this model: CBOW and Skip-gram. The CBOW model predicts a current word by using k words before and after the word W (t); the Skip-gram model is just the opposite, and uses word W (t) to predict k words before and after the word W (t).
Before training the model, words in the fraud message feature library need to be quantized and converted into word vectors. The number of the fraud short message words in the feature word library is the dimension of the vector, each word is endowed with a code by using a one-hot-vector, the position of the word is marked as '1', and the other positions are marked as '0'. For example, the word vector [0,0,0,0,1, …,0,0] for "congratulation", and the word vector for "winning" is [0,1,0,0,0, …,0,0 ].
Word2Vec is a neural network with one hidden layer. The input and output of the method are word vectors, and after the trained neural network converges, the weight from an input layer to a hidden layer is assigned to each word vector, so that each word can obtain a new vector with semantic meaning.
The specific implementation process of the Word2Vec algorithm can be further subdivided into:
counting the keywords in the fraud short message feature library, and assuming that m keywords exist;
firstly, converting a word into an n-dimensional word vector x by utilizing a one-hot-vector; take "winning" as an example:
"winning" → [0,0,0,0,1, …,0,0]
The hidden layer is provided with m neurons, and the known input layer is an n-dimensional vector and is fully connected with the hidden layer;
the hidden layer is connected with the output layer in a full mode, a softmax classifier is added in the computer output unit, a final vector W can be obtained through reverse transfer, and the final word vector, namely the vector W (i) of 1 m, can be obtained through multiplication of the final word vector and the initial word vector, namely x W;
x*w=W(i)=[Wi1Wi2… Wim]
and adding the word vectors corresponding to the fraud keywords appearing in each short message to obtain a text vector d belonging to the short message.
In this embodiment, the sentence vector module 23 is configured to convert the word vector into the sentence vector by using an LSTM algorithm.
The sentence vector module 23 specifically includes:
an arrangement unit for arranging the word vectors in a preset order, assuming that there are m word vectors xtForming;
an initialization unit for initializing model parameters Wf,Uf,bf,Wa,Ua,ba,Wi,Ui,bi,Wo,Uo,bo
A forgetting gate unit for changing xtPass into forgetting door ft+1And updating the weight W of the forgetting gatet+1,Ut+1,bt+1(ii) a Wherein,
ft+1=σ(Wt+1ht+Ut+1xt+1+bt+1);
wherein Wf,Uf,bfCoefficients and biases that are linear relationships; sigma is a sigmoid activation function;
an input gate unit for updating an input gate parameter itAnd at(ii) a Wherein,
it=σ(Wiht-1+Uixt+bi);at=tanh(Waht-1+Uaxt+ba);
Wa,Ua,ba,Wi,Ui,bicoefficients and biases that are linear relationships; sigma is a sigmoid activation function;
an output state unit for updating the model output state CtWherein
Ct=Ct-1⊙ft+it⊙at
as Hadamard product;
an output gate unit for updating an output gate parameter σtAnd htAnd outputs the predicted value of the current sequence indexWherein,
σt=σ(Woht-1+Uoxt+bo);
ht=ot⊙tanh(Ct);
specifically, a deep learning model for fraud short message recognition is built based on LSTM (long-short term memory artificial neural network), as shown in fig. 2.
LSTM, as a special case of RNN, can avoid the problem of gradient disappearance of conventional RNN. RNN assumes that the samples are sequence based. Such as from sequence index 1 to sequence index t. For any sequence index number t, its corresponding input is x in the corresponding sample sequence numberiAnd hidden state h at t-1 positiontThe determination is made in common. At any sequence index t, there is a corresponding model prediction output ot. By predicting the output otAnd training sequence true output ytAnd a loss function LtThe model can be trained in a similar way to softmax and then used to predict the output for some positions in the test sequence. Due to the problem of RNN gradient disappearance, the hidden structure of the sequence index position t is improved to avoidThe problem of gradient disappearance is avoided, and the special RNN is LSTM.
Hidden states h propagated forward at each sequence index position t instant, except as RNNtThere is another hidden state, generally called the cellular state, denoted Ct
In addition to cellular states, there are other structures in LSTM, which are generally referred to as gated structures. The gates of the LSTM at each sequence index position t typically include three types, a forgetting gate, an input gate, and an output gate.
The forgetting gate controls whether to forget, and in the LSTM, whether to forget the state of the hidden cell in the previous layer is controlled with a certain probability. Hidden state h of last sequence of inputt-1And the present sequence data xiThe output f of the forgetting gate is obtained by an activation function, generally sigmoidt. Output f due to sigmoidtIn [0,1 ]]And hence the output f heretRepresenting the probability of forgetting the state of a layer of hidden cells. Using the mathematical expression as:
ft+1=σ(Wt+1ht+Ut+1xt+1+bt+1);
wherein Wf,Uf,bfCoefficients and biases that are linear relationships; sigma is sigmoid activation function.
The input gate is responsible for processing the input of the current sequence position and consists of two parts, wherein the first part uses a sigmoid activation function, and the output is itThe second part uses the tanh activation function and outputs a bit atAnd multiplying the results of the two methods to update the cell state, wherein the mathematical expression is as follows:
it=σ(Wiht-1+Uixt+bi);
at=tanh(Waht-1+Uaxt+ba);
wherein, Wa,Ua,ba,Wi,Ui,biCoefficients and biases that are linear relationships; sigma is sigmoid activation function.
The results of both the forgetting gate and the entry gate contribute to the cell state Ct. Cell State CtIs composed of two parts, the first part is Ci-1And forget gate output ftThe second part is the input gate itAnd atThe product of (a) and (b), namely:
Ct=Ct-1⊙ft+it⊙at
wherein, it is Hadamard product.
Hidden state h of output gatetThe update of (2) is composed of two parts, the first part being otFrom the previous sequence of hidden states ht-1And the present sequence data xtAnd the activation function sigmoid, the second part being derived from the hidden state CtAnd tanh activation function, i.e.:
σt=σ(Woht-1+Uoxt+bo);
ht=ot⊙tanh(Ct)。
LSTM model with hidden state ht,CtParameter has Wf,Uf,bfWa,Ua,ba,Wi,Ui, bi,Wo,Uo,boAnd sequentially updating parameters according to the sequence of the forgetting gate, the input gate and the output gate, and outputting a current sequence index predicted value:
in this embodiment, the training module 24 is configured to use the vector as an input vector of the softmax classifier to train the deep learning model.
Wherein the softmax function formula is as follows:
wherein,is the output of the jth neuron in the lth layer,an output representing a jth neuron of a current layer; wherein,
in particular, one feature of the softmax function is that it takes as the output of each neuron the ratio of the input of that neuron to the sum of the inputs of all neurons in the current layer. This makes the output easier to interpret. The larger the output value of a neuron is, the higher the probability that the class corresponding to the neuron is a true class is.
The training module 24 specifically includes:
the selection unit is used for selecting the result with the maximum probability value as the output short message category;
the computing unit is used for computing a loss function and performing back propagation;
the adjusting unit is used for adjusting the weight of the deep learning model to a preset threshold value;
the input unit is used for inputting the test sample into the deep learning model and calculating the accuracy, the recall rate and the F value;
and the tuning unit is used for tuning the deep learning model by utilizing the self-learning capability of the neural network.
Specifically, the idea of the back propagation algorithm is to iteratively update all parameters by a gradient descent method, and the key point is to calculate partial derivatives of all parameters based on a loss function. There are two hidden states h in the LSTMtAnd CtTwo deltas are defined to propagate in opposite directions, i.e.
In this embodiment, the result with the maximum probability value of the hidden cell state is selected as the output of the short message, and then the loss function is calculated and the reverse propagation is performed. And then, adjusting the weight of the deep learning model to a preset threshold value, inputting the test sample into the deep learning model, calculating the accuracy, the recall rate and the F value, and then optimizing the deep learning model according to the calculated record. And training by the method to obtain the trained deep learning model.
In this embodiment, the recognition module 25 is configured to recognize the fraud short messages according to the output result of the trained deep learning softmax classifier.
Specifically, after the deep learning model is obtained through training, data is input into the deep learning model to identify the fraud short messages.
The specific embodiments described herein are merely illustrative of the spirit of the invention. Various modifications or additions may be made to the described embodiments or alternatives may be employed by those skilled in the art without departing from the spirit or ambit of the invention as defined in the appended claims.

Claims (10)

1. A method for recognizing fraud short messages by deep learning is characterized by comprising the following steps:
acquiring text data of a short message sample and performing word segmentation processing;
converting the text data after Word segmentation into Word vectors by adopting Word2 Vec;
converting the word vector into a sentence vector by adopting an LSTM algorithm;
taking the sentence vector as an input vector of a softmax classifier to train a deep learning model;
and recognizing fraud short messages according to the output result of the trained softmax classifier.
2. The method for recognizing fraud short messages by deep learning as claimed in claim 1, wherein the step of performing word segmentation processing on the text data of the obtained short messages specifically comprises:
collecting text data of all short message samples;
removing a non-text part in the text data by adopting a regularization method;
dividing the short message sample into a negative sample and a positive sample, and dividing the short message sample into a training sample and a testing sample according to a preset proportion;
performing word segmentation processing on the short message sample by adopting a crust word segmentation tool;
stop words are introduced to remove invalid words from the text.
3. The method of claim 2, wherein the step of training the deep learning model specifically comprises:
selecting the result with the maximum probability value as the output short message category;
calculating a loss function and performing back propagation;
adjusting the weight of the deep learning model to a preset threshold value;
inputting the test sample into the deep learning model, and calculating accuracy, recall rate and F value;
and optimizing the deep learning model by utilizing the self-learning capability of the neural network.
4. The method for recognizing fraud short messages by deep learning of claim 1, wherein said step of converting Word2Vec into Word vectors comprises:
counting the number m of the keywords in the fraud short message special library;
converting a word into an n-dimensional word vector x by using a one-hot-vector;
establishing a weight matrix w of n m to map the vector of n dimensions to the hidden neuron with dimension 1 m;
obtaining a vector W by back-propagation and a word vector W (i) of 1 m by multiplication with the word vector x;
and adding the word vectors corresponding to the fraud keywords appearing in each short message to obtain a text vector d of the short message.
5. The method for recognizing fraud short messages with deep learning of claim 1, wherein said step of converting said word vector into sentence vector using LSTM algorithm specifically comprises:
arranging the word vectors in a preset sequence, assuming that there are m word vectors xtForming;
initializing model parameters Wf,Uf,bf,Wa,Ua,ba,Wi,Ui,bi,Wo,Uo,bo
X is to betPass into forgetting door ft+1And updating the weight W of the forgetting gatet+1,Ut+1,bt+1(ii) a Wherein,
ft+1=σ(Wt+1ht+Ut+1xt+1+bt+1);
wherein, Wf,Uf,bfCoefficients and biases that are linear relationships; sigma is a sigmoid activation function;
updating input Gate parameter itAnd at(ii) a Wherein,
it=σ(Wiht-1+Uixt+bi);
at=tanh(Waht-1+Uaxt+ba);
Wa,Ua,ba,Wi,Ui,bicoefficients and biases that are linear relationships; sigma is a sigmoid activation function;
updating model output State CtWherein
Ct=Ct-1⊙ft+it⊙at
as Hadamard product;
updating output gate parameter σtAnd htAnd outputs the predicted value of the current sequence indexWherein,
σt=σ(Woht-1+Uoxt+bo);
ht=ot⊙tanh(Ct);
6. the method for identifying fraud short messages with deep learning of claim 1, wherein said softmax function formula is:
wherein,is the output of the jth neuron in the lth layer,an output representing a jth neuron of a current layer; wherein,
7. the method for recognizing fraud messages with deep learning of claim 1, further comprising the steps of:
and if the fraud short message is determined, intercepting the short message.
8. A system for recognizing fraud messages using deep learning, comprising:
the processing module is used for acquiring text data of the short message sample and performing word segmentation processing;
the Word vector module is used for converting the text data after Word segmentation into Word vectors by adopting Word2 Vec;
the sentence vector module is used for converting the word vector into a sentence vector by adopting an LSTM algorithm;
the training module is used for taking the sentence vector as an input vector of a softmax classifier so as to train a deep learning model;
and the recognition module is used for recognizing the fraud short messages according to the output result of the deep-learning softmax classifier after training.
9. The system of claim 8, wherein the processing module comprises:
the acquisition unit is used for acquiring text data of all short message samples;
the removing unit is used for removing the non-text part in the text data by adopting a regularization method;
the dividing unit is used for dividing the short message sample into a negative sample and a positive sample and dividing the short message sample into a training sample and a test sample according to a preset proportion;
the word segmentation unit is used for performing word segmentation processing on the short message sample by adopting a crust word segmentation tool;
and the introducing unit is used for introducing stop words to remove invalid words in the text.
10. The system of claim 8, wherein the training module specifically comprises:
the selection unit is used for selecting the result with the maximum probability value as the output short message category;
the computing unit is used for computing a loss function and performing back propagation;
the adjusting unit is used for adjusting the weight of the deep learning model to a preset threshold value;
the input unit is used for inputting the test sample into the deep learning model and calculating the accuracy, the recall rate and the F value;
and the tuning unit is used for tuning the deep learning model by utilizing the self-learning capability of the neural network.
CN201711205293.7A 2017-11-27 2017-11-27 A kind of method and system identifying fraud text message using deep learning Pending CN108566627A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711205293.7A CN108566627A (en) 2017-11-27 2017-11-27 A kind of method and system identifying fraud text message using deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711205293.7A CN108566627A (en) 2017-11-27 2017-11-27 A kind of method and system identifying fraud text message using deep learning

Publications (1)

Publication Number Publication Date
CN108566627A true CN108566627A (en) 2018-09-21

Family

ID=63529513

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711205293.7A Pending CN108566627A (en) 2017-11-27 2017-11-27 A kind of method and system identifying fraud text message using deep learning

Country Status (1)

Country Link
CN (1) CN108566627A (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109446528A (en) * 2018-10-30 2019-03-08 南京中孚信息技术有限公司 The recognition methods of new fraudulent gimmick and device
CN109492219A (en) * 2018-10-25 2019-03-19 山东省通信管理局 A kind of swindle website identification method analyzed based on tagsort and emotional semantic
CN109615377A (en) * 2018-12-13 2019-04-12 平安医疗健康管理股份有限公司 Repetition charge recognition methods, equipment, storage medium and device based on big data
CN109743732A (en) * 2018-12-20 2019-05-10 重庆邮电大学 Refuse messages method of discrimination based on improved CNN-LSTM
CN109801691A (en) * 2018-12-13 2019-05-24 平安科技(深圳)有限公司 Assault based on big data is gone to a doctor recognition methods, equipment, storage medium and device
CN109982272A (en) * 2019-02-13 2019-07-05 北京航空航天大学 A kind of fraud text message recognition methods and device
CN110110585A (en) * 2019-03-15 2019-08-09 西安电子科技大学 Intelligently reading realization method and system based on deep learning, computer program
CN110177179A (en) * 2019-05-16 2019-08-27 国家计算机网络与信息安全管理中心 A kind of swindle number identification method based on figure insertion
CN110198453A (en) * 2019-05-23 2019-09-03 武汉瓯越网视有限公司 Live content filter method, storage medium, equipment and system based on barrage
CN110781306A (en) * 2019-10-31 2020-02-11 山东师范大学 English text aspect layer emotion classification method and system
CN110913354A (en) * 2018-09-17 2020-03-24 阿里巴巴集团控股有限公司 Short message classification method and device and electronic equipment
CN111669757A (en) * 2020-06-15 2020-09-15 国家计算机网络与信息安全管理中心 Terminal fraud call identification method based on conversation text word vector
WO2021051585A1 (en) * 2019-09-17 2021-03-25 平安科技(深圳)有限公司 Method for constructing natural language processing system, electronic apparatus, and computer device
CN112906397A (en) * 2021-04-06 2021-06-04 南通大学 Short text entity disambiguation method
CN113923669A (en) * 2021-11-10 2022-01-11 恒安嘉新(北京)科技股份公司 Anti-fraud early warning method, device, equipment and medium for multi-source cross-platform fusion
CN114066490A (en) * 2022-01-17 2022-02-18 浙江鹏信信息科技股份有限公司 GoIP fraud nest point identification method, system and computer readable storage medium
CN114154490A (en) * 2020-08-18 2022-03-08 阿里巴巴集团控股有限公司 Model training method, title extracting method, device, electronic equipment and computer readable medium
CN118474682A (en) * 2024-07-15 2024-08-09 浙江三子智联科技有限公司 Service short message monitoring method and system based on big data

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9053431B1 (en) * 2010-10-26 2015-06-09 Michael Lamport Commons Intelligent control with hierarchical stacked neural networks
CN106202330A (en) * 2016-07-01 2016-12-07 北京小米移动软件有限公司 The determination methods of junk information and device
CN106294322A (en) * 2016-08-04 2017-01-04 哈尔滨工业大学 A kind of Chinese based on LSTM zero reference resolution method
CN106682089A (en) * 2016-11-26 2017-05-17 山东大学 RNNs-based method for automatic safety checking of short message

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9053431B1 (en) * 2010-10-26 2015-06-09 Michael Lamport Commons Intelligent control with hierarchical stacked neural networks
CN106202330A (en) * 2016-07-01 2016-12-07 北京小米移动软件有限公司 The determination methods of junk information and device
CN106294322A (en) * 2016-08-04 2017-01-04 哈尔滨工业大学 A kind of Chinese based on LSTM zero reference resolution method
CN106682089A (en) * 2016-11-26 2017-05-17 山东大学 RNNs-based method for automatic safety checking of short message

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
张英: "基于深度神经网络的微博短文本情感分析研究", 《中原工学院学位论文》 *

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110913354A (en) * 2018-09-17 2020-03-24 阿里巴巴集团控股有限公司 Short message classification method and device and electronic equipment
CN109492219A (en) * 2018-10-25 2019-03-19 山东省通信管理局 A kind of swindle website identification method analyzed based on tagsort and emotional semantic
CN109446528A (en) * 2018-10-30 2019-03-08 南京中孚信息技术有限公司 The recognition methods of new fraudulent gimmick and device
CN109615377A (en) * 2018-12-13 2019-04-12 平安医疗健康管理股份有限公司 Repetition charge recognition methods, equipment, storage medium and device based on big data
CN109801691A (en) * 2018-12-13 2019-05-24 平安科技(深圳)有限公司 Assault based on big data is gone to a doctor recognition methods, equipment, storage medium and device
CN109743732A (en) * 2018-12-20 2019-05-10 重庆邮电大学 Refuse messages method of discrimination based on improved CNN-LSTM
CN109982272A (en) * 2019-02-13 2019-07-05 北京航空航天大学 A kind of fraud text message recognition methods and device
CN109982272B (en) * 2019-02-13 2020-08-28 北京航空航天大学 Fraud short message identification method and device
CN110110585A (en) * 2019-03-15 2019-08-09 西安电子科技大学 Intelligently reading realization method and system based on deep learning, computer program
CN110177179A (en) * 2019-05-16 2019-08-27 国家计算机网络与信息安全管理中心 A kind of swindle number identification method based on figure insertion
CN110198453A (en) * 2019-05-23 2019-09-03 武汉瓯越网视有限公司 Live content filter method, storage medium, equipment and system based on barrage
WO2021051585A1 (en) * 2019-09-17 2021-03-25 平安科技(深圳)有限公司 Method for constructing natural language processing system, electronic apparatus, and computer device
CN110781306A (en) * 2019-10-31 2020-02-11 山东师范大学 English text aspect layer emotion classification method and system
CN111669757A (en) * 2020-06-15 2020-09-15 国家计算机网络与信息安全管理中心 Terminal fraud call identification method based on conversation text word vector
CN114154490A (en) * 2020-08-18 2022-03-08 阿里巴巴集团控股有限公司 Model training method, title extracting method, device, electronic equipment and computer readable medium
CN112906397A (en) * 2021-04-06 2021-06-04 南通大学 Short text entity disambiguation method
CN112906397B (en) * 2021-04-06 2021-11-19 南通大学 Short text entity disambiguation method
CN113923669A (en) * 2021-11-10 2022-01-11 恒安嘉新(北京)科技股份公司 Anti-fraud early warning method, device, equipment and medium for multi-source cross-platform fusion
CN113923669B (en) * 2021-11-10 2024-05-17 恒安嘉新(北京)科技股份公司 Multi-source cross-platform fusion anti-fraud early warning method, device, equipment and medium
CN114066490A (en) * 2022-01-17 2022-02-18 浙江鹏信信息科技股份有限公司 GoIP fraud nest point identification method, system and computer readable storage medium
CN118474682A (en) * 2024-07-15 2024-08-09 浙江三子智联科技有限公司 Service short message monitoring method and system based on big data

Similar Documents

Publication Publication Date Title
CN108566627A (en) A kind of method and system identifying fraud text message using deep learning
CN108595632B (en) Hybrid neural network text classification method fusing abstract and main body characteristics
CN109597997B (en) Comment entity and aspect-level emotion classification method and device and model training thereof
Du et al. Novel efficient RNN and LSTM-like architectures: Recurrent and gated broad learning systems and their applications for text classification
CN111061843B (en) Knowledge-graph-guided false news detection method
CN109934261B (en) Knowledge-driven parameter propagation model and few-sample learning method thereof
CN112883738A (en) Medical entity relation extraction method based on neural network and self-attention mechanism
CN111414461B (en) Intelligent question-answering method and system fusing knowledge base and user modeling
CN111985247B (en) Microblog user interest identification method and system based on multi-granularity text feature representation
CN107562784A (en) Short text classification method based on ResLCNN models
CN108536754A (en) Electronic health record entity relation extraction method based on BLSTM and attention mechanism
CN107122479A (en) A kind of user cipher conjecture system based on deep learning
CN112749274B (en) Chinese text classification method based on attention mechanism and interference word deletion
CN110851176B (en) Clone code detection method capable of automatically constructing and utilizing pseudo-clone corpus
CN111368920A (en) Quantum twin neural network-based binary classification method and face recognition method thereof
CN112906397B (en) Short text entity disambiguation method
CN113742733B (en) Method and device for extracting trigger words of reading and understanding vulnerability event and identifying vulnerability type
CN112015901A (en) Text classification method and device and warning situation analysis system
CN111309909A (en) Text emotion classification method based on hybrid model
EP3874412A1 (en) Computer architecture for multiplier-less machine learning
CN114547299A (en) Short text sentiment classification method and device based on composite network model
CN117593037A (en) Method for predicting completion capability of human-computer interaction user
Jeyakarthic et al. Optimal bidirectional long short term memory based sentiment analysis with sarcasm detection and classification on twitter data
CN113535928A (en) Service discovery method and system of long-term and short-term memory network based on attention mechanism
CN112560440A (en) Deep learning-based syntax dependence method for aspect-level emotion analysis

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20180921