CN114444485A - Cloud environment network equipment entity identification method - Google Patents

Cloud environment network equipment entity identification method Download PDF

Info

Publication number
CN114444485A
CN114444485A CN202210078810.3A CN202210078810A CN114444485A CN 114444485 A CN114444485 A CN 114444485A CN 202210078810 A CN202210078810 A CN 202210078810A CN 114444485 A CN114444485 A CN 114444485A
Authority
CN
China
Prior art keywords
character
cloud environment
sequence
network equipment
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210078810.3A
Other languages
Chinese (zh)
Other versions
CN114444485B (en
Inventor
陈兴蜀
郑涛
袁磊
刘朋
黄铁脉
廖志红
宋可儿
王海舟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan University
Original Assignee
Sichuan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan University filed Critical Sichuan University
Priority to CN202210078810.3A priority Critical patent/CN114444485B/en
Publication of CN114444485A publication Critical patent/CN114444485A/en
Application granted granted Critical
Publication of CN114444485B publication Critical patent/CN114444485B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses a cloud environment network equipment entity identification method, which comprises the steps of firstly, collecting original data of a text corpus corresponding to cloud environment network equipment information; manually making a small number of features to form a feature template according to the characteristics of the cloud environment network equipment information, and converting the input sentences into corresponding word vector sequences; then, extracting character-level features of each word aiming at the converted word vector sequence; combining the character-level feature vector and the local context feature as input information, and extracting semantic features of the input information; and finally, labeling the cloud environment network equipment entity for each character by using a conditional random field according to the semantic features of the obtained input information, and completing the identification of the cloud environment network equipment entity. The method can be used for realizing identification and classification of the network equipment entities in the cloud environment, and is convenient for quickly reconstructing the internal system architecture outside the closed cloud environment.

Description

Cloud environment network equipment entity identification method
Technical Field
The invention relates to the technical field of cloud environment network equipment entity identification, in particular to a cloud environment network equipment entity identification method based on a convolutional neural network, a bidirectional long-short term memory neural network and a conditional random field.
Background
Cloud environment: the cloud environment is a distributed computing system based on virtualization technology, the system architecture of the system can be mainly divided into three layers of an Infrastructure layer, a platform layer and a software Service layer, corresponding names are Infrastructure as a Service (IaaS) respectively, the system mainly comprises a computer server, communication equipment, storage equipment and the like, and the computing resources are unified and virtualized into computing resources, storage resources and network resources through the virtualization technology; platform as a Service (PaaS), which mainly comprises a supporting Platform for providing a whole set of application software for development, operation and operation for users; software as a Service (Software as a Service), which has a main function of providing an application mode of Software Service for a user through the internet, in which the user only needs to pay application use fees, and Software and hardware consumption and maintenance generated in the application development process are managed by a cloud Service provider.
Cloud environment network equipment entity identification: the cloud environment network device mainly refers to a computing resource device (CPU, memory and the like), a storage resource device (database and the like) and a network resource device (firewall, data center switch, VPN, router and the like). The entity identification is entity identification in a specific field in the field of named entity identification, and the main task is to identify various network equipment entities in the cloud environment from network security text data so as to identify and classify the network equipment entities in the cloud environment.
In recent years, a deep learning method based on a neural network has a good effect in named entity recognition in the general field, and completes feature extraction of data by building a multi-layer network structure, and then identifies and classifies the data. The more common deep Neural Network structures include a Recurrent Neural Network (RNN), a Convolutional Neural Network (CNN), and a Long Short-Term Memory Neural Network (LSTM). Although the deep neural network has a good effect in the field of naming identification in the general field, in a cloud environment, due to the fact that physical nodes and virtual equipment are difficult to distinguish after virtualization layer by layer, dynamically-changed cloud resources are difficult to locate, and various safety protection measures in the cloud, the whole appearance of a cloud environment network equipment entity is difficult to carve through traditional conventional port scanning, IP scanning and penetrating measures, and even can be found and disposed by the cloud safety protection measures rapidly. Therefore, it is necessary to effectively process the network security text information obtained by network detection through the deep neural network, and further complete identification and classification of various network device entity information in the cloud environment.
Disclosure of Invention
In view of the above problems, an object of the present invention is to provide a cloud environment network device entity identification method based on a deep neural network, which is implemented based on a convolutional neural network, a long-short term memory neural network, and a conditional random field, and can implement a cloud environment network device entity identification function while ensuring a high accuracy. The technical scheme is as follows:
a cloud environment network equipment entity identification method comprises
Step 1: collecting original data of a text corpus corresponding to the cloud environment network equipment information;
step 2: manually formulating a small amount of characteristics to form a characteristic template according to the characteristics of the cloud environment network equipment information; then extracting local context information, and converting the input sentence into a corresponding word vector sequence by using a pre-trained word vector file;
and step 3: performing convolution pooling on each word by using a convolution neural network aiming at the converted word vector sequence, and extracting the character-level characteristics of each word;
and 4, step 4: combining the character-level feature vector and the local context feature to serve as input information, transmitting the input information into a bidirectional long-short term memory neural network for training, and extracting semantic features of the input information;
and 5: and aiming at the semantic features of the obtained input information, carrying out cloud environment network equipment entity labeling on each character by using a conditional random field, and labeling cloud environment network equipment entity information in the sentence sequence, so as to obtain an optimal labeling sequence and further finish the identification of the cloud environment network equipment entity.
Further, the step 2 specifically comprises:
step 2.1: comprehensively considering the category information and the specific position information of the collected text data equipment, and manually making a characteristic template as follows:
T[-2,0],T[-1,0],T[0,0],T[1,0],T[2,0];
wherein T represents the current character, the first number in parentheses represents its position relative to the current character, and the second character represents the number of columns of the selected specific feature;
step 2.2: a set of feature functions is formulated according to the features:
Ft(Lloc-1,Lloc,w,loc)
wherein ,LlocRepresenting the current character label, Lloc-1Represents the next label, w represents the current character, and loc represents the current position;
step 2.3: giving a weight δ to all feature functionslocIf a feature function is activated, its weight δlocWill be added to the final cumulative value, the feature score value FtThe higher the value is, the more accurate the prediction result is; step 2.4: and training the original corpus of the cloud environment network equipment by using a CBOW model of word2vec to obtain a corresponding word vector file.
Further, the step 3 specifically includes:
step 3.1: setting different character vectors for characters of different types of equipment in a word embedding mode, and distinguishing the case and the type of the characters;
step 3.2: converting each word of the input sequence into a corresponding character vector, and filling placeholders on the basis of the longest word to make all character vector matrixes consistent in size;
step 3.3: local features of the character vector matrix are extracted by utilizing the convolution layer of the convolution neural network, dimension reduction features of the layers are pooled, and character-level features are further extracted.
Further, the step 4 specifically includes:
simultaneously using the extracted character-level features and the context local information of the input feature information as input, and splicing the semantic feature sequences output by the positive LSTM at the h moment
Figure BDA0003485226790000041
And reverse LSTM semantic feature vector sequences
Figure BDA0003485226790000042
Obtaining a complete semantic feature vector sequence
Figure BDA0003485226790000043
The preprocessing is done using the tanh activation function, and the output result of the LSTM hidden layer is obtained as:
Figure BDA0003485226790000044
wherein ,OuthRepresenting the output value, w, of the h-time character calculated by the LSTM hidden layerhIs a corresponding weight value, TkIs a bias vector.
Further, the step 5 specifically includes:
step 5.1: defining the input sequence as S ═ S1,...,sn]; wherein ,snIs the input vector for the nth character, L ═ L1,...,ln]Is a predicted tag sequence of sequence S;
step 5.2: calculating the label score corresponding to each character C at the moment T:
ScoreT=Wc×OutT+bc+FS
wherein, ScoreTEach character corresponding to T-timeTotal score value of label, Wc×OutTRepresenting the label score, W, of the LSTM hidden layer computationcRepresenting the weight matrix, Out, corresponding to each character CTThe output value of each character C at the moment T is calculated by an LSTM hidden layer; bcOffset vector of LSTM hidden layer input gate representing character C, FSRepresenting the weight values of the feature templates in the input sequence;
step 5.3: calculating the transfer score between tags:
defining a transition matrix U, and adding two special labels for starting and ending in the e labels, wherein the total score value of a label sequence is determined by the transition of each label, and the specific calculation formula is as follows:
Figure BDA0003485226790000045
wherein ,
Figure BDA0003485226790000046
representative label LiTransfer to Li+1Is the transfer score; score (L)i+1) Represents the current label Li+1A score of (a); l is a radical of an alcohol0 and LeRespectively representing the beginning and the end of two special labels, and the corresponding score value of the label is 0;
step 5.4: calculating the generation probability of the label sequence:
Figure BDA0003485226790000051
wherein, L' represents a tag sequence to be calculated except for a start tag and an end tag in the input sequence; XL represents the scalar space composed of total fractional value X and prediction label L;
step 5.5: obtaining a loss function by utilizing maximum likelihood estimation, so that the probability of generating a correct label sequence L is maximum under the condition of an input sequence S; the loss function is calculated by the formula:
Figure BDA0003485226790000052
wherein add represents an accumulation function;
step 5.6: obtaining the sequence L with the highest score through a gradient descent learning algorithm and a Viterbi algorithm*As a final result of the cloud environment network device entity identification, namely:
L*=argmaxX(S,L')。
the invention has the beneficial effects that: according to the method, a cloud environment characteristic template is manually established, and data text labeling of a cloud environment network equipment entity is completed through a convolutional neural network, a bidirectional long-short term memory neural network and a conditional random field; constructing a complete cloud environment network equipment entity identification method, and providing a theoretical method for comprehensively and completely depicting a cloud environment network equipment entity; the method can be used for effectively realizing the identification and classification of the network equipment entities in the cloud environment, and is convenient for quickly reconstructing the internal system architecture outside the closed cloud environment.
Drawings
Fig. 1 is a schematic diagram of a cloud environment network device entity identification method according to the present invention.
FIG. 2 is a schematic diagram of a neural network model designed by the present invention.
Detailed Description
The invention is described in further detail below with reference to the figures and specific embodiments. The method for identifying the cloud environment network equipment entity comprises the following steps of firstly collecting a data information text of the cloud environment network equipment entity, manually formulating a characteristic template by combining with proprietary information of cloud equipment, and then transmitting the characteristic template into a convolutional neural network and a bidirectional long-short term memory neural network for characteristic information extraction:
(1) and collecting original data of a text corpus corresponding to the cloud environment network equipment information.
Network text data related to cloud environment network equipment entities are obtained from various network security websites and cloud manufacturers, the data volume is 18920 pieces in total, ten thousand pieces of original data are selected as unmarked data and used as training word vector files, and the rest 8920 pieces of data are used for cloud environment network equipment entity identification marking corpora.
(2) Manually formulating a small amount of characteristics to form a characteristic template according to the characteristics of the cloud environment network equipment information; and then extracting local context information, and converting the input sentence into a corresponding word vector sequence by using the pre-trained word vector file.
The main body identified by the method is the equipment name, the software version number information and the category (computing resource equipment, storage resource equipment and network resource equipment) in the text data of the cloud environment network equipment; firstly, according to the information characteristics of the class to which the equipment belongs, a small number of characteristics are manually formulated to form a characteristic template.
The method comprises the following specific steps: the text data collected by comprehensive consideration is classified into category information and specific position information, and the characteristic template is manually set as T-2, 0],T[-1,0],T[0,0],T[1,0],T[2,0]. T denotes the current character, the first number in parentheses represents its position relative to the current character, and the second character denotes the number of columns of the particular feature selected, such as word and part of speech. A set of characteristic functions F is formulated according to the characteristicst(Lloc-1,LlocW, loc) in which LlocRepresenting the current character label, Lloc-1Representing the next label, w the current character, and loc the current position. All feature functions are then given a weight δlocIf a characteristic function is activated, its weight δlocWill be added to the final cumulative value, the feature score value FtHigher values indicate more accurate prediction results. The invention trains the original corpus of the cloud environment network equipment by using the CBOW model of word2vec to obtain a corresponding word vector file.
(3) And (4) carrying out convolution pooling on each word by utilizing a convolution neural network aiming at the converted word vector sequence, and extracting the character-level characteristics of the word.
After the word vector file corresponding to the original corpus is obtained, the invention adopts the convolutional neural network to extract the character-level characteristics in the word vector file, thereby improving the performance of the model. The specific operation is as follows: firstly, different character vectors are set for characters of different classes (computing resource equipment, storage resource equipment and network resource equipment) to which different equipment belongs in a word embedding mode, and are used for distinguishing the case and the type of the characters. Second, each word of the input sequence is converted to a corresponding character vector, and all character vector matrices are made uniform in size by filling placeholders with reference to the longest word. And finally, extracting local features of the character vector matrix by using the convolution layer of the convolution neural network, pooling layer dimension reduction features, and further extracting character level features.
(4) Combining the character-level feature vector and the local context feature as input information, transmitting the input information into a bidirectional long-short term memory neural network for training, and extracting semantic features of the input information.
The invention adopts a bidirectional long-short term memory neural network (BilSTM) training character feature vector sequence, and in order to improve the accuracy of cloud environment network equipment entity identification, the invention also takes context local information of input feature information as input at the same time and outputs the feature sequence by splicing the feature sequence output by the positive LSTM at the h moment
Figure BDA0003485226790000071
And inverse LSTM feature vector sequence
Figure BDA0003485226790000072
Obtaining complete characteristic vector sequence
Figure BDA0003485226790000073
The preprocessing is completed by using the tanh activation function, so that the output result of the hidden layer is obtained
Figure BDA0003485226790000074
wherein whIs a corresponding weight value, TkIs a bias vector.
(5) And aiming at the semantic features of the obtained input information, carrying out cloud environment network equipment entity labeling on each character by using a conditional random field, and labeling cloud environment network equipment entity information in the sentence sequence, so as to obtain an optimal labeling sequence and further finish the identification of the cloud environment network equipment entity.
And (5) by labeling the hidden layer output result obtained in the step (4), converting the cloud environment network equipment entity identification problem into a character sequence labeling problem. In order to improve the accuracy of cloud environment network equipment entity identification, the invention utilizes the chain random condition field CRF to calculate the probability of the whole tag sequence, thereby obtaining the globally optimal tag sequence, namely the optimal output result of cloud environment network equipment entity identification.
Specifically, assume that the input sequence of the present invention is S ═ S1,...,sn], wherein snIs the input vector for the nth character, L ═ L1,...,ln]Is the tag sequence of sequence S. In order to obtain the correct label corresponding to each character, it is assumed that each character in the sequence S has a corresponding score value, and the larger the value of the score value, the larger the value of the label correct value. Obviously, the larger the tag score value of the input sequence S, the higher the probability that the corresponding tag is correct. The label with the highest total value of the added label score values before and after the sequence L is the final prediction result of each character. Adding the label score calculated by the hidden layer of the bidirectional long and short term memory neural network (LSTM) in the step (4) and the weight value calculated by the feature template in the step (2) to obtain the label score corresponding to each character C at the moment T, wherein the specific calculation formula is as follows:
ScoreT=Wc×OutT+bc+FS
ScoreTrepresenting the total score value, W, of the label corresponding to each character at time Tc×OutTRepresenting the label score, W, of the LSTM hidden layer computationcRepresenting the weight matrix, Out, corresponding to each character CTIs the output value calculated by the LSTM hidden layer of each character C at the time T, bcOffset vector of LSTM hidden layer input gate representing character C, FSRepresenting the weight values of the feature templates in the input sequence.
Because the transition between labels is generated when the probability value of the whole label sequence is calculated by using the chain random condition field CRF, in order to calculate the transition score between the labels, the invention defines a transition matrix U, and two special labels of the beginning and the end are added in the e labels. Therefore, the total score value of a tag sequence is determined by the transfer of each tag, and the specific calculation formula is as follows:
Figure BDA0003485226790000081
wherein S is an input sequence, L is a predicted tag sequence,
Figure BDA0003485226790000082
representative label LiTransfer to Li+1I.e. the transition score. Score (L)i+1) Represents the current label Li+1The fraction of (c). L is a radical of an alcohol0 and LeRepresenting the beginning and the end of two special labels, respectively, with a score value of 0 for the label.
The probability of generation of a tag sequence L can be calculated given an input sequence S by the following formula:
Figure BDA0003485226790000091
wherein, L' represents the sequence of tags to be calculated except for the start and end tags in the input sequence. Then the invention obtains a loss function by utilizing maximum likelihood estimation, so that the probability of generating a correct label sequence L is maximum under the condition of an input sequence S. The loss function is calculated by the formula:
Figure BDA0003485226790000092
obtaining the sequence L with the highest score through a gradient descent learning algorithm and a Viterbi algorithm*As a final result of the cloud environment network device entity identification, namely:
L*=argmaxX(S,L')
according to the cloud environment network equipment entity identification method, various equipment entities in the cloud environment network are accurately identified from network text data by manually establishing characteristic templates of cloud environment computing resource equipment, storage resource equipment and network resource equipment and combining methods such as a convolutional neural network, a bidirectional long-short term memory neural network and a conditional random field.

Claims (5)

1. A cloud environment network equipment entity identification method is characterized by comprising
Step 1: collecting original data of a text corpus corresponding to the cloud environment network equipment information;
step 2: manually formulating a small quantity of characteristics to form a characteristic template according to the characteristics of the cloud environment network equipment information; then extracting local context information, and converting the input sentence into a corresponding word vector sequence by using a pre-trained word vector file;
and step 3: performing convolution pooling on each word by using a convolution neural network aiming at the converted word vector sequence, and extracting character-level features of each word;
and 4, step 4: combining the character-level feature vector and the local context feature to serve as input information, transmitting the input information into a bidirectional long-short term memory neural network for training, and extracting semantic features of the input information;
and 5: and aiming at the semantic features of the obtained input information, carrying out cloud environment network equipment entity labeling on each character by using a conditional random field, and labeling cloud environment network equipment entity information in the sentence sequence, so as to obtain an optimal labeling sequence and further finish the identification of the cloud environment network equipment entity.
2. The cloud environment network device entity identification method according to claim 1, wherein the step 2 specifically is:
step 2.1: comprehensively considering the category information and the specific position information of the collected text data equipment, and manually making a characteristic template as follows:
T[-2,0],T[-1,0],T[0,0],T[1,0],T[2,0];
wherein T represents the current character, the first number in parentheses represents its position relative to the current character, and the second character represents the number of columns of the selected specific feature;
step 2.2: a set of feature functions is formulated according to the features:
Ft(Lloc-1,Lloc,w,loc)
wherein ,LlocRepresenting the current character label, Lloc-1Represents the next label, w represents the current character, and loc represents the current position;
step 2.3: giving a weight δ to all feature functionslocIf a feature function is activated, its weight δlocWill be added to the final cumulative value, the feature score value FtThe higher the value is, the more accurate the prediction result is;
step 2.4: and training the original corpus of the cloud environment network equipment by using a CBOW model of word2vec to obtain a corresponding word vector file.
3. The cloud environment network device entity identification method according to claim 1, wherein the step 3 specifically is:
step 3.1: setting different character vectors for characters of different types of equipment in a word embedding mode, and distinguishing the case and the type of the characters;
step 3.2: converting each word of the input sequence into a corresponding character vector, and filling placeholders on the basis of the longest word to make all character vector matrixes consistent in size;
step 3.3: local features of the character vector matrix are extracted by utilizing the convolution layer of the convolution neural network, dimension reduction features of the layers are pooled, and character-level features are further extracted.
4. The cloud environment network device entity identification method according to claim 1, wherein the step 4 specifically is:
simultaneously using the extracted character-level features and the context local information of the input feature information as input, and splicing the semantic feature sequences output by the positive LSTM at the h moment
Figure FDA0003485226780000021
And reverse LSTM semantic feature vector sequences
Figure FDA0003485226780000022
Obtaining a complete semantic feature vector sequence
Figure FDA0003485226780000023
The preprocessing is done using the tanh activation function, and the output result of the LSTM hidden layer is obtained as:
Figure FDA0003485226780000024
wherein ,OuthRepresenting the output value, w, of the h-time character calculated by the LSTM hidden layerhIs a corresponding weight value, TkIs a bias vector.
5. The cloud environment network device entity identification method according to claim 4, wherein the step 5 specifically comprises:
step 5.1: defining the input sequence as S ═ S1,...,sn]; wherein ,snIs the input vector for the nth character, L ═ L1,...,ln]Is a predicted tag sequence of sequence S;
step 5.2: calculating the label score corresponding to each character C at the moment T:
ScoreT=Wc×OutT+bc+FS
wherein, ScoreTRepresents the total mark point value, W, corresponding to each character at the moment Tc×OutTRepresenting the label score, W, of the LSTM hidden layer computationcRepresenting the weight matrix, Out, corresponding to each character CTThe output value of each character C at the moment T is calculated by an LSTM hidden layer; bcOffset vector of input gate of LSTM hidden layer representing character C, FSRepresenting the weight values of the feature templates in the input sequence;
step 5.3: calculating the transfer score between tags:
defining a transition matrix U, and adding two special labels for starting and ending in the e labels, wherein the total score value of a label sequence is determined by the transition of each label, and the specific calculation formula is as follows:
Figure FDA0003485226780000031
wherein ,
Figure FDA0003485226780000032
representative label LiTransfer to Li+1Is the transition score; score (L)i+1) Represents the current label Li+1A fraction of (d); l is0 and LeRespectively representing the beginning and the end of two special labels, wherein the corresponding fraction value of the labels is 0;
step 5.4: calculating the generation probability of the label sequence:
Figure FDA0003485226780000033
wherein, L' represents a tag sequence to be calculated except for a start tag and an end tag in the input sequence; XL represents the scalar space composed of total fractional value X and prediction label L;
step 5.5: obtaining a loss function by utilizing maximum likelihood estimation, so that the probability of generating a correct label sequence L is maximum under the condition of an input sequence S; the loss function is calculated by the formula:
Figure FDA0003485226780000041
wherein add represents an accumulation function;
step 5.6: obtaining the sequence L with the highest score through a gradient descent learning algorithm and a Viterbi algorithm*As a cloud environment network deviceThe final result of the body identification, namely:
L*=argmaxX(S,L')。
CN202210078810.3A 2022-01-24 2022-01-24 Cloud environment network equipment entity identification method Active CN114444485B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210078810.3A CN114444485B (en) 2022-01-24 2022-01-24 Cloud environment network equipment entity identification method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210078810.3A CN114444485B (en) 2022-01-24 2022-01-24 Cloud environment network equipment entity identification method

Publications (2)

Publication Number Publication Date
CN114444485A true CN114444485A (en) 2022-05-06
CN114444485B CN114444485B (en) 2023-06-06

Family

ID=81368979

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210078810.3A Active CN114444485B (en) 2022-01-24 2022-01-24 Cloud environment network equipment entity identification method

Country Status (1)

Country Link
CN (1) CN114444485B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106569998A (en) * 2016-10-27 2017-04-19 浙江大学 Text named entity recognition method based on Bi-LSTM, CNN and CRF
CN109117472A (en) * 2018-11-12 2019-01-01 新疆大学 A kind of Uighur name entity recognition method based on deep learning
KR20190004541A (en) * 2017-07-04 2019-01-14 심동희 Apparatus for managing using could based platform of animal hospital management method for recognizing animal individual with based noseprint constructed with based it
CN109871535A (en) * 2019-01-16 2019-06-11 四川大学 A kind of French name entity recognition method based on deep neural network
CN111368996A (en) * 2019-02-14 2020-07-03 谷歌有限责任公司 Retraining projection network capable of delivering natural language representation
CN111444726A (en) * 2020-03-27 2020-07-24 河海大学常州校区 Method and device for extracting Chinese semantic information of long-time and short-time memory network based on bidirectional lattice structure
CN113190842A (en) * 2018-09-05 2021-07-30 甲骨文国际公司 Context-aware feature embedding using deep recurrent neural networks and anomaly detection of sequential log data

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106569998A (en) * 2016-10-27 2017-04-19 浙江大学 Text named entity recognition method based on Bi-LSTM, CNN and CRF
KR20190004541A (en) * 2017-07-04 2019-01-14 심동희 Apparatus for managing using could based platform of animal hospital management method for recognizing animal individual with based noseprint constructed with based it
CN113190842A (en) * 2018-09-05 2021-07-30 甲骨文国际公司 Context-aware feature embedding using deep recurrent neural networks and anomaly detection of sequential log data
CN109117472A (en) * 2018-11-12 2019-01-01 新疆大学 A kind of Uighur name entity recognition method based on deep learning
CN109871535A (en) * 2019-01-16 2019-06-11 四川大学 A kind of French name entity recognition method based on deep neural network
CN111368996A (en) * 2019-02-14 2020-07-03 谷歌有限责任公司 Retraining projection network capable of delivering natural language representation
CN111444726A (en) * 2020-03-27 2020-07-24 河海大学常州校区 Method and device for extracting Chinese semantic information of long-time and short-time memory network based on bidirectional lattice structure

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
WANWAN CAO等: "Power entity identification method in cloud environment", pages 2400 - 2405 *
周胜利 等: "基于评分卡――随机森林的云计算用户公共安全信誉模型研究", vol. 39, no. 05, pages 143 - 152 *

Also Published As

Publication number Publication date
CN114444485B (en) 2023-06-06

Similar Documents

Publication Publication Date Title
CN108920622B (en) Training method, training device and recognition device for intention recognition
CN110097085B (en) Lyric text generation method, training method, device, server and storage medium
WO2021179570A1 (en) Sequence labeling method and apparatus, and computer device and storage medium
CN110619051B (en) Question sentence classification method, device, electronic equipment and storage medium
CN111159485B (en) Tail entity linking method, device, server and storage medium
CN111666427A (en) Entity relationship joint extraction method, device, equipment and medium
CN110347790B (en) Text duplicate checking method, device and equipment based on attention mechanism and storage medium
CN113515632B (en) Text classification method based on graph path knowledge extraction
CN113312452A (en) Chapter-level text continuity classification method based on multi-task learning
CN113255320A (en) Entity relation extraction method and device based on syntax tree and graph attention machine mechanism
CN113051914A (en) Enterprise hidden label extraction method and device based on multi-feature dynamic portrait
CN113590810B (en) Abstract generation model training method, abstract generation device and electronic equipment
CN112926324A (en) Vietnamese event entity recognition method integrating dictionary and anti-migration
CN110334186A (en) Data query method, apparatus, computer equipment and computer readable storage medium
CN111881256B (en) Text entity relation extraction method and device and computer readable storage medium equipment
CN112650833A (en) API (application program interface) matching model establishing method and cross-city government affair API matching method
CN115408488A (en) Segmentation method and system for novel scene text
CN111368066B (en) Method, apparatus and computer readable storage medium for obtaining dialogue abstract
CN113609819B (en) Punctuation mark determination model and determination method
CN113486143A (en) User portrait generation method based on multi-level text representation and model fusion
CN113076744A (en) Cultural relic knowledge relation extraction method based on convolutional neural network
CN110717316B (en) Topic segmentation method and device for subtitle dialog flow
WO2023137903A1 (en) Reply statement determination method and apparatus based on rough semantics, and electronic device
CN110377753A (en) Relation extraction method and device based on relationship trigger word Yu GRU model
CN113434698B (en) Relation extraction model establishing method based on full-hierarchy attention and application thereof

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant